Efficient and effective text encoding for chinese llama and alpaca

Y Cui, Z Yang, X Yao - arxiv preprint arxiv:2304.08177, 2023 - arxiv.org
Large Language Models (LLMs), such as ChatGPT and GPT-4, have dramatically
transformed natural language processing research and shown promising strides towards …

Mer 2024: Semi-supervised learning, noise robustness, and open-vocabulary multimodal emotion recognition

Z Lian, H Sun, L Sun, Z Wen, S Zhang, S Chen… - Proceedings of the 2nd …, 2024 - dl.acm.org
Multimodal emotion recognition is an important research topic in artificial intelligence.
However, due to problems such as complex environments and inaccurate annotations …

Merbench: A unified evaluation benchmark for multimodal emotion recognition

Z Lian, L Sun, Y Ren, H Gu, H Sun, L Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Multimodal emotion recognition plays a crucial role in enhancing user experience in human-
computer interaction. Over the past few decades, researchers have proposed a series of …

An iteratively parallel generation method with the pre-filling strategy for document-level event extraction

G Huang, R Xu, Y Zeng, J Chen, Z Yang… - Proceedings of the …, 2023 - aclanthology.org
In document-level event extraction (DEE) tasks, a document typically contains many event
records with multiple event roles. Therefore, accurately extracting all event records is a big …

A Language Model-based Fine-Grained Address Resolution Framework in UAV Delivery System

S Luo, Y Yao, H Zhao, L Song - IEEE Journal of Selected …, 2024 - ieeexplore.ieee.org
Accurate address resolution plays a vital role in UAV delivery systems. Existing address
resolution systems heavily rely on user-provided Point of Interest (POI) information …

Gradual Syntactic Label Replacement for Language Model Pre-Training

Y Wang, Y Zhang, P Li, Y Liu - IEEE/ACM Transactions on …, 2023 - ieeexplore.ieee.org
Pre-training serves as a foundation of recent NLP models, where language modeling tasks
are performed over large texts. Typical models like BERT and GPT take the corpus as a …

Mitigating frequency bias and anisotropy in language model pre-training with syntactic smoothing

RD Martinez, Z Goriely, A Caines, P Buttery… - arxiv preprint arxiv …, 2024 - arxiv.org
Language models strongly rely on frequency information because they maximize the
likelihood of tokens during pre-training. As a consequence, language models tend to not …

Chinese Cyberbullying Detection Using XLNet and Deep Bi-LSTM Hybrid Model

S Chen, J Wang, K He - Information, 2024 - mdpi.com
The popularization of the internet and the widespread use of smartphones have led to a
rapid growth in the number of social media users. While information technology has brought …

Two Heads are Better than One: Zero-shot Cognitive Reasoning via Multi-LLM Knowledge Fusion

L Liu, D Zhang, S Li, G Zhou, E Cambria - Proceedings of the 33rd ACM …, 2024 - dl.acm.org
Cognitive reasoning holds a significant place within Natural Language Processing (NLP).
Yet, the exploration of zero-shot scenarios, which align more closely with real-life situations …

Can llm substitute human labeling? a case study of fine-grained chinese address entity recognition dataset for uav delivery

Y Yao, S Luo, H Zhao, G Deng, L Song - … of the ACM on Web Conference …, 2024 - dl.acm.org
We present CNER-UAV, a fine-grained C hinese N ame E ntity R ecognition dataset
specifically designed for the task of address resolution in U nmanned A erial V ehicle …