Language model tokenizers introduce unfairness between languages
Recent language models have shown impressive multilingual performance, even when not
explicitly trained for it. Despite this, there are concerns about the quality of their outputs …
explicitly trained for it. Despite this, there are concerns about the quality of their outputs …
Welm: A well-read pre-trained language model for chinese
Large Language Models pre-trained with self-supervised learning have demonstrated
impressive zero-shot generalization capabilities on a wide spectrum of tasks. In this work …
impressive zero-shot generalization capabilities on a wide spectrum of tasks. In this work …
Internlm-law: An open source chinese legal large language model
While large language models (LLMs) have showcased impressive capabilities, they struggle
with addressing legal queries due to the intricate complexities and specialized expertise …
with addressing legal queries due to the intricate complexities and specialized expertise …
Cbas: Character-level backdoor attacks against chinese pre-trained language models
X He, F Hao, T Gu, L Chang - ACM Transactions on Privacy and Security, 2024 - dl.acm.org
Pre-trained language models (PLMs) aim to assist computers in various domains to provide
natural and efficient language interaction and text processing capabilities. However, recent …
natural and efficient language interaction and text processing capabilities. However, recent …
Comparing explanation faithfulness between multilingual and monolingual fine-tuned language models
In many real natural language processing application scenarios, practitioners not only aim to
maximize predictive performance but also seek faithful explanations for the model …
maximize predictive performance but also seek faithful explanations for the model …
Enhancing pre-trained language models with Chinese character morphological knowledge
Z Zheng, X Wu, X Liu - Information Processing & Management, 2025 - Elsevier
Pre-trained language models (PLMs) have demonstrated success in Chinese natural
language processing (NLP) tasks by acquiring high-quality representations through …
language processing (NLP) tasks by acquiring high-quality representations through …
A comprehensive evaluation of parameter-efficient fine-tuning on software engineering tasks
Pre-trained models (PTMs) have achieved great success in various Software Engineering
(SE) downstream tasks following the``pre-train then fine-tune''paradigm. As fully fine-tuning …
(SE) downstream tasks following the``pre-train then fine-tune''paradigm. As fully fine-tuning …
Self-training improves few-shot learning in legal artificial intelligence tasks
Y Zhou, Y Qin, R Huang, Y Chen, C Lin… - Artificial Intelligence and …, 2024 - Springer
As the labeling costs in legal artificial intelligence tasks are expensive. Therefore, it
becomes a challenge to utilize low cost to train a robust model. In this paper, we propose a …
becomes a challenge to utilize low cost to train a robust model. In this paper, we propose a …
Enhance robustness of language models against variation attack through graph integration
The widespread use of pre-trained language models (PLMs) in natural language processing
(NLP) has greatly improved performance outcomes. However, these models' vulnerability to …
(NLP) has greatly improved performance outcomes. However, these models' vulnerability to …
PROTECT: Parameter-Efficient Tuning for Few-Shot Robust Chinese Text Correction
Non-normative texts and euphemisms are widely spread on the Internet, making it more
difficult to moderate the content. These phenomena result from misspelling errors or …
difficult to moderate the content. These phenomena result from misspelling errors or …