Understanding and mitigating language confusion in llms

K Marchisio, WY Ko, A Bérard, T Dehaze… - arxiv preprint arxiv …, 2024 - arxiv.org
We investigate a surprising limitation of LLMs: their inability to consistently generate text in a
user's desired language. We create the Language Confusion Benchmark (LCB) to evaluate …

Multilingual large language model: A survey of resources, taxonomy and frontiers

L Qin, Q Chen, Y Zhou, Z Chen, Y Li, L Liao… - arxiv preprint arxiv …, 2024 - arxiv.org
Multilingual Large Language Models are capable of using powerful Large Language
Models to handle and respond to queries in multiple languages, which achieves remarkable …

Analyzing and Adapting Large Language Models for Few-Shot Multilingual NLU: Are We There Yet?

E Razumovskaia, I Vulić, A Korhonen - arxiv preprint arxiv:2403.01929, 2024 - arxiv.org
Supervised fine-tuning (SFT), supervised instruction tuning (SIT) and in-context learning
(ICL) are three alternative, de facto standard approaches to few-shot learning. ICL has …

Large Language Model Instruction Following: A Survey of Progresses and Challenges

R Lou, K Zhang, W Yin - Computational Linguistics, 2024 - direct.mit.edu
Task semantics can be expressed by a set of input-output examples or a piece of textual
instruction. Conventional machine learning approaches for natural language processing …

Self-Alignment: Improving Alignment of Cultural Values in LLMs via In-Context Learning

R Choenni, E Shutova - arxiv preprint arxiv:2408.16482, 2024 - arxiv.org
Improving the alignment of Large Language Models (LLMs) with respect to the cultural
values that they encode has become an increasingly important topic. In this work, we study …

How Reliable Are Automatic Evaluation Methods for Instruction-Tuned LLMs?

E Doostmohammadi, O Holmström… - arxiv preprint arxiv …, 2024 - arxiv.org
Work on instruction-tuned Large Language Models (LLMs) has used automatic methods
based on text overlap and LLM judgments as cost-effective alternatives to human …

Quality or quantity? On data scale and diversity in adapting large language models for low-resource translation

V Iyer, B Malik, P Stepachev, P Chen… - arxiv preprint arxiv …, 2024 - arxiv.org
Despite the recent popularity of Large Language Models (LLMs) in Machine Translation
(MT), their performance in low-resource languages (LRLs) still lags significantly behind …

Getting More from Less: Large Language Models are Good Spontaneous Multilingual Learners

S Zhang, C Gao, W Zhu, J Chen, X Huang… - Proceedings of the …, 2024 - aclanthology.org
Abstract Recently, Large Language Models (LLMs) have shown impressive language
capabilities, while most of them have very unbalanced performance across different …

Analysis of Multi-Source Language Training in Cross-Lingual Transfer

SH Lim, T Yun, J Kim, J Choi, T Kim - arxiv preprint arxiv:2402.13562, 2024 - arxiv.org
The successful adaptation of multilingual language models (LMs) to a specific language-
task pair critically depends on the availability of data tailored for that condition. While cross …

A survey of multilingual large language models

L Qin, Q Chen, Y Zhou, Z Chen, Y Li, L Liao, M Li… - Patterns, 2025 - cell.com
Multilingual large language models (MLLMs) leverage advanced large language models to
process and respond to queries across multiple languages, achieving significant success in …