Prompting palm for translation: Assessing strategies and performance

D Vilar, M Freitag, C Cherry, J Luo, V Ratnakar… - arxiv preprint arxiv …, 2022 - arxiv.org
Large language models (LLMs) that have been trained on multilingual but not parallel text
exhibit a remarkable ability to translate between languages. We probe this ability in an in …

Task contamination: Language models may not be few-shot anymore

C Li, J Flanigan - Proceedings of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
Large language models (LLMs) offer impressive performance in various zero-shot and few-
shot tasks. However, their success in zero-shot or few-shot settings may be affected by task …

Large language models effectively leverage document-level context for literary translation, but critical errors persist

M Karpinska, M Iyyer - arxiv preprint arxiv:2304.03245, 2023 - arxiv.org
Large language models (LLMs) are competitive with the state of the art on a wide range of
sentence-level translation datasets. However, their ability to translate paragraphs and …

Aya model: An instruction finetuned open-access multilingual language model

A Üstün, V Aryabumi, ZX Yong, WY Ko… - arxiv preprint arxiv …, 2024 - arxiv.org
Recent breakthroughs in large language models (LLMs) have centered around a handful of
data-rich languages. What does it take to broaden access to breakthroughs beyond first …

Monolingual or multilingual instruction tuning: Which makes a better alpaca

P Chen, S Ji, N Bogoychev, A Kutuzov… - arxiv preprint arxiv …, 2023 - arxiv.org
Foundational large language models (LLMs) can be instruction-tuned to perform open-
domain question answering, facilitating applications like chat assistants. While such efforts …

Multilingual large language model: A survey of resources, taxonomy and frontiers

L Qin, Q Chen, Y Zhou, Z Chen, Y Li, L Liao… - arxiv preprint arxiv …, 2024 - arxiv.org
Multilingual Large Language Models are capable of using powerful Large Language
Models to handle and respond to queries in multiple languages, which achieves remarkable …

Machine translation with large language models: Prompting, few-shot learning, and fine-tuning with QLoRA

X Zhang, N Rajabi, K Duh, P Koehn - Proceedings of the Eighth …, 2023 - aclanthology.org
While large language models have made remarkable advancements in natural language
generation, their potential in machine translation, especially when fine-tuned, remains under …

Analyzing and Adapting Large Language Models for Few-Shot Multilingual NLU: Are We There Yet?

E Razumovskaia, I Vulić, A Korhonen - arxiv preprint arxiv:2403.01929, 2024 - arxiv.org
Supervised fine-tuning (SFT), supervised instruction tuning (SIT) and in-context learning
(ICL) are three alternative, de facto standard approaches to few-shot learning. ICL has …

Recent Advances in Interactive Machine Translation with Large Language Models

Y Wang, J Zhang, T Shi, D Deng, Y Tian… - IEEE …, 2024 - ieeexplore.ieee.org
This paper explores the role of Large Language Models (LLMs) in revolutionizing interactive
Machine Translation (MT), providing a comprehensive analysis across nine innovative …

Do gpts produce less literal translations?

V Raunak, A Menezes, M Post, HH Awadalla - arxiv preprint arxiv …, 2023 - arxiv.org
Large Language Models (LLMs) such as GPT-3 have emerged as general-purpose
language models capable of addressing many natural language generation or …