Recent advances in natural language processing via large pre-trained language models: A survey

B Min, H Ross, E Sulem, APB Veyseh… - ACM Computing …, 2023 - dl.acm.org
Large, pre-trained language models (PLMs) such as BERT and GPT have drastically
changed the Natural Language Processing (NLP) field. For numerous NLP tasks …

Pre-trained language models for text generation: A survey

J Li, T Tang, WX Zhao, JY Nie, JR Wen - ACM Computing Surveys, 2024 - dl.acm.org
Text Generation aims to produce plausible and readable text in human language from input
data. The resurgence of deep learning has greatly advanced this field, in particular, with the …

Neural machine translation for low-resource languages: A survey

S Ranathunga, ESA Lee, M Prifti Skenduli… - ACM Computing …, 2023 - dl.acm.org
Neural Machine Translation (NMT) has seen tremendous growth in the last ten years since
the early 2000s and has already entered a mature phase. While considered the most widely …

Multilingual speech translation with efficient finetuning of pretrained models

X Li, C Wang, Y Tang, C Tran, Y Tang, J Pino… - arxiv preprint arxiv …, 2020 - arxiv.org
We present a simple yet effective approach to build multilingual speech-to-text (ST)
translation by efficient transfer learning from pretrained speech encoder and text decoder …

Lightweight adapter tuning for multilingual speech translation

H Le, J Pino, C Wang, J Gu, D Schwab… - arxiv preprint arxiv …, 2021 - arxiv.org
Adapter modules were recently introduced as an efficient alternative to fine-tuning in NLP.
Adapter tuning consists in freezing pretrained parameters of a model and injecting …

Efficient hierarchical domain adaptation for pretrained language models

A Chronopoulou, ME Peters, J Dodge - arxiv preprint arxiv:2112.08786, 2021 - arxiv.org
The remarkable success of large language models has been driven by dense models
trained on massive unlabeled, unstructured corpora. These corpora typically contain text …

Multilingual unsupervised neural machine translation with denoising adapters

A Üstün, A Berard, L Besacier, M Gallé - arxiv preprint arxiv:2110.10472, 2021 - arxiv.org
We consider the problem of multilingual unsupervised machine translation, translating to
and from languages that only have monolingual data by using auxiliary parallel language …

MSP: Multi-stage prompting for making pre-trained language models better translators

Z Tan, X Zhang, S Wang, Y Liu - arxiv preprint arxiv:2110.06609, 2021 - arxiv.org
Prompting has recently been shown as a promising approach for applying pre-trained
language models to perform downstream tasks. We present Multi-Stage Prompting (MSP), a …

Zero-shot cross-lingual transfer of neural machine translation with multilingual pretrained encoders

G Chen, S Ma, Y Chen, L Dong, D Zhang, J Pan… - arxiv preprint arxiv …, 2021 - arxiv.org
Previous work mainly focuses on improving cross-lingual transfer for NLU tasks with a
multilingual pretrained encoder (MPE), or improving the performance on supervised …

Language and task arithmetic with parameter-efficient layers for zero-shot summarization

A Chronopoulou, J Pfeiffer, J Maynez, X Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
Parameter-efficient fine-tuning (PEFT) using labeled task data can significantly improve the
performance of large language models (LLMs) on the downstream task. However, there are …