Neural machine translation for low-resource languages: A survey
Neural Machine Translation (NMT) has seen tremendous growth in the last ten years since
the early 2000s and has already entered a mature phase. While considered the most widely …
the early 2000s and has already entered a mature phase. While considered the most widely …
[HTML][HTML] Progress in machine translation
After more than 70 years of evolution, great achievements have been made in machine
translation. Especially in recent years, translation quality has been greatly improved with the …
translation. Especially in recent years, translation quality has been greatly improved with the …
Finetuned language models are zero-shot learners
This paper explores a simple method for improving the zero-shot learning abilities of
language models. We show that instruction tuning--finetuning language models on a …
language models. We show that instruction tuning--finetuning language models on a …
[PDF][PDF] mt5: A massively multilingual pre-trained text-to-text transformer
L Xue - arxiv preprint arxiv:2010.11934, 2020 - fq.pkwyx.com
The recent" Text-to-Text Transfer Transformer"(T5) leveraged a unified text-to-text format and
scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this …
scale to attain state-of-the-art results on a wide variety of English-language NLP tasks. In this …
Exploring the limits of transfer learning with a unified text-to-text transformer
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-
tuned on a downstream task, has emerged as a powerful technique in natural language …
tuned on a downstream task, has emerged as a powerful technique in natural language …
[PDF][PDF] Unsupervised cross-lingual representation learning at scale
A Conneau - arxiv preprint arxiv:1911.02116, 2019 - fq.pkwyx.com
This paper shows that pretraining multilingual language models at scale leads to significant
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer …
performance gains for a wide range of cross-lingual transfer tasks. We train a Transformer …
[PDF][PDF] Multilingual denoising pre-training for neural machine translation
Y Liu - arxiv preprint arxiv:2001.08210, 2020 - fq.pkwyx.com
This paper demonstrates that multilingual denoising pre-training produces significant
performance gains across a wide variety of machine translation (MT) tasks. We present …
performance gains across a wide variety of machine translation (MT) tasks. We present …
Beyond english-centric multilingual machine translation
Existing work in translation demonstrated the potential of massively multilingual machine
translation by training a single model able to translate between any pair of languages …
translation by training a single model able to translate between any pair of languages …
Language-agnostic BERT sentence embedding
While BERT is an effective method for learning monolingual sentence embeddings for
semantic similarity and embedding based transfer learning (Reimers and Gurevych, 2019) …
semantic similarity and embedding based transfer learning (Reimers and Gurevych, 2019) …
Gpipe: Efficient training of giant neural networks using pipeline parallelism
Scaling up deep neural network capacity has been known as an effective approach to
improving model quality for several different machine learning tasks. In many cases …
improving model quality for several different machine learning tasks. In many cases …