Continual lifelong learning in natural language processing: A survey

M Biesialska, K Biesialska, MR Costa-Jussa - arxiv preprint arxiv …, 2020 - arxiv.org
Continual learning (CL) aims to enable information systems to learn from a continuous data
stream across time. However, it is difficult for existing deep learning architectures to learn a …

A survey on large language models with multilingualism: Recent advances and new frontiers

K Huang, F Mo, X Zhang, H Li, Y Li, Y Zhang… - arxiv preprint arxiv …, 2024 - arxiv.org
The rapid development of Large Language Models (LLMs) demonstrates remarkable
multilingual capabilities in natural language processing, attracting global attention in both …

Parameter-efficient transfer learning with diff pruning

D Guo, AM Rush, Y Kim - arxiv preprint arxiv:2012.07463, 2020 - arxiv.org
While task-specific finetuning of pretrained networks has led to significant empirical
advances in NLP, the large size of networks makes finetuning difficult to deploy in multi-task …

Simple, scalable adaptation for neural machine translation

A Bapna, N Arivazhagan, O Firat - arxiv preprint arxiv:1909.08478, 2019 - arxiv.org
Fine-tuning pre-trained Neural Machine Translation (NMT) models is the dominant approach
for adapting to new languages and domains. However, fine-tuning requires adapting and …

Recall and learn: Fine-tuning deep pretrained language models with less forgetting

S Chen, Y Hou, Y Cui, W Che, T Liu, X Yu - arxiv preprint arxiv …, 2020 - arxiv.org
Deep pretrained language models have achieved great success in the way of pretraining
first and then fine-tuning. But such a sequential transfer learning paradigm often confronts …

Explicit inductive bias for transfer learning with convolutional networks

LI Xuhong, Y Grandvalet… - … conference on machine …, 2018 - proceedings.mlr.press
In inductive transfer learning, fine-tuning pre-trained convolutional networks substantially
outperforms training from scratch. When using fine-tuning, the underlying assumption is that …

A survey of domain adaptation for machine translation

C Chu, R Wang - Journal of information processing, 2020 - jstage.jst.go.jp
Neural machine translation (NMT) is a deep learning based approach for machine
translation, which outperforms traditional statistical machine translation (SMT) and yields the …

Neural grammatical error correction systems with unsupervised pre-training on synthetic data

R Grundkiewicz, M Junczys-Dowmuntz… - 14th Workshop on …, 2019 - research.ed.ac.uk
Considerable effort has been made to address the data sparsity problem in neural
grammatical error correction. In this work, we propose a simple and surprisingly effective …

A constructive prediction of the generalization error across scales

JS Rosenfeld, A Rosenfeld, Y Belinkov… - arxiv preprint arxiv …, 2019 - arxiv.org
The dependency of the generalization error of neural networks on model and dataset size is
of critical importance both in practice and for understanding the theory of neural networks …

Domain adaptation and multi-domain adaptation for neural machine translation: A survey

D Saunders - Journal of Artificial Intelligence Research, 2022 - jair.org
The development of deep learning techniques has allowed Neural Machine Translation
(NMT) models to become extremely powerful, given sufficient training data and training time …