Continual lifelong learning in natural language processing: A survey

M Biesialska, K Biesialska, MR Costa-Jussa - arxiv preprint arxiv …, 2020 - arxiv.org
Continual learning (CL) aims to enable information systems to learn from a continuous data
stream across time. However, it is difficult for existing deep learning architectures to learn a …

Efficient methods for natural language processing: A survey

M Treviso, JU Lee, T Ji, B Aken, Q Cao… - Transactions of the …, 2023 - direct.mit.edu
Recent work in natural language processing (NLP) has yielded appealing results from
scaling model parameters and training data; however, using only scale to improve …

A survey of active learning for natural language processing

Z Zhang, E Strubell, E Hovy - arxiv preprint arxiv:2210.10109, 2022 - arxiv.org
In this work, we provide a survey of active learning (AL) for its applications in natural
language processing (NLP). In addition to a fine-grained categorization of query strategies …

Cold-start active learning through self-supervised language modeling

M Yuan, HT Lin, J Boyd-Graber - arxiv preprint arxiv:2010.09535, 2020 - arxiv.org
Active learning strives to reduce annotation costs by choosing the most critical examples to
label. Typically, the active learning strategy is contingent on the classification model. For …

Reinforcement learning based curriculum optimization for neural machine translation

G Kumar, G Foster, C Cherry, M Krikun - arxiv preprint arxiv:1903.00041, 2019 - arxiv.org
We consider the problem of making efficient use of heterogeneous training data in neural
machine translation (NMT). Specifically, given a training dataset with a sentence-level …

Graph policy network for transferable active learning on graphs

S Hu, Z **ong, M Qu, X Yuan… - Advances in Neural …, 2020 - proceedings.neurips.cc
Graph neural networks (GNNs) have been attracting increasing popularity due to their
simplicity and effectiveness in a variety of fields. However, a large number of labeled data is …

Active learning for abstractive text summarization

A Tsvigun, I Lysenko, D Sedashov, I Lazichny… - arxiv preprint arxiv …, 2023 - arxiv.org
Construction of human-curated annotated datasets for abstractive text summarization (ATS)
is very time-consuming and expensive because creating each instance requires a human …

Active learning approaches to enhancing neural machine translation

Y Zhao, RH Zhang, S Zhou, Z Zhang - Findings of the Association …, 2020 - aclanthology.org
Active learning is an efficient approach for mitigating data dependency when training neural
machine translation (NMT) models. In this paper, we explore new training frameworks by …

Reinforced curriculum learning on pre-trained neural machine translation models

M Zhao, H Wu, D Niu, X Wang - Proceedings of the AAAI Conference on …, 2020 - aaai.org
The competitive performance of neural machine translation (NMT) critically relies on large
amounts of training data. However, acquiring high-quality translation pairs requires expert …

Personalized estimation of engagement from videos using active learning with deep reinforcement learning

O Rudovic, HW Park, J Busche… - 2019 IEEE/CVF …, 2019 - ieeexplore.ieee.org
Perceiving users' engagement accurately is important for technologies that need to respond
to learners in a natural and intelligent way. In this paper, we address the problem of …