Neural machine translation: A review
F Stahlberg - Journal of Artificial Intelligence Research, 2020 - jair.org
The field of machine translation (MT), the automatic translation of written text from one
natural language into another, has experienced a major paradigm shift in recent years …
natural language into another, has experienced a major paradigm shift in recent years …
[PDF][PDF] Language models are unsupervised multitask learners
Natural language processing tasks, such as question answering, machine translation,
reading comprehension, and summarization, are typically approached with supervised …
reading comprehension, and summarization, are typically approached with supervised …
Adversarial NLI: A new benchmark for natural language understanding
We introduce a new large-scale NLI benchmark dataset, collected via an iterative,
adversarial human-and-model-in-the-loop procedure. We show that training models on this …
adversarial human-and-model-in-the-loop procedure. We show that training models on this …
Masked language modeling and the distributional hypothesis: Order word matters pre-training for little
A possible explanation for the impressive performance of masked language model (MLM)
pre-training is that such models have learned to represent the syntactic structures prevalent …
pre-training is that such models have learned to represent the syntactic structures prevalent …
Sparse, dense, and attentional representations for text retrieval
Dual encoders perform retrieval by encoding documents and queries into dense low-
dimensional vectors, scoring each document by its inner product with the query. We …
dimensional vectors, scoring each document by its inner product with the query. We …
Information-theoretic probing with minimum description length
To measure how well pretrained representations encode some linguistic property, it is
common to use accuracy of a probe, ie a classifier trained to predict the property from the …
common to use accuracy of a probe, ie a classifier trained to predict the property from the …
Backdoor learning for nlp: Recent advances, challenges, and future research directions
M Omar - arxiv preprint arxiv:2302.06801, 2023 - arxiv.org
Although backdoor learning is an active research topic in the NLP domain, the literature
lacks studies that systematically categorize and summarize backdoor attacks and defenses …
lacks studies that systematically categorize and summarize backdoor attacks and defenses …
Pre-training via paraphrasing
We introduce MARGE, a pre-trained sequence-to-sequence model learned with an
unsupervised multi-lingual multi-document paraphrasing objective. MARGE provides an …
unsupervised multi-lingual multi-document paraphrasing objective. MARGE provides an …
Probing the probing paradigm: Does probing accuracy entail task relevance?
Although neural models have achieved impressive results on several NLP benchmarks, little
is understood about the mechanisms they use to perform language tasks. Thus, much recent …
is understood about the mechanisms they use to perform language tasks. Thus, much recent …
Do attention heads in BERT track syntactic dependencies?
We investigate the extent to which individual attention heads in pretrained transformer
language models, such as BERT and RoBERTa, implicitly capture syntactic dependency …
language models, such as BERT and RoBERTa, implicitly capture syntactic dependency …