Analysis methods in neural language processing: A survey

Y Belinkov, J Glass - … of the Association for Computational Linguistics, 2019 - direct.mit.edu
The field of natural language processing has seen impressive progress in recent years, with
neural network models replacing many of the traditional systems. A plethora of new models …

PAWS: Paraphrase adversaries from word scrambling

Y Zhang, J Baldridge, L He - arxiv preprint arxiv:1904.01130, 2019 - arxiv.org
Existing paraphrase identification datasets lack sentence pairs that have high lexical overlap
without being paraphrases. Models trained on such data fail to distinguish pairs like flights …

Logic-guided data augmentation and regularization for consistent question answering

A Asai, H Hajishirzi - arxiv preprint arxiv:2004.10157, 2020 - arxiv.org
Many natural language questions require qualitative, quantitative or logical comparisons
between two entities or events. This paper addresses the problem of improving the accuracy …

We need to talk about standard splits

K Gorman, S Bedrick - Proceedings of the conference …, 2019 - pmc.ncbi.nlm.nih.gov
It is standard practice in speech & language technology to rank systems according to
performance on a test set held out for evaluation. However, few researchers apply statistical …

The language model understood the prompt was ambiguous: Probing syntactic uncertainty through generation

L Aina, T Linzen - arxiv preprint arxiv:2109.07848, 2021 - arxiv.org
Temporary syntactic ambiguities arise when the beginning of a sentence is compatible with
multiple syntactic analyses. We inspect to which extent neural language models (LMs) …

[HTML][HTML] Context-Aware Embedding Techniques for Addressing Meaning Conflation Deficiency in Morphologically Rich Languages Word Embedding: A Systematic …

MA Masethe, HD Masethe, SO Ojo - Computers, 2024 - mdpi.com
This systematic literature review aims to evaluate and synthesize the effectiveness of various
embedding techniques—word embeddings, contextual word embeddings, and context …

Does data augmentation improve generalization in NLP?

R Jha, C Lovering, E Pavlick - arxiv preprint arxiv:2004.15012, 2020 - arxiv.org
Neural models often exploit superficial features to achieve good performance, rather than
deriving more general features. Overcoming this tendency is a central challenge in areas …

Specification overfitting in artificial intelligence

B Roth, PH Luz de Araujo, Y **a… - Artificial Intelligence …, 2025 - Springer
Abstract Machine learning (ML) and artificial intelligence (AI) approaches are often criticized
for their inherent bias and for their lack of control, accountability, and transparency …

Do Pretrained Contextual Language Models Distinguish between Hebrew Homograph Analyses?

A Shmidman, CS Shmidman, D Bareket… - arxiv preprint arxiv …, 2024 - arxiv.org
Semitic morphologically-rich languages (MRLs) are characterized by extreme word
ambiguity. Because most vowels are omitted in standard texts, many of the words are …

[PDF][PDF] When does data augmentation help generalization in nlp

R Jha, C Lovering, E Pavlick - arxiv preprint arxiv:2004.15012, 2020 - ask.qcloudimg.com
Neural models often exploit superficial (“weak”) features to achieve good performance,
rather than deriving the more general (“strong”) features that we'd prefer a model to use …