Textbugger: Generating adversarial text against real-world applications

J Li, S Ji, T Du, B Li, T Wang - arxiv preprint arxiv:1812.05271, 2018 - arxiv.org
Deep Learning-based Text Understanding (DLTU) is the backbone technique behind
various applications, including question answering, machine translation, and text …

Hotflip: White-box adversarial examples for text classification

J Ebrahimi, A Rao, D Lowd, D Dou - arxiv preprint arxiv:1712.06751, 2017 - arxiv.org
We propose an efficient method to generate white-box adversarial examples to trick a
character-level neural classifier. We find that only a few manipulations are needed to greatly …

Learning what makes a difference from counterfactual examples and gradient supervision

D Teney, E Abbasnedjad, A van den Hengel - Computer Vision–ECCV …, 2020 - Springer
One of the primary challenges limiting the applicability of deep learning is its susceptibility to
learning spurious correlations rather than the underlying mechanisms of the task of interest …

How robust are character-based word embeddings in tagging and MT against wrod scramlbing or randdm nouse?

G Heigold, G Neumann, J van Genabith - arxiv preprint arxiv:1704.04441, 2017 - arxiv.org
This paper investigates the robustness of NLP against perturbed word forms. While neural
approaches can achieve (almost) human-like accuracy for certain tasks and conditions, they …

Robust training under linguistic adversity

Y Li, T Cohn, T Baldwin - Proceedings of the 15th Conference of …, 2017 - aclanthology.org
Deep neural networks have achieved remarkable results across many language processing
tasks, however they have been shown to be susceptible to overfitting and highly sensitive to …

Posterior differential regularization with f-divergence for improving model robustness

H Cheng, X Liu, L Pereira, Y Yu, J Gao - arxiv preprint arxiv:2010.12638, 2020 - arxiv.org
We address the problem of enhancing model robustness through regularization.
Specifically, we focus on methods that regularize the model posterior difference between …

Detecting textual adversarial examples based on distributional characteristics of data representations

N Liu, M Dras, WE Zhang - arxiv preprint arxiv:2204.13853, 2022 - arxiv.org
Although deep neural networks have achieved state-of-the-art performance in various
machine learning tasks, adversarial examples, constructed by adding small non-random …

Identifying adversarial attacks on text classifiers

Z **e, J Brophy, A Noack, W You, K Asthana… - arxiv preprint arxiv …, 2022 - arxiv.org
The landscape of adversarial attacks against text classifiers continues to grow, with new
attacks developed every year and many of them available in standard toolkits, such as …

Misspellings in Natural Language Processing: A survey

G Sperduti, A Moreo - arxiv preprint arxiv:2501.16836, 2025 - arxiv.org
This survey provides an overview of the challenges of misspellings in natural language
processing (NLP). While often unintentional, misspellings have become ubiquitous in digital …

What Learned Representations and Influence Functions Can Tell Us About Adversarial Examples

SM Tonni, M Dras - arxiv preprint arxiv:2309.10916, 2023 - arxiv.org
Adversarial examples, deliberately crafted using small perturbations to fool deep neural
networks, were first studied in image processing and more recently in NLP. While …