A survey of adversarial defenses and robustness in nlp
In the past few years, it has become increasingly evident that deep neural networks are not
resilient enough to withstand adversarial perturbations in input data, leaving them …
resilient enough to withstand adversarial perturbations in input data, leaving them …
Robust natural language processing: Recent advances, challenges, and future directions
Recent natural language processing (NLP) techniques have accomplished high
performance on benchmark data sets, primarily due to the significant improvement in the …
performance on benchmark data sets, primarily due to the significant improvement in the …
On evaluating adversarial robustness of large vision-language models
Large vision-language models (VLMs) such as GPT-4 have achieved unprecedented
performance in response generation, especially with visual inputs, enabling more creative …
performance in response generation, especially with visual inputs, enabling more creative …
Threats, attacks and defenses to federated learning: issues, taxonomy and perspectives
Abstract Empirical attacks on Federated Learning (FL) systems indicate that FL is fraught
with numerous attack surfaces throughout the FL execution. These attacks can not only …
with numerous attack surfaces throughout the FL execution. These attacks can not only …
Adversarial attacks on deep-learning models in natural language processing: A survey
With the development of high computational devices, deep neural networks (DNNs), in
recent years, have gained significant popularity in many Artificial Intelligence (AI) …
recent years, have gained significant popularity in many Artificial Intelligence (AI) …
Polyjuice: Generating counterfactuals for explaining, evaluating, and improving models
While counterfactual examples are useful for analysis and training of NLP models, current
generation methods either rely on manual labor to create very few counterfactuals, or only …
generation methods either rely on manual labor to create very few counterfactuals, or only …
Word-level textual adversarial attacking as combinatorial optimization
Adversarial attacks are carried out to reveal the vulnerability of deep neural networks.
Textual adversarial attacking is challenging because text is discrete and a small perturbation …
Textual adversarial attacking is challenging because text is discrete and a small perturbation …
Contextualized perturbation for textual adversarial attack
Adversarial examples expose the vulnerabilities of natural language processing (NLP)
models, and can be used to evaluate and improve their robustness. Existing techniques of …
models, and can be used to evaluate and improve their robustness. Existing techniques of …
Turn the combination lock: Learnable textual backdoor attacks via word substitution
Recent studies show that neural natural language processing (NLP) models are vulnerable
to backdoor attacks. Injected with backdoors, models perform normally on benign examples …
to backdoor attacks. Injected with backdoors, models perform normally on benign examples …
Explaining NLP models via minimal contrastive editing (MiCE)
Humans have been shown to give contrastive explanations, which explain why an observed
event happened rather than some other counterfactual event (the contrast case). Despite the …
event happened rather than some other counterfactual event (the contrast case). Despite the …