Measure and improve robustness in NLP models: A survey

X Wang, H Wang, D Yang - arxiv preprint arxiv:2112.08313, 2021 - arxiv.org
As NLP models achieved state-of-the-art performances over benchmarks and gained wide
applications, it has been increasingly important to ensure the safe deployment of these …

Zero-shot dense retrieval with momentum adversarial domain invariant representations

J **n, C **ong, A Srinivasan, A Sharma, D Jose… - arxiv preprint arxiv …, 2021 - arxiv.org
Dense retrieval (DR) methods conduct text retrieval by first encoding texts in the embedding
space and then matching them by nearest neighbor search. This requires strong locality …

Out-of-distribution generalization in natural language processing: Past, present, and future

L Yang, Y Song, X Ren, C Lyu, Y Wang… - Proceedings of the …, 2023 - aclanthology.org
Abstract Machine learning (ML) systems in natural language processing (NLP) face
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …

Prompt-based distribution alignment for domain generalization in text classification

C Jia, Y Zhang - Proceedings of the 2022 Conference on …, 2022 - aclanthology.org
Prompt-based learning (aka prompting) achieves high performance by bridging the gap
between the objectives of language modeling and downstream tasks. Domain …

Invariant language modeling

M Peyrard, SS Ghotra, M Josifoski, V Agarwal… - arxiv preprint arxiv …, 2021 - arxiv.org
Large pretrained language models are critical components of modern NLP pipelines. Yet,
they suffer from spurious correlations, poor out-of-domain generalization, and biases …

Learning list-level domain-invariant representations for ranking

R **an, H Zhuang, Z Qin, H Zamani… - Advances in …, 2023 - proceedings.neurips.cc
Abstract Domain adaptation aims to transfer the knowledge learned on (data-rich) source
domains to (low-resource) target domains, and a popular method is invariant representation …

To tune or not to tune? zero-shot models for legal case entailment

GM Rosa, RC Rodrigues, R de Alencar Lotufo… - Proceedings of the …, 2021 - dl.acm.org
There has been mounting evidence that pretrained language models fine-tuned on large
and diverse supervised datasets can transfer well to a variety of out-of-domain tasks. In this …

BERTAC: Enhancing transformer-based language models with adversarially pretrained convolutional neural networks

JH Oh, R Iida, J Kloetzer… - Proceedings of the 59th …, 2021 - aclanthology.org
Transformer-based language models (TLMs), such as BERT, ALBERT and GPT-3, have
shown strong performance in a wide range of NLP tasks and currently dominate the field of …

Adversarial weakly supervised domain adaptation for few shot sentiment analysis

SE Taher, M Shamsfard - 2021 7th International Conference on …, 2021 - ieeexplore.ieee.org
The ability of deep neural networks to generate state-of-the-art results on many NLP
problems has been apparent to everyone for some years now. However, when there is not …

DELIA: Diversity-Enhanced Learning for Instruction Adaptation in Large Language Models

Y Zeng, F Ren, X Zhou, Y Wang, Y Shao - arxiv preprint arxiv:2408.10841, 2024 - arxiv.org
Although instruction tuning is widely used to adjust behavior in Large Language Models
(LLMs), extensive empirical evidence and research indicates that it is primarily a process …