Measure and improve robustness in NLP models: A survey
As NLP models achieved state-of-the-art performances over benchmarks and gained wide
applications, it has been increasingly important to ensure the safe deployment of these …
applications, it has been increasingly important to ensure the safe deployment of these …
Zero-shot dense retrieval with momentum adversarial domain invariant representations
Dense retrieval (DR) methods conduct text retrieval by first encoding texts in the embedding
space and then matching them by nearest neighbor search. This requires strong locality …
space and then matching them by nearest neighbor search. This requires strong locality …
Out-of-distribution generalization in natural language processing: Past, present, and future
Abstract Machine learning (ML) systems in natural language processing (NLP) face
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …
Prompt-based distribution alignment for domain generalization in text classification
C Jia, Y Zhang - Proceedings of the 2022 Conference on …, 2022 - aclanthology.org
Prompt-based learning (aka prompting) achieves high performance by bridging the gap
between the objectives of language modeling and downstream tasks. Domain …
between the objectives of language modeling and downstream tasks. Domain …
Invariant language modeling
Large pretrained language models are critical components of modern NLP pipelines. Yet,
they suffer from spurious correlations, poor out-of-domain generalization, and biases …
they suffer from spurious correlations, poor out-of-domain generalization, and biases …
Learning list-level domain-invariant representations for ranking
Abstract Domain adaptation aims to transfer the knowledge learned on (data-rich) source
domains to (low-resource) target domains, and a popular method is invariant representation …
domains to (low-resource) target domains, and a popular method is invariant representation …
To tune or not to tune? zero-shot models for legal case entailment
There has been mounting evidence that pretrained language models fine-tuned on large
and diverse supervised datasets can transfer well to a variety of out-of-domain tasks. In this …
and diverse supervised datasets can transfer well to a variety of out-of-domain tasks. In this …
BERTAC: Enhancing transformer-based language models with adversarially pretrained convolutional neural networks
Transformer-based language models (TLMs), such as BERT, ALBERT and GPT-3, have
shown strong performance in a wide range of NLP tasks and currently dominate the field of …
shown strong performance in a wide range of NLP tasks and currently dominate the field of …
Adversarial weakly supervised domain adaptation for few shot sentiment analysis
The ability of deep neural networks to generate state-of-the-art results on many NLP
problems has been apparent to everyone for some years now. However, when there is not …
problems has been apparent to everyone for some years now. However, when there is not …
DELIA: Diversity-Enhanced Learning for Instruction Adaptation in Large Language Models
Y Zeng, F Ren, X Zhou, Y Wang, Y Shao - arxiv preprint arxiv:2408.10841, 2024 - arxiv.org
Although instruction tuning is widely used to adjust behavior in Large Language Models
(LLMs), extensive empirical evidence and research indicates that it is primarily a process …
(LLMs), extensive empirical evidence and research indicates that it is primarily a process …