Exploring the landscape of machine unlearning: A comprehensive survey and taxonomy

T Shaik, X Tao, H **e, L Li, X Zhu… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Machine unlearning (MU) is gaining increasing attention due to the need to remove or
modify predictions made by machine learning (ML) models. While training models have …

Revisiting out-of-distribution robustness in nlp: Benchmarks, analysis, and llms evaluations

L Yuan, Y Chen, G Cui, H Gao, F Zou… - Advances in …, 2023 - proceedings.neurips.cc
This paper reexamines the research on out-of-distribution (OOD) robustness in the field of
NLP. We find that the distribution shift settings in previous studies commonly lack adequate …

Fine-tuning large neural language models for biomedical natural language processing

R Tinn, H Cheng, Y Gu, N Usuyama, X Liu, T Naumann… - Patterns, 2023 - cell.com
Large neural language models have transformed modern natural language processing
(NLP) applications. However, fine-tuning such models for specific tasks remains challenging …

State-of-the-art generalisation research in NLP: a taxonomy and review

D Hupkes, M Giulianelli, V Dankers, M Artetxe… - arxiv preprint arxiv …, 2022 - arxiv.org
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …

Human parity on commonsenseqa: Augmenting self-attention with external attention

Y Xu, C Zhu, S Wang, S Sun, H Cheng, X Liu… - arxiv preprint arxiv …, 2021 - arxiv.org
Most of today's AI systems focus on using self-attention mechanisms and transformer
architectures on large amounts of diverse data to achieve impressive performance gains. In …

Metro: Efficient denoising pretraining of large scale autoencoding language models with model generated signals

P Bajaj, C **ong, G Ke, X Liu, D He, S Tiwary… - arxiv preprint arxiv …, 2022 - arxiv.org
We present an efficient method of pretraining large-scale autoencoding language models
using training signals generated by an auxiliary model. Originated in ELECTRA, this training …

UnitedQA: A hybrid approach for open domain question answering

H Cheng, Y Shen, X Liu, P He, W Chen… - arxiv preprint arxiv …, 2021 - arxiv.org
To date, most of recent work under the retrieval-reader framework for open-domain QA
focuses on either extractive or generative reader exclusively. In this paper, we study a hybrid …

Enhancing machine-generated text detection: adversarial fine-tuning of pre-trained language models

DH Lee, B Jang - IEEE Access, 2024 - ieeexplore.ieee.org
Advances in large language models (LLMs) have revolutionized the natural language
processing field. However, the text generated by LLMs can result in various issues, such as …

DIALKI: Knowledge identification in conversational systems through dialogue-document contextualization

Z Wu, BR Lu, H Hajishirzi, M Ostendorf - arxiv preprint arxiv:2109.04673, 2021 - arxiv.org
Identifying relevant knowledge to be used in conversational systems that are grounded in
long documents is critical to effective response generation. We introduce a knowledge …

Reeval: Automatic hallucination evaluation for retrieval-augmented large language models via transferable adversarial attacks

X Yu, H Cheng, X Liu, D Roth, J Gao - arxiv preprint arxiv:2310.12516, 2023 - arxiv.org
Despite remarkable advancements in mitigating hallucinations in large language models
(LLMs) by retrieval augmentation, it remains challenging to measure the reliability of LLMs …