A survey of adversarial defenses and robustness in nlp

S Goyal, S Doddapaneni, MM Khapra… - ACM Computing …, 2023 - dl.acm.org
In the past few years, it has become increasingly evident that deep neural networks are not
resilient enough to withstand adversarial perturbations in input data, leaving them …

[HTML][HTML] Large language models in radiology: fundamentals, applications, ethical considerations, risks, and future directions

TA D'Antonoli, A Stanzione, C Bluethgen… - Diagnostic and …, 2024 - ncbi.nlm.nih.gov
With the advent of large language models (LLMs), the artificial intelligence revolution in
medicine and radiology is now more tangible than ever. Every day, an increasingly large …

Large language models can be strong differentially private learners

X Li, F Tramer, P Liang, T Hashimoto - arxiv preprint arxiv:2110.05679, 2021 - arxiv.org
Differentially Private (DP) learning has seen limited success for building large deep learning
models of text, and straightforward attempts at applying Differentially Private Stochastic …

Five sources of bias in natural language processing

D Hovy, S Prabhumoye - Language and linguistics compass, 2021 - Wiley Online Library
Recently, there has been an increased interest in demographically grounded bias in natural
language processing (NLP) applications. Much of the recent work has focused on describing …

Null it out: Guarding protected attributes by iterative nullspace projection

S Ravfogel, Y Elazar, H Gonen, M Twiton… - arxiv preprint arxiv …, 2020 - arxiv.org
The ability to control for the kinds of information encoded in neural representation has a
variety of use cases, especially in light of the challenge of interpreting these models. We …

Pile of law: Learning responsible data filtering from the law and a 256gb open-source legal dataset

P Henderson, M Krass, L Zheng… - Advances in …, 2022 - proceedings.neurips.cc
One concern with the rise of large language models lies with their potential for significant
harm, particularly from pretraining on biased, obscene, copyrighted, and private information …

Predictive biases in natural language processing models: A conceptual framework and overview

D Shah, HA Schwartz, D Hovy - arxiv preprint arxiv:1912.11078, 2019 - arxiv.org
An increasing number of works in natural language processing have addressed the effect of
bias on the predicted outcomes, introducing mitigation techniques that act on different parts …

Information leakage in embedding models

C Song, A Raghunathan - Proceedings of the 2020 ACM SIGSAC …, 2020 - dl.acm.org
Embeddings are functions that map raw input data to low-dimensional vector
representations, while preserving important semantic information about the inputs. Pre …

Privacy-preserving prompt tuning for large language model services

Y Li, Z Tan, Y Liu - arxiv preprint arxiv:2305.06212, 2023 - arxiv.org
Prompt tuning provides an efficient way for users to customize Large Language Models
(LLMs) with their private data in the emerging LLM service scenario. However, the sensitive …

Personal llm agents: Insights and survey about the capability, efficiency and security

Y Li, H Wen, W Wang, X Li, Y Yuan, G Liu, J Liu… - arxiv preprint arxiv …, 2024 - arxiv.org
Since the advent of personal computing devices, intelligent personal assistants (IPAs) have
been one of the key technologies that researchers and engineers have focused on, aiming …