Security and privacy challenges of large language models: A survey

BC Das, MH Amini, Y Wu - ACM Computing Surveys, 2025 - dl.acm.org
Large language models (LLMs) have demonstrated extraordinary capabilities and
contributed to multiple fields, such as generating and summarizing text, language …

Machine unlearning: Taxonomy, metrics, applications, challenges, and prospects

N Li, C Zhou, Y Gao, H Chen, Z Zhang… - … on Neural Networks …, 2025 - ieeexplore.ieee.org
Personal digital data is a critical asset, and governments worldwide have enforced laws and
regulations to protect data privacy. Data users have been endowed with the “right to be …

Foundational challenges in assuring alignment and safety of large language models

U Anwar, A Saparov, J Rando, D Paleka… - arxiv preprint arxiv …, 2024 - arxiv.org
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …

Rethinking machine unlearning for large language models

S Liu, Y Yao, J Jia, S Casper, N Baracaldo… - arxiv preprint arxiv …, 2024 - arxiv.org
We explore machine unlearning (MU) in the domain of large language models (LLMs),
referred to as LLM unlearning. This initiative aims to eliminate undesirable data influence …

On protecting the data privacy of large language models (llms): A survey

B Yan, K Li, M Xu, Y Dong, Y Zhang, Z Ren… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are complex artificial intelligence systems capable of
understanding, generating and translating human language. They learn language patterns …

Don't make your llm an evaluation benchmark cheater

K Zhou, Y Zhu, Z Chen, W Chen, WX Zhao… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models~(LLMs) have greatly advanced the frontiers of artificial intelligence,
attaining remarkable improvement in model capacity. To assess the model performance, a …

Did the neurons read your book? document-level membership inference for large language models

M Meeus, S Jain, M Rei, YA de Montjoye - 33rd USENIX Security …, 2024 - usenix.org
With large language models (LLMs) poised to become embedded in our daily lives,
questions are starting to be raised about the data they learned from. These questions range …

Generalization or memorization: Data contamination and trustworthy evaluation for large language models

Y Dong, X Jiang, H Liu, Z **, B Gu, M Yang… - arxiv preprint arxiv …, 2024 - arxiv.org
Recent statements about the impressive capabilities of large language models (LLMs) are
usually supported by evaluating on open-access benchmarks. Considering the vast size and …

Muse: Machine unlearning six-way evaluation for language models

W Shi, J Lee, Y Huang, S Malladi, J Zhao… - arxiv preprint arxiv …, 2024 - arxiv.org
Language models (LMs) are trained on vast amounts of text data, which may include private
and copyrighted content. Data owners may request the removal of their data from a trained …

Black-box access is insufficient for rigorous ai audits

S Casper, C Ezell, C Siegmann, N Kolt… - The 2024 ACM …, 2024 - dl.acm.org
External audits of AI systems are increasingly recognized as a key mechanism for AI
governance. The effectiveness of an audit, however, depends on the degree of access …