Rethinking machine unlearning for large language models

S Liu, Y Yao, J Jia, S Casper, N Baracaldo… - Nature Machine …, 2025 - nature.com
We explore machine unlearning in the domain of large language models (LLMs), referred to
as LLM unlearning. This initiative aims to eliminate undesirable data influence (for example …

Negative preference optimization: From catastrophic collapse to effective unlearning

R Zhang, L Lin, Y Bai, S Mei - arxiv preprint arxiv:2404.05868, 2024 - arxiv.org
Large Language Models (LLMs) often memorize sensitive, private, or copyrighted data
during pre-training. LLM unlearning aims to eliminate the influence of undesirable data from …

Offset unlearning for large language models

JY Huang, W Zhou, F Wang, F Morstatter… - arxiv preprint arxiv …, 2024 - arxiv.org
Despite the strong capabilities of Large Language Models (LLMs) to acquire knowledge
from their training corpora, the memorization of sensitive information in the corpora such as …

Min-k%++: Improved baseline for detecting pre-training data from large language models

J Zhang, J Sun, E Yeats, Y Ouyang, M Kuo… - arxiv preprint arxiv …, 2024 - arxiv.org
The problem of pre-training data detection for large language models (LLMs) has received
growing attention due to its implications in critical issues like copyright violation and test data …

Machine unlearning in generative ai: A survey

Z Liu, G Dou, Z Tan, Y Tian, M Jiang - arxiv preprint arxiv:2407.20516, 2024 - arxiv.org
Generative AI technologies have been deployed in many places, such as (multimodal) large
language models and vision generative models. Their remarkable performance should be …

Soul: Unlocking the power of second-order optimization for llm unlearning

J Jia, Y Zhang, Y Zhang, J Liu, B Runwal… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have highlighted the necessity of effective unlearning
mechanisms to comply with data regulations and ethical AI practices. LLM unlearning aims …

Towards efficient and effective unlearning of large language models for recommendation

H Wang, J Lin, B Chen, Y Yang, R Tang… - Frontiers of Computer …, 2025 - Springer
Conclusion In this letter, we propose E2URec, the efficient and effective unlearning method
for LLMRec. Our method enables LLMRec to efficiently forget the specific data by only …

Large language model unlearning via embedding-corrupted prompts

CY Liu, Y Wang, J Flanigan, Y Liu - arxiv preprint arxiv:2406.07933, 2024 - arxiv.org
Large language models (LLMs) have advanced to encompass extensive knowledge across
diverse domains. Yet controlling what a large language model should not know is important …

Pre-text: Training language models on private federated data in the age of llms

C Hou, A Shrivastava, H Zhan, R Conway, T Le… - arxiv preprint arxiv …, 2024 - arxiv.org
On-device training is currently the most common approach for training machine learning
(ML) models on private, distributed user data. Despite this, on-device training has several …

Unlocking memorization in large language models with dynamic soft prompting

Z Wang, R Bao, Y Wu, J Taylor, C **ao, F Zheng… - arxiv preprint arxiv …, 2024 - arxiv.org
Pretrained large language models (LLMs) have revolutionized natural language processing
(NLP) tasks such as summarization, question answering, and translation. However, LLMs …