A comprehensive study of knowledge editing for large language models

N Zhang, Y Yao, B Tian, P Wang, S Deng… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have shown extraordinary capabilities in understanding
and generating text that closely mirrors human communication. However, a primary …

Can Editing LLMs Inject Harm?

C Chen, B Huang, Z Li, Z Chen, S Lai, X Xu… - arxiv preprint arxiv …, 2024 - arxiv.org
Knowledge editing has been increasingly adopted to correct the false or outdated
knowledge in Large Language Models (LLMs). Meanwhile, one critical but under-explored …

Propagating knowledge updates to lms through distillation

S Padmanabhan, Y Onoe, M Zhang… - Advances in Neural …, 2024 - proceedings.neurips.cc
Modern language models have the capacity to store and use immense amounts of
knowledge about real-world entities, but it remains unclear how to update such knowledge …

Can language models be instructed to protect personal information?

Y Chen, E Mendes, S Das, W Xu, A Ritter - arxiv preprint arxiv …, 2023 - arxiv.org
Large multimodal language models have proven transformative in numerous applications.
However, these models have been shown to memorize and leak pre-training data, raising …

Locking down the finetuned llms safety

M Zhu, L Yang, Y Wei, N Zhang, Y Zhang - arxiv preprint arxiv …, 2024 - arxiv.org
Fine-tuning large language models (LLMs) on additional datasets is often necessary to
optimize them for specific downstream tasks. However, existing safety alignment measures …

Fundamental Problems With Model Editing: How Should Rational Belief Revision Work in LLMs?

P Hase, T Hofweber, X Zhou, E Stengel-Eskin… - arxiv preprint arxiv …, 2024 - arxiv.org
The model editing problem concerns how language models should learn new facts about
the world over time. While empirical research on model editing has drawn widespread …

A survey on the memory mechanism of large language model based agents

Z Zhang, X Bo, C Ma, R Li, X Chen, Q Dai, J Zhu… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language model (LLM) based agents have recently attracted much attention from the
research and industry communities. Compared with original LLMs, LLM-based agents are …

Can llms be fooled? investigating vulnerabilities in llms

S Abdali, J He, CJ Barberan, R Anarfi - arxiv preprint arxiv:2407.20529, 2024 - arxiv.org
The advent of Large Language Models (LLMs) has garnered significant popularity and
wielded immense power across various domains within Natural Language Processing …

AmoebaLLM: Constructing Any-Shape Large Language Models for Efficient and Instant Deployment

Y Fu, Z Yu, J Li, J Qian, Y Zhang, X Yuan, D Shi… - arxiv preprint arxiv …, 2024 - arxiv.org
Motivated by the transformative capabilities of large language models (LLMs) across various
natural language tasks, there has been a growing demand to deploy these models …

Absinstruct: Eliciting abstraction ability from llms through explanation tuning with plausibility estimation

Z Wang, W Fan, Q Zong, H Zhang, S Choi… - arxiv preprint arxiv …, 2024 - arxiv.org
Abstraction ability is crucial in human intelligence, which can also benefit various tasks in
NLP study. Existing work shows that LLMs are deficient in abstract ability, and how to …