A comprehensive study of knowledge editing for large language models

N Zhang, Y Yao, B Tian, P Wang, S Deng… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have shown extraordinary capabilities in understanding
and generating text that closely mirrors human communication. However, a primary …

Editing large language models: Problems, methods, and opportunities

Y Yao, P Wang, B Tian, S Cheng, Z Li, S Deng… - arxiv preprint arxiv …, 2023 - arxiv.org
Despite the ability to train capable LLMs, the methodology for maintaining their relevancy
and rectifying errors remains elusive. To this end, the past few years have witnessed a surge …

Continual learning with pre-trained models: A survey

DW Zhou, HL Sun, J Ning, HJ Ye, DC Zhan - arxiv preprint arxiv …, 2024 - arxiv.org
Nowadays, real-world applications often face streaming data, which requires the learning
system to absorb new knowledge as data evolves. Continual Learning (CL) aims to achieve …

Deciphering the factors influencing the efficacy of chain-of-thought: Probability, memorization, and noisy reasoning

A Prabhakar, TL Griffiths, RT McCoy - arxiv preprint arxiv:2407.01687, 2024 - arxiv.org
Chain-of-Thought (CoT) prompting has been shown to enhance the multi-step reasoning
capabilities of Large Language Models (LLMs). However, debates persist about whether …

StructEval: Deepen and broaden large language model assessment via structured evaluation

B Cao, M Ren, H Lin, X Han, F Zhang, J Zhan… - arxiv preprint arxiv …, 2024 - arxiv.org
Evaluation is the baton for the development of large language models. Current evaluations
typically employ a single-item assessment paradigm for each atomic test objective, which …

How much can we forget about Data Contamination?

S Bordt, S Srinivas, V Boreiko… - arxiv preprint arxiv …, 2024 - arxiv.org
The leakage of benchmark data into the training data has emerged as a significant
challenge for evaluating the capabilities of large language models (LLMs). In this work, we …

[PDF][PDF] Editing large language models

N Zhang, Y Yao, S Deng - … Processing and the 3rd Conference of …, 2023 - aclanthology.org
Even with their impressive abilities, Large Language Models (LLMs) such as ChatGPT are
not immune to issues of factual or logically consistent. Concretely, the key concern is how to …

Realistic Continual Learning Approach using Pre-trained Models

N Nasri, C Gutiérrez-Álvarez, S Lafuente-Arroyo… - arxiv preprint arxiv …, 2024 - arxiv.org
Continual learning (CL) is crucial for evaluating adaptability in learning solutions to retain
knowledge. Our research addresses the challenge of catastrophic forgetting, where models …

An Efficient Replay for Class-Incremental Learning with Pre-trained Models

W Yin, Z Tan - arxiv preprint arxiv:2408.08084, 2024 - arxiv.org
In general class-incremental learning, researchers typically use sample sets as a tool to
avoid catastrophic forgetting during continuous learning. At the same time, researchers have …

Advancing Ultrasound Medical Continuous Learning with Task-Specific Generalization and Adaptability

C Zhu, J Lin, G Tan, N Zhu, K Li… - 2024 IEEE International …, 2024 - ieeexplore.ieee.org
As artificial intelligence progresses in the field of medical ultrasound image analysis,
mitigating catastrophic forgetting during continuous learning processes in disease diagnosis …