Can large language models identify authorship?

B Huang, C Chen, K Shu - arxiv preprint arxiv:2403.08213, 2024 - arxiv.org
The ability to accurately identify authorship is crucial for verifying content authenticity and
mitigating misinformation. Large Language Models (LLMs) have demonstrated an …

Agentreview: Exploring peer review dynamics with llm agents

Y **, Q Zhao, Y Wang, H Chen, K Zhu, Y **ao… - arxiv preprint arxiv …, 2024 - arxiv.org
Peer review is fundamental to the integrity and advancement of scientific publication.
Traditional methods of peer review analyses often rely on exploration and statistics of …

Editing conceptual knowledge for large language models

X Wang, S Mao, N Zhang, S Deng, Y Yao… - arxiv preprint arxiv …, 2024 - arxiv.org
Recently, there has been a growing interest in knowledge editing for Large Language
Models (LLMs). Current approaches and evaluations merely explore the instance-level …

Harmful fine-tuning attacks and defenses for large language models: A survey

T Huang, S Hu, F Ilhan, SF Tekin, L Liu - arxiv preprint arxiv:2409.18169, 2024 - arxiv.org
Recent research demonstrates that the nascent fine-tuning-as-a-service business model
exposes serious safety concerns--fine-tuning over a few harmful data uploaded by the users …

Mlake: Multilingual knowledge editing benchmark for large language models

Z Wei, J Deng, L Pang, H Ding, H Shen… - arxiv preprint arxiv …, 2024 - arxiv.org
The extensive utilization of large language models (LLMs) underscores the crucial necessity
for precise and contemporary knowledge embedded within their intrinsic parameters …

Can Knowledge Editing Really Correct Hallucinations?

B Huang, C Chen, X Xu, A Payani, K Shu - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) suffer from hallucinations, referring to the non-factual
information in generated content, despite their superior capacities across tasks. Meanwhile …

Locking down the finetuned llms safety

M Zhu, L Yang, Y Wei, N Zhang, Y Zhang - arxiv preprint arxiv …, 2024 - arxiv.org
Fine-tuning large language models (LLMs) on additional datasets is often necessary to
optimize them for specific downstream tasks. However, existing safety alignment measures …

Lisa: Lazy safety alignment for large language models against harmful fine-tuning attack

T Huang, S Hu, F Ilhan, SF Tekin, L Liu - arxiv preprint arxiv:2405.18641, 2024 - arxiv.org
Recent studies show that Large Language Models (LLMs) with safety alignment can be jail-
broken by fine-tuning on a dataset mixed with harmful data. First time in the literature, we …

Retrieval-enhanced knowledge editing in language models for multi-hop question answering

Y Shi, Q Tan, X Wu, S Zhong, K Zhou… - Proceedings of the 33rd …, 2024 - dl.acm.org
Large Language Models (LLMs) have shown proficiency in question-answering tasks but
often struggle to integrate real-time knowledge, leading to potentially outdated or inaccurate …

Political-llm: Large language models in political science

L Li, J Li, C Chen, F Gui, H Yang, C Yu, Z Wang… - arxiv preprint arxiv …, 2024 - arxiv.org
In recent years, large language models (LLMs) have been widely adopted in political
science tasks such as election prediction, sentiment analysis, policy impact assessment, and …