Security and privacy challenges of large language models: A survey

BC Das, MH Amini, Y Wu - ACM Computing Surveys, 2025 - dl.acm.org
Large language models (LLMs) have demonstrated extraordinary capabilities and
contributed to multiple fields, such as generating and summarizing text, language …

Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

A survey on hallucination in large language models: Principles, taxonomy, challenges, and open questions

L Huang, W Yu, W Ma, W Zhong, Z Feng… - ACM Transactions on …, 2025 - dl.acm.org
The emergence of large language models (LLMs) has marked a significant breakthrough in
natural language processing (NLP), fueling a paradigm shift in information acquisition …

Siren's song in the AI ocean: a survey on hallucination in large language models

Y Zhang, Y Li, L Cui, D Cai, L Liu, T Fu… - arxiv preprint arxiv …, 2023 - arxiv.org
While large language models (LLMs) have demonstrated remarkable capabilities across a
range of downstream tasks, a significant concern revolves around their propensity to exhibit …

Halueval: A large-scale hallucination evaluation benchmark for large language models

J Li, X Cheng, WX Zhao, JY Nie, JR Wen - arxiv preprint arxiv:2305.11747, 2023 - arxiv.org
Large language models (LLMs), such as ChatGPT, are prone to generate hallucinations, ie,
content that conflicts with the source or cannot be verified by the factual knowledge. To …

Evaluating correctness and faithfulness of instruction-following models for question answering

V Adlakha, P BehnamGhader, XH Lu… - Transactions of the …, 2024 - direct.mit.edu
Instruction-following models are attractive alternatives to fine-tuned approaches for question
answering (QA). By simply prepending relevant documents and an instruction to their input …

Cognitive mirage: A review of hallucinations in large language models

H Ye, T Liu, A Zhang, W Hua, W Jia - arxiv preprint arxiv:2309.06794, 2023 - arxiv.org
As large language models continue to develop in the field of AI, text generation systems are
susceptible to a worrisome phenomenon known as hallucination. In this study, we …

Hallucination detection: Robustly discerning reliable answers in large language models

Y Chen, Q Fu, Y Yuan, Z Wen, G Fan, D Liu… - Proceedings of the …, 2023 - dl.acm.org
Large language models (LLMs) have gained widespread adoption in various natural
language processing tasks, including question answering and dialogue systems. However …

Felm: Benchmarking factuality evaluation of large language models

Y Zhao, J Zhang, I Chern, S Gao… - Advances in Neural …, 2023 - proceedings.neurips.cc
Assessing factuality of text generated by large language models (LLMs) is an emerging yet
crucial research area, aimed at alerting users to potential errors and guiding the …

Factuality enhanced language models for open-ended text generation

N Lee, W **, P Xu, M Patwary… - Advances in …, 2022 - proceedings.neurips.cc
Pretrained language models (LMs) are susceptible to generate text with nonfactual
information. In this work, we measure and improve the factual accuracy of large-scale LMs …