Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM Computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

Survey on factuality in large language models: Knowledge, retrieval and domain-specificity

C Wang, X Liu, Y Yue, X Tang, T Zhang… - arxiv preprint arxiv …, 2023 - arxiv.org
This survey addresses the crucial issue of factuality in Large Language Models (LLMs). As
LLMs find applications across diverse domains, the reliability and accuracy of their outputs …

FaithDial: A Faithful Benchmark for Information-Seeking Dialogue

N Dziri, E Kamalloo, S Milton, O Zaiane… - Transactions of the …, 2022 - direct.mit.edu
The goal of information-seeking dialogue is to respond to seeker queries with natural
language utterances that are grounded on knowledge sources. However, dialogue systems …

Understanding factual errors in summarization: Errors, summarizers, datasets, error detectors

L Tang, T Goyal, AR Fabbri, P Laban, J Xu… - arxiv preprint arxiv …, 2022 - arxiv.org
The propensity of abstractive summarization models to make factual errors has been studied
extensively, including design of metrics to detect factual errors and annotation of errors in …

A survey on dialogue summarization: Recent advances and new frontiers

X Feng, X Feng, B Qin - arxiv preprint arxiv:2107.03175, 2021 - arxiv.org
Dialogue summarization aims to condense the original dialogue into a shorter version
covering salient information, which is a crucial way to reduce dialogue data overload …

AI hallucinations: a misnomer worth clarifying

N Maleki, B Padmanabhan… - 2024 IEEE conference on …, 2024 - ieeexplore.ieee.org
As large language models continue to advance in Artificial Intelligence (AI), text generation
systems have been shown to suffer from a problematic phenomenon often termed as" …

Zero-shot faithful factual error correction

KH Huang, HP Chan, H Ji - arxiv preprint arxiv:2305.07982, 2023 - arxiv.org
Faithfully correcting factual errors is critical for maintaining the integrity of textual knowledge
bases and preventing hallucinations in sequence-to-sequence models. Drawing on humans' …

Factkb: Generalizable factuality evaluation using language models enhanced with factual knowledge

S Feng, V Balachandran, Y Bai, Y Tsvetkov - arxiv preprint arxiv …, 2023 - arxiv.org
Evaluating the factual consistency of automatically generated summaries is essential for the
progress and adoption of reliable summarization systems. Despite recent advances, existing …

Embrace divergence for richer insights: A multi-document summarization benchmark and a case study on summarizing diverse information from news articles

KH Huang, P Laban, AR Fabbri, PK Choubey… - arxiv preprint arxiv …, 2023 - arxiv.org
Previous research in multi-document news summarization has typically concentrated on
collating information that all sources agree upon. However, to our knowledge, the …

Teaching language models to hallucinate less with synthetic tasks

E Jones, H Palangi, C Simões… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models (LLMs) frequently hallucinate on abstractive summarization tasks
such as document-based question-answering, meeting summarization, and clinical report …