Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM Computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

Exposing attention glitches with flip-flop language modeling

B Liu, J Ash, S Goel… - Advances in Neural …, 2024 - proceedings.neurips.cc
Why do large language models sometimes output factual inaccuracies and exhibit
erroneous reasoning? The brittleness of these models, particularly when executing long …

Hallucination detection: Robustly discerning reliable answers in large language models

Y Chen, Q Fu, Y Yuan, Z Wen, G Fan, D Liu… - Proceedings of the …, 2023 - dl.acm.org
Large language models (LLMs) have gained widespread adoption in various natural
language processing tasks, including question answering and dialogue systems. However …

Learning to break the loop: Analyzing and mitigating repetitions for neural text generation

J Xu, X Liu, J Yan, D Cai, H Li… - Advances in Neural …, 2022 - proceedings.neurips.cc
While large-scale neural language models, such as GPT2 and BART, have achieved
impressive results on various text generation tasks, they tend to get stuck in undesirable …

[PDF][PDF] Advances and challenges in multi-domain task-oriented dialogue policy optimization

M Rohmatillah, JT Chien - APSIPA Transactions on Signal …, 2023 - nowpublishers.com
Develo** a successful dialogue policy for a multi-domain task-oriented dialogue (MDTD)
system is a challenging task. Basically, a desirable dialogue policy acts as the decision …

Coarse-to-fine: a hierarchical diffusion model for molecule generation in 3d

B Qiang, Y Song, M Xu, J Gong, B Gao… - International …, 2023 - proceedings.mlr.press
Generating desirable molecular structures in 3D is a fundamental problem for drug
discovery. Despite the considerable progress we have achieved, existing methods usually …

Knn-lm does not improve open-ended text generation

S Wang, Y Song, A Drozdov, A Garimella… - arxiv preprint arxiv …, 2023 - arxiv.org
In this paper, we study the generation quality of interpolation-based retrieval-augmented
language models (LMs). These methods, best exemplified by the KNN-LM, interpolate the …

Tailoring language generation models under total variation distance

H Ji, P Ke, Z Hu, R Zhang, M Huang - arxiv preprint arxiv:2302.13344, 2023 - arxiv.org
The standard paradigm of neural language generation adopts maximum likelihood
estimation (MLE) as the optimizing method. From a distributional view, MLE in fact minimizes …