Survey of hallucination in natural language generation
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …
the development of sequence-to-sequence deep learning technologies such as Transformer …
Exposing attention glitches with flip-flop language modeling
Why do large language models sometimes output factual inaccuracies and exhibit
erroneous reasoning? The brittleness of these models, particularly when executing long …
erroneous reasoning? The brittleness of these models, particularly when executing long …
Hallucination detection: Robustly discerning reliable answers in large language models
Large language models (LLMs) have gained widespread adoption in various natural
language processing tasks, including question answering and dialogue systems. However …
language processing tasks, including question answering and dialogue systems. However …
Learning to break the loop: Analyzing and mitigating repetitions for neural text generation
While large-scale neural language models, such as GPT2 and BART, have achieved
impressive results on various text generation tasks, they tend to get stuck in undesirable …
impressive results on various text generation tasks, they tend to get stuck in undesirable …
[PDF][PDF] Advances and challenges in multi-domain task-oriented dialogue policy optimization
Develo** a successful dialogue policy for a multi-domain task-oriented dialogue (MDTD)
system is a challenging task. Basically, a desirable dialogue policy acts as the decision …
system is a challenging task. Basically, a desirable dialogue policy acts as the decision …
Coarse-to-fine: a hierarchical diffusion model for molecule generation in 3d
Generating desirable molecular structures in 3D is a fundamental problem for drug
discovery. Despite the considerable progress we have achieved, existing methods usually …
discovery. Despite the considerable progress we have achieved, existing methods usually …
Knn-lm does not improve open-ended text generation
In this paper, we study the generation quality of interpolation-based retrieval-augmented
language models (LMs). These methods, best exemplified by the KNN-LM, interpolate the …
language models (LMs). These methods, best exemplified by the KNN-LM, interpolate the …
Tailoring language generation models under total variation distance
The standard paradigm of neural language generation adopts maximum likelihood
estimation (MLE) as the optimizing method. From a distributional view, MLE in fact minimizes …
estimation (MLE) as the optimizing method. From a distributional view, MLE in fact minimizes …