Survey of hallucination in natural language generation

Z Ji, N Lee, R Frieske, T Yu, D Su, Y Xu, E Ishii… - ACM computing …, 2023 - dl.acm.org
Natural Language Generation (NLG) has improved exponentially in recent years thanks to
the development of sequence-to-sequence deep learning technologies such as Transformer …

Trustworthy llms: a survey and guideline for evaluating large language models' alignment

Y Liu, Y Yao, JF Ton, X Zhang, R Guo, H Cheng… - arxiv preprint arxiv …, 2023 - arxiv.org
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …

A stitch in time saves nine: Detecting and mitigating hallucinations of llms by validating low-confidence generation

N Varshney, W Yao, H Zhang, J Chen, D Yu - arxiv preprint arxiv …, 2023 - arxiv.org
Recently developed large language models have achieved remarkable success in
generating fluent and coherent text. However, these models often tend to'hallucinate'which …

Faithfulness in natural language generation: A systematic survey of analysis, evaluation and optimization methods

W Li, W Wu, M Chen, J Liu, X **ao, H Wu - arxiv preprint arxiv:2203.05227, 2022 - arxiv.org
Natural Language Generation (NLG) has made great progress in recent years due to the
development of deep learning techniques such as pre-trained language models. This …

Frequency-aware contrastive learning for neural machine translation

T Zhang, W Ye, B Yang, L Zhang, X Ren… - Proceedings of the …, 2022 - ojs.aaai.org
Low-frequency word prediction remains a challenge in modern neural machine translation
(NMT) systems. Recent adaptive training methods promote the output of infrequent words by …

ChatGPT incorrectness detection in software reviews

MH Tanzil, JY Khan, G Uddin - Proceedings of the IEEE/ACM 46th …, 2024 - dl.acm.org
We conducted a survey of 135 software engineering (SE) practitioners to understand how
they use Generative AI-based chatbots like ChatGPT for SE tasks. We find that they want to …

Prevent the language model from being overconfident in neural machine translation

M Miao, F Meng, Y Liu, XH Zhou, J Zhou - arxiv preprint arxiv:2105.11098, 2021 - arxiv.org
The Neural Machine Translation (NMT) model is essentially a joint language model
conditioned on both the source sentence and partial translation. Therefore, the NMT model …

Improving data augmentation for low resource speech-to-text translation with diverse paraphrasing

C Mi, L **e, Y Zhang - Neural Networks, 2022 - Elsevier
High quality end-to-end speech translation model relies on a large scale of speech-to-text
training data, which is usually scarce or even unavailable for some low-resource language …

One reference is not enough: Diverse distillation with reference selection for non-autoregressive translation

C Shao, X Wu, Y Feng - arxiv preprint arxiv:2205.14333, 2022 - arxiv.org
Non-autoregressive neural machine translation (NAT) suffers from the multi-modality
problem: the source sentence may have multiple correct translations, but the loss function is …

Attention calibration for transformer in neural machine translation

Y Lu, J Zeng, J Zhang, S Wu, M Li - … of the 59th Annual Meeting of …, 2021 - aclanthology.org
Attention mechanisms have achieved substantial improvements in neural machine
translation by dynamically selecting relevant inputs for different predictions. However, recent …