Cognitive mirage: A review of hallucinations in large language models

H Ye, T Liu, A Zhang, W Hua, W Jia - arxiv preprint arxiv:2309.06794, 2023 - arxiv.org
As large language models continue to develop in the field of AI, text generation systems are
susceptible to a worrisome phenomenon known as hallucination. In this study, we …

Digital forgetting in large language models: A survey of unlearning methods

A Blanco-Justicia, N Jebreel… - Artificial Intelligence …, 2025 - Springer
Large language models (LLMs) have become the state of the art in natural language
processing. The massive adoption of generative LLMs and the capabilities they have shown …

Mathdial: A dialogue tutoring dataset with rich pedagogical properties grounded in math reasoning problems

J Macina, N Daheim, SP Chowdhury, T Sinha… - arxiv preprint arxiv …, 2023 - arxiv.org
While automatic dialogue tutors hold great potential in making education personalized and
more accessible, research on such systems has been hampered by a lack of sufficiently …

Deep model fusion: A survey

W Li, Y Peng, M Zhang, L Ding, H Hu… - arxiv preprint arxiv …, 2023 - arxiv.org
Deep model fusion/merging is an emerging technique that merges the parameters or
predictions of multiple deep learning models into a single one. It combines the abilities of …

Detecting and mitigating hallucinations in multilingual summarisation

Y Qiu, Y Ziser, A Korhonen, EM Ponti… - arxiv preprint arxiv …, 2023 - arxiv.org
Hallucinations pose a significant challenge to the reliability of neural models for abstractive
summarisation. While automatically generated summaries may be fluent, they often lack …

Chat vector: A simple approach to equip LLMs with instruction following and model alignment in new languages

SC Huang, PZ Li, YC Hsu, KM Chen… - Proceedings of the …, 2024 - aclanthology.org
Recently, the development of open-source large language models (LLMs) has advanced
rapidly. Nevertheless, due to data constraints, the capabilities of most open-source LLMs are …

Configurable foundation models: Building llms from a modular perspective

C **ao, Z Zhang, C Song, D Jiang, F Yao, X Han… - arxiv preprint arxiv …, 2024 - arxiv.org
Advancements in LLMs have recently unveiled challenges tied to computational efficiency
and continual scalability due to their requirements of huge parameters, making the …

Dial beinfo for faithfulness: Improving factuality of information-seeking dialogue via behavioural fine-tuning

E Razumovskaia, I Vulić, P Marković… - Findings of the …, 2024 - aclanthology.org
Factual faithfulness is a crucial requirement in information-seeking dialogue: the system
should respond to the user queries so that the responses are meaningful and aligned with …

Model merging by uncertainty-based gradient matching

N Daheim, T Möllenhoff, EM Ponti, I Gurevych… - arxiv preprint arxiv …, 2023 - arxiv.org
Models trained on different datasets can be merged by a weighted-averaging of their
parameters, but why does it work and when can it fail? Here, we connect the inaccuracy of …

Climategpt: Towards ai synthesizing interdisciplinary research on climate change

D Thulke, Y Gao, P Pelser, R Brune, R Jalota… - arxiv preprint arxiv …, 2024 - arxiv.org
This paper introduces ClimateGPT, a model family of domain-specific large language
models that synthesize interdisciplinary research on climate change. We trained two 7B …