Cognitive mirage: A review of hallucinations in large language models
As large language models continue to develop in the field of AI, text generation systems are
susceptible to a worrisome phenomenon known as hallucination. In this study, we …
susceptible to a worrisome phenomenon known as hallucination. In this study, we …
Digital forgetting in large language models: A survey of unlearning methods
A Blanco-Justicia, N Jebreel… - Artificial Intelligence …, 2025 - Springer
Large language models (LLMs) have become the state of the art in natural language
processing. The massive adoption of generative LLMs and the capabilities they have shown …
processing. The massive adoption of generative LLMs and the capabilities they have shown …
Mathdial: A dialogue tutoring dataset with rich pedagogical properties grounded in math reasoning problems
While automatic dialogue tutors hold great potential in making education personalized and
more accessible, research on such systems has been hampered by a lack of sufficiently …
more accessible, research on such systems has been hampered by a lack of sufficiently …
Deep model fusion: A survey
Deep model fusion/merging is an emerging technique that merges the parameters or
predictions of multiple deep learning models into a single one. It combines the abilities of …
predictions of multiple deep learning models into a single one. It combines the abilities of …
Detecting and mitigating hallucinations in multilingual summarisation
Hallucinations pose a significant challenge to the reliability of neural models for abstractive
summarisation. While automatically generated summaries may be fluent, they often lack …
summarisation. While automatically generated summaries may be fluent, they often lack …
Chat vector: A simple approach to equip LLMs with instruction following and model alignment in new languages
Recently, the development of open-source large language models (LLMs) has advanced
rapidly. Nevertheless, due to data constraints, the capabilities of most open-source LLMs are …
rapidly. Nevertheless, due to data constraints, the capabilities of most open-source LLMs are …
Configurable foundation models: Building llms from a modular perspective
Advancements in LLMs have recently unveiled challenges tied to computational efficiency
and continual scalability due to their requirements of huge parameters, making the …
and continual scalability due to their requirements of huge parameters, making the …
Dial beinfo for faithfulness: Improving factuality of information-seeking dialogue via behavioural fine-tuning
E Razumovskaia, I Vulić, P Marković… - Findings of the …, 2024 - aclanthology.org
Factual faithfulness is a crucial requirement in information-seeking dialogue: the system
should respond to the user queries so that the responses are meaningful and aligned with …
should respond to the user queries so that the responses are meaningful and aligned with …
Model merging by uncertainty-based gradient matching
Models trained on different datasets can be merged by a weighted-averaging of their
parameters, but why does it work and when can it fail? Here, we connect the inaccuracy of …
parameters, but why does it work and when can it fail? Here, we connect the inaccuracy of …
Climategpt: Towards ai synthesizing interdisciplinary research on climate change
This paper introduces ClimateGPT, a model family of domain-specific large language
models that synthesize interdisciplinary research on climate change. We trained two 7B …
models that synthesize interdisciplinary research on climate change. We trained two 7B …