Unified hallucination detection for multimodal large language models
Despite significant strides in multimodal tasks, Multimodal Large Language Models (MLLMs)
are plagued by the critical issue of hallucination. The reliable detection of such …
are plagued by the critical issue of hallucination. The reliable detection of such …
Unfamiliar finetuning examples control how language models hallucinate
Large language models are known to hallucinate when faced with unfamiliar queries, but
the underlying mechanism that govern how models hallucinate are not yet fully understood …
the underlying mechanism that govern how models hallucinate are not yet fully understood …
Literature Review of AI Hallucination Research Since the Advent of ChatGPT: Focusing on Papers from arxiv
DM Park, HJ Lee - Informatization Policy, 2024 - koreascience.kr
Hallucination is a significant barrier to the utilization of large-scale language models or
multimodal models. In this study, we collected 654 computer science papers with" …
multimodal models. In this study, we collected 654 computer science papers with" …
Haloscope: Harnessing unlabeled llm generations for hallucination detection
The surge in applications of large language models (LLMs) has prompted concerns about
the generation of misleading or fabricated information, known as hallucinations. Therefore …
the generation of misleading or fabricated information, known as hallucinations. Therefore …
Critical Tokens Matter: Token-Level Contrastive Estimation Enhence LLM's Reasoning Capability
Large Language Models (LLMs) have exhibited remarkable performance on reasoning
tasks. They utilize autoregressive token generation to construct reasoning trajectories …
tasks. They utilize autoregressive token generation to construct reasoning trajectories …
Improving factuality in large language models via decoding-time hallucinatory and truthful comparators
Despite their remarkable capabilities, Large Language Models (LLMs) are prone to
generate responses that contradict verifiable facts, ie, unfaithful hallucination content …
generate responses that contradict verifiable facts, ie, unfaithful hallucination content …
OLAPH: Improving Factuality in Biomedical Long-form Question Answering
In the medical domain, numerous scenarios necessitate the long-form generation ability of
large language models (LLMs). Specifically, when addressing patients' questions, it is …
large language models (LLMs). Specifically, when addressing patients' questions, it is …
Open Problems in Machine Unlearning for AI Safety
As AI systems become more capable, widely deployed, and increasingly autonomous in
critical areas such as cybersecurity, biological research, and healthcare, ensuring their …
critical areas such as cybersecurity, biological research, and healthcare, ensuring their …
Pensieve: Retrospect-then-compare mitigates visual hallucination
Multi-modal Large Language Models (MLLMs) demonstrate remarkable success across
various vision-language tasks. However, they suffer from visual hallucination, where the …
various vision-language tasks. However, they suffer from visual hallucination, where the …
Multi-group Uncertainty Quantification for Long-form Text Generation
While large language models are rapidly moving towards consumer-facing applications,
they are often still prone to factual errors and hallucinations. In order to reduce the potential …
they are often still prone to factual errors and hallucinations. In order to reduce the potential …