Summary of a haystack: A challenge to long-context llms and rag systems
LLMs and RAG systems are now capable of handling millions of input tokens or more.
However, evaluating the output quality of such systems on long-context tasks remains …
However, evaluating the output quality of such systems on long-context tasks remains …
Debate on graph: a flexible and reliable reasoning framework for large language models
Large Language Models (LLMs) may suffer from hallucinations in real-world applications
due to the lack of relevant knowledge. In contrast, knowledge graphs encompass extensive …
due to the lack of relevant knowledge. In contrast, knowledge graphs encompass extensive …
Logical Consistency of Large Language Models in Fact-checking
In recent years, large language models (LLMs) have demonstrated significant success in
performing varied natural language tasks such as language translation, question-answering …
performing varied natural language tasks such as language translation, question-answering …
An evaluation framework for attributed information retrieval using large language models
With the growing success of Large Language models (LLMs) in information-seeking
scenarios, search engines are now adopting generative approaches to provide answers …
scenarios, search engines are now adopting generative approaches to provide answers …
Evaluation of Attribution Bias in Retrieval-Augmented Large Language Models
Attributing answers to source documents is an approach used to enhance the verifiability of
a model's output in retrieval augmented generation (RAG). Prior work has mainly focused on …
a model's output in retrieval augmented generation (RAG). Prior work has mainly focused on …
Attributed Question Answering for Preconditions in the Dutch Law
In this paper, we address the problem of answering questions about preconditions in the
law, eg “When can the court terminate the guardianship of a natural person?”. When …
law, eg “When can the court terminate the guardianship of a natural person?”. When …
Enhancing rag-retrieval to improve llms robustness and resilience to hallucinations
Abstract The use of Retrieval-Augmented Generation (RAG) models has become a powerful
method for generating natural language by incorporating information retrieval. However, the …
method for generating natural language by incorporating information retrieval. However, the …
Counter-intuitive: Large Language Models Can Better Understand Knowledge Graphs Than We Thought
Although the method of enhancing large language models'(LLMs') reasoning ability and
reducing their hallucinations through the use of knowledge graphs (KGs) has received …
reducing their hallucinations through the use of knowledge graphs (KGs) has received …
Post-Hoc Answer Attribution for Grounded and Trustworthy Long Document Comprehension: Task, Insights, and Challenges
Attributing answer text to its source document for information-seeking questions is crucial for
building trustworthy, reliable, and accountable systems. We formulate a new task of post-hoc …
building trustworthy, reliable, and accountable systems. We formulate a new task of post-hoc …
Inner-Probe: Discovering Copyright-related Data Generation in LLM Architecture
Large Language Models (LLMs) utilize extensive knowledge databases and show powerful
text generation ability. However, their reliance on high-quality copyrighted datasets raises …
text generation ability. However, their reliance on high-quality copyrighted datasets raises …