Large language models can be easily distracted by irrelevant context

F Shi, X Chen, K Misra, N Scales… - International …, 2023 - proceedings.mlr.press
Large language models have achieved impressive performance on various natural
language processing tasks. However, so far they have been evaluated primarily on …

Branch-train-merge: Embarrassingly parallel training of expert language models

M Li, S Gururangan, T Dettmers, M Lewis… - arxiv preprint arxiv …, 2022 - arxiv.org
We present Branch-Train-Merge (BTM), a communication-efficient algorithm for
embarrassingly parallel training of large language models (LLMs). We show it is possible to …

Improving machine reading comprehension with general reading strategies

K Sun, D Yu, D Yu, C Cardie - arxiv preprint arxiv:1810.13441, 2018 - arxiv.org
Reading strategies have been shown to improve comprehension levels, especially for
readers lacking adequate prior knowledge. Just as the process of knowledge accumulation …

EXAMS: A multi-subject high school examinations dataset for cross-lingual and multilingual question answering

M Hardalov, T Mihaylov, D Zlatkova, Y Dinkov… - arxiv preprint arxiv …, 2020 - arxiv.org
We propose EXAMS--a new benchmark dataset for cross-lingual and multilingual question
answering for high school examinations. We collected more than 24,000 high-quality high …

Improving question answering by commonsense-based pre-training

W Zhong, D Tang, N Duan, M Zhou, J Wang… - … Processing and Chinese …, 2019 - Springer
Although neural network approaches achieve remarkable success on a variety of NLP tasks,
many of them struggle to answer questions that require commonsense knowledge. We …

An Empirical Study of Retrieval-Augmented Code Generation: Challenges and Opportunities

Z Yang, S Chen, C Gao, Z Li, X Hu, K Liu… - ACM Transactions on …, 2025 - dl.acm.org
Code generation aims to automatically generate code snippets of specific programming
language according to natural language descriptions. The continuous advancements in …

Careful selection of knowledge to solve open book question answering

P Banerjee, KK Pal, A Mitra, C Baral - arxiv preprint arxiv:1907.10738, 2019 - arxiv.org
Open book question answering is a type of natural language based QA (NLQA) where
questions are expected to be answered with respect to a given set of open book facts, and …

Towards teachable reasoning systems: Using a dynamic memory of user feedback for continual system improvement

BD Mishra, O Tafjord, P Clark - arxiv preprint arxiv:2204.13074, 2022 - arxiv.org
Our goal is a teachable reasoning system for question-answering (QA), where a user can
interact with faithful answer explanations, and correct its errors so that the system improves …

Think you have solved direct-answer question answering? try arc-da, the direct-answer AI2 reasoning challenge

S Bhakthavatsalam, D Khashabi, T Khot… - arxiv preprint arxiv …, 2021 - arxiv.org
We present the ARC-DA dataset, a direct-answer (" open response"," freeform") version of
the ARC (AI2 Reasoning Challenge) multiple-choice dataset. While ARC has been …

Improving question answering with external knowledge

X Pan, K Sun, D Yu, J Chen, H Ji, C Cardie… - arxiv preprint arxiv …, 2019 - arxiv.org
We focus on multiple-choice question answering (QA) tasks in subject areas such as
science, where we require both broad background knowledge and the facts from the given …