Teach me to explain: A review of datasets for explainable natural language processing

S Wiegreffe, A Marasović - arxiv preprint arxiv:2102.12060, 2021 - arxiv.org
Explainable NLP (ExNLP) has increasingly focused on collecting human-annotated textual
explanations. These explanations are used downstream in three ways: as data …

Generating literal and implied subquestions to fact-check complex claims

J Chen, A Sriram, E Choi, G Durrett - arxiv preprint arxiv:2205.06938, 2022 - arxiv.org
Verifying complex political claims is a challenging task, especially when politicians use
various tactics to subtly misrepresent the facts. Automatic fact-checking systems fall short …

Conditional generation with a question-answering blueprint

S Narayan, J Maynez, RK Amplayo… - Transactions of the …, 2023 - direct.mit.edu
The ability to convey relevant and faithful information is critical for many tasks in conditional
generation and yet remains elusive for neural seq-to-seq models whose outputs often reveal …

Iseeq: Information seeking question generation using dynamic meta-information retrieval and knowledge graphs

M Gaur, K Gunaratna, V Srinivasan, H ** - Proceedings of the AAAI …, 2022 - ojs.aaai.org
Abstract Conversational Information Seeking (CIS) is a relatively new research area within
conversational AI that attempts to seek information from end-users in order to understand …

Peek across: Improving multi-document modeling via cross-document question-answering

A Caciularu, ME Peters, J Goldberger, I Dagan… - arxiv preprint arxiv …, 2023 - arxiv.org
The integration of multi-document pre-training objectives into language models has resulted
in remarkable improvements in multi-document downstream tasks. In this work, we propose …

Explanatory artificial intelligence (YAI): human-centered explanations of explainable AI and complex data

F Sovrano, F Vitali - Data Mining and Knowledge Discovery, 2024 - Springer
In this paper we introduce a new class of software tools engaged in delivering successful
explanations of complex processes on top of basic Explainable AI (XAI) software systems …

An objective metric for explainable AI: how and why to estimate the degree of explainability

F Sovrano, F Vitali - Knowledge-Based Systems, 2023 - Elsevier
This paper presents a new method for objectively measuring the explainability of textual
information, such as the outputs of Explainable AI (XAI). We introduce a metric called …

A question answering framework for decontextualizing user-facing snippets from scientific documents

B Newman, L Soldaini, R Fok, A Cohan… - arxiv preprint arxiv …, 2023 - arxiv.org
Many real-world applications (eg, note taking, search) require extracting a sentence or
paragraph from a document and showing that snippet to a human outside of the source …

Design choices for crowdsourcing implicit discourse relations: Revealing the biases introduced by task design

V Pyatkin, F Yung, MCJ Scholman… - Transactions of the …, 2023 - direct.mit.edu
Disagreement in natural language annotation has mostly been studied from a perspective of
biases introduced by the annotators and the annotation frameworks. Here, we propose to …

Asking it all: Generating contextualized questions for any semantic role

V Pyatkin, P Roit, J Michael, R Tsarfaty… - arxiv preprint arxiv …, 2021 - arxiv.org
Asking questions about a situation is an inherent step towards understanding it. To this end,
we introduce the task of role question generation, which, given a predicate mention and a …