Teach me to explain: A review of datasets for explainable natural language processing

S Wiegreffe, A Marasović - arxiv preprint arxiv:2102.12060, 2021 - arxiv.org
Explainable NLP (ExNLP) has increasingly focused on collecting human-annotated textual
explanations. These explanations are used downstream in three ways: as data …

The'Problem'of Human Label Variation: On Ground Truth in Data, Modeling and Evaluation

B Plank - arxiv preprint arxiv:2211.02570, 2022 - arxiv.org
Human variation in labeling is often considered noise. Annotation projects for machine
learning (ML) aim at minimizing human label variation, with the assumption to maximize …

Bridging the gap: A survey on integrating (human) feedback for natural language generation

P Fernandes, A Madaan, E Liu, A Farinhas… - Transactions of the …, 2023 - direct.mit.edu
Natural language generation has witnessed significant advancements due to the training of
large language models on vast internet-scale datasets. Despite these advancements, there …

Learning from disagreement: A survey

AN Uma, T Fornaciari, D Hovy, S Paun, B Plank… - Journal of Artificial …, 2021 - jair.org
Abstract Many tasks in Natural Language Processing (NLP) and Computer Vision (CV) offer
evidence that humans disagree, from objective tasks such as part-of-speech tagging to more …

The prism alignment project: What participatory, representative and individualised human feedback reveals about the subjective and multicultural alignment of large …

HR Kirk, A Whitefield, P Röttger, A Bean… - arxiv preprint arxiv …, 2024 - arxiv.org
Human feedback plays a central role in the alignment of Large Language Models (LLMs).
However, open questions remain about the methods (how), domains (where), people (who) …

Minicheck: Efficient fact-checking of llms on grounding documents

L Tang, P Laban, G Durrett - arxiv preprint arxiv:2404.10774, 2024 - arxiv.org
Recognizing if LLM output can be grounded in evidence is central to many tasks in NLP:
retrieval-augmented generation, summarization, document-grounded dialogue, and more …

We're afraid language models aren't modeling ambiguity

A Liu, Z Wu, J Michael, A Suhr, P West, A Koller… - arxiv preprint arxiv …, 2023 - arxiv.org
Ambiguity is an intrinsic feature of natural language. Managing ambiguity is a key part of
human language understanding, allowing us to anticipate misunderstanding as …

The PRISM alignment dataset: What participatory, representative and individualised human feedback reveals about the subjective and multicultural alignment of large …

HR Kirk, A Whitefield, P Rottger… - Advances in …, 2025 - proceedings.neurips.cc
Human feedback is central to the alignment of Large Language Models (LLMs). However,
open questions remain about the methods (how), domains (where), people (who) and …

Culturally aware natural language inference

J Huang, D Yang - Findings of the Association for Computational …, 2023 - aclanthology.org
Humans produce and consume language in a particular cultural context, which includes
knowledge about specific norms and practices. A listener's awareness of the cultural context …

Investigating reasons for disagreement in natural language inference

NJ Jiang, MC Marneffe - Transactions of the Association for …, 2022 - direct.mit.edu
We investigate how disagreement in natural language inference (NLI) annotation arises. We
developed a taxonomy of disagreement sources with 10 categories spanning 3 high-level …