The'Problem'of Human Label Variation: On Ground Truth in Data, Modeling and Evaluation

B Plank - arxiv preprint arxiv:2211.02570, 2022 - arxiv.org
Human variation in labeling is often considered noise. Annotation projects for machine
learning (ML) aim at minimizing human label variation, with the assumption to maximize …

NLPositionality: Characterizing design biases of datasets and models

S Santy, JT Liang, RL Bras, K Reinecke… - arxiv preprint arxiv …, 2023 - arxiv.org
Design biases in NLP systems, such as performance differences for different populations,
often stem from their creator's positionality, ie, views and lived experiences shaped by …

Two contrasting data annotation paradigms for subjective NLP tasks

P Röttger, B Vidgen, D Hovy… - arxiv preprint arxiv …, 2021 - arxiv.org
Labelled data is the foundation of most natural language processing tasks. However,
labelling data is difficult and there often are diverse valid beliefs about what the correct data …

Dices dataset: Diversity in conversational ai evaluation for safety

L Aroyo, A Taylor, M Diaz, C Homan… - Advances in …, 2023 - proceedings.neurips.cc
Abstract Machine learning approaches often require training and evaluation datasets with a
clear separation between positive and negative examples. This requirement overly …

Is your toxicity my toxicity? exploring the impact of rater identity on toxicity annotation

N Goyal, ID Kivlichan, R Rosen… - Proceedings of the ACM …, 2022 - dl.acm.org
Machine learning models are commonly used to detect toxicity in online conversations.
These models are trained on datasets annotated by human raters. We explore how raters' …

Evaluation gaps in machine learning practice

B Hutchinson, N Rostamzadeh, C Greer… - Proceedings of the …, 2022 - dl.acm.org
Forming a reliable judgement of a machine learning (ML) model's appropriateness for an
application ecosystem is critical for its responsible use, and requires considering a broad …

SemEval-2023 task 11: Learning with disagreements (LeWiDi)

E Leonardelli, A Uma, G Abercrombie… - arxiv preprint arxiv …, 2023 - arxiv.org
NLP datasets annotated with human judgments are rife with disagreements between the
judges. This is especially true for tasks depending on subjective judgments such as …

I beg to differ: how disagreement is handled in the annotation of legal machine learning data sets

D Braun - Artificial intelligence and law, 2024 - Springer
Legal documents, like contracts or laws, are subject to interpretation. Different people can
have different interpretations of the very same document. Large parts of judicial branches all …

Perspectivist approaches to natural language processing: a survey

S Frenda, G Abercrombie, V Basile, A Pedrani… - Language Resources …, 2024 - Springer
Abstract In Artificial Intelligence research, perspectivism is an approach to machine learning
that aims at leveraging data annotated by different individuals in order to model varied …

[HTML][HTML] A comprehensive analysis of the role of artificial intelligence and machine learning in modern digital forensics and incident response

D Dunsin, MC Ghanem, K Ouazzane… - Forensic Science …, 2024 - Elsevier
In the dynamic landscape of digital forensics, the integration of Artificial Intelligence (AI) and
Machine Learning (ML) stands as a transformative technology, poised to amplify the …