The'Problem'of Human Label Variation: On Ground Truth in Data, Modeling and Evaluation
B Plank - arxiv preprint arxiv:2211.02570, 2022 - arxiv.org
Human variation in labeling is often considered noise. Annotation projects for machine
learning (ML) aim at minimizing human label variation, with the assumption to maximize …
learning (ML) aim at minimizing human label variation, with the assumption to maximize …
NLPositionality: Characterizing design biases of datasets and models
Design biases in NLP systems, such as performance differences for different populations,
often stem from their creator's positionality, ie, views and lived experiences shaped by …
often stem from their creator's positionality, ie, views and lived experiences shaped by …
Two contrasting data annotation paradigms for subjective NLP tasks
Labelled data is the foundation of most natural language processing tasks. However,
labelling data is difficult and there often are diverse valid beliefs about what the correct data …
labelling data is difficult and there often are diverse valid beliefs about what the correct data …
Dices dataset: Diversity in conversational ai evaluation for safety
Abstract Machine learning approaches often require training and evaluation datasets with a
clear separation between positive and negative examples. This requirement overly …
clear separation between positive and negative examples. This requirement overly …
Is your toxicity my toxicity? exploring the impact of rater identity on toxicity annotation
Machine learning models are commonly used to detect toxicity in online conversations.
These models are trained on datasets annotated by human raters. We explore how raters' …
These models are trained on datasets annotated by human raters. We explore how raters' …
Evaluation gaps in machine learning practice
Forming a reliable judgement of a machine learning (ML) model's appropriateness for an
application ecosystem is critical for its responsible use, and requires considering a broad …
application ecosystem is critical for its responsible use, and requires considering a broad …
SemEval-2023 task 11: Learning with disagreements (LeWiDi)
NLP datasets annotated with human judgments are rife with disagreements between the
judges. This is especially true for tasks depending on subjective judgments such as …
judges. This is especially true for tasks depending on subjective judgments such as …
I beg to differ: how disagreement is handled in the annotation of legal machine learning data sets
D Braun - Artificial intelligence and law, 2024 - Springer
Legal documents, like contracts or laws, are subject to interpretation. Different people can
have different interpretations of the very same document. Large parts of judicial branches all …
have different interpretations of the very same document. Large parts of judicial branches all …
Perspectivist approaches to natural language processing: a survey
Abstract In Artificial Intelligence research, perspectivism is an approach to machine learning
that aims at leveraging data annotated by different individuals in order to model varied …
that aims at leveraging data annotated by different individuals in order to model varied …
[HTML][HTML] A comprehensive analysis of the role of artificial intelligence and machine learning in modern digital forensics and incident response
In the dynamic landscape of digital forensics, the integration of Artificial Intelligence (AI) and
Machine Learning (ML) stands as a transformative technology, poised to amplify the …
Machine Learning (ML) stands as a transformative technology, poised to amplify the …