Learning from disagreement: A survey
Abstract Many tasks in Natural Language Processing (NLP) and Computer Vision (CV) offer
evidence that humans disagree, from objective tasks such as part-of-speech tagging to more …
evidence that humans disagree, from objective tasks such as part-of-speech tagging to more …
A checklist to combat cognitive biases in crowdsourcing
Recent research has demonstrated that cognitive biases such as the confirmation bias or the
anchoring effect can negatively affect the quality of crowdsourced data. In practice, however …
anchoring effect can negatively affect the quality of crowdsourced data. In practice, however …
[PDF][PDF] Hacred: A large-scale relation extraction dataset toward hard cases in practical applications
Relation extraction (RE) is an essential topic in natural language processing and has
attracted extensive attention. Current RE approaches achieve fantastic results on common …
attracted extensive attention. Current RE approaches achieve fantastic results on common …
Efficient elicitation approaches to estimate collective crowd answers
When crowdsourcing the creation of machine learning datasets, statistical distributions that
capture diverse answers can represent ambiguous data better than a single best answer …
capture diverse answers can represent ambiguous data better than a single best answer …
A citizen science approach for analyzing social media with crowdsourcing
Social media have the potential to provide timely information about emergency situations
and sudden events. However, finding relevant information among the millions of posts being …
and sudden events. However, finding relevant information among the millions of posts being …
Establishing annotation quality in multi-label annotations
In many linguistic fields requiring annotated data, multiple interpretations of a single item are
possible. Multi-label annotations more accurately reflect this possibility. However, allowing …
possible. Multi-label annotations more accurately reflect this possibility. However, allowing …
Human rationales as attribution priors for explainable stance detection
S Jayaram, E Allaway - Proceedings of the 2021 Conference on …, 2021 - aclanthology.org
As NLP systems become better at detecting opinions and beliefs from text, it is important to
ensure not only that models are accurate but also that they arrive at their predictions in ways …
ensure not only that models are accurate but also that they arrive at their predictions in ways …
Collect, measure, repeat: Reliability factors for responsible AI data collection
The rapid entry of machine learning approaches in our daily activities and high-stakes
domains demands transparency and scrutiny of their fairness and reliability. To help gauge …
domains demands transparency and scrutiny of their fairness and reliability. To help gauge …
Human-annotated rationales and explainable text classification: a survey
Asking annotators to explain “why” they labeled an instance yields annotator rationales:
natural language explanations that provide reasons for classifications. In this work, we …
natural language explanations that provide reasons for classifications. In this work, we …
Capturing perspectives of crowdsourced annotators in subjective learning tasks
In most classification models, it has been assumed to have a single ground truth label for
each data point. However, subjective tasks like toxicity classification can lead to genuine …
each data point. However, subjective tasks like toxicity classification can lead to genuine …