Whose opinions do language models reflect?
Abstract Language models (LMs) are increasingly being used in open-ended contexts,
where the opinions they reflect in response to subjective queries can have a profound …
where the opinions they reflect in response to subjective queries can have a profound …
Auditing large language models: a three-layered approach
Large language models (LLMs) represent a major advance in artificial intelligence (AI)
research. However, the widespread use of LLMs is also coupled with significant ethical and …
research. However, the widespread use of LLMs is also coupled with significant ethical and …
Toxigen: A large-scale machine-generated dataset for adversarial and implicit hate speech detection
Toxic language detection systems often falsely flag text that contains minority group
mentions as toxic, as those groups are often the targets of online hate. Such over-reliance …
mentions as toxic, as those groups are often the targets of online hate. Such over-reliance …
From pretraining data to language models to downstream tasks: Tracking the trails of political biases leading to unfair NLP models
Language models (LMs) are pretrained on diverse data sources, including news, discussion
forums, books, and online encyclopedias. A significant portion of this data includes opinions …
forums, books, and online encyclopedias. A significant portion of this data includes opinions …
Towards measuring the representation of subjective global opinions in language models
Large language models (LLMs) may not equitably represent diverse global perspectives on
societal issues. In this paper, we develop a quantitative framework to evaluate whose …
societal issues. In this paper, we develop a quantitative framework to evaluate whose …
Evaluating the social impact of generative ai systems in systems and society
Generative AI systems across modalities, ranging from text, image, audio, and video, have
broad social impacts, but there exists no official standard for means of evaluating those …
broad social impacts, but there exists no official standard for means of evaluating those …
Bridging the gap: A survey on integrating (human) feedback for natural language generation
Natural language generation has witnessed significant advancements due to the training of
large language models on vast internet-scale datasets. Despite these advancements, there …
large language models on vast internet-scale datasets. Despite these advancements, there …
A survey of language model confidence estimation and calibration
Language models (LMs) have demonstrated remarkable capabilities across a wide range of
tasks in various domains. Despite their impressive performance, the reliability of their output …
tasks in various domains. Despite their impressive performance, the reliability of their output …
The'Problem'of Human Label Variation: On Ground Truth in Data, Modeling and Evaluation
B Plank - arxiv preprint arxiv:2211.02570, 2022 - arxiv.org
Human variation in labeling is often considered noise. Annotation projects for machine
learning (ML) aim at minimizing human label variation, with the assumption to maximize …
learning (ML) aim at minimizing human label variation, with the assumption to maximize …
Two contrasting data annotation paradigms for subjective NLP tasks
Labelled data is the foundation of most natural language processing tasks. However,
labelling data is difficult and there often are diverse valid beliefs about what the correct data …
labelling data is difficult and there often are diverse valid beliefs about what the correct data …