Ai alignment: A comprehensive survey
AI alignment aims to make AI systems behave in line with human intentions and values. As
AI systems grow more capable, the potential large-scale risks associated with misaligned AI …
AI systems grow more capable, the potential large-scale risks associated with misaligned AI …
A survey on fairness in large language models
Large language models (LLMs) have shown powerful performance and development
prospect and are widely deployed in the real world. However, LLMs can capture social …
prospect and are widely deployed in the real world. However, LLMs can capture social …
Gpt-4 technical report
We report the development of GPT-4, a large-scale, multimodal model which can accept
image and text inputs and produce text outputs. While less capable than humans in many …
image and text inputs and produce text outputs. While less capable than humans in many …
Five sources of bias in natural language processing
Recently, there has been an increased interest in demographically grounded bias in natural
language processing (NLP) applications. Much of the recent work has focused on describing …
language processing (NLP) applications. Much of the recent work has focused on describing …
Language (technology) is power: A critical survey of" bias" in nlp
We survey 146 papers analyzing" bias" in NLP systems, finding that their motivations are
often vague, inconsistent, and lacking in normative reasoning, despite the fact that …
often vague, inconsistent, and lacking in normative reasoning, despite the fact that …
BBQ: A hand-built bias benchmark for question answering
It is well documented that NLP models learn social biases, but little work has been done on
how these biases manifest in model outputs for applied tasks like question answering (QA) …
how these biases manifest in model outputs for applied tasks like question answering (QA) …
Bold: Dataset and metrics for measuring biases in open-ended language generation
Recent advances in deep learning techniques have enabled machines to generate
cohesive open-ended text when prompted with a sequence of words as context. While these …
cohesive open-ended text when prompted with a sequence of words as context. While these …
Superglue: A stickier benchmark for general-purpose language understanding systems
In the last year, new models and methods for pretraining and transfer learning have driven
striking performance improvements across a range of language understanding tasks. The …
striking performance improvements across a range of language understanding tasks. The …
CrowS-pairs: A challenge dataset for measuring social biases in masked language models
Pretrained language models, especially masked language models (MLMs) have seen
success across many NLP tasks. However, there is ample evidence that they use the cultural …
success across many NLP tasks. However, there is ample evidence that they use the cultural …
BERT for coreference resolution: Baselines and analysis
We apply BERT to coreference resolution, achieving strong improvements on the OntoNotes
(+ 3.9 F1) and GAP (+ 11.5 F1) benchmarks. A qualitative analysis of model predictions …
(+ 3.9 F1) and GAP (+ 11.5 F1) benchmarks. A qualitative analysis of model predictions …