Societal biases in language generation: Progress and challenges

E Sheng, KW Chang, P Natarajan, N Peng - arxiv preprint arxiv …, 2021 - arxiv.org
Technology for language generation has advanced rapidly, spurred by advancements in pre-
training large models on massive amounts of data and the need for intelligent agents to …

A survey on fairness in large language models

Y Li, M Du, R Song, X Wang, Y Wang - arxiv preprint arxiv:2308.10149, 2023 - arxiv.org
Large language models (LLMs) have shown powerful performance and development
prospect and are widely deployed in the real world. However, LLMs can capture social …

Beyond the imitation game: Quantifying and extrapolating the capabilities of language models

A Srivastava, A Rastogi, A Rao, AAM Shoeb… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models demonstrate both quantitative improvement and new qualitative
capabilities with increasing scale. Despite their potentially transformative impact, these new …

On the opportunities and risks of foundation models

R Bommasani, DA Hudson, E Adeli, R Altman… - arxiv preprint arxiv …, 2021 - arxiv.org
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …

A survey of data augmentation approaches for NLP

SY Feng, V Gangal, J Wei, S Chandar… - arxiv preprint arxiv …, 2021 - arxiv.org
Data augmentation has recently seen increased interest in NLP due to more work in low-
resource domains, new tasks, and the popularity of large-scale neural networks that require …

Language (technology) is power: A critical survey of" bias" in nlp

SL Blodgett, S Barocas, H Daumé III… - arxiv preprint arxiv …, 2020 - arxiv.org
We survey 146 papers analyzing" bias" in NLP systems, finding that their motivations are
often vague, inconsistent, and lacking in normative reasoning, despite the fact that …

Persistent anti-muslim bias in large language models

A Abid, M Farooqi, J Zou - Proceedings of the 2021 AAAI/ACM …, 2021 - dl.acm.org
It has been observed that large-scale language models capture undesirable societal biases,
eg relating to race and gender; yet religious bias has been relatively unexplored. We …

Learning the difference that makes a difference with counterfactually-augmented data

D Kaushik, E Hovy, ZC Lipton - arxiv preprint arxiv:1909.12434, 2019 - arxiv.org
Despite alarm over the reliance of machine learning systems on so-called spurious patterns,
the term lacks coherent meaning in standard statistical frameworks. However, the language …

A multiscale visualization of attention in the transformer model

J Vig - arxiv preprint arxiv:1906.05714, 2019 - arxiv.org
The Transformer is a sequence model that forgoes traditional recurrent architectures in favor
of a fully attention-based approach. Besides improving performance, an advantage of using …

Trustworthy LLMs: A survey and guideline for evaluating large language models' alignment

Y Liu, Y Yao, JF Ton, X Zhang, RGH Cheng… - arxiv preprint arxiv …, 2023 - arxiv.org
Ensuring alignment, which refers to making models behave in accordance with human
intentions [1, 2], has become a critical task before deploying large language models (LLMs) …