A survey on fairness in large language models

Y Li, M Du, R Song, X Wang, Y Wang - arxiv preprint arxiv:2308.10149, 2023 - arxiv.org
Large Language Models (LLMs) have shown powerful performance and development
prospects and are widely deployed in the real world. However, LLMs can capture social …

Fairness in large language models: A taxonomic survey

Z Chu, Z Wang, W Zhang - ACM SIGKDD explorations newsletter, 2024 - dl.acm.org
Large Language Models (LLMs) have demonstrated remarkable success across various
domains. However, despite their promising performance in numerous real-world …

[PDF][PDF] Bias and fairness in large language models: A survey

IO Gallegos, RA Rossi, J Barrow, MM Tanjim… - Computational …, 2024 - direct.mit.edu
Rapid advancements of large language models (LLMs) have enabled the processing,
understanding, and generation of human-like text, with increasing integration into systems …

A survey on large language models for critical societal domains: Finance, healthcare, and law

ZZ Chen, J Ma, X Zhang, N Hao, A Yan… - arxiv preprint arxiv …, 2024 - arxiv.org
In the fast-evolving domain of artificial intelligence, large language models (LLMs) such as
GPT-3 and GPT-4 are revolutionizing the landscapes of finance, healthcare, and law …

Disclosure and mitigation of gender bias in llms

X Dong, Y Wang, PS Yu, J Caverlee - arxiv preprint arxiv:2402.11190, 2024 - arxiv.org
Large Language Models (LLMs) can generate biased responses. Yet previous direct
probing techniques contain either gender mentions or predefined gender stereotypes, which …

Hi guys or hi folks? benchmarking gender-neutral machine translation with the GeNTE corpus

A Piergentili, B Savoldi, D Fucci, M Negri… - arxiv preprint arxiv …, 2023 - arxiv.org
Gender inequality is embedded in our communication practices and perpetuated in
translation technologies. This becomes particularly apparent when translating into …

Mitigating social biases of pre-trained language models via contrastive self-debiasing with double data augmentation

Y Li, M Du, R Song, X Wang, M Sun, Y Wang - Artificial Intelligence, 2024 - Elsevier
Abstract Pre-trained Language Models (PLMs) have been shown to inherit and even amplify
the social biases contained in the training corpus, leading to undesired stereotype in real …

A prompt response to the demand for automatic gender-neutral translation

B Savoldi, A Piergentili, D Fucci, M Negri… - arxiv preprint arxiv …, 2024 - arxiv.org
Gender-neutral translation (GNT) that avoids biased and undue binary assumptions is a
pivotal challenge for the creation of more inclusive translation technologies. Advancements …

Test suites task: Evaluation of gender fairness in MT with MuST-SHE and INES

B Savoldi, M Gaido, M Negri, L Bentivogli - arxiv preprint arxiv …, 2023 - arxiv.org
As part of the WMT-2023" Test suites" shared task, in this paper we summarize the results of
two test suites evaluations: MuST-SHE-WMT23 and INES. By focusing on the en-de and de …

Tackling bias in pre-trained language models: Current trends and under-represented societies

V Yogarajan, G Dobbie, TT Keegan… - arxiv preprint arxiv …, 2023 - arxiv.org
The benefits and capabilities of pre-trained language models (LLMs) in current and future
innovations are vital to any society. However, introducing and using LLMs comes with …