Typology of risks of generative text-to-image models

C Bird, E Ungless, A Kasirzadeh - Proceedings of the 2023 AAAI/ACM …, 2023 - dl.acm.org
This paper investigates the direct risks and harms associated with modern text-to-image
generative models, such as DALL-E and Midjourney, through a comprehensive literature …

The illusion of artificial inclusion

W Agnew, AS Bergman, J Chien, M Díaz… - Proceedings of the …, 2024 - dl.acm.org
Human participants play a central role in the development of modern artificial intelligence
(AI) technology, in psychological science, and in user research. Recent advances in …

Gender bias and stereotypes in large language models

H Kotek, R Dockum, D Sun - Proceedings of the ACM collective …, 2023 - dl.acm.org
Large Language Models (LLMs) have made substantial progress in the past several months,
shattering state-of-the-art benchmarks in many domains. This paper investigates LLMs' …

Foundational challenges in assuring alignment and safety of large language models

U Anwar, A Saparov, J Rando, D Paleka… - arxiv preprint arxiv …, 2024 - arxiv.org
This work identifies 18 foundational challenges in assuring the alignment and safety of large
language models (LLMs). These challenges are organized into three different categories …

Aya dataset: An open-access collection for multilingual instruction tuning

S Singh, F Vargus, D Dsouza, BF Karlsson… - arxiv preprint arxiv …, 2024 - arxiv.org
Datasets are foundational to many breakthroughs in modern artificial intelligence. Many
recent achievements in the space of natural language processing (NLP) can be attributed to …

Do llms exhibit human-like response biases? a case study in survey design

L Tjuatja, V Chen, T Wu, A Talwalkwar… - Transactions of the …, 2024 - direct.mit.edu
One widely cited barrier to the adoption of LLMs as proxies for humans in subjective tasks is
their sensitivity to prompt wording—but interestingly, humans also display sensitivities to …

Collective constitutional ai: Aligning a language model with public input

S Huang, D Siddarth, L Lovitt, TI Liao… - Proceedings of the …, 2024 - dl.acm.org
There is growing consensus that language model (LM) developers should not be the sole
deciders of LM behavior, creating a need for methods that enable the broader public to …

Having beer after prayer? measuring cultural bias in large language models

T Naous, MJ Ryan, A Ritter, W Xu - arxiv preprint arxiv:2305.14456, 2023 - arxiv.org
As the reach of large language models (LMs) expands globally, their ability to cater to
diverse cultural contexts becomes crucial. Despite advancements in multilingual …

The prism alignment project: What participatory, representative and individualised human feedback reveals about the subjective and multicultural alignment of large …

HR Kirk, A Whitefield, P Röttger, A Bean… - arxiv preprint arxiv …, 2024 - arxiv.org
Human feedback plays a central role in the alignment of Large Language Models (LLMs).
However, open questions remain about the methods (how), domains (where), people (who) …

Does writing with language models reduce content diversity?

V Padmakumar, H He - arxiv preprint arxiv:2309.05196, 2023 - arxiv.org
Large language models (LLMs) have led to a surge in collaborative writing with model
assistance. As different users incorporate suggestions from the same model, there is a risk of …