Safetyprompts: a systematic review of open datasets for evaluating and improving large language model safety

P Röttger, F Pernisi, B Vidgen, D Hovy - arxiv preprint arxiv:2404.05399, 2024 - arxiv.org
The last two years have seen a rapid growth in concerns around the safety of large
language models (LLMs). Researchers and practitioners have met these concerns by …

Culturally aware and adapted nlp: A taxonomy and a survey of the state of the art

CC Liu, I Gurevych, A Korhonen - arxiv preprint arxiv:2406.03930, 2024 - arxiv.org
The surge of interest in culturally aware and adapted Natural Language Processing (NLP)
has inspired much recent research. However, the lack of common understanding of the …

Kmmlu: Measuring massive multitask language understanding in korean

G Son, H Lee, S Kim, S Kim, N Muennighoff… - arxiv preprint arxiv …, 2024 - arxiv.org
We propose KMMLU, a new Korean benchmark with 35,030 expert-level multiple-choice
questions across 45 subjects ranging from humanities to STEM. While prior Korean …

Culturebank: An online community-driven knowledge base towards culturally aware language technologies

W Shi, R Li, Y Zhang, C Ziems, R Horesh… - arxiv preprint arxiv …, 2024 - arxiv.org
To enhance language models' cultural awareness, we design a generalizable pipeline to
construct cultural knowledge bases from different online communities on a massive scale …

Social bias evaluation for large language models requires prompt variations

R Hida, M Kaneko, N Okazaki - arxiv preprint arxiv:2407.03129, 2024 - arxiv.org
Warning: This paper contains examples of stereotypes and biases. Large Language Models
(LLMs) exhibit considerable social biases, and various studies have tried to evaluate and …

CaLMQA: Exploring culturally specific long-form question answering across 23 languages

S Arora, M Karpinska, HT Chen, I Bhattacharjee… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are used for long-form question answering (LFQA), which
requires them to generate paragraph-length answers to complex questions. While LFQA has …

Exploring cross-cultural differences in English hate speech annotations: From dataset construction to analysis

N Lee, C Jung, J Myung, J **… - arxiv preprint arxiv …, 2023 - arxiv.org
Warning: this paper contains content that may be offensive or upsetting. Most hate speech
datasets neglect the cultural diversity within a single language, resulting in a critical …

Survey of cultural awareness in language models: Text and beyond

S Pawar, J Park, J **, A Arora, J Myung… - arxiv preprint arxiv …, 2024 - arxiv.org
Large-scale deployment of large language models (LLMs) in various applications, such as
chatbots and virtual assistants, requires LLMs to be culturally sensitive to the user to ensure …

Ask LLMs Directly," What shapes your bias?": Measuring Social Bias in Large Language Models

J Shin, H Song, H Lee, S Jeong, JC Park - arxiv preprint arxiv:2406.04064, 2024 - arxiv.org
Social bias is shaped by the accumulation of social perceptions towards targets across
various demographic identities. To fully understand such social bias in large language …

Do Multilingual Large Language Models Mitigate Stereotype Bias?

S Nie, M Fromm, C Welch, R Görge, A Karimi… - arxiv preprint arxiv …, 2024 - arxiv.org
While preliminary findings indicate that multilingual LLMs exhibit reduced bias compared to
monolingual ones, a comprehensive understanding of the effect of multilingual training on …