Generative agents: Interactive simulacra of human behavior
Believable proxies of human behavior can empower interactive applications ranging from
immersive environments to rehearsal spaces for interpersonal communication to prototy** …
immersive environments to rehearsal spaces for interpersonal communication to prototy** …
Evaluating the social impact of generative ai systems in systems and society
Generative AI systems across modalities, ranging from text, image, audio, and video, have
broad social impacts, but there exists no official standard for means of evaluating those …
broad social impacts, but there exists no official standard for means of evaluating those …
[PDF][PDF] Ai transparency in the age of llms: A human-centered research roadmap
The rise of powerful large language models (LLMs) brings about tremendous opportunities
for innovation but also looming risks for individuals and society at large. We have reached a …
for innovation but also looming risks for individuals and society at large. We have reached a …
Towards responsible development of generative AI for education: An evaluation-driven approach
A major challenge facing the world is the provision of equitable and universal access to
quality education. Recent advances in generative AI (gen AI) have created excitement about …
quality education. Recent advances in generative AI (gen AI) have created excitement about …
" I'm Not Sure, But...": Examining the Impact of Large Language Models' Uncertainty Expression on User Reliance and Trust
Widely deployed large language models (LLMs) can produce convincing yet incorrect
outputs, potentially misleading users who may rely on them as if they were correct. To …
outputs, potentially misleading users who may rely on them as if they were correct. To …
The art of saying no: Contextual noncompliance in language models
Chat-based language models are designed to be helpful, yet they should not comply with
every user request. While most existing work primarily focuses on refusal of" unsafe" …
every user request. While most existing work primarily focuses on refusal of" unsafe" …
Findings of wassa 2024 shared task on empathy and personality detection in interactions
This paper presents the results of the WASSA 2024 shared task on predicting empathy,
emotion, and personality in conversations and reactions to news articles. Participating teams …
emotion, and personality in conversations and reactions to news articles. Participating teams …
Grounding or guesswork? large language models are presumptive grounders
Effective conversation requires common ground: a shared understanding between the
participants. Common ground, however, does not emerge spontaneously in conversation …
participants. Common ground, however, does not emerge spontaneously in conversation …
From" AI" to Probabilistic Automation: How Does Anthropomorphization of Technical Systems Descriptions Influence Trust?
In this paper we investigate how people's level of trust (as reported through self-assessment)
in so-called “AI”(artificial intelligence) is influenced by anthropomorphizing language in …
in so-called “AI”(artificial intelligence) is influenced by anthropomorphizing language in …
Cognitive Dissonance: Why Do Language Model Outputs Disagree with Internal Representations of Truthfulness?
Neural language models (LMs) can be used to evaluate the truth of factual statements in two
ways: they can be either queried for statement probabilities, or probed for internal …
ways: they can be either queried for statement probabilities, or probed for internal …