The ethics of advanced ai assistants

I Gabriel, A Manzini, G Keeling, LA Hendricks… - arxiv preprint arxiv …, 2024 - arxiv.org
This paper focuses on the opportunities and the ethical and societal risks posed by
advanced AI assistants. We define advanced AI assistants as artificial agents with natural …

Evaluating frontier models for dangerous capabilities

M Phuong, M Aitchison, E Catt, S Cogan… - arxiv preprint arxiv …, 2024 - arxiv.org
To understand the risks posed by a new AI system, we must understand what it can and
cannot do. Building on prior work, we introduce a programme of new" dangerous capability" …

[PDF][PDF] Assessing audio hallucination in large multimodal models

S Hanamaki, N Kirishima, S Narumi - 2024 - osf.io
Speech recognition systems have become increasingly integral in various applications, from
virtual assistants to automated transcription services, necessitating the development of …

[PDF][PDF] Llms achieve adult human performance on higher-order theory of mind tasks

W Street, JO Siy, G Keeling, A Baranes… - arxiv preprint arxiv …, 2024 - academia.edu
This paper examines the extent to which large language models (LLMs) have developed
higher-order theory of mind (ToM); the human ability to reason about multiple mental and …

Epistemic injustice in generative ai

J Kay, A Kasirzadeh, S Mohamed - … of the AAAI/ACM Conference on AI …, 2024 - ojs.aaai.org
This paper investigates how generative AI can potentially undermine the integrity of
collective knowledge and the processes we rely on to acquire, assess, and trust information …

Safety case template for frontier AI: A cyber inability argument

A Goemans, MD Buhl, J Schuett, T Korbak… - arxiv preprint arxiv …, 2024 - arxiv.org
Frontier artificial intelligence (AI) systems pose increasing risks to society, making it
essential for developers to provide assurances about their safety. One approach to offering …

The Code That Binds Us: Navigating the Appropriateness of Human-AI Assistant Relationships

A Manzini, G Keeling, L Alberts, S Vallor… - Proceedings of the …, 2024 - ojs.aaai.org
The development of increasingly agentic and human-like AI assistants, capable of
performing a wide range of tasks on user's behalf over time, has sparked heightened interest …

[PDF][PDF] Personhood credentials: Artificial intelligence and the value of privacy-preserving tools to distinguish who is real online

S Adler, Z Hitzig, S Jain, C Brewer… - arxiv preprint arxiv …, 2024 - openreview.net
Anonymity is an important principle online. However, malicious actors have long used
misleading identities to conduct fraud, spread disinformation, and carry out other deceptive …

Lies, Damned Lies, and Distributional Language Statistics: Persuasion and Deception with Large Language Models

CR Jones, BK Bergen - arxiv preprint arxiv:2412.17128, 2024 - arxiv.org
Large Language Models (LLMs) can generate content that is as persuasive as human-
written text and appear capable of selectively producing deceptive outputs. These …

Grounding fallacies misrepresenting scientific publications in evidence

M Glockner, Y Hou, P Nakov, I Gurevych - arxiv preprint arxiv:2408.12812, 2024 - arxiv.org
Health-related misinformation claims often falsely cite a credible biomedical publication as
evidence, which superficially appears to support the false claim. The publication does not …