Take Caution in Using LLMs as Human Surrogates: Scylla Ex Machina

Y Gao, D Lee, G Burtch, S Fazelpour - arxiv preprint arxiv:2410.19599, 2024 - arxiv.org
Recent studies suggest large language models (LLMs) can exhibit human-like reasoning,
aligning with human behavior in economic experiments, surveys, and political discourse …

Testing the limits of fine-tuning to improve reasoning in vision language models

LMS Buschoff, K Voudouris, E Akata, M Bethge… - arxiv preprint arxiv …, 2025 - arxiv.org
Pre-trained vision language models still fall short of human visual cognition. In an effort to
improve visual cognition and align models with human behavior, we introduce visual stimuli …

NeuroAI for AI Safety

P Mineault, N Zanichelli, JZ Peng, A Arkhipov… - arxiv preprint arxiv …, 2024 - arxiv.org
As AI systems become increasingly powerful, the need for safe AI has become more
pressing. Humans are an attractive model for AI safety: as the only known agents capable of …

The potential--and the pitfalls--of using pre-trained language models as cognitive science theories

RS Shah, S Varma - arxiv preprint arxiv:2501.12651, 2025 - arxiv.org
Many studies have evaluated the cognitive alignment of Pre-trained Language Models
(PLMs), ie, their correspondence to adult performance across a range of cognitive domains …

" All that Glitters": Approaches to Evaluations with Unreliable Model and Human Annotations

M Hardy - arxiv preprint arxiv:2411.15634, 2024 - arxiv.org
" Gold" and" ground truth" human-mediated labels have error. The effects of this error can
escape commonly reported metrics of label quality or obscure questions of accuracy, bias …

[PDF][PDF] Мнемические эффекты при опосредствовании деятельности цифровыми технологиями

ДВ Ушаков - ipran.ru
Стремительное развитие и внедрение в жизнь цифровых технологий не только
трансформирует общество, но и ставит новые вопросы перед академической …