Ai alignment: A comprehensive survey

J Ji, T Qiu, B Chen, B Zhang, H Lou, K Wang… - arxiv preprint arxiv …, 2023 - arxiv.org
AI alignment aims to make AI systems behave in line with human intentions and values. As
AI systems grow more capable, so do risks from misalignment. To provide a comprehensive …

Natural language reasoning, a survey

F Yu, H Zhang, P Tiwari, B Wang - ACM Computing Surveys, 2024 - dl.acm.org
This survey article proposes a clearer view of Natural Language Reasoning (NLR) in the
field of Natural Language Processing (NLP), both conceptually and practically …

Large language models can be easily distracted by irrelevant context

F Shi, X Chen, K Misra, N Scales… - International …, 2023 - proceedings.mlr.press
Large language models have achieved impressive performance on various natural
language processing tasks. However, so far they have been evaluated primarily on …

Scaling data-constrained language models

N Muennighoff, A Rush, B Barak… - Advances in …, 2023 - proceedings.neurips.cc
The current trend of scaling language models involves increasing both parameter count and
training dataset size. Extrapolating this trend suggests that training dataset size may soon be …

Holistic evaluation of language models

P Liang, R Bommasani, T Lee, D Tsipras… - arxiv preprint arxiv …, 2022 - arxiv.org
Language models (LMs) are becoming the foundation for almost all major language
technologies, but their capabilities, limitations, and risks are not well understood. We present …

Emergent world representations: Exploring a sequence model trained on a synthetic task

K Li, AK Hopkins, D Bau, F Viégas, H Pfister… - ICLR, 2023 - par.nsf.gov
Language models show a surprising range of capabilities, but the source of their apparent
competence is unclear. Do these networks just memorize a collection of surface statistics, or …

Language models meet world models: Embodied experiences enhance language models

J **ang, T Tao, Y Gu, T Shu, Z Wang… - Advances in neural …, 2023 - proceedings.neurips.cc
While large language models (LMs) have shown remarkable capabilities across numerous
tasks, they often struggle with simple reasoning and planning in physical environments …

Selection-inference: Exploiting large language models for interpretable logical reasoning

A Creswell, M Shanahan, I Higgins - arxiv preprint arxiv:2205.09712, 2022 - arxiv.org
Large language models (LLMs) have been shown to be capable of impressive few-shot
generalisation to new tasks. However, they still tend to perform poorly on multi-step logical …

Same task, more tokens: the impact of input length on the reasoning performance of large language models

M Levy, A Jacoby, Y Goldberg - arxiv preprint arxiv:2402.14848, 2024 - arxiv.org
This paper explores the impact of extending input lengths on the capabilities of Large
Language Models (LLMs). Despite LLMs advancements in recent times, their performance …

The unreliability of explanations in few-shot prompting for textual reasoning

X Ye, G Durrett - Advances in neural information processing …, 2022 - proceedings.neurips.cc
Does prompting a large language model (LLM) like GPT-3 with explanations improve in-
context learning? We study this question on two NLP tasks that involve reasoning over text …