Reasoning with large language models, a survey

A Plaat, A Wong, S Verberne, J Broekens… - arxiv preprint arxiv …, 2024 - arxiv.org
Scaling up language models to billions of parameters has opened up possibilities for in-
context learning, allowing instruction tuning and few-shot learning on tasks that the model …

Training large language models to reason in a continuous latent space

S Hao, S Sukhbaatar, DJ Su, X Li, Z Hu… - arxiv preprint arxiv …, 2024 - arxiv.org
Large language models (LLMs) are restricted to reason in the" language space", where they
typically express the reasoning process with a chain-of-thought (CoT) to solve a complex …

Dualformer: Controllable fast and slow thinking by learning with randomized reasoning traces

DJ Su, S Sukhbaatar, M Rabbat, Y Tian… - The Thirteenth …, 2024 - openreview.net
In human cognition theory, human thinking is governed by two systems: the fast and intuitive
System 1 and the slower but more deliberative System 2. Recent studies have shown that …

Longwriter: Unleashing 10,000+ word generation from long context llms

Y Bai, J Zhang, X Lv, L Zheng, S Zhu, L Hou… - arxiv preprint arxiv …, 2024 - arxiv.org
Current long context large language models (LLMs) can process inputs up to 100,000
tokens, yet struggle to generate outputs exceeding even a modest length of 2,000 words …

Thinking llms: General instruction following with thought generation

T Wu, J Lan, W Yuan, J Jiao, J Weston… - arxiv preprint arxiv …, 2024 - arxiv.org
LLMs are typically trained to answer user questions or follow instructions similarly to how
human experts respond. However, in the standard alignment framework they lack the basic …

[PDF][PDF] Numinamath: The largest public dataset in ai4maths with 860k pairs of competition math problems and solutions

J Li, E Beeching, L Tunstall, B Lipkin… - Hugging Face …, 2024 - faculty.bicmr.pku.edu.cn
Numina is an open AI4Maths initiative dedicated to advancing both artificial and human
intelligence in the field of mathematics. In this paper, we present the NuminaMath dataset, a …

An Overview of Large Language Models for Statisticians

W Ji, W Yuan, E Getzen, K Cho, MI Jordan… - arxiv preprint arxiv …, 2025 - arxiv.org
Large Language Models (LLMs) have emerged as transformative tools in artificial
intelligence (AI), exhibiting remarkable capabilities across diverse tasks such as text …

Disentangling memory and reasoning ability in large language models

M **, W Luo, S Cheng, X Wang, W Hua… - arxiv preprint arxiv …, 2024 - arxiv.org
Large Language Models (LLMs) have demonstrated strong performance in handling
complex tasks requiring both extensive knowledge and reasoning abilities. However, the …

VipAct: Visual-perception enhancement via specialized vlm agent collaboration and tool-use

Z Zhang, R Rossi, T Yu, F Dernoncourt… - arxiv preprint arxiv …, 2024 - arxiv.org
While vision-language models (VLMs) have demonstrated remarkable performance across
various tasks combining textual and visual information, they continue to struggle with fine …

Think More, Hallucinate Less: Mitigating Hallucinations via Dual Process of Fast and Slow Thinking

X Cheng, J Li, WX Zhao, JR Wen - arxiv preprint arxiv:2501.01306, 2025 - arxiv.org
Large language models (LLMs) demonstrate exceptional capabilities, yet still face the
hallucination issue. Typical text generation approaches adopt an auto-regressive generation …