Augmented language models: a survey

G Mialon, R Dessì, M Lomeli, C Nalmpantis… - arxiv preprint arxiv …, 2023 - arxiv.org
This survey reviews works in which language models (LMs) are augmented with reasoning
skills and the ability to use tools. The former is defined as decomposing a potentially …

Towards revealing the mystery behind chain of thought: a theoretical perspective

G Feng, B Zhang, Y Gu, H Ye, D He… - Advances in Neural …, 2023 - proceedings.neurips.cc
Recent studies have discovered that Chain-of-Thought prompting (CoT) can dramatically
improve the performance of Large Language Models (LLMs), particularly when dealing with …

Transformers as statisticians: Provable in-context learning with in-context algorithm selection

Y Bai, F Chen, H Wang, C **ong… - Advances in neural …, 2023 - proceedings.neurips.cc
Neural sequence models based on the transformer architecture have demonstrated
remarkable\emph {in-context learning}(ICL) abilities, where they can perform new tasks …

Transformers learn to implement preconditioned gradient descent for in-context learning

K Ahn, X Cheng, H Daneshmand… - Advances in Neural …, 2023 - proceedings.neurips.cc
Several recent works demonstrate that transformers can implement algorithms like gradient
descent. By a careful construction of weights, these works show that multiple layers of …

Foundation models for decision making: Problems, methods, and opportunities

S Yang, O Nachum, Y Du, J Wei, P Abbeel… - arxiv preprint arxiv …, 2023 - arxiv.org
Foundation models pretrained on diverse data at scale have demonstrated extraordinary
capabilities in a wide range of vision and language tasks. When such models are deployed …

Transformers learn shortcuts to automata

B Liu, JT Ash, S Goel, A Krishnamurthy… - arxiv preprint arxiv …, 2022 - arxiv.org
Algorithmic reasoning requires capabilities which are most naturally understood through
recurrent models of computation, like the Turing machine. However, Transformer models …

[PDF][PDF] Chain of thought empowers transformers to solve inherently serial problems

Z Li, H Liu, D Zhou, T Ma - arxiv preprint arxiv:2402.12875, 2024 - academia.edu
Instructing the model to generate a sequence of intermediate steps, aka, a chain of thought
(CoT), is a highly effective method to improve the accuracy of large language models (LLMs) …

Teaching arithmetic to small transformers

N Lee, K Sreenivasan, JD Lee, K Lee… - arxiv preprint arxiv …, 2023 - arxiv.org
Large language models like GPT-4 exhibit emergent capabilities across general-purpose
tasks, such as basic arithmetic, when trained on extensive text data, even though these tasks …

In-context convergence of transformers

Y Huang, Y Cheng, Y Liang - arxiv preprint arxiv:2310.05249, 2023 - arxiv.org
Transformers have recently revolutionized many domains in modern machine learning and
one salient discovery is their remarkable in-context learning capability, where models can …

Learning transformer programs

D Friedman, A Wettig, D Chen - Advances in Neural …, 2023 - proceedings.neurips.cc
Recent research in mechanistic interpretability has attempted to reverse-engineer
Transformer models by carefully inspecting network weights and activations. However, these …