Dissociating language and thought in large language models
Large language models (LLMs) have come closest among all models to date to mastering
human language, yet opinions about their linguistic and cognitive capabilities remain split …
human language, yet opinions about their linguistic and cognitive capabilities remain split …
Code as policies: Language model programs for embodied control
Large language models (LLMs) trained on code-completion have been shown to be capable
of synthesizing simple Python programs from docstrings [1]. We find that these code-writing …
of synthesizing simple Python programs from docstrings [1]. We find that these code-writing …
Human-like systematic generalization through a meta-learning neural network
The power of human language and thought arises from systematic compositionality—the
algebraic ability to understand and produce novel combinations from known components …
algebraic ability to understand and produce novel combinations from known components …
Beyond the imitation game: Quantifying and extrapolating the capabilities of language models
Language models demonstrate both quantitative improvement and new qualitative
capabilities with increasing scale. Despite their potentially transformative impact, these new …
capabilities with increasing scale. Despite their potentially transformative impact, these new …
Sugarcrepe: Fixing hackable benchmarks for vision-language compositionality
In the last year alone, a surge of new benchmarks to measure $\textit {compositional} $
understanding of vision-language models have permeated the machine learning ecosystem …
understanding of vision-language models have permeated the machine learning ecosystem …
Socratic models: Composing zero-shot multimodal reasoning with language
Large pretrained (eg," foundation") models exhibit distinct capabilities depending on the
domain of data they are trained on. While these domains are generic, they may only barely …
domain of data they are trained on. While these domains are generic, they may only barely …
Measuring and narrowing the compositionality gap in language models
We investigate the ability of language models to perform compositional reasoning tasks
where the overall solution depends on correctly composing the answers to sub-problems …
where the overall solution depends on correctly composing the answers to sub-problems …
From machine learning to robotics: Challenges and opportunities for embodied intelligence
Machine learning has long since become a keystone technology, accelerating science and
applications in a broad range of domains. Consequently, the notion of applying learning …
applications in a broad range of domains. Consequently, the notion of applying learning …
A survey of zero-shot generalisation in deep reinforcement learning
The study of zero-shot generalisation (ZSG) in deep Reinforcement Learning (RL) aims to
produce RL algorithms whose policies generalise well to novel unseen situations at …
produce RL algorithms whose policies generalise well to novel unseen situations at …
Large language models as general pattern machines
We observe that pre-trained large language models (LLMs) are capable of autoregressively
completing complex token sequences--from arbitrary ones procedurally generated by …
completing complex token sequences--from arbitrary ones procedurally generated by …