Syntactic structure from deep learning

T Linzen, M Baroni - Annual Review of Linguistics, 2021 - annualreviews.org
Modern deep neural networks achieve impressive performance in engineering applications
that require extensive linguistic skills, such as machine translation. This success has …

Neurocomputational models of language processing

JT Hale, L Campanelli, J Li, S Bhattasali… - Annual Review of …, 2022 - annualreviews.org
Efforts to understand the brain bases of language face the Map** Problem: At what level
do linguistic computations and representations connect to human neurobiology? We review …

Brains and algorithms partially converge in natural language processing

C Caucheteux, JR King - Communications biology, 2022 - nature.com
Deep learning algorithms trained to predict masked words from large amount of text have
recently been shown to generate activations similar to those of the human brain. However …

Why does surprisal from larger transformer-based language models provide a poorer fit to human reading times?

BD Oh, W Schuler - Transactions of the Association for Computational …, 2023 - direct.mit.edu
This work presents a linguistic analysis into why larger Transformer-based pre-trained
language models with more parameters and lower perplexity nonetheless yield surprisal …

A hierarchy of linguistic predictions during natural language comprehension

M Heilbron, K Armeni, JM Schoffelen… - Proceedings of the …, 2022 - National Acad Sciences
Understanding spoken language requires transforming ambiguous acoustic streams into a
hierarchy of representations, from phonemes to meaning. It has been suggested that the …

Localizing syntactic predictions using recurrent neural network grammars

JR Brennan, C Dyer, A Kuncoro, JT Hale - Neuropsychologia, 2020 - Elsevier
Brain activity in numerous perisylvian brain regions is modulated by the expectedness of
linguistic stimuli. We leverage recent advances in computational parsing models to test what …

Interpreting and improving natural-language processing (in machines) with natural language-processing (in the brain)

M Toneva, L Wehbe - Advances in neural information …, 2019 - proceedings.neurips.cc
Neural networks models for NLP are typically implemented without the explicit encoding of
language rules and yet they are able to break one performance record after another. This …

Lossy‐context surprisal: An information‐theoretic model of memory effects in sentence processing

R Futrell, E Gibson, RP Levy - Cognitive science, 2020 - Wiley Online Library
A key component of research on human sentence processing is to characterize the
processing difficulty associated with the comprehension of words in context. Models that …

Synchronous, but not entrained: exogenous and endogenous cortical rhythms of speech and language processing

L Meyer, Y Sun, AE Martin - Language, Cognition and …, 2020 - Taylor & Francis
Research on speech processing is often focused on a phenomenon termed “entrainment”,
whereby the cortex shadows rhythmic acoustic information with oscillatory activity …

Disentangling syntax and semantics in the brain with deep networks

C Caucheteux, A Gramfort… - … conference on machine …, 2021 - proceedings.mlr.press
The activations of language transformers like GPT-2 have been shown to linearly map onto
brain activity during speech comprehension. However, the nature of these activations …