Dissociating language and thought in large language models

K Mahowald, AA Ivanova, IA Blank, N Kanwisher… - Trends in Cognitive …, 2024 - cell.com
Large language models (LLMs) have come closest among all models to date to mastering
human language, yet opinions about their linguistic and cognitive capabilities remain split …

Syntactic structure from deep learning

T Linzen, M Baroni - Annual Review of Linguistics, 2021 - annualreviews.org
Modern deep neural networks achieve impressive performance in engineering applications
that require extensive linguistic skills, such as machine translation. This success has …

Unnatural language inference

K Sinha, P Parthasarathi, J Pineau… - arxiv preprint arxiv …, 2020 - arxiv.org
Recent investigations into the inner-workings of state-of-the-art large-scale pre-trained
Transformer-based Natural Language Understanding (NLU) models indicate that they …

Language models as models of language

R Millière - arxiv preprint arxiv:2408.07144, 2024 - arxiv.org
This chapter critically examines the potential contributions of modern language models to
theoretical linguistics. Despite their focus on engineering goals, these models' ability to …

Square one bias in NLP: Towards a multi-dimensional exploration of the research manifold

S Ruder, I Vulić, A Søgaard - arxiv preprint arxiv:2206.09755, 2022 - arxiv.org
The prototypical NLP experiment trains a standard architecture on labeled English data and
optimizes for accuracy, without accounting for other dimensions such as fairness …

Multi-VALUE: A framework for cross-dialectal English NLP

C Ziems, W Held, J Yang, J Dhamala, R Gupta… - arxiv preprint arxiv …, 2022 - arxiv.org
Dialect differences caused by regional, social, and economic factors cause performance
discrepancies for many groups of language technology users. Inclusive and equitable …

Studying the inductive biases of RNNs with synthetic variations of natural languages

S Ravfogel, Y Goldberg, T Linzen - arxiv preprint arxiv:1903.06400, 2019 - arxiv.org
How do typological properties such as word order and morphological case marking affect
the ability of neural sequence models to acquire the syntax of a language? Cross-linguistic …

Analyzing and interpreting neural networks for NLP: A report on the first BlackboxNLP workshop

A Alishahi, G Chrupała, T Linzen - Natural Language Engineering, 2019 - cambridge.org
The Empirical Methods in Natural Language Processing (EMNLP) 2018 workshop
BlackboxNLP was dedicated to resources and techniques specifically developed for …

Adversarial removal of demographic attributes revisited

M Barrett, Y Kementchedjhieva, Y Elazar… - Proceedings of the …, 2019 - aclanthology.org
Elazar and Goldberg (2018) showed that protected attributes can be extracted from the
representations of a debiased neural network for mention detection at above-chance levels …

How to plant trees in language models: Data and architectural effects on the emergence of syntactic inductive biases

A Mueller, T Linzen - arxiv preprint arxiv:2305.19905, 2023 - arxiv.org
Accurate syntactic representations are essential for robust generalization in natural
language. Recent work has found that pre-training can teach language models to rely on …