Dissociating language and thought in large language models
Large language models (LLMs) have come closest among all models to date to mastering
human language, yet opinions about their linguistic and cognitive capabilities remain split …
human language, yet opinions about their linguistic and cognitive capabilities remain split …
[PDF][PDF] Language model behavior: A comprehensive survey
Transformer language models have received widespread public attention, yet their
generated text is often surprising even to NLP researchers. In this survey, we discuss over …
generated text is often surprising even to NLP researchers. In this survey, we discuss over …
Call for Papers--The BabyLM Challenge: Sample-efficient pretraining on a developmentally plausible corpus
We present the call for papers for the BabyLM Challenge: Sample-efficient pretraining on a
developmentally plausible corpus. This shared task is intended for participants with an …
developmentally plausible corpus. This shared task is intended for participants with an …
[HTML][HTML] Pre-trained models: Past, present and future
Large-scale pre-trained models (PTMs) such as BERT and GPT have recently achieved
great success and become a milestone in the field of artificial intelligence (AI). Owing to …
great success and become a milestone in the field of artificial intelligence (AI). Owing to …
A primer in BERTology: What we know about how BERT works
Transformer-based models have pushed state of the art in many areas of NLP, but our
understanding of what is behind their success is still limited. This paper is the first survey of …
understanding of what is behind their success is still limited. This paper is the first survey of …
What artificial neural networks can tell us about human language acquisition
Rapid progress in machine learning for natural language processing has the potential to
transform debates about how humans learn language. However, the learning environments …
transform debates about how humans learn language. However, the learning environments …
Experience grounds language
Language understanding research is held back by a failure to relate language to the
physical world it describes and to the social interactions it facilitates. Despite the incredible …
physical world it describes and to the social interactions it facilitates. Despite the incredible …
COGS: A compositional generalization challenge based on semantic interpretation
Natural language is characterized by compositionality: the meaning of a complex expression
is constructed from the meanings of its constituent parts. To facilitate the evaluation of the …
is constructed from the meanings of its constituent parts. To facilitate the evaluation of the …
Syntactic structure from deep learning
Modern deep neural networks achieve impressive performance in engineering applications
that require extensive linguistic skills, such as machine translation. This success has …
that require extensive linguistic skills, such as machine translation. This success has …
A systematic assessment of syntactic generalization in neural language models
While state-of-the-art neural network models continue to achieve lower perplexity scores on
language modeling benchmarks, it remains unknown whether optimizing for broad …
language modeling benchmarks, it remains unknown whether optimizing for broad …