Language in brains, minds, and machines

G Tuckute, N Kanwisher… - Annual Review of …, 2024 - annualreviews.org
It has long been argued that only humans could produce and understand language. But
now, for the first time, artificial language models (LMs) achieve this feat. Here we survey the …

A primer in BERTology: What we know about how BERT works

A Rogers, O Kovaleva, A Rumshisky - Transactions of the Association …, 2021 - direct.mit.edu
Transformer-based models have pushed state of the art in many areas of NLP, but our
understanding of what is behind their success is still limited. This paper is the first survey of …

Bertology meets biology: Interpreting attention in protein language models

J Vig, A Madani, LR Varshney, C **ong… - arxiv preprint arxiv …, 2020 - arxiv.org
Transformer architectures have proven to learn useful representations for protein
classification and generation tasks. However, these representations present challenges in …

Implicit representations of meaning in neural language models

BZ Li, M Nye, J Andreas - arxiv preprint arxiv:2106.00737, 2021 - arxiv.org
Does the effectiveness of neural language models derive entirely from accurate modeling of
surface word co-occurrence statistics, or do these models represent and reason about the …

Short-text semantic similarity (stss): Techniques, challenges and future perspectives

ZH Amur, Y Kwang Hooi, H Bhanbhro, K Dahri… - Applied Sciences, 2023 - mdpi.com
In natural language processing, short-text semantic similarity (STSS) is a very prominent
field. It has a significant impact on a broad range of applications, such as question …

Word meaning in minds and machines.

BM Lake, GL Murphy - Psychological review, 2023 - psycnet.apa.org
Abstract Machines have achieved a broad and growing set of linguistic competencies,
thanks to recent progress in Natural Language Processing (NLP). Psychologists have …

Let's Play Mono-Poly: BERT Can Reveal Words' Polysemy Level and Partitionability into Senses

A Garí Soler, M Apidianaki - Transactions of the Association for …, 2021 - direct.mit.edu
Pre-trained language models (LMs) encode rich information about linguistic structure but
their knowledge about lexical polysemy remains unclear. We propose a novel experimental …

A comparative evaluation and analysis of three generations of Distributional Semantic Models

A Lenci, M Sahlgren, P Jeuniaux… - Language resources …, 2022 - Springer
Distributional semantics has deeply changed in the last decades. First, predict models stole
the thunder from traditional count ones, and more recently both of them were replaced in …

Topic modeling with contextualized word representation clusters

L Thompson, D Mimno - arxiv preprint arxiv:2010.12626, 2020 - arxiv.org
Clustering token-level contextualized word representations produces output that shares
many similarities with topic models for English text collections. Unlike clusterings of …

Analysis and evaluation of language models for word sense disambiguation

D Loureiro, K Rezaee, MT Pilehvar… - Computational …, 2021 - direct.mit.edu
Transformer-based language models have taken many fields in NLP by storm. BERT and its
derivatives dominate most of the existing evaluation benchmarks, including those for Word …