Probing classifiers: Promises, shortcomings, and advances

Y Belinkov - Computational Linguistics, 2022 - direct.mit.edu
Probing classifiers have emerged as one of the prominent methodologies for interpreting
and analyzing deep neural network models of natural language processing. The basic idea …

Analysis methods in neural language processing: A survey

Y Belinkov, J Glass - … of the Association for Computational Linguistics, 2019 - direct.mit.edu
The field of natural language processing has seen impressive progress in recent years, with
neural network models replacing many of the traditional systems. A plethora of new models …

On the opportunities and risks of foundation models

R Bommasani, DA Hudson, E Adeli, R Altman… - arxiv preprint arxiv …, 2021 - arxiv.org
AI is undergoing a paradigm shift with the rise of models (eg, BERT, DALL-E, GPT-3) that are
trained on broad data at scale and are adaptable to a wide range of downstream tasks. We …

Bertology meets biology: Interpreting attention in protein language models

J Vig, A Madani, LR Varshney, C **ong… - arxiv preprint arxiv …, 2020 - arxiv.org
Transformer architectures have proven to learn useful representations for protein
classification and generation tasks. However, these representations present challenges in …

Compositionality decomposed: How do neural networks generalise?

D Hupkes, V Dankers, M Mul, E Bruni - Journal of Artificial Intelligence …, 2020 - jair.org
Despite a multitude of empirical studies, little consensus exists on whether neural networks
are able to generalise compositionally, a controversy that, in part, stems from a lack of …

Visualisation and'diagnostic classifiers' reveal how recurrent and recursive neural networks process hierarchical structure

D Hupkes, S Veldhoen, W Zuidema - Journal of Artificial Intelligence …, 2018 - jair.org
We investigate how neural networks can learn and process languages with hierarchical,
compositional semantics. To this end, we define the artifical task of processing nested …

Towards faithful model explanation in nlp: A survey

Q Lyu, M Apidianaki, C Callison-Burch - Computational Linguistics, 2024 - direct.mit.edu
End-to-end neural Natural Language Processing (NLP) models are notoriously difficult to
understand. This has given rise to numerous efforts towards model explainability in recent …

State-of-the-art generalisation research in NLP: a taxonomy and review

D Hupkes, M Giulianelli, V Dankers, M Artetxe… - arxiv preprint arxiv …, 2022 - arxiv.org
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …

Semantic structure in deep learning

E Pavlick - Annual Review of Linguistics, 2022 - annualreviews.org
Deep learning has recently come to dominate computational linguistics, leading to claims of
human-level performance in a range of language processing tasks. Like much previous …

Making transformers solve compositional tasks

S Ontanon, J Ainslie, V Cvicek, Z Fisher - arxiv preprint arxiv:2108.04378, 2021 - arxiv.org
Several studies have reported the inability of Transformer models to generalize
compositionally, a key type of generalization in many NLP tasks such as semantic parsing …