Semantic structure in deep learning
E Pavlick - Annual Review of Linguistics, 2022 - annualreviews.org
Deep learning has recently come to dominate computational linguistics, leading to claims of
human-level performance in a range of language processing tasks. Like much previous …
human-level performance in a range of language processing tasks. Like much previous …
Masked language modeling and the distributional hypothesis: Order word matters pre-training for little
A possible explanation for the impressive performance of masked language model (MLM)
pre-training is that such models have learned to represent the syntactic structures prevalent …
pre-training is that such models have learned to represent the syntactic structures prevalent …
What BERT is not: Lessons from a new suite of psycholinguistic diagnostics for language models
A Ettinger - Transactions of the Association for Computational …, 2020 - direct.mit.edu
Pre-training by language modeling has become a popular and successful approach to NLP
tasks, but we have yet to understand exactly what linguistic capacities these pre-training …
tasks, but we have yet to understand exactly what linguistic capacities these pre-training …
[PDF][PDF] Linguistic Knowledge and Transferability of Contextual Representations
NF Liu - arxiv preprint arxiv:1903.08855, 2019 - fq.pkwyx.com
Contextual word representations derived from large-scale neural language models are
successful across a diverse set of NLP tasks, suggesting that they encode useful and …
successful across a diverse set of NLP tasks, suggesting that they encode useful and …
Interpreting graph neural networks for NLP with differentiable edge masking
Graph neural networks (GNNs) have become a popular approach to integrating structural
inductive biases into NLP models. However, there has been little work on interpreting them …
inductive biases into NLP models. However, there has been little work on interpreting them …
BLiMP: The benchmark of linguistic minimal pairs for English
Abstract We introduce The Benchmark of Linguistic Minimal Pairs (BLiMP), a challenge set
for evaluating the linguistic knowledge of language models (LMs) on major grammatical …
for evaluating the linguistic knowledge of language models (LMs) on major grammatical …
State-of-the-art generalisation research in NLP: a taxonomy and review
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …
Local interpretations for explainable natural language processing: A survey
As the use of deep learning techniques has grown across various fields over the past
decade, complaints about the opaqueness of the black-box models have increased …
decade, complaints about the opaqueness of the black-box models have increased …
The emergence of number and syntax units in LSTM language models
Recent work has shown that LSTMs trained on a generic language modeling objective
capture syntax-sensitive generalizations such as long-distance number agreement. We …
capture syntax-sensitive generalizations such as long-distance number agreement. We …
Understanding by understanding not: Modeling negation in language models
Negation is a core construction in natural language. Despite being very successful on many
tasks, state-of-the-art pre-trained language models often handle negation incorrectly. To …
tasks, state-of-the-art pre-trained language models often handle negation incorrectly. To …