On the effectiveness of parameter-efficient fine-tuning
Fine-tuning pre-trained models has been ubiquitously proven to be effective in a wide range
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …
Superglue: A stickier benchmark for general-purpose language understanding systems
In the last year, new models and methods for pretraining and transfer learning have driven
striking performance improvements across a range of language understanding tasks. The …
striking performance improvements across a range of language understanding tasks. The …
Revisiting few-sample BERT fine-tuning
This paper is a study of fine-tuning of BERT contextual representations, with focus on
commonly observed instabilities in few-sample scenarios. We identify several factors that …
commonly observed instabilities in few-sample scenarios. We identify several factors that …
BLiMP: The benchmark of linguistic minimal pairs for English
Abstract We introduce The Benchmark of Linguistic Minimal Pairs (BLiMP), a challenge set
for evaluating the linguistic knowledge of language models (LMs) on major grammatical …
for evaluating the linguistic knowledge of language models (LMs) on major grammatical …
Investigating reasons for disagreement in natural language inference
We investigate how disagreement in natural language inference (NLI) annotation arises. We
developed a taxonomy of disagreement sources with 10 categories spanning 3 high-level …
developed a taxonomy of disagreement sources with 10 categories spanning 3 high-level …
Nas-bench-nlp: neural architecture search benchmark for natural language processing
Neural Architecture Search (NAS) is a promising and rapidly evolving research area.
Training a large number of neural networks requires an exceptional amount of …
Training a large number of neural networks requires an exceptional amount of …
[PDF][PDF] Investigating BERT's Knowledge of Language: Five Analysis Methods with NPIs
A Warstadt - arxiv preprint arxiv:1909.02597, 2019 - alexwarstadt.github.io
Though state-of-the-art sentence representation models can perform tasks requiring
significant knowledge of grammar, it is an open question how best to evaluate their …
significant knowledge of grammar, it is an open question how best to evaluate their …
KLEJ: Comprehensive benchmark for Polish language understanding
P Rybak, R Mroczkowski, J Tracz, I Gawlik - arxiv preprint arxiv …, 2020 - arxiv.org
In recent years, a series of Transformer-based models unlocked major improvements in
general natural language understanding (NLU) tasks. Such a fast pace of research would …
general natural language understanding (NLU) tasks. Such a fast pace of research would …
Can neural networks acquire a structural bias from raw linguistic data?
We evaluate whether BERT, a widely used neural network for sentence processing,
acquires an inductive bias towards forming structural generalizations through pretraining on …
acquires an inductive bias towards forming structural generalizations through pretraining on …
Contextual embeddings: When are they worth it?
We study the settings for which deep contextual embeddings (eg, BERT) give large
improvements in performance relative to classic pretrained embeddings (eg, GloVe), and an …
improvements in performance relative to classic pretrained embeddings (eg, GloVe), and an …