On the effectiveness of parameter-efficient fine-tuning

Z Fu, H Yang, AMC So, W Lam, L Bing… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
Fine-tuning pre-trained models has been ubiquitously proven to be effective in a wide range
of NLP tasks. However, fine-tuning the whole model is parameter inefficient as it always …

Superglue: A stickier benchmark for general-purpose language understanding systems

A Wang, Y Pruksachatkun, N Nangia… - Advances in neural …, 2019 - proceedings.neurips.cc
In the last year, new models and methods for pretraining and transfer learning have driven
striking performance improvements across a range of language understanding tasks. The …

Revisiting few-sample BERT fine-tuning

T Zhang, F Wu, A Katiyar, KQ Weinberger… - arxiv preprint arxiv …, 2020 - arxiv.org
This paper is a study of fine-tuning of BERT contextual representations, with focus on
commonly observed instabilities in few-sample scenarios. We identify several factors that …

BLiMP: The benchmark of linguistic minimal pairs for English

A Warstadt, A Parrish, H Liu, A Mohananey… - Transactions of the …, 2020 - direct.mit.edu
Abstract We introduce The Benchmark of Linguistic Minimal Pairs (BLiMP), a challenge set
for evaluating the linguistic knowledge of language models (LMs) on major grammatical …

Investigating reasons for disagreement in natural language inference

NJ Jiang, MC Marneffe - Transactions of the Association for …, 2022 - direct.mit.edu
We investigate how disagreement in natural language inference (NLI) annotation arises. We
developed a taxonomy of disagreement sources with 10 categories spanning 3 high-level …

Nas-bench-nlp: neural architecture search benchmark for natural language processing

N Klyuchnikov, I Trofimov, E Artemova… - IEEE …, 2022 - ieeexplore.ieee.org
Neural Architecture Search (NAS) is a promising and rapidly evolving research area.
Training a large number of neural networks requires an exceptional amount of …

[PDF][PDF] Investigating BERT's Knowledge of Language: Five Analysis Methods with NPIs

A Warstadt - arxiv preprint arxiv:1909.02597, 2019 - alexwarstadt.github.io
Though state-of-the-art sentence representation models can perform tasks requiring
significant knowledge of grammar, it is an open question how best to evaluate their …

KLEJ: Comprehensive benchmark for Polish language understanding

P Rybak, R Mroczkowski, J Tracz, I Gawlik - arxiv preprint arxiv …, 2020 - arxiv.org
In recent years, a series of Transformer-based models unlocked major improvements in
general natural language understanding (NLU) tasks. Such a fast pace of research would …

Can neural networks acquire a structural bias from raw linguistic data?

A Warstadt, SR Bowman - arxiv preprint arxiv:2007.06761, 2020 - arxiv.org
We evaluate whether BERT, a widely used neural network for sentence processing,
acquires an inductive bias towards forming structural generalizations through pretraining on …

Contextual embeddings: When are they worth it?

S Arora, A May, J Zhang, C Ré - arxiv preprint arxiv:2005.09117, 2020 - arxiv.org
We study the settings for which deep contextual embeddings (eg, BERT) give large
improvements in performance relative to classic pretrained embeddings (eg, GloVe), and an …