Understanding Dataset Difficulty with -Usable Information

K Ethayarajh, Y Choi… - … Conference on Machine …, 2022 - proceedings.mlr.press
Estimating the difficulty of a dataset typically involves comparing state-of-the-art models to
humans; the bigger the performance gap, the harder the dataset is said to be. However, this …

Schrödinger's tree—On syntax and neural language models

A Kulmizev, J Nivre - Frontiers in Artificial Intelligence, 2022 - frontiersin.org
In the last half-decade, the field of natural language processing (NLP) has undergone two
major transitions: the switch to neural networks as the primary modeling paradigm and the …

Ravel: Evaluating interpretability methods on disentangling language model representations

J Huang, Z Wu, C Potts, M Geva, A Geiger - arxiv preprint arxiv …, 2024 - arxiv.org
Individual neurons participate in the representation of multiple high-level concepts. To what
extent can different interpretability methods successfully disentangle these roles? To help …

Receval: Evaluating reasoning chains via correctness and informativeness

A Prasad, S Saha, X Zhou, M Bansal - arxiv preprint arxiv:2304.10703, 2023 - arxiv.org
Multi-step reasoning ability is fundamental to many natural language tasks, yet it is unclear
what constitutes a good reasoning chain and how to evaluate them. Most existing methods …

A closer look at how fine-tuning changes BERT

Y Zhou, V Srikumar - arxiv preprint arxiv:2106.14282, 2021 - arxiv.org
Given the prevalence of pre-trained contextualized representations in today's NLP, there
have been many efforts to understand what information they contain, and why they seem to …

Probing for the usage of grammatical number

K Lasri, T Pimentel, A Lenci, T Poibeau… - arxiv preprint arxiv …, 2022 - arxiv.org
A central quest of probing is to uncover how pre-trained models encode a linguistic property
within their representations. An encoding, however, might be spurious-ie, the model might …

Probing for constituency structure in neural language models

D Arps, Y Samih, L Kallmeyer, H Sajjad - arxiv preprint arxiv:2204.06201, 2022 - arxiv.org
In this paper, we investigate to which extent contextual neural language models (LMs)
implicitly learn syntactic structure. More concretely, we focus on constituent structure as …

[PDF][PDF] 自然语言处理中的探针可解释方法综述

鞠天杰, 刘功申, 张倬胜, 张茹 - 计算机学报, 2024 - cjc.ict.ac.cn
摘要随着大规模预训练模型的广泛应用, 自然语言处理的多个领域(如文本分类和机器翻译)
取得了长足的发展. 然而, 受限于预训练模型的“黑盒” 特性, 其内部的决策模式以及编码的知识 …

When classifying grammatical role, BERT doesn't care about word order... except when it matters

I Papadimitriou, R Futrell, K Mahowald - arxiv preprint arxiv:2203.06204, 2022 - arxiv.org
Because meaning can often be inferred from lexical semantics alone, word order is often a
redundant cue in natural language. For example, the words chopped, chef, and onion are …

Gaussian process probes (GPP) for uncertainty-aware probing

Z Wang, A Ku, J Baldridge… - Advances in neural …, 2023 - proceedings.neurips.cc
Understanding which concepts models can and cannot represent has been fundamental to
many tasks: from effective and responsible use of models to detecting out of distribution …