From anecdotal evidence to quantitative evaluation methods: A systematic review on evaluating explainable ai

M Nauta, J Trienes, S Pathak, E Nguyen… - ACM Computing …, 2023 - dl.acm.org
The rising popularity of explainable artificial intelligence (XAI) to understand high-performing
black boxes raised the question of how to evaluate explanations of machine learning (ML) …

On the explainability of natural language processing deep models

JE Zini, M Awad - ACM Computing Surveys, 2022 - dl.acm.org
Despite their success, deep networks are used as black-box models with outputs that are not
easily explainable during the learning and the prediction phases. This lack of interpretability …

ERASER: A benchmark to evaluate rationalized NLP models

J DeYoung, S Jain, NF Rajani, E Lehman… - arxiv preprint arxiv …, 2019 - arxiv.org
State-of-the-art models in NLP are now predominantly based on deep neural networks that
are opaque in terms of how they come to make predictions. This limitation has increased …

Bertology meets biology: Interpreting attention in protein language models

J Vig, A Madani, LR Varshney, C **ong… - arxiv preprint arxiv …, 2020 - arxiv.org
Transformer architectures have proven to learn useful representations for protein
classification and generation tasks. However, these representations present challenges in …

Interpreting graph neural networks for NLP with differentiable edge masking

MS Schlichtkrull, N De Cao, I Titov - arxiv preprint arxiv:2010.00577, 2020 - arxiv.org
Graph neural networks (GNNs) have become a popular approach to integrating structural
inductive biases into NLP models. However, there has been little work on interpreting them …

Survey of low-resource machine translation

B Haddow, R Bawden, AVM Barone, J Helcl… - Computational …, 2022 - direct.mit.edu
We present a survey covering the state of the art in low-resource machine translation (MT)
research. There are currently around 7,000 languages spoken in the world and almost all …

The elephant in the interpretability room: Why use attention as explanation when we have saliency methods?

J Bastings, K Filippova - arxiv preprint arxiv:2010.05607, 2020 - arxiv.org
There is a recent surge of interest in using attention as explanation of model predictions,
with mixed evidence on whether attention can be used as such. While attention conveniently …

Do feature attribution methods correctly attribute features?

Y Zhou, S Booth, MT Ribeiro, J Shah - Proceedings of the AAAI …, 2022 - ojs.aaai.org
Feature attribution methods are popular in interpretable machine learning. These methods
compute the attribution of each input feature to represent its importance, but there is no …

Width & depth pruning for vision transformers

F Yu, K Huang, M Wang, Y Cheng, W Chu… - Proceedings of the AAAI …, 2022 - ojs.aaai.org
Transformer models have demonstrated their promising potential and achieved excellent
performance on a series of computer vision tasks. However, the huge computational cost of …

Invariant rationalization

S Chang, Y Zhang, M Yu… - … Conference on Machine …, 2020 - proceedings.mlr.press
Selective rationalization improves neural network interpretability by identifying a small
subset of input features {—} the rationale {—} that best explains or supports the prediction. A …