From anecdotal evidence to quantitative evaluation methods: A systematic review on evaluating explainable ai
The rising popularity of explainable artificial intelligence (XAI) to understand high-performing
black boxes raised the question of how to evaluate explanations of machine learning (ML) …
black boxes raised the question of how to evaluate explanations of machine learning (ML) …
On the explainability of natural language processing deep models
Despite their success, deep networks are used as black-box models with outputs that are not
easily explainable during the learning and the prediction phases. This lack of interpretability …
easily explainable during the learning and the prediction phases. This lack of interpretability …
ERASER: A benchmark to evaluate rationalized NLP models
State-of-the-art models in NLP are now predominantly based on deep neural networks that
are opaque in terms of how they come to make predictions. This limitation has increased …
are opaque in terms of how they come to make predictions. This limitation has increased …
Bertology meets biology: Interpreting attention in protein language models
Transformer architectures have proven to learn useful representations for protein
classification and generation tasks. However, these representations present challenges in …
classification and generation tasks. However, these representations present challenges in …
Interpreting graph neural networks for NLP with differentiable edge masking
Graph neural networks (GNNs) have become a popular approach to integrating structural
inductive biases into NLP models. However, there has been little work on interpreting them …
inductive biases into NLP models. However, there has been little work on interpreting them …
Survey of low-resource machine translation
We present a survey covering the state of the art in low-resource machine translation (MT)
research. There are currently around 7,000 languages spoken in the world and almost all …
research. There are currently around 7,000 languages spoken in the world and almost all …
The elephant in the interpretability room: Why use attention as explanation when we have saliency methods?
There is a recent surge of interest in using attention as explanation of model predictions,
with mixed evidence on whether attention can be used as such. While attention conveniently …
with mixed evidence on whether attention can be used as such. While attention conveniently …
Do feature attribution methods correctly attribute features?
Feature attribution methods are popular in interpretable machine learning. These methods
compute the attribution of each input feature to represent its importance, but there is no …
compute the attribution of each input feature to represent its importance, but there is no …
Width & depth pruning for vision transformers
Transformer models have demonstrated their promising potential and achieved excellent
performance on a series of computer vision tasks. However, the huge computational cost of …
performance on a series of computer vision tasks. However, the huge computational cost of …
Invariant rationalization
Selective rationalization improves neural network interpretability by identifying a small
subset of input features {—} the rationale {—} that best explains or supports the prediction. A …
subset of input features {—} the rationale {—} that best explains or supports the prediction. A …