From anecdotal evidence to quantitative evaluation methods: A systematic review on evaluating explainable ai

M Nauta, J Trienes, S Pathak, E Nguyen… - ACM Computing …, 2023 - dl.acm.org
The rising popularity of explainable artificial intelligence (XAI) to understand high-performing
black boxes raised the question of how to evaluate explanations of machine learning (ML) …

Robust explainability: A tutorial on gradient-based attribution methods for deep neural networks

IE Nielsen, D Dera, G Rasool… - IEEE Signal …, 2022 - ieeexplore.ieee.org
The rise in deep neural networks (DNNs) has led to increased interest in explaining their
predictions. While many methods for this exist, there is currently no consensus on how to …

Multiresolution knowledge distillation for anomaly detection

M Salehi, N Sadjadi, S Baselizadeh… - Proceedings of the …, 2021 - openaccess.thecvf.com
Unsupervised representation learning has proved to be a critical component of anomaly
detection/localization in images. The challenges to learn such a representation are two-fold …

Explainable deep learning: A field guide for the uninitiated

G Ras, N Xie, M Van Gerven, D Doran - Journal of Artificial Intelligence …, 2022 - jair.org
Deep neural networks (DNNs) are an indispensable machine learning tool despite the
difficulty of diagnosing what aspects of a model's input drive its decisions. In countless real …

Definitions, methods, and applications in interpretable machine learning

WJ Murdoch, C Singh, K Kumbier… - Proceedings of the …, 2019 - National Acad Sciences
Machine-learning models have demonstrated great success in learning complex patterns
that enable them to make predictions about unobserved data. In addition to using models for …

Sanity checks for saliency maps

J Adebayo, J Gilmer, M Muelly… - Advances in neural …, 2018 - proceedings.neurips.cc
Saliency methods have emerged as a popular tool to highlight features in an input deemed
relevant for the prediction of a learned model. Several saliency methods have been …

Interpreting graph neural networks for NLP with differentiable edge masking

MS Schlichtkrull, N De Cao, I Titov - arXiv preprint arXiv:2010.00577, 2020 - arxiv.org
Graph neural networks (GNNs) have become a popular approach to integrating structural
inductive biases into NLP models. However, there has been little work on interpreting them …

[HTML][HTML] Analysis of explainers of black box deep neural networks for computer vision: A survey

V Buhrmester, D Münch, M Arens - Machine Learning and Knowledge …, 2021 - mdpi.com
Deep Learning is a state-of-the-art technique to make inference on extensive or complex
data. As a black box model due to their multilayer nonlinear structure, Deep Neural …

Assessing the trustworthiness of saliency maps for localizing abnormalities in medical imaging

N Arun, N Gaw, P Singh, K Chang… - Radiology: Artificial …, 2021 - pubs.rsna.org
Purpose To evaluate the trustworthiness of saliency maps for abnormality localization in
medical imaging. Materials and Methods Using two large publicly available radiology …

Which explanation should i choose? a function approximation perspective to characterizing post hoc explanations

T Han, S Srinivas, H Lakkaraju - Advances in neural …, 2022 - proceedings.neurips.cc
A critical problem in the field of post hoc explainability is the lack of a common foundational
goal among methods. For example, some methods are motivated by function approximation …