Robust explainability: A tutorial on gradient-based attribution methods for deep neural networks

IE Nielsen, D Dera, G Rasool… - IEEE Signal …, 2022 - ieeexplore.ieee.org
The rise in deep neural networks (DNNs) has led to increased interest in explaining their
predictions. While many methods for this exist, there is currently no consensus on how to …

[PDF][PDF] Human-AI Complementarity in Hybrid Intelligence Systems: A Structured Literature Review.

P Hemmer, M Schemmer, M Vössing, N Kühl - PACIS, 2021 - researchgate.net
Hybrid Intelligence is an emerging concept that emphasizes the complementary nature of
human intelligence and artificial intelligence (AI). One key requirement for collaboration …

Explanations can reduce overreliance on ai systems during decision-making

H Vasconcelos, M Jörke… - Proceedings of the …, 2023 - dl.acm.org
Prior work has identified a resilient phenomenon that threatens the performance of human-
AI decision-making teams: overreliance, when people agree with an AI, even when it is …

Are explanations helpful? a comparative study of the effects of explanations in ai-assisted decision-making

X Wang, M Yin - Proceedings of the 26th International Conference on …, 2021 - dl.acm.org
This paper contributes to the growing literature in empirical evaluation of explainable AI
(XAI) methods by presenting a comparison on the effects of a set of established XAI methods …

Debugging tests for model explanations

J Adebayo, M Muelly, I Liccardi, B Kim - arXiv preprint arXiv:2011.05429, 2020 - arxiv.org
We investigate whether post-hoc model explanations are effective for diagnosing model
errors--model debugging. In response to the challenge of explaining a model's prediction, a …

Learning and evaluating graph neural network explanations based on counterfactual and factual reasoning

J Tan, S Geng, Z Fu, Y Ge, S Xu, Y Li… - Proceedings of the ACM …, 2022 - dl.acm.org
Structural data well exists in Web applications, such as social networks in social media,
citation networks in academic websites, and threads data in online forums. Due to the …

Post hoc explanations may be ineffective for detecting unknown spurious correlation

J Adebayo, M Muelly, H Abelson… - International conference on …, 2022 - openreview.net
We investigate whether three types of post hoc model explanations–feature attribution,
concept activation, and training point ranking–are effective for detecting a model's reliance …

Hydrological concept formation inside long short-term memory (LSTM) networks

T Lees, S Reece, F Kratzert, D Klotz… - Hydrology and Earth …, 2021 - hess.copernicus.org
Neural networks have been shown to be extremely effective rainfall-runoff models, where
the river discharge is predicted from meteorological inputs. However, the question remains …

The effectiveness of feature attribution methods and its correlation with automatic evaluation scores

G Nguyen, D Kim, A Nguyen - Advances in Neural …, 2021 - proceedings.neurips.cc
Explaining the decisions of an Artificial Intelligence (AI) model is increasingly critical in many
real-world, high-stake applications. Hundreds of papers have either proposed new feature …

Visual correspondence-based explanations improve AI robustness and human-AI team accuracy

MR Taesiri, G Nguyen… - Advances in Neural …, 2022 - proceedings.neurips.cc
Explaining artificial intelligence (AI) predictions is increasingly important and even
imperative in many high-stake applications where humans are the ultimate decision-makers …