Explainability of artificial intelligence methods, applications and challenges: A comprehensive survey

W Ding, M Abdel-Basset, H Hawash, AM Ali - Information Sciences, 2022 - Elsevier
The continuous advancement of Artificial Intelligence (AI) has been revolutionizing the
strategy of decision-making in different life domains. Regardless of this achievement, AI …

Leveraging explanations in interactive machine learning: An overview

S Teso, Ö Alkan, W Stammer, E Daly - Frontiers in Artificial …, 2023 - frontiersin.org
Explanations have gained an increasing level of interest in the AI and Machine Learning
(ML) communities in order to improve model transparency and allow users to form a mental …

A perspective on explanations of molecular prediction models

GP Wellawatte, HA Gandhi, A Seshadri… - Journal of Chemical …, 2023 - ACS Publications
Chemists can be skeptical in using deep learning (DL) in decision making, due to the lack of
interpretability in “black-box” models. Explainable artificial intelligence (XAI) is a branch of …

{Meta-Sift}: How to Sift Out a Clean Subset in the Presence of Data Poisoning?

Y Zeng, M Pan, H Jahagirdar, M Jin, L Lyu… - 32nd USENIX Security …, 2023 - usenix.org
External data sources are increasingly being used to train machine learning (ML) models as
the data demand increases. However, the integration of external data into training poses …

Concept-level debugging of part-prototype networks

A Bontempelli, S Teso, K Tentori, F Giunchiglia… - arXiv preprint arXiv …, 2022 - arxiv.org
Part-prototype Networks (ProtoPNets) are concept-based classifiers designed to achieve the
same performance as black-box models without compromising transparency. ProtoPNets …

A rationale-centric framework for human-in-the-loop machine learning

J Lu, L Yang, B Mac Namee, Y Zhang - arXiv preprint arXiv:2203.12918, 2022 - arxiv.org
We present a novel rationale-centric framework with human-in-the-loop--Rationales-centric
Double-robustness Learning (RDL)--to boost model out-of-distribution performance in few …

Studying How to Efficiently and Effectively Guide Models with Explanations

S Rao, M Böhle, A Parchami-Araghi… - Proceedings of the …, 2023 - openaccess.thecvf.com
Despite being highly performant, deep neural networks might base their decisions on
features that spuriously correlate with the provided labels, thus hurting generalization. To …

A typology for exploring the mitigation of shortcut behaviour

F Friedrich, W Stammer, P Schramowski… - Nature Machine …, 2023 - nature.com
As machine learning models become larger, and are increasingly trained on large and
uncurated datasets in weakly supervised mode, it becomes important to establish …

Identifying spurious correlations and correcting them with an explanation-based learning

MT Hagos, KM Curran, B Mac Namee - arXiv preprint arXiv:2211.08285, 2022 - arxiv.org
Identifying spurious correlations learned by a trained model is at the core of refining a
trained model and building a trustworthy model. We present a simple method to identify …

Targeted Activation Penalties Help CNNs Ignore Spurious Signals

D Zhang, M Williams, F Toni - Proceedings of the AAAI Conference on …, 2024 - ojs.aaai.org
Neural networks (NNs) can learn to rely on spurious signals in the training data, leading to
poor generalisation. Recent methods tackle this problem by training NNs with additional …