Understanding the (extra-) ordinary: Validating deep model decisions with prototypical concept-based explanations

M Dreyer, R Achtibat, W Samek… - Proceedings of the …, 2024 - openaccess.thecvf.com
Ensuring both transparency and safety is critical when deploying Deep Neural Networks
(DNNs) in high-risk applications such as medicine. The field of explainable AI (XAI) has …

Labeling neural representations with inverse recognition

K Bykov, L Kopf, S Nakajima… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Deep Neural Networks (DNNs) demonstrated remarkable capabilities in learning
complex hierarchical data representations, but the nature of these representations remains …

Disentangled explanations of neural network predictions by finding relevant subspaces

P Chormai, J Herrmann, KR Müller… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Explainable AI aims to overcome the black-box nature of complex ML models like neural
networks by generating explanations for their predictions. Explanations often take the form of …

From hope to safety: Unlearning biases of deep models via gradient penalization in latent space

M Dreyer, F Pahde, CJ Anders, W Samek… - Proceedings of the …, 2024 - ojs.aaai.org
Deep Neural Networks are prone to learning spurious correlations embedded in the training
data, leading to potentially biased predictions. This poses risks when deploying these …

Manipulating feature visualizations with gradient slingshots

D Bareeva, MMC Höhne, A Warnecke, L Pirch… - arXiv preprint arXiv …, 2024 - arxiv.org
Deep Neural Networks (DNNs) are capable of learning complex and versatile
representations, however, the semantic nature of the learned concepts remains unknown. A …

[HTML][HTML] Preemptively pruning Clever-Hans strategies in deep neural networks

L Linhardt, KR Müller, G Montavon - Information Fusion, 2024 - Elsevier
Robustness has become an important consideration in deep learning. With the help of
explainable AI, mismatches between an explained model's decision strategy and the user's …

Finding Spurious Correlations with Function-Semantic Contrast Analysis

K Bykov, L Kopf, MMC Höhne - World Conference on Explainable Artificial …, 2023 - Springer
In the field of Computer Vision (CV), the degree to which two objects, eg two classes, share
a common conceptual meaning, known as semantic similarity, is closely linked to the visual …

Visualizing the diversity of representations learned by Bayesian neural networks

D Grinwald, K Bykov, S Nakajima… - arXiv preprint arXiv …, 2022 - arxiv.org
Explainable Artificial Intelligence (XAI) aims to make learning machines less opaque, and
offers researchers and practitioners various tools to reveal the decision-making strategies of …

CoSy: Evaluating Textual Explanations of Neurons

L Kopf, PL Bommer, A Hedström, S Lapuschkin… - arXiv preprint arXiv …, 2024 - arxiv.org
A crucial aspect of understanding the complex nature of Deep Neural Networks (DNNs) is
the ability to explain learned concepts within their latent representations. While various …

Mark my words: Dangers of watermarked images in imagenet

K Bykov, KR Müller, MMC Höhne - European Conference on Artificial …, 2023 - Springer
The utilization of pre-trained networks, especially those trained on ImageNet, has become a
common practice in Computer Vision. However, prior research has indicated that a …