Labeling neural representations with inverse recognition

K Bykov, L Kopf, S Nakajima… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Deep Neural Networks (DNNs) demonstrated remarkable capabilities in learning
complex hierarchical data representations, but the nature of these representations remains …

Influenciæ: A library for tracing the influence back to the data-points

A Picard, L Hervier, T Fel, D Vigouroux - World Conference on Explainable …, 2024 - Springer
In today's AI-driven world, understanding model behavior is becoming more important than
ever. While libraries abound for doing so via traditional XAI methods, the domain of …

Manipulating feature visualizations with gradient slingshots

D Bareeva, MMC Höhne, A Warnecke, L Pirch… - arXiv preprint arXiv …, 2024 - arxiv.org
Deep Neural Networks (DNNs) are capable of learning complex and versatile
representations, however, the semantic nature of the learned concepts remains unknown. A …

AttributionScanner: A Visual Analytics System for Model Validation with Metadata-Free Slice Finding

X Xuan, JP Ono, L Gou, KL Ma, L Ren - arXiv preprint arXiv:2401.06462, 2024 - arxiv.org
Data slice finding is an emerging technique for validating machine learning (ML) models by
identifying and analyzing subgroups in a dataset that exhibit poor performance, often …

CoSy: Evaluating Textual Explanations of Neurons

L Kopf, PL Bommer, A Hedström, S Lapuschkin… - arXiv preprint arXiv …, 2024 - arxiv.org
A crucial aspect of understanding the complex nature of Deep Neural Networks (DNNs) is
the ability to explain learned concepts within their latent representations. While various …

GIFT: A Framework for Global Interpretable Faithful Textual Explanations of Vision Classifiers

É Zablocki, V Gerard, A Cardiel, E Gaussier… - arXiv preprint arXiv …, 2024 - arxiv.org
Understanding deep models is crucial for deploying them in safety-critical applications. We
introduce GIFT, a framework for deriving post-hoc, global, interpretable, and faithful textual …

Explaining deep neural networks by leveraging intrinsic methods

B La Rosa - arXiv preprint arXiv:2407.12243, 2024 - arxiv.org
Despite their impact on the society, deep neural networks are often regarded as black-box
models due to their intricate structures and the absence of explanations for their decisions …

Guidelines to explain machine learning algorithms

F Boisnard, R Boumazouza, M Ducoffe, T Fel, E Glize… - 2023 - hal.science
In the rapidly evolving and increasingly complex field of Artificial Intelligence (AI),
understanding and interpreting the decision‐making process of models is crucial. This …

Feature Accentuation: Explaining'what'features respond to in natural images

CJ Hamblin, FEL Thomas, S Saha, T Konkle… - openreview.net
Efforts to decode neural network vision models necessitate a comprehensive grasp of both
the spatial and semantic facets governing feature responses within images. Most research …