Leveraging explanations in interactive machine learning: An overview

S Teso, Ö Alkan, W Stammer, E Daly - Frontiers in Artificial …, 2023 - frontiersin.org
Explanations have gained an increasing level of interest in the AI and Machine Learning
(ML) communities in order to improve model transparency and allow users to form a mental …

Learning bottleneck concepts in image classification

B Wang, L Li, Y Nakashima… - Proceedings of the ieee …, 2023 - openaccess.thecvf.com
Interpreting and explaining the behavior of deep neural networks is critical for many tasks.
Explainable AI provides a way to address this challenge, mostly by providing per-pixel …

Glancenets: Interpretable, leak-proof concept-based models

E Marconato, A Passerini… - Advances in Neural …, 2022 - proceedings.neurips.cc
There is growing interest in concept-based models (CBMs) that combine high-performance
and interpretability by acquiring and reasoning with a vocabulary of high-level concepts. A …

Neural systematic binder

G Singh, Y Kim, S Ahn - arXiv preprint arXiv:2211.01177, 2022 - arxiv.org
The key to high-level cognition is believed to be the ability to systematically manipulate and
compose knowledge pieces. While token-like structured knowledge representations are …

Not all neuro-symbolic concepts are created equal: Analysis and mitigation of reasoning shortcuts

E Marconato, S Teso, A Vergari… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Neuro-Symbolic (NeSy) predictive models hold the promise of improved
compliance with given constraints, systematic generalization, and interpretability, as they …

Interpretable concept bottlenecks to align reinforcement learning agents

Q Delfosse, S Sztwiertnia, M Rothermel… - arXiv preprint arXiv …, 2024 - arxiv.org
Goal misalignment, reward sparsity and difficult credit assignment are only a few of the many
issues that make it difficult for deep reinforcement learning (RL) agents to learn optimal …

Learning to intervene on concept bottlenecks

D Steinmann, W Stammer, F Friedrich… - arXiv preprint arXiv …, 2023 - arxiv.org
While traditional deep learning models often lack interpretability, concept bottleneck models
(CBMs) provide inherent explanations via their concept representations. Specifically, they …

Interpretability is in the mind of the beholder: A causal framework for human-interpretable representation learning

E Marconato, A Passerini, S Teso - Entropy, 2023 - mdpi.com
Research on Explainable Artificial Intelligence has recently started exploring the idea of
producing explanations that, rather than being expressed in terms of low-level features, are …

ConceptHash: Interpretable Fine-Grained Hashing via Concept Discovery

KW Ng, X Zhu, YZ Song… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Existing fine-grained hashing methods typically lack code interpretability as they compute
hash code bits holistically using both global and local features. To address this limitation we …

Learning Concept-Based Visual Causal Transition and Symbolic Reasoning for Visual Planning

Y Qian, P Yu, YN Wu, W Wang, L Fan - arXiv preprint arXiv:2310.03325, 2023 - arxiv.org
Visual planning simulates how humans make decisions to achieve desired goals in the form
of searching for visual causal transitions between an initial visual state and a final visual …