Discover and cure: Concept-aware mitigation of spurious correlation

S Wu, M Yuksekgonul, L Zhang… - … Conference on Machine …, 2023 - proceedings.mlr.press
Deep neural networks often rely on spurious correlations to make predictions, which hinders
generalization beyond training environments. For instance, models that associate cats with …

Adversarial attacks in explainable machine learning: A survey of threats against models and humans

J Vadillo, R Santana, JA Lozano - … Reviews: Data Mining and …, 2025 - Wiley Online Library
Reliable deployment of machine learning models such as neural networks continues to be
challenging due to several limitations. Some of the main shortcomings are the lack of …

Concept-level debugging of part-prototype networks

A Bontempelli, S Teso, K Tentori, F Giunchiglia… - arXiv preprint arXiv …, 2022 - arxiv.org
Part-prototype Networks (ProtoPNets) are concept-based classifiers designed to achieve the
same performance as black-box models without compromising transparency. ProtoPNets …

Not all neuro-symbolic concepts are created equal: Analysis and mitigation of reasoning shortcuts

E Marconato, S Teso, A Vergari… - Advances in Neural …, 2024 - proceedings.neurips.cc
Abstract Neuro-Symbolic (NeSy) predictive models hold the promise of improved
compliance with given constraints, systematic generalization, and interpretability, as they …

[HTML][HTML] Human-in-the-loop machine learning: Reconceptualizing the role of the user in interactive approaches

O Gómez-Carmona, D Casado-Mansilla… - Internet of Things, 2024 - Elsevier
The rise of intelligent systems and smart spaces has opened up new opportunities for
human–machine collaborations. Interactive Machine Learning (IML) contribute to fostering …

Interpretable concept bottlenecks to align reinforcement learning agents

Q Delfosse, S Sztwiertnia, M Rothermel… - arXiv preprint arXiv …, 2024 - arxiv.org
Goal misalignment, reward sparsity and difficult credit assignment are only a few of the many
issues that make it difficult for deep reinforcement learning (RL) agents to learn optimal …

EXMOS: Explanatory Model Steering Through Multifaceted Explanations and Data Configurations

A Bhattacharya, S Stumpf, L Gosak, G Stiglic… - Proceedings of the CHI …, 2024 - dl.acm.org
Explanations in interactive machine-learning systems facilitate debugging and improving
prediction models. However, the effectiveness of various global model-centric and data …

Neuro-symbolic continual learning: Knowledge, reasoning shortcuts and concept rehearsal

E Marconato, G Bontempo, E Ficarra… - arXiv preprint arXiv …, 2023 - arxiv.org
We introduce Neuro-Symbolic Continual Learning, where a model has to solve a sequence
of neuro-symbolic tasks, that is, it has to map sub-symbolic inputs to high-level concepts and …

Studying How to Efficiently and Effectively Guide Models with Explanations

S Rao, M Böhle, A Parchami-Araghi… - Proceedings of the …, 2023 - openaccess.thecvf.com
Despite being highly performant, deep neural networks might base their decisions on
features that spuriously correlate with the provided labels, thus hurting generalization. To …

Learning to intervene on concept bottlenecks

D Steinmann, W Stammer, F Friedrich… - arXiv preprint arXiv …, 2023 - arxiv.org
While traditional deep learning models often lack interpretability, concept bottleneck models
(CBMs) provide inherent explanations via their concept representations. Specifically, they …