A survey on the explainability of supervised machine learning

N Burkart, MF Huber - Journal of Artificial Intelligence Research, 2021 - jair.org
Predictions obtained by, eg, artificial neural networks have a high accuracy but humans
often perceive the models as black boxes. Insights about the decision making are mostly …

Machine vision methods for analyzing social interactions

AA Robie, KM Seagraves… - Journal of …, 2017 - journals.biologists.com
Recent developments in machine vision methods for automatic, quantitative analysis of
social behavior have immensely improved both the scale and level of resolution with which …

Efficient dataset distillation using random feature approximation

N Loo, R Hasani, A Amini… - Advances in Neural …, 2022 - proceedings.neurips.cc
Dataset distillation compresses large datasets into smaller synthetic coresets which retain
performance with the aim of reducing the storage and computational burden of processing …

Understanding black-box predictions via influence functions

PW Koh, P Liang - International conference on machine …, 2017 - proceedings.mlr.press
How can we explain the predictions of a black-box model? In this paper, we use influence
functions—a classic technique from robust statistics—to trace a model's prediction through …

Tide: A general toolbox for identifying object detection errors

D Bolya, S Foley, J Hays, J Hoffman - … , Glasgow, UK, August 23–28, 2020 …, 2020 - Springer
We introduce TIDE, a framework and associated toolbox (https://dbolya. github. io/tide/) for
analyzing the sources of error in object detection and instance segmentation algorithms …

The dangers of post-hoc interpretability: Unjustified counterfactual explanations

T Laugel, MJ Lesot, C Marsala, X Renard… - arXiv preprint arXiv …, 2019 - arxiv.org
Post-hoc interpretability approaches have been proven to be powerful tools to generate
explanations for the predictions made by a trained black-box model. However, they create …

Dataset pruning: Reducing training data by examining generalization influence

S Yang, Z Xie, H Peng, M Xu, M Sun, P Li - arXiv preprint arXiv …, 2022 - arxiv.org
The great success of deep learning heavily relies on increasingly larger training data, which
comes at a price of huge computational and infrastructural costs. This poses crucial …

Auditing black-box models for indirect influence

P Adler, C Falk, SA Friedler, T Nix, G Rybeck… - … and Information Systems, 2018 - Springer
Data-trained predictive models see widespread use, but for the most part they are used as
black boxes which output a prediction or score. It is therefore hard to acquire a deeper …

Fairtest: Discovering unwarranted associations in data-driven applications

F Tramer, V Atlidakis, R Geambasu… - 2017 IEEE European …, 2017 - ieeexplore.ieee.org
In a world where traditional notions of privacy are increasingly challenged by the myriad
companies that collect and analyze our data, it is important that decision-making entities are …

Comparison-based inverse classification for interpretability in machine learning

T Laugel, MJ Lesot, C Marsala, X Renard… - … and Management of …, 2018 - Springer
In the context of post-hoc interpretability, this paper addresses the task of explaining the
prediction of a classifier, considering the case where no information is available, neither on …