Accuracy measures for the comparison of classifiers

V Labatut, H Cherifi - arXiv preprint arXiv:1207.3790, 2012 - arxiv.org
The selection of the best classification algorithm for a given dataset is a very widespread
problem. It is also a complex one, in the sense it requires to make several important …

[PDF][PDF] The case against accuracy estimation for comparing induction algorithms.

FJ Provost, T Fawcett, R Kohavi - ICML, 1998 - ai.stanford.edu
We analyze critically the use of classification accuracy to compare classifiers on natural data
sets, providing a thorough investigation using ROC analysis, standard machine learning …

[PDF][PDF] A unified view of performance metrics: Translating threshold choice into expected classification loss

J Hernández-Orallo, P Flach, C Ferri Ramírez - Journal of Machine …, 2012 - jmlr.org
Many performance metrics have been introduced in the literature for the evaluation of
classification performance, each of them with different origins and areas of application …

Quantifying trends accurately despite classifier error and class imbalance

G Forman - Proceedings of the 12th ACM SIGKDD international …, 2006 - dl.acm.org
This paper promotes a new task for supervised machine learning research: quantification-
the pursuit of learning methods for accurately estimating the class distribution of a test set …

Insights into performance fitness and error metrics for machine learning

MZ Naser, A Alavi - arXiv preprint arXiv:2006.00887, 2020 - arxiv.org
Machine learning (ML) is the field of training machines to achieve high level of cognition and
perform human-like analysis. Since ML is a data-driven approach, it seemingly fits into our …

[PDF][PDF] The effect of class distribution on classifier learning: an empirical study

GM Weiss, F Provost - 2001 - storm.cis.fordham.edu
In this article we analyze the effect of class distribution on classifier learning. We begin by
describing the different ways in which class distribution affects learning and how it affects the …

Are we learning yet? a meta review of evaluation failures across machine learning

T Liao, R Taori, ID Raji, L Schmidt - Thirty-fifth Conference on …, 2021 - openreview.net
Many subfields of machine learning share a common stumbling block: evaluation. Advances
in machine learning often evaporate under closer scrutiny or turn out to be less widely …

The choice of scaling technique matters for classification performance

LBV de Amorim, GDC Cavalcanti, RMO Cruz - Applied Soft Computing, 2023 - Elsevier
Dataset scaling, also known as normalization, is an essential preprocessing step in a
machine learning pipeline. It is aimed at adjusting attributes scales in a way that they all vary …

Model selection via the AUC

S Rosset - Proceedings of the twenty-first international conference …, 2004 - dl.acm.org
We present a statistical analysis of the AUC as an evaluation criterion for classification
scoring models. First, we consider significance tests for the difference between AUC scores …

Classifier calibration: a survey on how to assess and improve predicted class probabilities

T Silva Filho, H Song, M Perello-Nieto… - Machine Learning, 2023 - Springer
This paper provides both an introduction to and a detailed overview of the principles and
practice of classifier calibration. A well-calibrated classifier correctly quantifies the level of …