The assessment of the classification performance can be based on class indices, such as sensitivity, specificity and precision, which describe the classification results achieved on each modelled class. However, in several situations, it is useful to represent the global classification performance with a single number. Therefore, several measures have been introduced in literature to deal with this assessment, accuracy being the most known and used. These metrics have been proposed to generally face binary classification tasks and can behave differently depending on the classification scenario. In this study, different global measures of classification performances are compared by means of results achieved on an extended set of real multivariate datasets. The systematic comparison is carried out through multivariate analysis. Further investigations are then derived on specific indices to understand how the presence of unbalanced classes and the number of modelled classes can influence their behaviour. Finally, this work introduces a set of benchmark values based on different random classification scenarios. These benchmark thresholds can serve as the initial criterion to accept or reject a classification model on the basis of its performance
Ballabio, D., Grisoni, F., Todeschini, R. (2018). Multivariate comparison of classification performance measures. CHEMOMETRICS AND INTELLIGENT LABORATORY SYSTEMS, 174, 33-44 [10.1016/j.chemolab.2017.12.004].
Multivariate comparison of classification performance measures
Ballabio, D
;Grisoni, F;Todeschini, R
2018
Abstract
The assessment of the classification performance can be based on class indices, such as sensitivity, specificity and precision, which describe the classification results achieved on each modelled class. However, in several situations, it is useful to represent the global classification performance with a single number. Therefore, several measures have been introduced in literature to deal with this assessment, accuracy being the most known and used. These metrics have been proposed to generally face binary classification tasks and can behave differently depending on the classification scenario. In this study, different global measures of classification performances are compared by means of results achieved on an extended set of real multivariate datasets. The systematic comparison is carried out through multivariate analysis. Further investigations are then derived on specific indices to understand how the presence of unbalanced classes and the number of modelled classes can influence their behaviour. Finally, this work introduces a set of benchmark values based on different random classification scenarios. These benchmark thresholds can serve as the initial criterion to accept or reject a classification model on the basis of its performanceFile | Dimensione | Formato | |
---|---|---|---|
classification_measures.pdf
Solo gestori archivio
Tipologia di allegato:
Publisher’s Version (Version of Record, VoR)
Dimensione
980.42 kB
Formato
Adobe PDF
|
980.42 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.