4.5 Article

General Performance Score for classification problems

Journal

APPLIED INTELLIGENCE
Volume 52, Issue 10, Pages 12049-12063

Publisher

SPRINGER
DOI: 10.1007/s10489-021-03041-7

Keywords

Performance metrics; Binary classification; Multi-class classification; Combination of information; Explainability

Funding

  1. Madrid Autonomous Community [IND2018/TIC-9665]
  2. Spanish Science and Innovation, under the Retos-Colaboracion program: SABERMED [RTC-2017-6253-1]
  3. Retos-Investigacion program: MODAS-IN [RTI-2018-094269-B-I00]

Ask authors/readers for more resources

This paper introduces several performance metrics for evaluating the performance of Machine Learning models in classification problems. It proposes a methodological approach, the General Performance Score (GPS), for constructing performance metrics for binary and multi-class classification problems. The GPS combines multiple individual metrics to obtain a conservative combination. Experimental results show that the metrics built using the proposed method improve the stability and explainability compared to alternative metrics.
Several performance metrics are currently available to evaluate the performance of Machine Learning (ML) models in classification problems. ML models are usually assessed using a single measure because it facilitates the comparison between several models. However, there is no silver bullet since each performance metric emphasizes a different aspect of the classification. Thus, the choice depends on the particular requirements and characteristics of the problem. An additional problem arises in multi-class classification problems, since most of the well-known metrics are only directly applicable to binary classification problems. In this paper, we propose the General Performance Score (GPS), a methodological approach to build performance metrics for binary and multi-class classification problems. The basic idea behind GPS is to combine a set of individual metrics, penalising low values in any of them. Thus, users can combine several performance metrics that are relevant in the particular problem based on their preferences obtaining a conservative combination. Different GPS-based performance metrics are compared with alternatives in classification problems using real and simulated datasets. The metrics built using the proposed method improve the stability and explainability of the usual performance metrics. Finally, the GPS brings benefits in both new research lines and practical usage, where performance metrics tailored for each particular problem are considered.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available