4.7 Article

AECID: Asymmetric entropy for classifying imbalanced data

Journal

INFORMATION SCIENCES
Volume 467, Issue -, Pages 373-397

Publisher

ELSEVIER SCIENCE INC
DOI: 10.1016/j.ins.2018.07.076

Keywords

Asymmetric entropy; Offset split criterion; Class imbalance learning (CIL); IBA; B42

Ask authors/readers for more resources

In class imbalance problems, it is often more important and expensive to recognize examples from the minority class than from the majority. Standard entropies are known to exhibit poor performance towards the rare class since they take their maximal value for the uniform distribution. To deal with this issue, the present paper introduced a novel adaption of the decision-tree algorithm to imbalanced data situations. We focused, more specifically, on how to let the split criterion discriminate the minority-class examples on a binary-classification problem. Our algorithm uses a new asymmetric entropy measure, termed AECID, which adjusts the most uncertain class distribution to the prior class distribution and includes it in the evaluation of a node impurity. Unlike most competitive split, which include only the prior imbalanced class distribution in their formula, the proposed entropy is customizable with an adjustable concavity to take into account the specificities of each data-set and to better comply with the users' requirements. Extensive experiments were conducted on thirty-six real life imbalanced data-sets to apprise the effectiveness of the proposed approach. Furthermore, the comparative results prove that the new proposal outperforms various algorithmic, data level and ensemble approaches that have been already proposed for imbalanced learning. (C) 2018 Elsevier Inc. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available