4.7 Article

Entropy-based fuzzy support vector machine for imbalanced datasets

期刊

KNOWLEDGE-BASED SYSTEMS
卷 115, 期 -, 页码 87-99

出版社

ELSEVIER
DOI: 10.1016/j.knosys.2016.09.032

关键词

Information entropy; Fuzzy support vector machine; Imbalanced dataset; Pattern recognition

资金

  1. Natural Science Foundations of China [61672227, 61272198]
  2. Innovation Program of Shanghai Municipal Education Commission [14ZZ054]
  3. Shanghai Key Laboratory of Intelligent Information Processing of China [IIPL-2012-003]
  4. Fundamental Research Funds for the Central Universities
  5. Provincial Key Laboratory for Computer Information Processing Technology of Soochow University [KJS1323]

向作者/读者索取更多资源

Imbalanced problem occurs when the size of the positive class is much smaller than that of the negative one. Positive class usually refers to the main interest of the classification task. Although conventional Support Vector Machine (SVM) results in relatively robust classification performance on imbalanced datasets, it treats all samples with the same importance leading to the decision surface biasing toward the negative class. To overcome this inherent drawback, Fuzzy SVM (FSVM) is proposed by applying fuzzy membership to training samples such that different samples provide different contributions to the classifier. However, how to evaluate an appropriate fuzzy membership is the main issue to FSVM. In this paper, we propose a novel fuzzy membership evaluation which determines the fuzzy membership based on the class certainty of samples. That is, the samples with higher class certainty are assigned to larger fuzzy memberships. As the entropy is utilized to measure the class certainty, the fuzzy membership evaluation is named as entropy-based fuzzy membership evaluation. Therefore, the Entropy-based FSVM (EFSVM) is proposed by using the entropy-based fuzzy membership. EFSVM can pay more attention to the samples with higher class certainty, i.e. enhancing the importance of samples with high class certainty. Meanwhile, EFSVM guarantees the importance of the positive class by assigning positive samples to relatively large fuzzy memberships. The contributions of this work are: (1) proposing a novel entropy-based fuzzy membership evaluation method which enhances the importance of certainty samples, (2) guaranteeing the importance of the positive samples to result in a more flexible decision surface. Experiments on imbalanced datasets validate that EFSV outperforms the compared algorithms. (C) 2016 Elsevier B.V. All rights reserved.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据