4.7 Article

Large-Margin Label-Calibrated Support Vector Machines for Positive and Unlabeled Learning

期刊

出版社

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TNNLS.2019.2892403

关键词

Support vector machines; Calibration; Training; Training data; Data models; Learning systems; Intserv networks; Label calibration; large margin; positive and unlabeled learning (PU Learning); support vector machines (SVMs)

资金

  1. NSF of China [61602246, U1713208]
  2. NSF of Jiangsu Province [BK20171430]
  3. Fundamental Research Funds for the Central Universities [30918011319]
  4. State Key Laboratory of Integrated Services Networks, Xidian University [ISN1903]
  5. 111 Program [AH92005]
  6. Summit of the Six Top Talents Program [DZXX-027]
  7. Program for Changjiang Scholars
  8. Lift Program for Young Talents of Jiangsu Province
  9. CAST Lift Program for Young Talents
  10. Australian Research Council [FL-170100117, DE-1901014738, DP-180103424]

向作者/读者索取更多资源

Positive and unlabeled learning (PU learning) aims to train a binary classifier based on only PU data. Existing methods usually cast PU learning as a label noise learning problem or a cost-sensitive learning problem. However, none of them fully take the data distribution information into consideration when designing the model, which hinders them from acquiring more encouraging performance. In this paper, we argue that the clusters formed by positive examples and potential negative examples in the feature space should be critically utilized to establish the PU learning model, especially when the negative data are not explicitly available. To this end, we introduce a hat loss to discover the margin between data clusters, a label calibration regularizer to amend the biased decision boundary to the potentially correct one, and propose a novel discriminative PU classifier termed Large-margin Label-calibrated Support Vector Machines (LLSVM). Our LLSVM classifier can work properly in the absence of negative training examples and effectively achieve the max-margin effect between positive and negative classes. Theoretically, we derived the generalization error bound of LLSVM which reveals that the introduction of PU data does help to enhance the algorithm performance. Empirically, we compared LLSVM with state-of-the-art PU methods on various synthetic and practical data sets, and the results confirm that the proposed LLSVM is more effective than other compared methods on dealing with PU learning tasks.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.7
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据