4.6 Article

MSSBoost: A new multiclass boosting to semi-supervised learning

Journal

NEUROCOMPUTING
Volume 314, Issue -, Pages 251-266

Publisher

ELSEVIER
DOI: 10.1016/j.neucom.2018.06.047

Keywords

Multiclass classification; Semi-supervised learning; Similarity learning; Boosting

Funding

  1. IPM [CS1396-4-69]

Ask authors/readers for more resources

In this article, we focus on the multiclass classification problem to semi-supervised learning. Semi-supervised learning is a learning task from both labeled and unlabeled data points. We formulate the multiclass semi-supervised classification problem as an optimization problem. In this formulation, we combine the classifier predictions, based on the labeled data, and the pairwise similarity. The goal here is to minimize the inconsistency between classifier predictions and the pairwise similarity. A boosting algorithm is proposed to solve the multiclass classification problem directly. The proposed multiclass approach uses a new multiclass formulation to loss function, which includes two terms. The first term is the multiclass margin cost of the labeled data and the second term is a regularization term on unlabeled data. The regularization term is used to minimize the inconsistency between the pairwise similarity and the classifier predictions. It in fact assigns the soft labels weighted with the similarity between unlabeled and labeled examples. First, the gradient descent approach is used to solve the resulting optimization problem and derive a boosting algorithm, named MSSBoost. The derived algorithm also uses a learning optimal similarity function for a given data. The second approach to solve the optimization problem is to apply the coordinate gradient descent. The resulting algorithm is called CD-MSSB. We also use a variation of CD-MSSB in the experiments. The results of our experiments on a number of UCI and real-world text classification benchmark datasets show that MSSBoost and CD-MSSB outperform the state-of-the-art boosting methods to multiclass semi-supervised learning. Another observation is that the proposed methods exploit the informative unlabeled data. (C) 2018 Elsevier B.V. All rights reserved.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available