4.8 Article

Good Practice in Large-Scale Learning for Image Classification

Publisher

IEEE COMPUTER SOC
DOI: 10.1109/TPAMI.2013.146

Keywords

Large scale; fine-grained visual categorization; image classification; ranking; SVM; stochastic learning

Funding

  1. QUAERO project
  2. OSEO
  3. French State agency for innovation
  4. European integrated project AXES
  5. ERC grant ALLEGRO
  6. GARGANTUA project
  7. Mastodons program of CNRS

Ask authors/readers for more resources

We benchmark several SVM objective functions for large-scale image classification. We consider one-versus-rest, multiclass, ranking, and weighted approximate ranking SVMs. A comparison of online and batch methods for optimizing the objectives shows that online methods perform as well as batch methods in terms of classification accuracy, but with a significant gain in training speed. Using stochastic gradient descent, we can scale the training to millions of images and thousands of classes. Our experimental evaluation shows that ranking-based algorithms do not outperform the one-versus-rest strategy when a large number of training examples are used. Furthermore, the gap in accuracy between the different algorithms shrinks as the dimension of the features increases. We also show that learning through cross-validation the optimal rebalancing of positive and negative examples can result in a significant improvement for the one-versus-rest strategy. Finally, early stopping can be used as an effective regularization strategy when training with online algorithms. Following these good practices, we were able to improve the state of the art on a large subset of 10K classes and 9M images of ImageNet from 16.7 percent Top-1 accuracy to 19.1 percent.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.8
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available