4.7 Article

Coarse-to-Fine Description for Fine-Grained Visual Categorization

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 25, Issue 10, Pages 4858-4872

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2016.2599102

Keywords

-

Funding

  1. National High Technology Research and Development Program of China [2014AA015202]
  2. National Nature Science Foundation of China [61525206, 61428207, 61572050, 91538111, 61429201]
  3. Beijing Advanced Innovation Center for Imaging Technology [BAICIT-2016009]
  4. ARO [W911NF-15-1-0290]
  5. Faculty Research Gift Awards by NEC Laboratories of America
  6. Blippar

Ask authors/readers for more resources

Recent years have witnessed the significant advance in fine-grained visual categorization, which targets to classify the objects belonging to the same species. To capture enough subtle visual differences and build discriminative visual description, most of the existing methods heavily rely on the artificial part annotations, which are expensive to collect in real applications. Motivated to conquer this issue, this paper proposes a multilevel coarse-to-fine object description. This novel description only requires the original image as input, but could automatically generate visual descriptions discriminative enough for fine-grained visual categorization. This description is extracted from five sources representing coarse-to-fine visual clues: 1) original image is used as the source of global visual clue; 2) object bounding boxes are generated using convolutional neural network (CNN); 3) with the generated bounding box, foreground is segmented using the proposed k nearest neighbour-based co-segmentation algorithm; and 4) two types of part segmentations are generated by dividing the foreground with an unsupervised part learning strategy. The final description is generated by feeding these sources into CNN models and concatenating their outputs. Experiments on two public benchmark data sets show the impressive performance of this coarse-to-fine description, i.e., classification accuracy achieves 82.5% on CUB-200-2011, and 86.9% on fine-grained visual categorization-Aircraft, respectively, which outperform many recent works.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available