4.7 Article

Low-Resolution Face Recognition in the Wild via Selective Knowledge Distillation

Journal

IEEE TRANSACTIONS ON IMAGE PROCESSING
Volume 28, Issue 4, Pages 2051-2062

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TIP.2018.2883743

Keywords

Face recognition in the wild; two-stream architecture; knowledge distillation; CNNs

Funding

  1. National Key Research and Development Plan [2016YFC0801005]
  2. National Natural Science Foundation of China [61772513, 61672072]
  3. Beijing Nova Program [Z181100006218063]
  4. International Cooperation Project of the Institute of Information Engineering, Chinese Academy of Sciences [Y7Z0511101]
  5. Youth Innovation Promotion Association, CAS

Ask authors/readers for more resources

Typically, the deployment of face recognition models in the wild needs to identify low-resolution faces with extremely low computational cost. To address this problem, a feasible solution is compressing a complex face model to achieve higher speed and lower memory at the cost of minimal performance drop. Inspired by that, this paper proposes a learning approach to recognize low-resolution faces via selective knowledge distillation. In this approach, a two-stream convolutional neural network (CNN) is first initialized to recognize high-resolution faces and resolution-degraded faces with a teacher stream and a student stream, respectively. The teacher stream is represented by a complex CNN for high-accuracy recognition, and the student stream is represented by a much simpler CNN for low-complexity recognition. To avoid significant performance drop at the student stream, we then selectively distil the most informative facial features from the teacher stream by solving a sparse graph optimization problem, which are then used to regularize the fine-tuning process of the student stream. In this way, the student stream is actually trained by simultaneously handling two tasks with limited computational resources: approximating the most informative facial cues via feature regression, and recovering the missing facial cues via low-resolution face classification. Experimental results show that the student stream performs impressively in recognizing low-resolution faces and costs only 0.15-MB memory and runs at 418 faces per second on CPU and 9433 faces per second on GPU.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.7
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available