4.6 Article

Residual error based knowledge distillation

期刊

NEUROCOMPUTING
卷 433, 期 -, 页码 154-161

出版社

ELSEVIER
DOI: 10.1016/j.neucom.2020.10.113

关键词

Model compression; Knowledge distillation; Residual learning

资金

  1. National Natural Science dation of China [61572354]

向作者/读者索取更多资源

Knowledge distillation is a popular method for model compression, transferring knowledge from a deep teacher model to a shallower student model. The RKD method introduces an assistant model to further distill knowledge, achieving appealing results on popular classification datasets.
Knowledge distillation (KD) is one of the most popular ways for model compression. The key idea is to transfer the knowledge from a deep teacher model (T) to a shallower student (S). However, existing methods suffer from performance degradation due to the substantial gap between the learning capacities of S and T. To remedy this problem, this paper proposes Residual error based Knowledge Distillation (RKD), which further distills the knowledge by introducing an assistant model(A). Specifically, S is trained to mimic the feature maps of T, and A aids this process by learning the residual error between them. In this way, S and A complement with each other to get better knowledge from T. Furthermore, we devise an effective method to derive S and A from a given model without increasing the total computational cost. Extensive experiments show that our approach achieves appealing results on popular classification data sets, CIFAR-100 and ImageNet, surpassing state-of-the-art methods and keep strong robustness to adversarial samples. CO 2020 Published by Elsevier B.V.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.6
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据