4.5 Article

Designing Efficient DNNs via Hardware-Aware Neural Architecture Search and Beyond

Publisher

IEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC
DOI: 10.1109/TCAD.2021.3100249

Keywords

Hardware; Computer architecture; Search problems; Runtime; Adaptation models; Computational modeling; Neural networks; Adaptive inference; deep neural networks (DNNs); hardware performance modeling; neural architecture search (NAS)

Funding

  1. Ministry of Education, Singapore [MOE2019T2-1-071, MOE2019-T1-001-072]
  2. Nanyang Technological University, Singapore [M4082282, M4082087]

Ask authors/readers for more resources

This article introduces GoldenNAS, a hardware-aware neural architecture search framework, for automating the design of efficient deep neural networks. The authors propose techniques such as dynamic channel scaling, progressive space shrinking, and hardware performance modeling to improve the efficiency and accuracy of the search. They employ the evolutionary algorithm for searching and introduce adaptive batch normalization and self-knowledge distillation techniques for improving network accuracy.
Hardware systems integrated with deep neural networks (DNNs) are deemed to pave the way for future artificial intelligence (AI). However, manually designing efficient DNNs involves nontrivial computation resources since significant trial-and-errors are required to finalize the network configuration. To this end, we, in this article, introduce a novel hardware-aware neural architecture search (NAS) framework, namely, GoldenNAS, to automate the design of efficient DNNs. To begin with, we present a novel technique, called dynamic channel scaling, to enable the channel-level search since the number of channels has non-negligible impacts on both accuracy and efficiency. Besides, we introduce an efficient progressive space shrinking method to raise the awareness of the search space toward target hardware and alleviate the search overheads as well. Moreover, we propose an effective hardware performance modeling method to approximate the runtime latency of DNNs upon target hardware, which is further integrated into GoldenNAS to avoid the tedious on-device measurements. Then, we employ the evolutionary algorithm (EA) to search for the optimal operator/channel configurations of DNNs, denoted as GoldenNets. Finally, to enable the depthwise adaptiveness of GoldenNets under dynamic environments, we propose the adaptive batch normalization (ABN) technique, followed by the self-knowledge distillation (SKD) approach to improve the accuracy of adaptive subnetworks. We conduct extensive experiments directly on ImageNet, which clearly demonstrate the advantages of GoldenNAS over existing state-of-the-art approaches.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.5
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available