4.5 Article

NeST: A Neural Network Synthesis Tool Based on a Grow-and-Prune Paradigm

期刊

IEEE TRANSACTIONS ON COMPUTERS
卷 68, 期 10, 页码 1487-1497

出版社

IEEE COMPUTER SOC
DOI: 10.1109/TC.2019.2914438

关键词

Architecture synthesis; grow-and-prune paradigm; network parameters; neural network

资金

  1. US National Science Foundation [CNS-1617640]

向作者/读者索取更多资源

Deep neural networks (DNNs) have begun to have a pervasive impact on various applications of machine learning. However, the problem of finding an optimal DNN architecture for large applications is challenging. Common approaches go for deeper and larger DNN architectures but may incur substantial redundancy. To address these problems, we introduce a network growth algorithm that complements network pruning to learn both weights and compact DNN architectures during training. We propose a DNN synthesis tool (NeST) that combines both methods to automate the generation of compact and accurate DNNs. NeSTstarts with a randomly initialized sparse network called the seed architecture. It iteratively tunes the architecture with gradient-based growth and magnitude-based pruning of neurons and connections. Our experimental results show that NeST yields accurate, yet very compact DNNs, with a wide range of seed architecture selection. For the LeNet-300-100 (LeNet-5) architecture, we reduce network parameters by 70.2 x (74.3 x) and floating-point operations (FLOPs) by 79.4 x (43.7 x). For the AlexNet, VGG-16, and ResNet-50 architectures, we reduce network parameters (FLOPs) by 15.7 x (4.6 x), 33.2 x (8.9 x), and 4.1 x (2.1 x) respectively. NeST's grow-and-prune paradigm delivers significant additional parameter and FLOPs reduction relative to pruning-only methods.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.5
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据