4.3 Article

Block pruning residual networks using Multi-Armed Bandits

出版社

TAYLOR & FRANCIS LTD
DOI: 10.1080/0952813X.2023.2247412

关键词

Deep learning; Neural network compression; Multi-armed bandits

向作者/读者索取更多资源

This paper proposes a new block pruning method for residual networks based on the Multi-Armed Bandits framework. The method identifies and prunes less important residual blocks to reduce model size and computational cost. Experimental results show the effectiveness of the method.
Recently, deep neural networks have been adopted in many application domains showing impressive performances that sometimes even exceed those of humans. However, the considerable size of deep learning models and their computational cost limits their use in mobile and embedded systems. Pruning is an important technique that has been used in the literature to compress a trained model by removing some of its elements (weights, neurons, filters, etc.) that do not notably contribute to its performance. In this paper, we propose a new block pruning method for residual networks that removes holes layers of the network based on the Multi-Armed Bandits (MAB) framework. The proposed method takes advantage of the existence of multiple trainable paths in residual models to identify and prune the less important residual blocks for inference. To evaluate our approach, we conducted extensive experiments using three different MAB algorithms, namely $\varepsilon $e-greedy, Win Stay Lose Shift (WSLS), and the Upper Confidence Bound (UCB). The obtained results on different datasets show that our block pruning method can considerably reduce the model size and FLOPs with little to no impact on overall accuracy. We conducted a comparative analysis of our method with three existing techniques to demonstrate that our approach is highly competitive in terms of accuracy, while significantly reducing both model size and FLOPs. Moreover, our experimental results demonstrate that when our block pruning method is combined with a filter-level pruning technique, it leads to even greater reductions in model size and FLOPs while maintaining a relatively high accuracy.

作者

我是这篇论文的作者
点击您的名字以认领此论文并将其添加到您的个人资料中。

评论

主要评分

4.3
评分不足

次要评分

新颖性
-
重要性
-
科学严谨性
-
评价这篇论文

推荐

暂无数据
暂无数据