Back to Search Start Over

A New Network Pruning Framework Based on Rewind.

Authors :
Yu, Haoran
Zhang, Weiwei
Zhou, Hongbo
Ma, Xin
Source :
International Journal of Pattern Recognition & Artificial Intelligence. Sep2022, Vol. 36 Issue 12, p1-19. 19p.
Publication Year :
2022

Abstract

Model pruning is one of the main methods of deep neural network model compression. However, the existing model pruning methods are inefficient, there is still a lot of redundancy in the network, and the pruning has a great impact on the accuracy. In addition, the traditional pruning process usually needs to fine-tune the network to restore accuracy, and fine-tuning has a limited effect on accuracy recovery, so it is difficult to achieve a high level of accuracy. In this paper, we propose a new neural network pruning framework: the channel sparsity is realized by introducing a scale factor, and the sparse network is pruned by setting a global threshold, which greatly improves the efficiency of pruning. After pruning, this paper proposes a rewind method to restore the accuracy, that is, save the weight after training, and then reload it on the network for training to restore the accuracy. In addition, we also study the best rewind point of the three networks. The experimental results show that our method significantly reduces the number of parameters and FLOPs without affecting or even improving the accuracy, and the rewind method proposed by us achieves a better accuracy recovery effect than fine-tuning. At the same time, we find that the epoch with the highest accuracy is the best rewind point, and the accuracy is the highest after saving its corresponding weight and retraining the model. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
02180014
Volume :
36
Issue :
12
Database :
Academic Search Index
Journal :
International Journal of Pattern Recognition & Artificial Intelligence
Publication Type :
Academic Journal
Accession number :
159689003
Full Text :
https://doi.org/10.1142/S0218001422590285