1. Learning sparse reparameterization with layer-wise continuous sparsification.
- Author
-
Wang, Xiaodong, Huang, Yaxiang, Zeng, Xianxian, Guo, Jianlan, and Chen, Yuqiang
- Subjects
- *
ARTIFICIAL neural networks , *MACHINE learning , *HTTP (Computer network protocol) , *WEIGHT training , *LOTTERY tickets , *PARAMETERIZATION - Abstract
Sparse reparameterization in Deep Neural Networks (DNNs) aims to achieve a better tradeoff between the network parameter count and performance. Recently, the lottery ticket hypothesis suppose that excellent sub-networks ("winning tickets") exist in dense randomly-initialized networks. These sparse sub-networks trained from scratch are able to reach the performance of their dense counterparts. Compared with Iterative Magnitude Pruning that relies on pruning strategies, the Continuous Sparsification algorithm learns the "winning tickets" with gradient-based methods, achieving better performance. In this paper, we propose Layer-wise Continuous Sparsification (LCS) scheme for finding sparse sub-networks, in which the parameterized relaxation of step functions used to remove network parameters in each layer is integrated into the DNN loss as an optimization objective. LCS utilizes a family of sigmoid functions to asynchronously filter important per-layer weights throughout training, yielding sparser and better sub-networks. Experiments show that our method surpasses state-of-the-art methods for sparse reparameterization. Additionally, the proposed method can be utilized as a regularization technique to further improve the accuracy of dense networks 1 1 Our code is publicly available at https://github.com/RiyaoDong/LCS.. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF