Back to Search Start Over

EDropout: Energy-Based Dropout and Pruning of Deep Neural Networks

Authors :
Shahrokh Valaee
Hojjat Salehinejad
Source :
IEEE transactions on neural networks and learning systems. 33(10)
Publication Year :
2021

Abstract

Dropout is a well-known regularization method by sampling a sub-network from a larger deep neural network and training different sub-networks on different subsets of the data. Inspired by the dropout concept, we propose EDropout as an energy-based framework for pruning neural networks in classification tasks. In this approach, a set of binary pruning state vectors (population) represents a set of corresponding sub-networks from an arbitrary original neural network. An energy loss function assigns a scalar energy loss value to each pruning state. The energy-based model (EBM) stochastically evolves the population to find states with lower energy loss. The best pruning state is then selected and applied to the original network. Similar to dropout, the kept weights are updated using backpropagation in a probabilistic model. The EBM again searches for better pruning states and the cycle continuous. This procedure is a switching between the energy model, which manages the pruning states, and the probabilistic model, which updates the kept weights, in each iteration. The population can dynamically converge to a pruning state. This can be interpreted as dropout leading to pruning the network. From an implementation perspective, unlike most of the pruning methods, EDropout can prune neural networks without manually modifying the network architecture code. We have evaluated the proposed method on different flavors of ResNets, AlexNet, l₁ pruning, ThinNet, ChannelNet, and SqueezeNet on the Kuzushiji, Fashion, CIFAR-10, CIFAR-100, Flowers, and ImageNet data sets, and compared the pruning rate and classification performance of the models. The networks trained with EDropout on average achieved a pruning rate of more than 50% of the trainable parameters with approximately

Details

ISSN :
21622388
Volume :
33
Issue :
10
Database :
OpenAIRE
Journal :
IEEE transactions on neural networks and learning systems
Accession number :
edsair.doi.dedup.....f2cfffa7aed3cdfea5a53749e7ee4af6