Back to Search Start Over

Ising-dropout: A Regularization Method for Training and Compression of Deep Neural Networks

Authors :
Hojjat Salehinejad
Shahrokh Valaee
Source :
ICASSP
Publication Year :
2019
Publisher :
IEEE, 2019.

Abstract

Overfitting is a major problem in training machine learning models, specifically deep neural networks. This problem may be caused by imbalanced datasets and initialization of the model parameters, which conforms the model too closely to the training data and negatively affects the generalization performance of the model for unseen data. The original dropout is a regularization technique to drop hidden units randomly during training. In this paper, we propose an adaptive technique to wisely drop the visible and hidden units in a deep neural network using Ising energy of the network. The preliminary results show that the proposed approach can keep the classification performance competitive to the original network while eliminating optimization of unnecessary network parameters in each training cycle. The dropout state of units can also be applied to the trained (inference) model. This technique could compress the network in terms of number of parameters up to 41.18% and 55.86% for the classification task on the MNIST and Fashion-MNIST datasets, respectively.<br />This paper is accepted at 44th IEEE International Conference on Acoustics, Speech and Signal Processing (IEEE ICASSP), 2019

Details

Database :
OpenAIRE
Journal :
ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
Accession number :
edsair.doi.dedup.....da31f440f5f906c4731a7f09d2dc1e15