Back to Search Start Over

Layer Flexible Adaptive Computation Time

Authors :
Abdolghani Ebrahimi
Lida Zhang
Diego Klabjan
Source :
IJCNN
Publication Year :
2021
Publisher :
IEEE, 2021.

Abstract

Deep recurrent neural networks perform well on sequence data and are the model of choice. However, it is a daunting task to decide the structure of the networks, i.e. the number of layers, especially considering different computational needs of a sequence. We propose a layer flexible recurrent neural network with adaptive computation time, and expand it to a sequence to sequence model. Different from the adaptive computation time model, our model has a dynamic number of transmission states which vary by step and sequence. We evaluate the model on a financial data set and Wikipedia language modeling. Experimental results show the performance improvement of 7 % to 12 % and indicate the model's ability to dynamically change the number of layers along with the computational steps.

Details

Database :
OpenAIRE
Journal :
2021 International Joint Conference on Neural Networks (IJCNN)
Accession number :
edsair.doi...........5f944c2fa3fefbae2db352d44a15041b
Full Text :
https://doi.org/10.1109/ijcnn52387.2021.9534317