Back to Search Start Over

Time series forecasting by recurrent product unit neural networks

Authors :
César Hervás-Martínez
A. Castaño
Pedro Antonio Gutiérrez
Francisco Fernández-Navarro
Maria Angeles Cruz
Source :
Brújula, Universidad Loyola Andalucía
Publication Year :
2016
Publisher :
Springer Science and Business Media LLC, 2016.

Abstract

Time series forecasting (TSF) consists on estimating models to predict future values based on previously observed values of time series, and it can be applied to solve many real-world problems. TSF has been traditionally tackled by considering autoregressive neural networks (ARNNs) or recurrent neural networks (RNNs), where hidden nodes are usually configured using additive activation functions, such as sigmoidal functions. ARNNs are based on a short-term memory of the time series in the form of lagged time series values used as inputs, while RNNs include a long-term memory structure. The objective of this paper is twofold. First, it explores the potential of multiplicative nodes for ARNNs, by considering product unit (PU) activation functions, motivated by the fact that PUs are specially useful for modelling highly correlated features, such as the lagged time series values used as inputs for ARNNs. Second, it proposes a new hybrid RNN model based on PUs, by estimating the PU outputs from the combination of a long-term reservoir and the short-term lagged time series values. A complete set of experiments with 29 data sets shows competitive performance for both model proposals, and a set of statistical tests confirms that they achieve the state of the art in TSF, with specially promising results for the proposed hybrid RNN. The experiments in this paper show that the recurrent model is very competitive for relatively large time series, where longer forecast horizons are required, while the autoregressive model is a good selection if the data set is small or if a low computational cost is needed.

Details

ISSN :
14333058 and 09410643
Volume :
29
Database :
OpenAIRE
Journal :
Neural Computing and Applications
Accession number :
edsair.doi.dedup.....a74c5a699b3094d38bbe2a685bd98e8f
Full Text :
https://doi.org/10.1007/s00521-016-2494-2