Back to Search Start Over

Time-Smoothed Gradients for Online Forecasting

Authors :
Zhu, Tianhao
Aydore, Sergul
Publication Year :
2019

Abstract

Here, we study different update rules in stochastic gradient descent (SGD) for online forecasting problems. The selection of the learning rate parameter is critical in SGD. However, it may not be feasible to tune this parameter in online learning. Therefore, it is necessary to have an update rule that is not sensitive to the selection of the learning parameter. Inspired by the local regret metric that we introduced previously, we propose to use time-smoothed gradients within SGD update. Using the public data set-- GEFCom2014, we validate that our approach yields more stable results than the other existing approaches. Furthermore, we show that such a simple approach is computationally efficient compared to the alternatives.<br />ICML 2019, time series workshop

Details

Language :
English
Database :
OpenAIRE
Accession number :
edsair.doi.dedup.....f9034613083225b218e60074e74399fb