Back to Search Start Over

Applying and dissecting LSTM neural networks and regularized learning for dynamic inferential modeling.

Authors :
Li, Jicheng
Qin, S. Joe
Source :
Computers & Chemical Engineering. Jul2023, Vol. 175, pN.PAG-N.PAG. 1p.
Publication Year :
2023

Abstract

Deep learning models such as the long short-term memory (LSTM) network have been applied for dynamic inferential modeling. However, many studies apply LSTM as a black-box approach without examining the necessity and usefulness of the internal LSTM gates for inferential modeling. In this paper, we use LSTM as a state space realization and compare it with linear state space modeling and statistical learning methods, including N4SID, partial least squares, the Lasso, and support vector regression. Two case studies on an industrial 660 MW boiler and a debutanizer column process indicate that LSTM underperforms all other methods. LSTM is shown to be capable of outperforming linear methods for a simulated reactor process with severely excited nonlinearity in the data. By dissecting the sub-components of a simple LSTM model, the effectiveness of the LSTM gates and nonlinear activation functions is scrutinized. • LSTM is implemented as a nonlinear Kalman filter for dynamic inferential modeling. • LSTM is benchmarked with N4SID, PLS, Lasso, and SVR statistical learning methods. • Results on a boiler emission and debutanizer and a CSTR process are reported. • The results show that LSTM underperform N4SID and statistical learning methods. • With severe nonlinearity in the simulated CSTR data, LSTM underperforms SVR. [ABSTRACT FROM AUTHOR]

Details

Language :
English
ISSN :
00981354
Volume :
175
Database :
Academic Search Index
Journal :
Computers & Chemical Engineering
Publication Type :
Academic Journal
Accession number :
163795551
Full Text :
https://doi.org/10.1016/j.compchemeng.2023.108264