17 results on '"long-term dependency"'
Search Results
2. Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks
- Author
-
Cristian Ubal, Gustavo Di-Giorgi, Javier E. Contreras-Reyes, and Rodrigo Salas
- Subjects
long-term dependency ,Hurst exponent ,fractional differentiation ,recurrent neural networks ,Computer engineering. Computer hardware ,TK7885-7895 - Abstract
Long-term dependence is an essential feature for the predictability of time series. Estimating the parameter that describes long memory is essential to describing the behavior of time series models. However, most long memory estimation methods assume that this parameter has a constant value throughout the time series, and do not consider that the parameter may change over time. In this work, we propose an automated methodology that combines the estimation methodologies of the fractional differentiation parameter (and/or Hurst parameter) with its application to Recurrent Neural Networks (RNNs) in order for said networks to learn and predict long memory dependencies from information obtained in nonlinear time series. The proposal combines three methods that allow for better approximation in the prediction of the values of the parameters for each one of the windows obtained, using Recurrent Neural Networks as an adaptive method to learn and predict the dependencies of long memory in Time Series. For the RNNs, we have evaluated four different architectures: the Simple RNN, LSTM, the BiLSTM, and the GRU. These models are built from blocks with gates controlling the cell state and memory. We have evaluated the proposed approach using both synthetic and real-world data sets. We have simulated ARFIMA models for the synthetic data to generate several time series by varying the fractional differentiation parameter. We have evaluated the proposed approach using synthetic and real datasets using Whittle’s estimates of the Hurst parameter classically obtained in each window. We have simulated ARFIMA models in such a way that the synthetic data generate several time series by varying the fractional differentiation parameter. The real-world IPSA stock option index and Tree Ringtime series datasets were evaluated. All of the results show that the proposed approach can predict the Hurst exponent with good performance by selecting the optimal window size and overlap change.
- Published
- 2023
- Full Text
- View/download PDF
3. Modeling air quality PM2.5 forecasting using deep sparse attention-based transformer networks.
- Author
-
Zhang, Z. and Zhang, S.
- Abstract
Air quality forecasting is of great importance in environmental protection, government decision-making, people's daily health, etc. Existing research methods have failed to effectively modeling long-term and complex relationships in time series PM2.5 data and exhibited low precision in long-term prediction. To address this issue, in this paper a new lightweight deep learning model using sparse attention-based Transformer networks (STN) consisting of encoder and decoder layers, in which a multi-head sparse attention mechanism is adopted to reduce the time complexity, is proposed to learn long-term dependencies and complex relationships from time series PM2.5 data for modeling air quality forecasting. Extensive experiments on two real-world datasets in China, i.e., Beijing PM2.5 dataset and Taizhou PM2.5 dataset, show that our proposed method not only has relatively small time complexity, but also outperforms state-of-the-art methods, demonstrating the effectiveness of the proposed STN method on both short-term and long-term air quality prediction tasks. In particular, on singe-step PM2.5 forecasting tasks our proposed method achieves R
2 of 0.937 and reduces RMSE to 19.04 µg/m3 and MAE to 11.13 µg/m3 on Beijing PM2.5 dataset. Also, our proposed method obtains R2 of 0.924 and reduces RMSE to 5.79 µg/m3 and MAE to 3.76 µg/m3 on Taizhou PM2.5 dataset. For long-term time step prediction, our proposed method still performs best among all used methods on multi-step PM2.5 forecasting results for the next 6, 12, 24, and 48 h on two real-world datasets. [ABSTRACT FROM AUTHOR]- Published
- 2023
- Full Text
- View/download PDF
4. Predicting the Long-Term Dependencies in Time Series Using Recurrent Artificial Neural Networks.
- Author
-
Ubal, Cristian, Di-Giorgi, Gustavo, Contreras-Reyes, Javier E., and Salas, Rodrigo
- Subjects
RECURRENT neural networks ,TIME series analysis ,STOCK options ,PARAMETER estimation ,TIME management ,ARTIFICIAL neural networks - Abstract
Long-term dependence is an essential feature for the predictability of time series. Estimating the parameter that describes long memory is essential to describing the behavior of time series models. However, most long memory estimation methods assume that this parameter has a constant value throughout the time series, and do not consider that the parameter may change over time. In this work, we propose an automated methodology that combines the estimation methodologies of the fractional differentiation parameter (and/or Hurst parameter) with its application to Recurrent Neural Networks (RNNs) in order for said networks to learn and predict long memory dependencies from information obtained in nonlinear time series. The proposal combines three methods that allow for better approximation in the prediction of the values of the parameters for each one of the windows obtained, using Recurrent Neural Networks as an adaptive method to learn and predict the dependencies of long memory in Time Series. For the RNNs, we have evaluated four different architectures: the Simple RNN, LSTM, the BiLSTM, and the GRU. These models are built from blocks with gates controlling the cell state and memory. We have evaluated the proposed approach using both synthetic and real-world data sets. We have simulated ARFIMA models for the synthetic data to generate several time series by varying the fractional differentiation parameter. We have evaluated the proposed approach using synthetic and real datasets using Whittle's estimates of the Hurst parameter classically obtained in each window. We have simulated ARFIMA models in such a way that the synthetic data generate several time series by varying the fractional differentiation parameter. The real-world IPSA stock option index and Tree Ringtime series datasets were evaluated. All of the results show that the proposed approach can predict the Hurst exponent with good performance by selecting the optimal window size and overlap change. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
5. PASPP Medical Transformer for Medical Image Segmentation
- Author
-
Lai, Hong-Phuc, Tran, Thi-Thao, Pham, Van-Truong, Kacprzyk, Janusz, Series Editor, Gomide, Fernando, Advisory Editor, Kaynak, Okyay, Advisory Editor, Liu, Derong, Advisory Editor, Pedrycz, Witold, Advisory Editor, Polycarpou, Marios M., Advisory Editor, Rudas, Imre J., Advisory Editor, Wang, Jun, Advisory Editor, Saraswat, Mukesh, editor, Chowdhury, Chandreyee, editor, Kumar Mandal, Chintan, editor, and Gandomi, Amir H., editor
- Published
- 2023
- Full Text
- View/download PDF
6. Layer-wise-residual-driven approach for soft sensing in composite dynamic system based on slow and fast time-varying latent variables.
- Author
-
Zhang, Zhengxuan, Yang, Xu, Huang, Jian, and Shardt, Yuri A.W.
- Subjects
- *
LATENT variables , *LEAST squares , *DYNAMICAL systems , *REGRESSION analysis , *INDUSTRIALISM - Abstract
Driven by the requirements for a comprehensive understanding of composite dynamic systems in industrial processes, this paper investigates a new soft sensor for quality prediction based on slow and fast time-varying latent variables extraction using layer-wise residuals. First, the slow feature partial least squares were expanded into long-term dependency by introducing explicit expressions of the potential state of the process into the objective function. Then, the multilayer regression model for exploring composite dynamics driven by layer-wise residuals is developed using a serial structure that can extract both slow and fast time-varying latent variables that are completely orthogonal. Finally, the exponential-weighted partial least squares are proposed for extracting fast time-varying dynamic latent variables by learning the exponential decay properties of the time-series data correlation. Case studies on the industrial debutanizer and sulfur recovery unit show that the prediction accuracy of the proposed approach outperforms traditional methods. • The composite dynamics with slow time-varying overall trends accompanied by fast time-varying fluctuations can be described. • The slow time-varying quality-related LVs based on the assumption of long-term dependency are extracted for quality prediction. • The EWMA was integrated into the dynamic inner model to extract fast time-varying dynamic LVs with weak autocorrelation. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
7. A Video Summarization Model Based on Deep Reinforcement Learning with Long-Term Dependency.
- Author
-
Wang, Xu, Li, Yujie, Wang, Haoyu, Huang, Longzhao, and Ding, Shuxue
- Subjects
- *
VIDEO summarization , *VIDEOS , *REINFORCEMENT learning - Abstract
Deep summarization models have succeeded in the video summarization field based on the development of gated recursive unit (GRU) and long and short-term memory (LSTM) technology. However, for some long videos, GRU and LSTM cannot effectively capture long-term dependencies. This paper proposes a deep summarization network with auxiliary summarization losses to address this problem. We introduce an unsupervised auxiliary summarization loss module with LSTM and a swish activation function to capture the long-term dependencies for video summarization, which can be easily integrated with various networks. The proposed model is an unsupervised framework for deep reinforcement learning that does not depend on any labels or user interactions. Additionally, we implement a reward function ( R (S) ) that jointly considers the consistency, diversity, and representativeness of generated summaries. Furthermore, the proposed model is lightweight and can be successfully deployed on mobile devices and enhance the experience of mobile users and reduce pressure on server operations. We conducted experiments on two benchmark datasets and the results demonstrate that our proposed unsupervised approach can obtain better summaries than existing video summarization methods. Furthermore, the proposed algorithm can generate higher F scores with a nearly 6.3% increase on the SumMe dataset and a 2.2% increase on the TVSum dataset compared to the DR-DSN model. [ABSTRACT FROM AUTHOR]
- Published
- 2022
- Full Text
- View/download PDF
8. Temporal Difference-Based Graph Transformer Networks For Air Quality PM2.5 Prediction: A Case Study in China
- Author
-
Zhen Zhang, Shiqing Zhang, Xiaoming Zhao, Linjian Chen, and Jun Yao
- Subjects
air quality prediction ,deep learning ,temporal difference ,graph attention ,transformer ,long-term dependency ,Environmental sciences ,GE1-350 - Abstract
Air quality PM2.5 prediction is an effective approach for providing early warning of air pollution. This paper proposes a new deep learning model called temporal difference-based graph transformer networks (TDGTN) to learn long-term temporal dependencies and complex relationships from time series PM2.5 data for air quality PM2.5 prediction. The proposed TDGTN comprises of encoder and decoder layers associated with the developed graph attention mechanism. In particular, considering the similarity of different time moments and the importance of temporal difference between two adjacent moments for air quality PM2.5prediction, we first construct graph-structured data from original time series PM2.5 data at different moments without explicit graph structure. Then we improve the self-attention mechanism with the temporal difference information, and develop a new graph attention mechanism. Finally, the developed graph attention mechanism is embedded into the encoder and decoder layers of the proposed TDGTN to learn long-term temporal dependencies and complex relationships from a graph prospective on air quality PM2.5 prediction tasks. Experiment results on two collected real-world datasets in China, such as Beijing and Taizhou PM2.5 datasets, show that the proposed method outperforms other used methods on both short-term and long-term air quality PM2.5 prediction tasks.
- Published
- 2022
- Full Text
- View/download PDF
9. LEISN: A long explicit–implicit spatio-temporal network for traffic flow forecasting.
- Author
-
Lai, Qiang and Chen, Peng
- Subjects
- *
TRAFFIC flow , *TRAFFIC estimation , *RECURRENT neural networks - Abstract
Recent studies have shown that it is necessary to further improve the prediction accuracy of complex and dynamic traffic flow from two angles: temporal dependence and spatial dependence. Although many spatio-temporal networks are extracting features from both temporal and spatial dependencies, there are still two problems that need to be addressed. Firstly, existing methods often focus on modeling local spatial dependence relationship between adjacent nodes, but neglect non-local spatial dependence relationship between distant nodes. Secondly, while recent work has explored various methods such as convolutional and recurrent neural networks for modeling temporal dependence, they may not fully capture the long-term temporal dependence in traffic data. Therefore, addressing the challenges of long-term dependency, explicit spatial dependency, and implicit spatial dependency, this paper proposes a Long-term Explicit–Implicit Spatio-Temporal Network (LEISN) for traffic flow prediction. A Long-term Dependency Module has been designed to store hidden states generated from multiple previous time steps, facilitating the transmission of long-term features. Based on this, two graph convolution-based spatial feature extraction branches are designed to extract explicit spatial features based on the adjacency matrix generated by spatial topology and implicit spatial features based on the implicit adjacency matrix generated by trend similarity, respectively. Then all spatio-temporal features are fused to produce the next state. Additionally, a new encoder framework, LENSI-ED, was proposed based on LEISN. Comparative experiments are conducted on four datasets, and the results showed that our model has advantages over existing methods. The proposed model addresses the issues of local and non-local spatial dependence relationships and long-term temporal dependence in traffic flow forecasting. • A long explicit–implicit spatio-temporal network for traffic flow prediction is proposed. • A new encoder framework named LENSI-ED is proposed based on LEISN. • Comparative experiments are conducted to show the advantages of the neural network. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
10. Exploiting Long-Term Dependency for Topic Sentiment Analysis
- Author
-
Faliang Huang, Changan Yuan, Yingzhou Bi, and Jianbo Lu
- Subjects
Sentiment analysis ,topic detection ,probabilistic graphical model ,long-term dependency ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Most existing unsupervised approaches to detect topic sentiment in social texts consider only the text sequences in corpus and put aside social dynamics, as leads to algorithm’s disability to discover true sentiment of social users. To address the issue, a probabilistic graphical model LDTSM (Long-term Dependence Topic-Sentiment Mixture) is proposed, which introduces dependency distance and uses the dynamics of social media to achieve the perfect combination of inheriting historical topic sentiment and fitting topic sentiment distribution underlying in current social texts. Extensive experiments on real-world SinaWeibo datasets show that LDTSM significantly outperforms JST, TUS-LDA and dNJST in terms of sentiment classification accuracy, with better inference convergence, and topic and sentiment evolution analysis results demonstrate that our approach is promising.
- Published
- 2020
- Full Text
- View/download PDF
11. AESGRU: An Attention-Based Temporal Correlation Approach for End-to-End Machine Health Perception
- Author
-
Weiting Zhang, Dong Yang, Hongchao Wang, Jun Zhang, and Mikael Gidlund
- Subjects
Health perception ,temporal correlation ,gated recurrent unit networks ,long-term dependency ,attention mechanism ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Accurate and real-time perception of the operating status of rolling bearings, which constitute a key component of rotating machinery, is of vital significance. However, most existing solutions not only require substantial expertise to conduct feature engineering, but also seldom consider the temporal correlation of sensor sequences, ultimately leading to complex modeling processes. Therefore, we present a novel model, named Attention-based Equitable Segmentation Gated Recurrent Unit Networks (AESGRU), to improve diagnostic accuracy and model-building efficiency. Specifically, our proposed AESGRU consists of two modules, an equitable segmentation approach and an improved deep model. We first transform the original dataset into time-series segments with temporal correlation, so that the model enables end-to-end learning from the strongly correlated data. Then, we deploy a single-layer bidirectional GRU network, which is enhanced by attention mechanism, to capture the long-term dependency of sensor segments and focus limited attention resources on those informative sampling points. Finally, our experimental results show that the proposed approach outperforms previous approaches in terms of the accuracy.
- Published
- 2019
- Full Text
- View/download PDF
12. Deeply Exploiting Long-Term View Dependency for 3D Shape Recognition
- Author
-
Yong Xu, Chaoda Zheng, Ruotao Xu, and Yuhui Quan
- Subjects
3D shape recognition ,long-term dependency ,multi-view deep learning ,view aggregation ,Electrical engineering. Electronics. Nuclear engineering ,TK1-9971 - Abstract
Recognition of 3D shapes is a fundamental task in computer vision. In recent years, view-based deep learning has emerged as an effective approach for 3D shape recognition. Most existing view-based methods treat the views of an object as an unordered set, which ignores the dynamic relations among the views, e.g. sequential semantic dependencies. In this paper, modeling the views of an object by a sequence, we aim at exploiting the long-term dependencies among different views for shape recognition, which is done by constructing a sequence-aware view aggregation module based on the bi-directional Long Short-Term Memory network. It is shown that our view aggregation module not only captures the bi-directional dependencies in view sequences, but also enjoys the robustness to circular shifts of input sequences. Incorporating the aggregation module into a standard convolutional network architecture, we develop an effective method for 3D shape classification and retrieval. Our method was evaluated on the ModelNet40/10 and ShapeNetCore55 datasets. The results show the encouraging performance gain from exploiting long-term dependencies in view sequences, as well as the superior performance of our method compared to the existing ones.
- Published
- 2019
- Full Text
- View/download PDF
13. Exemplar-based video colorization with long-term spatiotemporal dependency.
- Author
-
Chen, Siqi, Li, Xueming, Zhang, Xianlin, Wang, Mingdao, Zhang, Yu, Han, Jiatong, and Zhang, Yue
- Subjects
- *
PROBLEM solving , *VIDEOS , *COLOR - Abstract
Exemplar-based video colorization is an essential technique for applications like old movie restoration. Although recent methods perform well in still scenes or scenes with regular movement, they always lack robustness in moving scenes due to their weak ability to model long-term dependency both spatially and temporally, leading to color fading, color discontinuity, or other artifacts. To solve this problem, we propose an exemplar-based video colorization framework with long-term spatiotemporal dependency. To enhance the long-term spatial dependency, a parallelized CNN-Transformer block and a double-head non-local operation are designed. The proposed CNN-Transformer block can better incorporate the long-term spatial dependency with local texture and structural features, and the double-head non-local operation further exploits the performance of the augmented feature. While for the long-term temporal dependency enhancement, we further introduce the novel Linkage subnet. The Linkage subnet propagates motion information across adjacent frame blocks and helps to maintain temporal continuity. Experiments demonstrate that our model outperforms recent state-of-the-art methods both quantitatively and qualitatively. Also, our model can generate more colorful, realistic and stabilized results, especially for scenes where objects change greatly and irregularly. [ABSTRACT FROM AUTHOR]
- Published
- 2024
- Full Text
- View/download PDF
14. Long-term sequence dependency capture for spatiotemporal graph modeling.
- Author
-
Huang, Longji, Huang, Jianbin, Chen, Peiji, Li, He, and Cui, Jiangtao
- Subjects
- *
MACHINE learning , *SMART cities , *TIME series analysis , *FORECASTING , *MEMORY , *DEEP learning - Abstract
Long term dependency capture is essentially important for time series prediction and spatial–temporal forecasting. In recent years, many deep learning-based forecasting methods have been proposed, leading to rapid development in this area. We systematically reviewed long-term dependency capture methods, including temporal dependency in sequence (named intra-sequence temporal dependency), temporal dependency out of sequence (named inter-sequence temporal dependency, in this scenario the long-term dependencies are split by many subsequences). Because the batch technique is widely adopted in machine learning and deep learning, the range of temporal capturing ability for many proposed methods is intra-sequence temporal dependency, which limits the capacity of long-term dependency capture. Aiming at the above problems, we designed three type memory mechanisms (i.e., a temporal encoding memory mechanism, a cross-sequence memory mechanism and a query-key based memory mechanism) to solve those long term dependency problems. Moreover, based on the cross-sequence memory mechanism and query-key architecture, an Attention-based Long-Term Dependency Capture model (ALTDC) is proposed for long-term dependency modeling and further solves the temporal dependency coherence problem. ALTDC includes temporal Transformer and spatial Transformer. The temporal Transformer adopts multi-head attention mechanism in temporal dimension and takes into consideration the relative position encoding. The spatial Transformer leverages attention mechanism in spatial dimension, and utilizes learnable position encoding and graph convolution to capture spatial relationships. Experiments demonstrate that the proposed model outperforms the state-of-art baselines on real world time-series datasets and spatial–temporal datasets. [ABSTRACT FROM AUTHOR]
- Published
- 2023
- Full Text
- View/download PDF
15. AESGRU: An Attention-Based Temporal Correlation Approach for End-to-End Machine Health Perception
- Author
-
Mikael Gidlund, Jun Zhang, Dong Yang, Hongchao Wang, and Weiting Zhang
- Subjects
Gated Recurrent Unit Networks ,Feature engineering ,Dependency (UML) ,General Computer Science ,Long-term Dependency ,Computer science ,media_common.quotation_subject ,temporal correlation ,Machine learning ,computer.software_genre ,Datorteknik ,Health Perception ,End-to-end principle ,Component (UML) ,Perception ,Attention Mechanism ,General Materials Science ,Segmentation ,Computer Engineering ,gated recurrent unit networks ,media_common ,Health perception ,business.industry ,Communication Systems ,General Engineering ,long-term dependency ,Key (cryptography) ,Artificial intelligence ,lcsh:Electrical engineering. Electronics. Nuclear engineering ,Temporal Correlation ,business ,Focus (optics) ,attention mechanism ,computer ,lcsh:TK1-9971 ,Kommunikationssystem - Abstract
Accurate and real-time perception of the operating status of rolling bearings, which constitute a key component of rotating machinery, is of vital significance. However, most existing solutions not only require substantial expertise to conduct feature engineering, but also seldom consider the temporal correlation of sensor sequences, ultimately leading to complex modeling processes. Therefore, we present a novel model, named Attention-based Equitable Segmentation Gated Recurrent Unit Networks (AESGRU), to improve diagnostic accuracy and model-building efficiency. Specifically, our proposed AESGRU consists of two modules, an equitable segmentation approach and an improved deep model. We first transform the original dataset into time-series segments with temporal correlation, so that the model enables end-to-end learning from the strongly correlated data. Then, we deploy a single-layer bidirectional GRU network, which is enhanced by attention mechanism, to capture the long-term dependency of sensor segments and focus limited attention resources on those informative sampling points. Finally, our experimental results show that the proposed approach outperforms previous approaches in terms of the accuracy. NIIT
- Published
- 2019
16. Capital Market Hypotheses and Their Statistical Implications: A Comparative Study
- Author
-
Petras, Petr, Krištoufek, Ladislav, and Křehlík, Tomáš
- Subjects
Teorie koherentních trhů ,Dlouhodobá paměť ,Efficient Market Hypothesis ,Teorie efektivních trhů ,Long-term dependency ,Fractal Market Hypothesis ,Teorie fraktálních trhů ,Coherent Market Hypothesis ,Rescaled Range Analysis - Abstract
In this bachelor thesis we focus on different Market Hypotheses. Specifically on Efficient Market Hypothesis, Fractal Market Hypothesis and Coherent Market Hypothesis. In the first part of the work we provide description of researched hypotheses and methods used for testing. In the second part of the work we run test on time series of share markets, gold markets and currency markets and test if our hypotheses can provide explanation about price changes on those markets. For Efficient Market Hypothesis we wonder if prices are following random walk (via augmented Dickey-Fuller test), if residuals are normally distributed (via Shapiro-Wilk and Jarque-Bera tests) and if residuals are uncorrelated (via Box-Pierce test). For Fractal Market Hypothesis we are trying to find value of Hurst exponent via Rescaled Range analysis. This exponent describes if time series are persistent or not. And for Coherent Market Hypothesis we develop simple method for testing if some time periods can yield above-average revenues, thanks to increased mean and decreased standard deviation. After that we find out what are consequences of short time series and different frequencies for obtaining data points and we learn that some hypotheses describes different time periods or lengths better and are not so good for different ones. Powered...
- Published
- 2014
17. Ising model in finance: from microscopic rules to macroscopic phenomena
- Author
-
Dvořák, Pavel, Krištoufek, Ladislav, and Kukačka, Jiří
- Subjects
financial markets ,shluky volatility ,econophysics ,ferromagnetism ,dlouhá paměť ,feromagnetismus ,long-term dependency ,stylizovaná fakta ,Ising model ,stylized facts ,volatility clustering ,finanční trhy ,Isingův model - Abstract
The main objective of this thesis is to inspect the abilities of the Ising model to exhibit selected statistical properties, or stylized facts, that are common to a wide range of financial assets. The investigated properties are heteroskedasticity of returns, rapidly decaying linear autocorrelation, volatility clustering, heavy tails, negative skewness and non-Gaussianity of the return distribution. In the first part of the thesis, we test the presence of these stylized facts in S&P 500 daily returns over the last 30 years. The main part of the thesis is dedicated to the Ising model-based simulations and to discussion of the results. New features such as Poisson process governed lag or magnetisation dependent trading activity are incorporated in the model. We conclude that the Ising model is able to convincingly replicate most of the examined statistical properties while even more satisfactory results can be obtained with appropriate tuning. 1
- Published
- 2012
Catalog
Discovery Service for Jio Institute Digital Library
For full access to our library's resources, please sign in.