2023
DOI: 10.1016/j.neunet.2023.06.044
|View full text |Cite
|
Sign up to set email alerts
|

GBT: Two-stage transformer framework for non-stationary time series forecasting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
5
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
7
1

Relationship

0
8

Authors

Journals

citations
Cited by 11 publications
(5 citation statements)
references
References 11 publications
0
5
0
Order By: Relevance
“…This study introduces an innovative framework for time-series forecasting algo rithms, the Two-stage Transformer [10]. It optimizes the traditional Transformer forecast ing method's structure by segmenting it into an Auto-Regression Stage and a Self-Regres sion Stage.…”
Section: Two-stage Transformermentioning
confidence: 99%
See 1 more Smart Citation
“…This study introduces an innovative framework for time-series forecasting algo rithms, the Two-stage Transformer [10]. It optimizes the traditional Transformer forecast ing method's structure by segmenting it into an Auto-Regression Stage and a Self-Regres sion Stage.…”
Section: Two-stage Transformermentioning
confidence: 99%
“…However, these Transformer model variants are computationally intensive and characterized by complex parameters, leading to significant memory usage and computational resource consumption. The GBT (Two-stage Transformer) [10] addresses the critical issue of overfitting associated with improper decoder input initialization by innovatively dividing the traditional Transformer architecture into two segments: the Auto-Regression and Self-Regression stages. This approach enables the model to enhance its predictive capacity while maintaining low temporal and spatial complexity.…”
Section: Introductionmentioning
confidence: 99%
“…To mitigate the decline in predictive performance caused by non-stationary data, classical statistical methods such as ARMA and ARIMA [20] can achieve stationarity of time series through difference operations. However, these traditional models cannot compete with deep-learning-based models in capturing dynamic changes in data and solving long-sequence time series forecasting problems [21]. With deep learning models, some researchers have attempted to alleviate the non-stationarity in the original time series by normalizing the global data distribution of related time series into a specific distribution with mean 0 and variance 1, commonly referred to as Z-score standardization [15,22].…”
Section: Introductionmentioning
confidence: 99%
“…To address the issue of non-stationarity in deep learning models due to differences in the statistical properties of input windows, Shen proposed a two-level transformer framework, which exhibited excellent stability [21]. On the other hand, Liu effectively mitigated the overfitting phenomenon caused by non-stationary data such as weather and electricity consumption predictions by optimizing the attention mechanism in the transformer model [24].…”
Section: Introductionmentioning
confidence: 99%
“…There is an effort to exploit the emergent advances in deep-learning machine models (such as Transformers and GANs) for time series analysis [11,12]. Despite the great success achieved by these models in processing natural languages and generating images and videos, Long short-term memory (LSTM) and gated recurrent unit (GRU) networks still maintain their popularity for time series forecasting [13,14,15]. Their popularity could be attributed to their ability to effectively capture temporal dependencies, which is crucial for accurately predicting future values in a time series.…”
Section: Introductionmentioning
confidence: 99%