2022
DOI: 10.48550/arxiv.2205.13504
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Are Transformers Effective for Time Series Forecasting?

Abstract: Recently, there has been a surge of Transformer-based solutions for the time series forecasting (TSF) task, especially for the challenging long-term TSF problem. Transformer architecture relies on self-attention mechanisms to effectively extract the semantic correlations between paired elements in a long sequence, which is permutation-invariant and "anti-ordering" to some extent. However, in time series modeling, we are to extract the temporal relations among an ordering set of continuous points. Consequently,… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
36
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
6
1
1
1

Relationship

0
9

Authors

Journals

citations
Cited by 30 publications
(54 citation statements)
references
References 13 publications
1
36
0
Order By: Relevance
“…In addition to linear regression (LR), seven deep learning methods, namely hierarchical interpolation for time series forecasting (N-HiTS) [66], temporal convolutional network (TCN) [67], Transformer (TF) [68], NLinear [69], long short-term memory (LSTM) [70], gated recurrent unit (GRU) [71], and temporal fusion transformer (TFT) [53], were investigated to develop predictive models and compare their performance. N-HiTS and NLinear were improved to support producing probabilistic forecasts based on quantile regression.…”
Section: Studied Machine Learning Methodsmentioning
confidence: 99%
“…In addition to linear regression (LR), seven deep learning methods, namely hierarchical interpolation for time series forecasting (N-HiTS) [66], temporal convolutional network (TCN) [67], Transformer (TF) [68], NLinear [69], long short-term memory (LSTM) [70], gated recurrent unit (GRU) [71], and temporal fusion transformer (TFT) [53], were investigated to develop predictive models and compare their performance. N-HiTS and NLinear were improved to support producing probabilistic forecasts based on quantile regression.…”
Section: Studied Machine Learning Methodsmentioning
confidence: 99%
“…These models show that sequence dependencies can be modeled for long-term forecasting [ 27 ], but rely heavily on periodicity. In general, our findings suggest that the correlation between subsequences is crucial.…”
Section: Related Workmentioning
confidence: 99%
“…Both modules are used to replace self-attention and cross-attention modules, with a time complexity of O (L). These Transformer-based models have demonstrated outstanding performance in learning long-term sequence dependencies [38]. CNN-based methods are commonly used to extract local temporal features by leveraging convolution kernels.…”
Section: Tsf Modelsmentioning
confidence: 99%
“…The overall structure of MLGN is shown in figure 2. Inspired by traditional time series decomposition algorithms [46] and deep learning models [24,30,37,38], we design a multi-scale sequence decomposition (MSDecomp) block to separate complex patterns of input series. Then we utilize seasonal component prediction module to predict seasonal information and trend component prediction module to predict trend information respectively.…”
Section: Mlgn Frameworkmentioning
confidence: 99%