2020
DOI: 10.48550/arxiv.2012.07436
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting

Abstract: Many real-world applications require the prediction of long sequence time-series, such as electricity consumption planning. Long sequence time-series forecasting (LSTF) demands a high prediction capacity of the model, which is the ability to capture precise long-range dependency coupling between output and input efficiently. Recent studies have shown the potential of Transformer to increase the prediction capacity. However, there are several severe issues with Transformer that prevent it from being directly ap… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
74
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 64 publications
(92 citation statements)
references
References 48 publications
0
74
0
Order By: Relevance
“…On using more sophisticated Neural Network architectures: In this paper, to simulate DAEs over a long-time horizon, we employed a modified version of the conventional fully −0.5 0.0 0.5 connected neural network architecture. However, we realize that other architectures may increase our ability to simulate long-time dependencies [23]. Thus, in our future work, we plan to implement DAE-PINN using neural networks that can generalize well to unseen events.…”
Section: Discussionmentioning
confidence: 99%
“…On using more sophisticated Neural Network architectures: In this paper, to simulate DAEs over a long-time horizon, we employed a modified version of the conventional fully −0.5 0.0 0.5 connected neural network architecture. However, we realize that other architectures may increase our ability to simulate long-time dependencies [23]. Thus, in our future work, we plan to implement DAE-PINN using neural networks that can generalize well to unseen events.…”
Section: Discussionmentioning
confidence: 99%
“…Besides the above related works on CTR prediction task, plenty of works aim to improve the efficiency and effectiveness of transformer. Reformer [13] and Informer [37] are the most relevant Each 𝑑-dimensional vector is converted to a 𝑚-length signature vector. Each random rotation here can be regarded as one "hash function".…”
Section: Related Workmentioning
confidence: 99%
“…Vaswani et al completely abandoned RNN and CNN to build transformer which is entirely based on the fully-connected layer and attention mechanism [35]. In 2020, Zhou H et al proposed probsparse self-attention mechanism and informer, and this model achieved high prediction capacity in the long sequence time-series forecasting [41]. Transformer is divided into two parts: encoding and decoding.…”
Section: Attention and Transformermentioning
confidence: 99%