2020
DOI: 10.48550/arxiv.2009.14799
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MQTransformer: Multi-Horizon Forecasts with Context Dependent and Feedback-Aware Attention

Abstract: Recent advances in neural forecasting have produced major improvements in accuracy for probabilistic demand prediction. In this work, we propose novel improvements to the current state of the art by incorporating changes inspired by recent advances in Transformer architectures for Natural Language Processing. We develop a novel decoder-encoder attention for contextalignment, improving forecasting accuracy by allowing the network to study its own history based on the context for which it is producing a forecast… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
13
0

Year Published

2021
2021
2022
2022

Publication Types

Select...
5

Relationship

1
4

Authors

Journals

citations
Cited by 5 publications
(13 citation statements)
references
References 18 publications
0
13
0
Order By: Relevance
“…Considering the selected features are the same for all these methods, we believe the lower SD is due to the dynamic dropout probability mechanism in DemandNet's prediction model. Overall, the two proposed models, DemandNet-LSTM and DemandNet-GRU, consistently outperform the baseline NN models [20,19], especially under longer horizon windows (40 and 80 days) due to their ability to incorporate state closure/open policy in their prediction. Considering all three evaluation metrics, DemandNet-GRU is best suited for the US consumer spending dataset due to reported higher confidence and accuracy.…”
Section: Resultsmentioning
confidence: 93%
See 1 more Smart Citation
“…Considering the selected features are the same for all these methods, we believe the lower SD is due to the dynamic dropout probability mechanism in DemandNet's prediction model. Overall, the two proposed models, DemandNet-LSTM and DemandNet-GRU, consistently outperform the baseline NN models [20,19], especially under longer horizon windows (40 and 80 days) due to their ability to incorporate state closure/open policy in their prediction. Considering all three evaluation metrics, DemandNet-GRU is best suited for the US consumer spending dataset due to reported higher confidence and accuracy.…”
Section: Resultsmentioning
confidence: 93%
“…• MQ-Transformer [20]: A transformer-based deep learning model that benefits from static features and future predefined labels. It also provides a state-of-art prediction in multiple datasets.…”
Section: Methods For Comparisonmentioning
confidence: 99%
“…This is typically achieved by two avenues. The first approach, including Salinas et al (2020Salinas et al ( , 2019 Eisenach et al (2022). Other classes of works include variance reduced training (Lu et al, 2021) or domain adaptaton based techniques (Jin et al, 2022).…”
Section: Related Workmentioning
confidence: 99%
“…Fortunately, quantile regression [14,13], which has been successfully used for robustly modeling probabilistic outputs, comes to rescue. The incorporation of the quantile regression component to various sequential neural network backbones has been shown to be particularly effective with recent advances in deep learning [29,9,18,6]. Obtaining a full probabilistic prediction (i.e., the ability to query a forecast at an arbitrary quantile) usually requires generating multiple quantiles at once.…”
Section: Introductionmentioning
confidence: 99%