2020
DOI: 10.48550/arxiv.2005.10111
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

The Effectiveness of Discretization in Forecasting: An Empirical Study on Neural Time Series Models

Abstract: Time series modeling techniques based on deep learning have seen many advancements in recent years, especially in data-abundant settings and with the central aim of learning global models that can extract patterns across multiple time series. While the crucial importance of appropriate data pre-processing and scaling has often been noted in prior work, most studies focus on improving model architectures. In this paper we empirically investigate the effect of data input and output transformations on the predict… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
2
2

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 23 publications
(48 reference statements)
0
5
0
Order By: Relevance
“…In forecasting, a prominent example is the DeepAR model (Salinas et al, 2019b), which uses a recurrent neural network architecture and assumes the probability distribution to be from a standard probability density function (e.g., negative binomial or Student's t). Variations are possible, with either non-standard output distributions in forecasting such as the multinomial distribution (Rabanser et al, 2020) or via representing the probability density as cumulative distribution function (Salinas et al, 2019a) or the quantile function (Gasthaus et al, 2019).…”
Section: Deep Probabilistic Forecasting Models 52mentioning
confidence: 99%

Forecasting: theory and practice

Petropoulos,
Apiletti,
Assimakopoulos
et al. 2020
Preprint
Self Cite
“…In forecasting, a prominent example is the DeepAR model (Salinas et al, 2019b), which uses a recurrent neural network architecture and assumes the probability distribution to be from a standard probability density function (e.g., negative binomial or Student's t). Variations are possible, with either non-standard output distributions in forecasting such as the multinomial distribution (Rabanser et al, 2020) or via representing the probability density as cumulative distribution function (Salinas et al, 2019a) or the quantile function (Gasthaus et al, 2019).…”
Section: Deep Probabilistic Forecasting Models 52mentioning
confidence: 99%

Forecasting: theory and practice

Petropoulos,
Apiletti,
Assimakopoulos
et al. 2020
Preprint
Self Cite
“…To reduce the impact of noise and uncertainty on the results, and facilitate the identification of patterns and trends in the data, measured water quality was discretized (as shown in Table S2) before conducting Spearman correlation analysis, as discretization has been proven to be an effective data preprocessing technique. 30,31 The latent variables included in the SEM analysis will be determined based on the results of the reliability and validity tests. Reliability measurements were conducted using Cronbach's alpha coefficient, while the validity of the data was assessed through exploratory factor analysis (EFA) and confirmatory factor analysis (CFA).…”
Section: Discussionmentioning
confidence: 99%
“…Spearman correlations were employed to explore the associations between two variables. To reduce the impact of noise and uncertainty on the results, and facilitate the identification of patterns and trends in the data, measured water quality was discretized (as shown in Table S2) before conducting Spearman correlation analysis, as discretization has been proven to be an effective data preprocessing technique. , …”
Section: Methodsmentioning
confidence: 99%
“…We model the base of the predictive distribution with a discrete binned distribution to make it robust to extreme values and adaptable to the variety of real work distributions. As described by Rabanser et al (2020), we discretise the real axis between two points into n bins. A NN is trained to predict the probability of the next point falling in each of these bins, as shown in Figure 1a.…”
Section: Spliced Binned-pareto Distributionmentioning
confidence: 99%
“…While some methods have been proposed for tail estimation, they assume little to no time-varying components Siffer et al (2017). To obtain a forecast robust to extreme values and adjustable to many shapes of distributions, discrete binned distributions can be used Rabanser et al (2020), these are parametrised by a Neural Network (NN), which have been shown capable of capturing complex long-term time dependencies in forecasting (Benidis et al, 2020). We address these two challenges by combining a binned distribution with Generalised Pareto distributions for the tails, all three distributions parameterised by a single NN, allowing us to jointly model time dependencies in the base distribution and in the tails.…”
Section: Introductionmentioning
confidence: 99%