2017
DOI: 10.3390/a10040114
|View full text |Cite
|
Sign up to set email alerts
|

Variable Selection in Time Series Forecasting Using Random Forests

Abstract: Time series forecasting using machine learning algorithms has gained popularity recently. Random forest is a machine learning algorithm implemented in time series forecasting; however, most of its forecasting properties have remained unexplored. Here we focus on assessing the performance of random forests in one-step forecasting using two large datasets of short time series with the aim to suggest an optimal set of predictor variables. Furthermore, we compare its performance to benchmarking methods. The first … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
61
0

Year Published

2018
2018
2023
2023

Publication Types

Select...
7
3

Relationship

2
8

Authors

Journals

citations
Cited by 121 publications
(62 citation statements)
references
References 54 publications
1
61
0
Order By: Relevance
“…is the moving average operator and [17,18]. Box and Jenkins (1970) created the building blocks of ARIMA, breaking down the prediction process into three iterative steps: identification, estimation, and validation-as seen in Figure 1 [3,19,20].…”
Section: Autoregressive Integrated Moving Average (Arima)mentioning
confidence: 99%
“…is the moving average operator and [17,18]. Box and Jenkins (1970) created the building blocks of ARIMA, breaking down the prediction process into three iterative steps: identification, estimation, and validation-as seen in Figure 1 [3,19,20].…”
Section: Autoregressive Integrated Moving Average (Arima)mentioning
confidence: 99%
“…The RF model provides an OOB data-based unbiased estimation error for the test dataset [61]. Performance of the algorithm depends on the selected parameters, such as the number of trees [52,62], splitting at each node of each tree [60,63], and the number of examples in each cell, below which the cell is not split [38], but equals the default value of the nodesize [64]. In this study, the default value was used as recommended in the literature.…”
Section: Discussionmentioning
confidence: 99%
“…S1 (see Supplement). The seasonality pattern is obvious in the sample autocorrelation function (ACF) of the original time series and reduced in the sample ACF of the deseasonalized time series, while the estimates of the Hurst parameter (H ) of the Hurst-Kolmogorov process (for its definition see Supplement; see also Tyralis et al, 2018), when the latter is fitted to the deseasonalized time series as described in Tyralis and Koutsoyiannis (2011), have a median value of 0.75 and, therefore, indicate significant long-range dependence. We note that the parameter H is commonly used in the literature for measuring this dependence under the established assumption that the latter is present in the various geophysical processes.…”
Section: Methodsmentioning
confidence: 99%