Augmented Out-of-sample Comparison Method for Time Series - - PowerPoint PPT Presentation

augmented out of sample comparison method for time series
SMART_READER_LITE
LIVE PREVIEW

Augmented Out-of-sample Comparison Method for Time Series - - PowerPoint PPT Presentation

Augmented Out-of-sample Comparison Method for Time Series Forecasting Techniques Igor Ilic, Berk Gorgulu, Mucahit Cevik Data Science Lab Ryerson University http://www.datasciencelab.ca Background Methods Results Conclusions I NTRODUCTION


slide-1
SLIDE 1

Augmented Out-of-sample Comparison Method for Time Series Forecasting Techniques

Igor Ilic, Berk Gorgulu, Mucahit Cevik

Data Science Lab Ryerson University http://www.datasciencelab.ca

slide-2
SLIDE 2

Background Methods Results Conclusions

INTRODUCTION

  • Problems with current comparison techniques for time series models:

⇤ one shot comparison ⇤ lack of robustness of conventional methods for comparing different

regression methods (e.g random train-test splits)

⇤ the testing usually does not correctly reflect real-world situations

  • An augmented out-of-sample model comparison method:

⇤ more flexible and robust technique ⇤ takes the spatio-temporal nature of time series into account

  • Effectiveness of the method are tested using ARIMA, LSTM, and GRU

models on Turkish electricity consumption data

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 2 / 10

slide-3
SLIDE 3

Background Methods Results Conclusions

LITERATURE REVIEW

  • Single interval tests (Valipour et al. (2013), Kane et al. (2014))

⇤ Single time step is used as the test data ⇤ Simple to implement and translates to real-world tests ⇤ Susceptible to lucky one-shot tests

  • Multiple datasets (Zhang (2003), Merh et al. (2010))

⇤ Predict the future data points for many datasets ⇤ Computationally intensive, requires a lot of data

  • Random Test Interval Sampling (Huang et al. (2015))

⇤ Single dataset is sufficient for testing ⇤ The predictions are not made for the immediate future, not suitable for

comparing methods that are data dependent (e.g. ARIMA)

  • Augmented Training (Tashman (2000))

⇤ Rolling method for training a model ⇤ Fast ad hoc way to train a model with the best hyper parameters

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 3 / 10

slide-4
SLIDE 4

Background Methods Results Conclusions

AUGMENTED OUT-OF-SAMPLE COMPARISON METHOD

  • The dataset is used to obtain the train-test sets, a forecast interval, and

the number of tests.

  • After each test, the model is updated to include the test data.
  • Since the model is merely updated on a small portion of data, we do not

have to retrain the entire model as found in rolling horizons ⇒ Leads to a significant speedup in testing

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 4 / 10

slide-5
SLIDE 5

Background Methods Results Conclusions

ALGORITHM

Algorithm 1: Augmented Out-Of-Sample Testing

Input : Dataset sorted by ascending date as D, algorithm as f, test interval length as `, number of tests as n Output: Array of predicted values and real values

1 T , U ← TrainTestSplit(D); 2 model ← TrainUsing(f, T ); 3 {Cj}n j=1 ← Split(U, n); 4 results ← ∅; 5 for i = 1 . . . n do 6

testingData ← RetrieveFirst(`, Ci);

7

testResults ← TestUsing(model, testingData);

8

results ← results ∪ testResults;

9

model ← UpdateUsing(model, Ci);

10 end 11 return results;

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 5 / 10

slide-6
SLIDE 6

Background Methods Results Conclusions

EXPERIMENTAL SETUP

  • Dataset:

⇤ The Turkish electricity dataset: Five years worth of hourly data ⇤ Daily and weekly seasonality ⇤ Covariates included: day of week, hour of day ⇤ Covariates excluded: holidays, weather information, electricity pricing

  • Models:

⇤ Two RNN algorithms: LSTM and GRU ⇤ SARIMAX: Seasonal ARIMA with Regressors ⇤ A naive baseline model: use last week’s data to predict current week

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 6 / 10

slide-7
SLIDE 7

Background Methods Results Conclusions

COMPARING MODEL PERFORMANCES - 1

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 7 / 10

slide-8
SLIDE 8

Background Methods Results Conclusions

COMPARING MODEL PERFORMANCES - 2

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 8 / 10

slide-9
SLIDE 9

Background Methods Results Conclusions

SUMMARY

  • The augmented out-of-sample method alleviates many shortcomings of

the standard approaches.

  • By allowing for more testing on the same dataset, in a realistic manner to

real-world training, augmented out-of-sample comparison is able to determine the best algorithm.

  • Our augmented out-of-sample model comparison method show that

neural networks outperform classical models such as ARIMA to predict electricity consumption rates in Turkey.

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 10 / 10

slide-10
SLIDE 10

Background Methods Results Conclusions

SUMMARY RESULTS

Table: MAPE by Prediction Forecast Interval with 95% Error Bounds Hours Baseline (%) SARIMAX (%) LSTM (%) GRU (%) 6 6.8 ± 1.8 2.6 ± 0.5 1.8 ± 0.4 1.6 ± 0.3 24 6.8 ± 0.9 5.4 ± 0.5 2.6 ± 0.2 1.9 ± 0.1 48 6.2 ± 0.6 7.9 ± 0.5 3.2 ± 0.2 3.3 ± 0.2

  • RNN architectures outperform classical algorithms
  • The gap widens as the forecast interval grows

Ilic et al. (Ryerson DSL) Augmented Out-of-sample Comparison 9 / 10