This vignette covers ces()
and
auto.ces()
functions, which are part of smooth package.
Let’s load the necessary packages:
ces()
function allows constructing Complex Exponential
Smoothing either with no seasonality, or with simple / partial / full
seasonality. A simple call for ces()
results in estimation
of non-seasonal model:
For the same series from M3 dataset ces()
can be
constructed using:
## Time elapsed: 0.04 seconds
## Model estimated: CES(n)
## a0 + ia1: 1.9981+1.0034i
## Initial values were produced using backcasting.
##
## Loss function type: likelihood; Loss function value: 249.4613
## Error standard deviation: 1.4914
## Sample size: 138
## Number of estimated parameters: 3
## Number of degrees of freedom: 135
## Information criteria:
## AIC AICc BIC BICc
## 504.9227 505.1018 513.7045 514.1457
##
## Forecast errors:
## MPE: 0%; sCE: 0.7%; Asymmetry: -5%; MAPE: 0.4%
## MASE: 0.857; sMAE: 0.4%; sMSE: 0%; rMAE: 0.329; rRMSE: 0.338
This output is very similar to ones printed out by es()
function. The only difference is complex smoothing parameter values
which are printed out instead of persistence vector in
es()
.
If we want automatic model selection, then we use
auto.ces()
function:
## Time elapsed: 0.05 seconds
## Model estimated: CES(n)
## a0 + ia1: 1.9981+1.0034i
## Initial values were produced using backcasting.
##
## Loss function type: likelihood; Loss function value: 249.4613
## Error standard deviation: 1.4914
## Sample size: 138
## Number of estimated parameters: 3
## Number of degrees of freedom: 135
## Information criteria:
## AIC AICc BIC BICc
## 504.9227 505.1018 513.7045 514.1457
##
## 95% parametric prediction interval was constructed
## 100% of values are in the prediction interval
## Forecast errors:
## MPE: 0%; sCE: 0.7%; Asymmetry: -5%; MAPE: 0.4%
## MASE: 0.857; sMAE: 0.4%; sMSE: 0%; rMAE: 0.329; rRMSE: 0.338
Note that prediction interval are too narrow and do not include 95% of values. This is because CES is pure additive model and it cannot take into account possible heteroscedasticity.
If for some reason we want to optimise initial values then we call:
## Time elapsed: 0.1 seconds
## Model estimated: CES(n)
## a0 + ia1: 2.007+1.0035i
## Initial values were optimised.
##
## Loss function type: likelihood; Loss function value: 249.3039
## Error standard deviation: 1.5009
## Sample size: 138
## Number of estimated parameters: 5
## Number of degrees of freedom: 133
## Information criteria:
## AIC AICc BIC BICc
## 508.6077 509.0622 523.2440 524.3638
##
## 95% semiparametric prediction interval was constructed
## 100% of values are in the prediction interval
## Forecast errors:
## MPE: 0%; sCE: 0.7%; Asymmetry: -5.7%; MAPE: 0.4%
## MASE: 0.857; sMAE: 0.4%; sMSE: 0%; rMAE: 0.329; rRMSE: 0.337
Now let’s introduce some artificial exogenous variables:
ces()
allows using exogenous variables and different
types of prediction interval in exactly the same manner as
es()
:
## Time elapsed: 0.16 seconds
## Model estimated: CES(n)
## a0 + ia1: 1.9981+1.0034i
## Initial values were produced using backcasting.
##
## Loss function type: likelihood; Loss function value: 249.4613
## Error standard deviation: 1.4914
## Sample size: 138
## Number of estimated parameters: 3
## Number of degrees of freedom: 135
## Information criteria:
## AIC AICc BIC BICc
## 504.9227 505.1018 513.7045 514.1457
##
## 95% parametric prediction interval was constructed
## 100% of values are in the prediction interval
## Forecast errors:
## MPE: 0%; sCE: 0.7%; Asymmetry: -5.3%; MAPE: 0.4%
## MASE: 0.857; sMAE: 0.4%; sMSE: 0%; rMAE: 0.329; rRMSE: 0.338