Currently “smooth” function contains simulation function for Exponential Smoothing (es) and SARIMA (ssarima). It is planned to produce other functions for other state-space models. This is part of smooth package.
Let’s load the necessary package:
We start with an Exponential Smoothing. For example, monthly data generated from ETS(A,N,N), 120 observations:
The resulting ourSimulation
object contains:
ourSimulation$model
– name of ETS model used in simulation;
ourSimulation$data
– vector of simulated data;
ourSimulation$states
– matrix of states, where columns
contain different states and rows corresponds to time;
ourSimulation$persistence
– vector of smoothing parameters
used in simulation (in our case generated randomly);
ourSimulation$residuals
– vector of errors generated in the
simulation; ourSimulation$occurrence
– vector of demand
occurrences (zeroes and ones, in our case only ones);
ourSimulation$logLik
– true likelihood function for the
used generating model.
We can plot produced data, states or residuals in order to see what was generated. This is done using:
If only one time series has been generated, we can use a simpler command in order to plot it:
Now let’s use more complicated model and be more specific, providing persistence vector:
ourSimulation <- sim.es("MAdM", frequency=12, obs=120, phi=0.95, persistence=c(0.1,0.05,0.01))
plot(ourSimulation)
High values of smoothing parameters are not advised for models with multiplicative components, because they may lead to explosive data. As for randomizer the default values seem to work fine in the majority of cases, but if we want, we can intervene and ask for something specific (for example, some values taken from some estimated model):
ourSimulation <- sim.es("MAdM", frequency=12, obs=120, phi=0.95, persistence=c(0.1,0.05,0.01), randomizer="rlnorm", meanlog=0, sdlog=0.015)
plot(ourSimulation)
It is advised to use lower values for sdlog
and
sd
for models with multiplicative components. Once again,
using higher values may lead to data with explosive behaviour.
If we need intermittent data, we can define probability of occurrences. And it also makes sense to use pure multiplicative models and specify initials in this case:
ourSimulation <- sim.es("MNN", frequency=12, obs=120, probability=0.2, initial=10, persistence=0.1)
plot(ourSimulation)
If we want to have several time series generated using the same
parameters then we can use nsim
parameter:
ourSimulation <- sim.es("MNN", frequency=12, obs=120, probability=0.2, initial=10, persistence=0.1, nsim=50)
We will have the same set of returned values, but with one more
dimension. So, for example, we will end up with matrix for
ourSimulation$data
and array for
ourSimulation$states
.
There is also simulate()
function that allows to
simulate data from estimated ETS model. For example:
x <- ts(rnorm(100,120,15),frequency=12)
ourModel <- es(x, h=18, silent=TRUE)
ourData <- simulate(ourModel, nsim=50, obs=100)
Now let’s compare the original data and the simulated one:
As we see the level is the same and variance is similar for both series. Achievement unlocked!
Let’s start from something simple. By default
sim.ssarima()
generates data from ARIMA(0,1,1). Let’s
produce 10 time series with 120 observations of that:
The resulting ourSimulation
object contains:
ourSimulation$model
– name of ARIMA model used in
simulation; ourSimulation$AR
– matrix with generated or
provided AR parameters, ourSimulation$MA
– matrix with MA
parameters, ourSimulation$AR
– vector with constant values
(one for each time series), ourSimulation$initial
– matrix
with initial values, ourSimulation$data
– matrix of
simulated data (if we had nsim=1
, then that would be a
vector); ourSimulation$states
– array of states, where
columns contain different states, rows corresponds to time and last
dimension is for each time series; ourSimulation$residuals
– vector of errors generated in the simulation;
ourSimulation$occurrence
– vector of demand occurrences
(zeroes and ones, in our case only ones);
ourSimulation$logLik
– true likelihood function for the
used generating model.
Similarly to sim.es()
, we can plot produced data, states
or residuals for each time series in order to see what was generated.
Here’s an example:
Now let’s use more complicated model. For example, data from SARIMA(0,1,1)(1,0,2)_12 with drift can be generated using:
ourSimulation <- sim.ssarima(orders=list(ar=c(0,1),i=c(1,0),ma=c(1,2)), lags=c(1,12), constant=TRUE, obs=120)
plot(ourSimulation)
If we want to provide some specific parameters, then we should follow the structure: from lower lag to lag from lower order to higher order. For example, the same model with predefined MA terms will be:
ourSimulation <- sim.ssarima(orders=list(ar=c(0,1),i=c(1,0),ma=c(1,2)), lags=c(1,12), constant=TRUE, MA=c(0.5,0.2,0.3), obs=120)
ourSimulation
## Data generated from: SARIMA(0,1,1)[1](1,0,2)[12] with drift
## Number of generated series: 1
## AR parameters:
## Lag 12
## AR(1) -0.0091
## MA parameters:
## Lag 1 Lag 12
## MA(1) 0.5 0.2
## MA(2) 0.0 0.3
## Constant value: 26.6793
## True likelihood: -635.2447
We can create time series with several frequencies For example, some sort of daily series from SARIMA(1,0,2)(0,1,1)_7(1,0,1)_30 can be generated with a command:
ourSimulation <- sim.ssarima(orders=list(ar=c(1,0,1),i=c(0,1,0),ma=c(2,1,1)), lags=c(1,7,30), obs=360)
ourSimulation
## Data generated from: SARIMA(1,0,2)[1](0,1,1)[7](1,0,1)[30]
## Number of generated series: 1
## AR parameters:
## Lag 1 Lag 30
## AR(1) 0.342 -0.5547
## MA parameters:
## Lag 1 Lag 7 Lag 30
## MA(1) 0.1576 -0.8113 -0.6115
## MA(2) 0.7728 0.0000 0.0000
## Constant value: 0
## True likelihood: -1600.7486
sim.ssarima
also supports intermittent data, which is
defined via probability
parameter, similar to
sim.es()
:
ourSimulation <- sim.ssarima(orders=list(ar=c(1,0),i=c(0,1),ma=c(2,1)), lags=c(1,7), obs=120, probability=0.2)
ourSimulation
## Data generated from: iSARIMA(1,0,2)[1](0,1,1)[7]
## Number of generated series: 1
## AR parameters:
## Lag 1
## AR(1) 0.1711
## MA parameters:
## Lag 1 Lag 7
## MA(1) 0.4893 0.9475
## MA(2) 0.3642 0.0000
## Constant value: 0
## True likelihood: -594.2268
Finally we can use simulate()
function in a similar
manner as with sim.es()
. For example:
x <- ts(100 + c(1:100) + rnorm(100,0,15),frequency=12)
ourModel <- auto.ssarima(x, h=18, silent=TRUE)
ourData <- simulate(ourModel, nsim=50, obs=100)
If we don’t want to use everything from the estimated function and
need, for example, to use orders only, then we can extract them using
orders()
and lags()
functions:
Now let’s compare the original data and one of series from the simulated array:
As we see series demonstrate similarities in dynamics and have similar variances.
As usual we start from a simple model. By default
sim.ces()
generates data from CES(n). If we do not provide
specific parameters, then they will be automatically generated:
The resulting ourSimulation
object contains:
ourSimulation$model
– name of CES model used in simulation;
ourSimulation$A
– vector of complex smoothing parameters A,
ourSimulation$B
– vector of complex smoothing parameters A
(if “partial” or “full” seasonal model was used in the simulation),
ourSimulation$initial
– array with initial values,
ourSimulation$data
– matrix of simulated data (if we had
nsim=1
, then that would be a vector);
ourSimulation$states
– array of states, where columns
contain different states, rows corresponds to time and last dimension is
for each time series; ourSimulation$residuals
– vector of
errors generated in the simulation;
ourSimulation$occurrence
– vector of demand occurrences
(zeroes and ones, in our case only ones);
ourSimulation$logLik
– true likelihood function for the
used generating model.
Similarly to other simulate functions in “smooth”, we can plot produced data, states or residuals for each time series in order to see what was generated. Here’s an example:
We can also see a brief summary of our simulated data:
## Data generated from: CES(n)
## Number of generated series: 1
## Smoothing parameter a: 1.8221+0.9579i
## True likelihood: -415.1624
We can produce one out of three possible seasonal CES models. For example, Let’s generate data from “Simple CES”, which does not have a level:
Now let’s be more creative and mess around with the generated initial values of the previous model. We will make some of them equal to zero and regenerate the data:
ourSimulation$initial[c(1:5,20:24),] <- 0
ourSimulation <- sim.ces("s",frequency=24, obs=120, nsim=1, initial=ourSimulation$initial, randomizer="rt", df=4)
plot(ourSimulation)
The resulting generated series has properties close to the ones that solar irradiation data has: changing amplitude of seasonality without changes in level. We have also chosen a random number generator, based on Student distribution rather than normal. This is done just in order to show what can be done using simulate functions in “smooth”.
We can also produce CES with so called “partial” seasonality, which corresponds to CES(n) with additive seasonal components. Let’s produce 10 of such time series:
Finally, the most complicated CES model is the one with “full” seasonality, implying that there are two complex exponential smoothing models inside: one for seasonal and the other for non-seasonal part:
The generated smoothing parameters may sometimes lead to explosive behaviour and produce meaningless time series. That is why it is advised to use parameters of a model fitted to time series of interest. For example, here we generate something crazy and then simulate the data:
x <- ts(rnorm(120,0,5) + rep(runif(12,-50,50),10)*rep(c(1:10),each=12) ,frequency=12)
ourModel <- ces(x, seasonality="s", h=18, silent=TRUE)
ourData <- simulate(ourModel, nsim=50, obs=100)
Comparing the original data with the simulated one, we will see some similarities:
There is also sim.gum()
function in the package that
generates data from GUM model. If we do not provide specific parameters,
then they will be automatically generated. This model is very flexible
and taking that there is a lot of parameters to set, in general it is
not advised to leave such parameters as measurement
,
transition
and persistence
blank. The good
practice is to simulate data from a fitted model:
x <- ts(100 + rnorm(120,0,5) + rep(runif(12,-50,50),10)*rep(c(1:10),each=12) ,frequency=12)
ourModel <- auto.gum(x, h=18, silent=TRUE)
ourData <- simulate(ourModel, nsim=50)
Comparing the original data with the simulated one, we will see some similarities:
Note that GUM is still an ongoing research and its properties are currently understudied.
Now that there is a model underlying simple moving averages, we can simulate the data for it. Here how it can be done:
As usual, you can use simulate function as well: