Professional Documents
Culture Documents
6 Seasonality & Forecasting
6 Seasonality & Forecasting
6 Seasonality & Forecasting
Examples:
Example:
ARIMA models provide an improvement over past efforts to “deseasonalize” the time-series
through the use of dummy variables. The latter approach simply employs multiple regres-
sion with dummy variables to estimate the seasonal variance of each phase of the cycle.
This estimated variance is then subtracted from the time-series. This approach, however, is
inadequate (see McCleary & Hay, p.81)
Seasonal Non-Stationarity (dfn): Trend or drift in steps rather than as “observation to ob-
servation steps.
Example:
(1 − B 12 )yt = θ0 or,
yt = yt−12 + θ0
yt = φ12 yt−12 + at
yt = (1 − θ12 B 12 )at
= at − θ1 2at−12
Example:
(1 − B)(1 − B 12 ) = (1 − B 12 )(1 − B) = (1 − B − B 12 + B 13 )
zt∗ = yt − yt−1
zt ∗ = zt − zt−12
Finally,
• Invertibility conditions
−1 < φ, φs < 1
−1 < θ, θs < 1
6.2.2 Identification
The same logic applies here, as before — know how the different models behave, in theory,
then just “match up . . . ”
Large and nearly equal values of the ACF at all seasonal lags.
The PACF will decay from seasonal lag to seasonal lag, with the rate of decay de-
termined by;
θs , θ2s , θ3s , · · · , θQs
• ARIMA(P, 0, 0)s has a decaying ACF, with the rate of decay determined by;
φs , φ2s , φ3s , · · · , φP s
The PACF has P spikes at the first P seasonal lags (with all other successive lags
expected to be zero).
McCleary & Hay contend that small order will work, if you can find it at all [Clear up
from M&H, pp. 87-9]
Example: Hog Production (see Pindyck & Rubinfeld, pp. 457-9, Figures 15.14 & 15.15)
6.3 Forecasting
6.3.1 Background
However, an entire cottage industry has sprung up around the issue of forecasting. A more
interesting question is, why has there been so little use of ARIMA methods in the area of
forecasting?
6.3.2 Definitions
Forecasting: The likelihood of future events given current & past information.
Ex Post Forecast: Knowing all values of the exogenous and endogenous variables with cer-
tainty, so you can evaluate accuracy exactly.
Ex Ante Forecast: A prediction beyond the estimated period using explanatory variables
with and without certainty.
T1 T2 T 3 (present)
time (t)
- - -
Estimation Ex Post Ex Ante
Period Forecast Period Forecast Period
Conditional Forecast: Forecasts themselves are used to predict values. Values for one or
more explanatory variables are unknown, so that forecasts must be used to produces the
forecast on the dependent variable.
Example:
If monthly sales, s(t), are linearly related to 2 variables, X1 & X2 , but with lags of 3 &
4 months respectively then,
s(t) = a0 + a1 X1 (t − 3) + a2 X2 (t − 4) + ²t
If this equation is estimated, the results can be used to produce an unconditional forecast of
s(t) 1,2, and 3 months into the future.
To get the 3 month forecast of s(t), we would use the current value of X1 & last month’s
value of X2 , both of which are known.
Pindyck & Rubin present a general development for the general model:
(1 − φ1 B)yt = at (3)
−1
yt = (1 − φ1 B) at
Recall,
(1 − B)−1 = 1 + B + B 2 + · · · + B k + · · ·
So that,
Since yt = yt − θ0 ,
⇒ yt = yt + θ0
Then,
E[yt ] = E[yt ] + θ0
= 0 + θ0
= θ0 (Mean)
“Extrapolating” yields:
yt (1) = E[yt+1 ] = 0
..
.
yt (n) = E[yt+n ] = 0
This implies that an unconditional forecast will always have the same value . . . the series
mean.
As l becomes large, the mean is the best prediction. This is not surprising since the se-
ries is stationary.
Conditional Forecast for AR(1) — will depend on previous forecasts
Assuming that all the ψ weights & past shocks are known, we can use the conditional expectation
as a forecast of yt+1 .
So, q q
−1.96 V ar(2) < yt (2) < 1.96 V ar(2)
Moreover,
V ar(n) = (1 + ψ12 + ψ22 + · · · + ψn−1
2
)σa2
V ar(n) will be even larger if uncertainty exists — resulting from a poor fit of coefficient
estimates.
McCleary & Hay and Pindyck & Rubinfeld both develop “profiles” for difficult ARIMA
models. They show that the nature of the point forecasts and confidence intervals are not
all the same. They vary with respect to:
• Conditional forecasts
yt (1) = E[at+1 ] = 0
yt (2) = E[at+2 ] = 0
..
.
yt (n) = E[at+n ] = 0
Because the ψ weights are zero, the variance around these point estimate forecasts
is constant for all lead times. Therefore,
V ar(1) = σa2
V ar(2) = σa2
..
.
V ar(n) = σa2
For this profile, the conditional and unconditional expectations are identical. The
best forecast is the process mean. Moreover, the history of the process results in no
improvement to the prediction.
After only two or three steps into the future the confidence intervals become so large
as to make the interval forecasts meaningless.
3. ARIMA(2,0,0) — Autoregressive
As the lead time increases, the forecasts regress to the process mean and confidence
intervals at each point. Forecasts tend to increase with increases in lead time (with
the confidence intervals increasing at a rate determined by φ1 ).
For lead times > 1, yt (n) equals the process mean; V ar(n) remains constant.
5. ARIMA(1,1,0) — Mixed (Autoregressive-Moving Average)
• Interest Rates
• Hog Production
• Both McCleary & Hay and Pindyck & Rubinfeld stress the short-run value of ARIMA
models for forecasting purposes.
This is because M A(q) processes have a memory of only q periods. Actual data can
be used only if l < q. Only the most recent observations have much impact on the
AR forecast since exponential weights and invertibility will wipe out the effect of long
memory (seasonality is something of an exception here).
Also, for most ARIMA models the last observation often has the most impact. If
the last observation is an extreme value its presence can be problematic.
Applications: