Professional Documents
Culture Documents
Econometrics III, Summary
Econometrics III, Summary
The illustration shows why the process is called a Methods for estimating a trend
random walk with drift. The process consists of two Regression model or filtering (method of moving
components: average)
Time series plot provides an indication of a possible
◦ Deterministic growth path (drift): Y0 + δ⋅t regression model
◦ Accumulated disturbance variables (random walk): Seasonality Estimation Methods (In Course)
Σui Modeling with dummy variables
Seasonality mostly known from problem / data (e.g. Q
Goals of the time series analysis: Description of data)
time series Determination of linear trend and exponent. Ft = 317.9641 + 0.081953 * t + 0.000106 * t2 - 5.77E-
Goal: To make internal relationships visible Growth 08 * t3 + Σαit
Descriptive methods: Graphics of the time series Linear growth (approximately constant absolute
("time series plot") / decomposition of the time series Warning: auto-correlated residuals (Durbin-
increases)
into components / lagged scatter plot (LS03) / Ft = β0 + β1 * t + Et Watson)
correlogram (LS03) Durbin-Watson stat. = 0.231636 (residuals highly
Exponential growth (roughly constant relative growth)
Time series plot: Provides information about trends, Ft = β0 * β1t + Et correlated)
seasonal effects, variability and "outliers". Usual -> As close as possible to 2
Linearization of exponential growth -> linear trend
representation: X: time and Y: measured variable log Ft = log β0 + t * log β1 + log Et
Trend determination through filtering / smoothing
Properties of time series plot Filtering a time series (moving average)
Example: Ln (4.6) = 1.526 => e1.526 = 4.6
Misinterpretations depending on: choice of scales Regression models: A trend can be estimated by filtering the time series.
(transformed? Units?) / Choice of section (zero point? Prerequisite: Seasonal fluctuation has been adjusted
Start? End?) The regressor variable is the variable t of the observed or does not exist. For filtering, a new time series is
points in time. generated with an additive, linear filter:
Time series plot is a unique realization of a random Parameters can be determined using conventional
process: The next realization can reject hypotheses / OLS estimators.
Only under certain conditions (LS03) can one-off
realizations be considered representative. Conclusion: Determination of linear growth with regression
Time series plots are important analytical tools, but model in EViews:
should be used with caution.
The "ai" are called weights. The smoothed series is
identified as a trend. The deviations from the
Modeling of time series - goal: understanding &
stochastic residual term. Disadvantage: Trend is not
defining model relationships
described by the functional equation.
(Given a time series y1, y2, y3, ...)
Linear filtering: simple moving average
Methods: Descriptive modeling (classic, "older" time
series analysis) / Stochastic modeling: specification of
If the additive, linear filter applies ...
a stochastic process (Yt) t = 1,2,3, ...
p = q (symmetry) and sum of the "ai" = 1
Process: Theoretical considerations, analysis of
(standardization)
empirical data to uncover the laws, definition of the
Sum of "ai" = sum of all weights
model type, estimation of the model parameters,
Same weights: ai = 1 / (2p + 1) = 1 / (2q + 1)
testing of the model quality
... it is simply the same
average
Forecast with time series - goal: forecast of
unknown values with associated forecast intervals
The basis is an estimated and validated model
Procedure of an analysis The remaining time series "Residual" only contains
seasonality and remaining term. The number of weights is odd n = 2p + 1 = 2q + 1
The number of weights "n" means length or Order of
the filter is lower.
At the edges p = q moving average values cannot be
calculated
Example: Real rate return with moving average: If δ ≠ 0 -> increases expected value> 0: E (Yt) = δ * t
(= drift)
3 different filters: p = 1 => n = 3 / p = 5 => n = 11 / p =
11 => n = 23 Drift -> mean increases with time
n = "length", depending on p / the larger, the smoother
Simulation example in Excel:
Depending on the choice of α, past observations are Weak, white noise (relevant for us)
weighted differently. The smaller α, the less strongly
Available if εt = serial uncorrelated. May have
"more recent" points in time are weighted (but still the
dependencies in third or higher moments and may not
strongest). The larger α, the less strongly "less
be completely independent.
current" points in time are weighted (the most obvious
at all times).
Special stochastic process - random walk with
drift
Example: Exponential filtering of a real rate of
return with eviews and α = 0.2
Yt = δ + Yt-1 + ut (δ as constant and ut white noise)
Eviews: "Smoothing method": Trend and seasonality Problem: Everything OK except Durbin Watson (0.56)
Determination of expected value and variance (with indicates a problem. Reason: trend. If time series are
can be included in exponential smoothing.
constant = 0 and Y0 = 0); Substitution by repeated not stationary -> usual test procedures provide
Holt-Winters method. At the beginning, you don't insertion; Yt = ut + ut-1 + ... + u1 (Yt is omitted), acc. distorted estimators; leads to false regression.
know whether it's additive or multiplicative. Test both Calculation rule for E and Var.
and select where “Root Mean Squared Error (RMSE)” Stationarity important prerequisite for analysis; if not
stationary -> make stationary! (Remove trend).
Weak and steady process Although all theoretical auto-correlations are positive,
some estimates will be negative. Be careful when
Weakly stationary (also covariance-stationary) if: interpreting correlograms.
- μt = E (Yt) (strict, i.e. for all t, independent of time) Lag operator and characteristic polynomial AR (p)
The expected value is constant
- Autocovariance depends only on k (the distance). A lag operator is often used to simplify the notation of
- It follows: variance is constant an AR (p) process. The lag operator L shifts the index
- Weakly stationary if your first and second moments of a variable by 1 in the past. LYt = Yt-1
do not depend on time. (E.g. black stat. = Black white
noise) Multiple use of L corresponds to multiple shift: LsYt =
Yt-s (specific case identity I: L0 = I -> L0Yt = Yt)
Types of instationarities:
AR (p) model in the notation with lag operator:
Trends (not mean stat.), Heteroscedasticity (not Yt = α + φ1Yt-1 + φ2Yt-2 + ... + φpYt-p + ut
variance-stationary), or both, long periodicities Yt = α + φ1LYt + φ2L2Yt + ... + φpLpYt + ut
(business cycles) Yt = α + (φ1L + φ2L2 + ... + φpLp) Yt + ut
(1- φ1L - φ2L2 - ... - φpLp) Yt = α + ut
How can stationary time series be recognized? ---> Prob everywhere 0.000 (no white noise)
Attention: The same correlograms can differ from one
Consideration of the time series plots (danger of another. Time series come!
deception) / autocorrelations and partial Characteristic polynomial:
autocorrelation function of the time series / tests for AR (p) process – definition Φ (L) = (1 - φ1L - φ2L2 - ... - φpLp)
stationarity (LS05) / window technology (different "cut Autoregressive process -> variable is regressed on Φ (L) Yt = α + ut
out windows" variabilities significant?) itself. Model for the dependence of Yt on the past.
MA (q) process - definition (moving average)
Lagged scatter plot (e.g. logarithmic lynx data) AR (1) process: Yt = α + φYt-1 + ut (φ = Beta at
regression number) Yt is the weighted sum of the error terms
Is linear dependence on Yt and Yt-1 recognizable? AR (p) process: Yt = α + φ1Yt-1 + φ2Yt-2 + ... + φpYt- MA (1) process: Yt = α + ut + θut-1
Y axis: normal data / shift X axis by time unit p + ut MA (q) process:
In a year with few lynxes, it has little in the following Yt = α + ut + θ1ut-1 + θ2ut-2 + ... + θqut-q
year Error term is white noise The error term is white noise
A negative sign at φ leads to stronger swings.
Autocorrelation function (AC function) φ = 1 -> no longer stationary (up to 1 limit, stationary Example MA (1) process / properties
<1) Time series are shifted upwards on average.
Measurement of dependence on Yt and Yt-k (Eviews Negative sign at θ -> Stronger debauchery
provides value). MA (∞) representation and moments AR (1) Time series are also stationary for θ = +/- 1.
Thumb rules for the AC function of a stationary processes
process: By recursive insertion, the AR (1) representation AR (∞) representation (inversion) and moments of
Mind. N = 30 data points / only lags with k <n / 4 results in the MA (∞) representation: Yt = α + φYt-1 + an MA (1) process By recursive insertion, the MA (1)
(purists) or Trust k <n / 2 (pragmatist) / Only lags k <10 ut ..... (algebraic conversion) representation results in the AR (∞) representation:
* log10 (n). Equal distances and no gaps! (Holidays Yt = α + ut + θt-1 .... --->
are a problem).
Prob> 0.05 -> assume H0 (example: lag 8 to lag 12; For k> 1, all autocorrelations are 0.
white noise (not significantly different)).
Lag operator and characteristic polynomial MA (q)
Patterns for periodicity and non-stationary time
series: A lag operator is often used to simplify the notation of
an MA (q) process.
(for positive φ all theoretical autocorrelations are> 0)
For the n-1 possible empirical autocorrelations (finally) MA (q) model in the notation with lag operator:
the following sum applies (without proof):
Yt = α + θ1ut-1 + θ 2ut-2 + ... + θqut-q
Yt = α + ut + θ 1Lut + θ 2L2ut + ... + θqLqut
Yt = α + (1 + θ 1L + θ2L2 + ... + θqLq) ut
Characteristic polynomial:
Θ (L) = (1 - θ1L - θ2L2 - ... - θqLq)
Yt = α + Θ (L) ut
Coefficient β is well estimated. Stay true to ---> Residuals are autocorrelated (2nd order) with lag
expectations. 2
Standard error is distorted. Correction with flu: Cochrane-Orcutt model with lag 2
AR; φ1 = -0.8 und MA; θ1 = 0.3 (1. Bild LS05 - Trends and Unit Root Test / ARIMA Models
ACF, 2. Bild PACF)
Deterministic vs. Stochastic trend
Properties of the ACF function of AR (p)
Fundamental differences:
Stochastic trend:
Yt = f (t) + ut ut ̴ i.i.d.
f (t): function of time t
Frequently linear trend:
Yt = δ + Yt-1 + ∑ui
Y0, δ determine growth path (drift)
moments:
Properties of the PACF function of MA (q)
E (Yt) = Y0 + δ * t
Properties of the PACF function of MA (q)
Var (Yt) = t * σu2
Non-canceling, damped, exponential or sine function.
Random walk with drift is not stationary (not constant
Root of the characteristic polynomial real =
mean, not constant variance)) not variance / trend
exponential function. Root of the characteristic
stationary
polynomial complex = sine function.
Stochastic trend Random Walk (B)
Theoretical ACF and PACF different ARMA (1,1)
processes
Yt = Yt-1 + ut
AR; φ1 = 0.8 and MA; θ1 = 0.3 (1st picture ACF, 2nd
Representation as an MA (∞) process
picture PACF)
Yt = Yt-1 + ∑ui
moments:
The new time series is white noise. Original time St = Current share price
series is stationary. (white noise is stationary)
<- if φ goes to 1 it becomes unstable For statistical analysis it is assumed that rt is normally
Example differential time series: GDP growth distributed. This leads to two problems:
Original time series = difference stationary 1) A normally distributed random variable takes values
between minus infinity and plus infinity, but rt is
New time series = growth rate (log return) compared to capped with -1 to plus infinity (-1 if St = 0 and St-1 ≠ 0)
previous quarter's Eviews: gdp_log = log (gdp) - log
(gdp (-4)) // Yt = (1-L4) log (BIPt) = log (BIPt) -log 2) Multi-period returns are not normally distributed,
AR (1) process is stationary if φ applies (BIPt- 4) New time series is stationary. even if single-period returns are normally distributed
(follows from the probability calculation).
Comparison: A: AR (1), B: Random Walk, C: Integrated time series
Random Walk with Drift, D: Linear trend -> Problems can be almost solved with steady returns,
A stochastic process Yt means integrated from the since these are viewed as more normally distributed.
order d if it has to be differentiated d times so that it
becomes stationary. Continuous (“log return”) versus discrete returns
RW drift long in one direction without returning to the The formation of the difference is described for
mean of the time series -> trend simplification with the difference operator ∆: The
difference operator ∆ forms the difference: ∆1 = ∆ Yt =
Yt- Yt-1 -> Multiple use of ∆ corresponds to multiple
difference formation. Example of second difference:
Distance from 1-year GD in DJ since 1900 in percent Unconditional variance: Varu (Yt) = 2 / (1- φ2) =
0.33 / (1-0.62) = 0.52
3 Stylized Facts from Financial Time Series - 3.
Leptokurtosis
Further developments: