Professional Documents
Culture Documents
Compiled Notes: Mscfe 610 Econometrics
Compiled Notes: Mscfe 610 Econometrics
Compiled Notes
Module 4
MScFE 610
Econometrics
Table of Contents
Introduction............................................................................................................. 4
Unit 2:
Post-script ............................................................................................................... 14
Unit 3:
Bibliography....................................................................................................................... 24
Unit 1: Introduction
In this module we turn to a central aspect of interest to the financial engineer, namely risk. In
econometric terms, we link this to the volatility of returns, which we characterize with the
statistical concept of variance, or its square root, standard deviation. Importantly, this volatility is
not constant over time.
Almost all returns on stocks or stock market indices show periods of relative tranquility followed
by periods of high volatility, as we observed in the daily log returns on the S&P500.
Figure 1
During the Global Financial Crisis of 2008, the daily log returns on the S&P500 were far more
volatile than in the rest of the sample. Similarly, there are periods in 2010, 2011, 2015/16 and
possibly in 2018 where the volatility was obviously larger than in the rest of the sample.
This means that this time series does not have constant variance. Hence, according to our
definitions in Module 3, the series does not qualify as weakly stationary. In terms of our definitions
in Module 2, this means that the series is not homoscedastic, but heteroscedastic. If the mean is
constant, as it seems to be in the S&P500 log returns, simple linear method will still be consistent,
but not efficient.
In this module, we will build models that simultaneously model the conditional expectation and
conditional variance of the data-generating process.
An accessible treatment of applications and the generic theoretical issues is available in Enders
(2014) chapter 3. A condensed treatment of the main concerns for the financial engineer is
available in Tsay (2010) chapter 3, and a deep treatment of the theoretical issues is available in
Hamilton (1992).
We assume that the conditional expectation of the process is modeled as an ARMA process, but
now with errors that are not white noise (that is, they do not have constant variance). For
expositional simplicity, we assume that an AR(1) process is sufficient to capture the time series
properties of the mean of the process. Note that nothing would change if we assumed a more
complicated time series process for the conditional mean, but all the important aspects of this
choice remains the same as in the previous module:
Now, however, we assume that the error process 𝑢𝑢𝑡𝑡 has the following structure:
Firstly, the expectation (conditional and unconditional 1) of the error remains zero, so only its
variance deviates from white noise:
Secondly, for the model of the conditional mean to be consistent, we still require the level of the
errors to be uncorrelated over time:
We allow the conditional variance 𝔼𝔼𝑡𝑡 (𝑢𝑢𝑡𝑡2 ) = 𝜎𝜎𝑡𝑡2 to be auto-correlated. The subscript 𝑡𝑡 now denotes
that the time 𝑡𝑡 conditional variance might depend on the moment of observation 𝑡𝑡. Unfortunately,
this variance is not observable, so we have to build a model that we can use to estimate its time
path. The various assumptions we make on this model leads to the different types of ARCH
models. A large number of different models have been developed for different applications.
1
The notation 𝔼𝔼𝑡𝑡−1 (𝑧𝑧𝑡𝑡 ) means 𝔼𝔼(𝑧𝑧𝑡𝑡 |𝐹𝐹𝑡𝑡−1 ) where 𝐹𝐹𝑡𝑡−1 refers to all information known up to period 𝑡𝑡.
The simplest version which we will call an ARCH(1). We thus assume that the following process
generates the innovations and their conditional variance (this specification is due to Engle (1982)):
2
𝑢𝑢𝑡𝑡 = 𝜀𝜀𝑡𝑡 �𝜔𝜔 + 𝛼𝛼1 𝑢𝑢𝑡𝑡−1 ,
where 𝜀𝜀𝑡𝑡 is a white noise, as defined in Module 3, that is independent of 𝑢𝑢𝑡𝑡 with unit variance
𝜎𝜎𝜀𝜀2 = 1.
Note what this means for the conditional and unconditional variance. The period 𝑡𝑡 − 1 conditional
variance of 𝑢𝑢𝑡𝑡 is:
2
𝔼𝔼𝑡𝑡−1 (𝑢𝑢𝑡𝑡2 ) = 𝔼𝔼𝑡𝑡−1 �𝜀𝜀𝑡𝑡 �𝜔𝜔 + 2
𝛼𝛼1 𝑢𝑢𝑡𝑡−1 �
2 )�
= 𝜎𝜎𝜀𝜀2 �𝜔𝜔 + 𝛼𝛼1 𝔼𝔼𝑡𝑡−1 (𝑢𝑢𝑡𝑡−1
2
= 𝜔𝜔 + 𝛼𝛼1 𝑢𝑢𝑡𝑡−1.
This is an AR(1) process in 𝑢𝑢𝑡𝑡2 , i.e. it is a stationary, mean-reverting process (if 𝛼𝛼1 < 1).
2
𝔼𝔼(𝑢𝑢𝑡𝑡2 ) = 𝔼𝔼 �𝜀𝜀𝑡𝑡 �𝜔𝜔 + 2
𝛼𝛼1 𝑢𝑢𝑡𝑡−1 �
2 )
= 𝜔𝜔 + 𝛼𝛼1 𝔼𝔼(𝑢𝑢𝑡𝑡−1
𝜔𝜔
= ,
1 − 𝛼𝛼1
2 ).
where the last step used the definition of a stationarity: 𝔼𝔼(𝑢𝑢𝑡𝑡2 ) = 𝔼𝔼(𝑢𝑢𝑡𝑡−1 Note, for this to be
positive, we will require 𝛼𝛼1 < 1 . In practice, stronger restrictions are necessary for large sample
results to hold. You can read up on this in Tsay (2010).
You should check that this assumed process indeed satisfies all the requirements of an
unconditional white noise process: 𝔼𝔼(𝑢𝑢𝑡𝑡 ) = 𝔼𝔼𝑡𝑡−1 (𝑢𝑢𝑡𝑡 ) = 0 and 𝔼𝔼(𝑢𝑢𝑡𝑡 𝑢𝑢𝑡𝑡−𝑠𝑠 ) = 0 ∀ 𝑠𝑠 > 0.
We can easily extend this to an ARCH(𝑝𝑝) process if we assume the conditional variance follows an
AR(𝑝𝑝) process:
2 2
𝑢𝑢𝑡𝑡 = 𝜀𝜀𝑡𝑡 �𝜔𝜔 + 𝛼𝛼1 𝑢𝑢𝑡𝑡−1 + ⋯ + 𝛼𝛼𝑝𝑝 𝑢𝑢𝑡𝑡−𝑝𝑝
or a GARCH(𝑝𝑝, 𝑞𝑞) process if we assume that the conditional variance follows an ARMA(𝑝𝑝, 𝑞𝑞)
process (Bollerslev, 1986):
𝑝𝑝 𝑞𝑞
2
ℎ𝑡𝑡 = 𝜔𝜔 + � 𝛼𝛼𝑖𝑖 𝑢𝑢𝑡𝑡−𝑖𝑖 + � 𝛽𝛽𝑗𝑗 ℎ𝑡𝑡−𝑗𝑗
𝑖𝑖=1 𝑗𝑗=1
The conditions for stationarity and for large sample theory to hold are available in the texts
mentioned in the introduction and in the original papers cited.
There are a number of limitations of the ARCH model, some of which are addressed by the
extensions to the ARCH model we consider (see Tsay, 2010).
1 The model assumes symmetric effects: whether a shock to the process is positive or
negative, the impact on the conditional variance is the same. This is not representative for
financial series – negative shocks tend to be more destabilizing to the variability of the
returns on a financial asset than positive shocks. Consider what happened in the Global
Financial Crisis of 2008. The EGARCH model mentioned below is a generalization built to
deal with this weakness.
2 ARCH models allowing only AR properties in the conditional variance process tend to
predict slow mean reversion in the conditional variance. In practice, there can be sharp
spikes in volatility that vanish quickly.
3 The ARCH model struggles to deal with excess kurtosis (relative to normal distributions)
due to fundamental mathematical constraints that must hold for the derivations to be
well defined. This can be rectified somewhat by using distributions other than the normal
ones (with fatter tails and more kurtosis).
4 The ARCH model allows only the shock to the conditional expectation to feed to
persistence in the conditional variance but does not allow changes in the conditional
variance to affect the conditional expectation. In practice, we may expect the expected
return on an asset to be depressed in periods of high volatility. The GARCH-M process is
built to allow this feedback.
5 The model imposes a specific, mechanical process that generates the conditional variance
time path. It does not explain why the conditional variance follows this process. For the
skeptical financial engineer, this is not too troublesome if the interest is in near future
projections rather than fundamental analysis. For the long run investor, however, this is
not desirable.
1 −ε2𝑡𝑡
𝐿𝐿𝑡𝑡 = � � exp � �
√2πσ2 2σ2
{ε𝑡𝑡 } – independent ⟹the likelihood of the joint realizations is the product of the individual
likelihoods.
𝑇𝑇
−ε2𝑡𝑡
1
L = �� � exp � 2 �
√2πσ2 2σ
𝑡𝑡=1
We apply the natural logarithm to both sides as it is easier to work with the sum as compared to
the product.
𝑇𝑇
𝑇𝑇 𝑇𝑇 1
𝑙𝑙𝑙𝑙𝑙𝑙 = − ln(2𝜋𝜋) − 𝑙𝑙𝑙𝑙𝜎𝜎 2 − 2 � 𝜀𝜀𝑡𝑡2 (1)
2 2 2𝜎𝜎
𝑡𝑡=1
The goal is to select the distributional parameters so as to maximize the likelihood of drawing the
observed sample.
𝑇𝑇
𝑇𝑇 𝑇𝑇 1
𝑙𝑙𝑙𝑙𝑙𝑙 = − ln(2𝜋𝜋) − 𝑙𝑙𝑙𝑙𝜎𝜎 2 − 2 �(𝑦𝑦𝑡𝑡 − 𝛽𝛽𝑥𝑥𝑡𝑡 )2 (𝟑𝟑)
2 2 2𝜎𝜎
𝑡𝑡=1
𝑇𝑇
𝜕𝜕𝜕𝜕𝜕𝜕𝜕𝜕 𝑇𝑇 1
2
= − 2 + 4 �(𝑦𝑦𝑡𝑡 − 𝛽𝛽𝑥𝑥𝑡𝑡 )2 (𝟒𝟒)
𝜕𝜕𝜎𝜎 2𝜎𝜎 2𝜎𝜎
𝑡𝑡=1
and
𝑇𝑇
𝜕𝜕𝜕𝜕𝜕𝜕𝜕𝜕 1
= 2 �(𝑦𝑦𝑡𝑡 𝑥𝑥𝑡𝑡 − 𝛽𝛽𝑥𝑥𝑡𝑡2 ). (5)
𝜕𝜕𝜕𝜕 𝜎𝜎
𝑡𝑡=1
𝜕𝜕𝜕𝜕𝜕𝜕𝜕𝜕
=0 (6)
𝜕𝜕𝜎𝜎 2
and
𝜕𝜕𝜕𝜕𝜕𝜕𝜕𝜕
= 0. (7)
𝜕𝜕𝜕𝜕
Solving for the values of σ2 and β yield the maximum value of lnL
∑ 𝜀𝜀𝑡𝑡2
𝜎𝜎� 2 = (𝟖𝟖)
𝑇𝑇
∑ 𝑥𝑥𝑡𝑡 𝑦𝑦𝑡𝑡
𝛽𝛽̂ = (𝟗𝟗)
∑ 𝑥𝑥𝑡𝑡2
The results are similar to the ones provided by OLS estimates. It is easy to find the optimal
solutions when there is a linear model.
𝜀𝜀𝑡𝑡 = 𝑣𝑣𝑡𝑡 (ℎ𝑡𝑡 )0.5 (the conditional variance is not constant) (10)
𝑇𝑇
1 −𝜀𝜀𝑡𝑡2
𝐿𝐿 = � � � 𝑒𝑒𝑒𝑒𝑒𝑒 � � (11)
�2𝜋𝜋ℎ𝑡𝑡 2ℎ𝑡𝑡
𝑡𝑡=1
𝑇𝑇 𝑇𝑇
𝑇𝑇 ε2𝑡𝑡
lnL = − ln(2π) − 0.5 � 𝑙𝑙𝑙𝑙(ℎ𝑡𝑡 − 0.5 � � � (12)
2 ℎ𝑡𝑡
𝑡𝑡=1 𝑡𝑡=1
2 (13)
ℎ𝑡𝑡 = 𝑎𝑎0 + 𝑎𝑎1 𝜀𝜀𝑡𝑡−1
𝑇𝑇 𝑇𝑇
𝑇𝑇 − 1 𝑦𝑦𝑡𝑡 − β𝑥𝑥𝑡𝑡
lnL = − ln(2π) − 0.5 �{l n(α0 } + α1 ε2𝑡𝑡−1 ) − 0.5 � � � (14)
2 α0 + α1 ε2𝑡𝑡−1
𝑡𝑡=2 𝑡𝑡=2
• Computers are able to select parameter values that maximize log likelihood
The academic world had more success in modeling volatility then in modeling financial assets
prices and returns. Engle and Bollerslev GARCH framework is still popular today after more than
three decades. We will focus more on the financial interpretation of omega, alpha and beta
parameters in GARCH(1,1).
2 2
𝜎𝜎𝑡𝑡2 = 𝜔𝜔 + 𝛼𝛼𝜀𝜀𝑡𝑡−1 + 𝛽𝛽𝜎𝜎𝑡𝑡−1
𝛼𝛼 is a reaction parameter. High 𝛼𝛼 is generally associated with spiky or nervous market while low 𝛼𝛼
indicates stable market. Alexander (2008) considers that for daily data used in GARCH model, 𝛼𝛼
generally ranges from 0.05 (stable market) to 0.10 for a spiky one.
𝛽𝛽 indicates volatility persistence (how much from past volatility is transferred into current
volatility). High beta means high persistency and therefore volatility clustering appears. 𝛽𝛽
generally takes values from 0.85 to 0.98 for daily data used in the model according to Alexander
(2008).
Low 𝛼𝛼 values are associated with high 𝛽𝛽 and high 𝛼𝛼 is connected to low 𝛽𝛽.
𝜔𝜔
-> unconditional variance
1−𝛼𝛼−𝛽𝛽
Halls-Moore (2017) presents how ARMA-GARCH approach can be used in algorithmic trading.
https://www.quantstart.com/articles/arima-garch-trading-strategy-on-the-sp500-stock-market-
index-using-r
Post-script
You may also experiment with simulating a GARCH(1,1) process in Python as follows:
• GARCH models are popular quantitative models for projecting mean and short-term
volatility of interest rates, commodities, stock prices, cryptocurrencies or stock market
indexes. Most GARCH research papers focus on stock and FX, forecasting, and other
types of assets being less popular.
• Univariate Stochastic Volatility models like GARCH are generally used to predict next
period evolution (𝑡𝑡 + 1) volatility and return. It is generally easier to predict a trend than
to predict a certain level of price / return.
• The models are used to capture the so-called volatility clustering phenomenon in financial
markets. Volatility clustering means that “large changes tend to be followed by large
changes, of either sign, and small changes tend to be followed by small changes”
(Mandelbrot, 1963). Consequently, high volatility is usually followed by a high-volatility
period, while low volatility is usually followed by a low-volatility period in the financial
markets.
• The strategy is carried out on a rolling basis. GARCH parameters are estimated based on
previous k days EUR/USD daily returns. So 𝑘𝑘 days are used as a window to estimate
GARCH parameters. The goal is to forecast next day (𝑡𝑡 + 1) return value.
• The strategy is focused on the conditional mean equation (ARMA part) in order to
forecast next period (t+1) FX return, while the conditional variance equation is used for
risk management.
• If the predicted EUR/USD > 1 -> EUR is expected to appreciate against USD -> Buy EUR
and sell USD.
• If the predicted EUR/USD <1 -> EUR is expected to depreciate against USD -> Sell EUR
and buy USD.
• The quant trader indicates the level of 𝑘𝑘 (it can be 500 or 1,000 for example). If 𝑘𝑘 is too
low, the algorithm cannot estimate GARCH parameters. If 𝑘𝑘 is too high, there could be
structural breaks in the data not correctly captured by parameters. Therefore, GARCH
parameters can vary from one period to another.
• The GARCH model may not converge in certain cases (the solution is not unique). In this
case the algorithm can indicate “buy EUR”, which is simply a guess.
• The mentioned algorithm is basic and there is room for further improvement. Ways for
improving the algorithm include introducing signals generated by:
Halls-Moore (2017) presents how ARMA-GARCH approach can be used in algorithmic trading.
https://www.quantstart.com/articles/arima-garch-trading-strategy-on-the-sp500-stock-market-
index-using-r
We forecast the evolution of EUR/USD daily return using a GARCH(1,1) model. GARCH models
are popular quantitative models for projecting high-frequency financial asset returns with
stochastic volatility. The following example shows how to use the GARCH approach in FX
forecasting.
Rugarch package may not work in older versions of R as truncnorm package cannot be loaded.
Null values are removed from the analysis
Here are links to download the data and code that can also be found in the Additional Files folder
under the M4 Resources section.
https://masters.wqu.org/pluginfile.php/175327/mod_folder/content/0/DEXUSEU.xlsx?forcedow
nload=1
https://masters.wqu.org/pluginfile.php/175327/mod_folder/content/0/EURUSD_GARCH_Foreca
st.txt?forcedownload=1
R code:
install.packages("rugarch")
library("rugarch")
plot(fx)
#Forecast FX
garch11_forecast<-ugarchforecast(garch11_fit, nahead=15)
garch11_forecast
Abstracting from ARMA properties of some return series 𝑟𝑟𝑡𝑡 , a simple version, the GARCH(1,1)-M
is given by the data generating process:
2
𝑟𝑟𝑡𝑡 = 𝜇𝜇 + 𝑐𝑐𝜎𝜎𝑡𝑡−1 + 𝑢𝑢𝑡𝑡
2
ℎ𝑡𝑡 = 𝜔𝜔 + 𝛼𝛼1 𝑢𝑢𝑡𝑡−1 + 𝛽𝛽1 ℎ𝑡𝑡−1
If 𝑐𝑐 is significantly positive, it can be considered a “risk premium” for the asset in times when it is
more volatile. Other alternatives are to use just the conditional standard deviation 𝜎𝜎𝑡𝑡 or the log of
the variance in the mean equation. The most appropriate data-generating process will depend on
the economics behind the data-generating process you are trying to model.
To model this, we can use the threshold-GARCH model (TGARCH) where we add a dummy
variable 𝑑𝑑𝑡𝑡−1 which is zero when 𝜀𝜀𝑡𝑡−1 > 0 and equal to 1 when 𝜀𝜀𝑡𝑡−1 < 0.
The simplest case, considering only the variance equation in ℎ𝑡𝑡 , is the following TGARCH(1,1)
model:
2 2
ℎ𝑡𝑡 = 𝜔𝜔 + 𝛼𝛼1 𝑢𝑢𝑡𝑡−1 + 𝛾𝛾1 𝑑𝑑𝑡𝑡−1 𝑢𝑢𝑡𝑡−1 + 𝛽𝛽1 ℎ𝑡𝑡−1
2 2
Thus, if 𝜀𝜀𝑡𝑡−1 > 0 (so 𝑢𝑢𝑡𝑡−1 > 0), the impact of 𝑢𝑢𝑡𝑡−1 on ℎ𝑡𝑡 is 𝛼𝛼1 . If 𝜀𝜀𝑡𝑡−1 < 0, the impact of 𝑢𝑢𝑡𝑡−1 on ℎ𝑡𝑡 is
(𝛼𝛼1 + 𝛾𝛾1 ). If 𝛾𝛾1 is positive and statistically significant, it means that negative shocks have larger
(positive) impacts on volatility than positive shocks.
The EGARCH model employs a logarithmic transformation to ℎ𝑡𝑡 to ensure positive variances (in all
the other models, all the estimated coefficients need to be positive, which is not the case here) as
well as working in the levels of the residuals (not their squares) and additionally, standardizing by
their standard deviation �ℎ𝑡𝑡 .
𝑢𝑢𝑡𝑡−1 𝑢𝑢𝑡𝑡−1
ln(ℎ𝑡𝑡 ) = 𝜔𝜔 + 𝛼𝛼1 + 𝛾𝛾1 � � + 𝛽𝛽1 𝑙𝑙𝑙𝑙(ℎ𝑡𝑡−1 )
�ℎ𝑡𝑡−1 �ℎ𝑡𝑡−1
Again, we have leverage effects: if 𝑢𝑢𝑡𝑡−1 > 0 its impact on ln(ℎ𝑡𝑡 ) is 𝛼𝛼1 + 𝛾𝛾1 . If 𝑢𝑢𝑡𝑡−1 < 0 its impact on
ln(ℎ𝑡𝑡 ) is −𝛼𝛼1 + 𝛾𝛾1 .
Enders (2014) warns, however, that it may be very difficult to forecast with this model.
If we were to augment and integrate, i.e. ARIMA(𝑝𝑝, 𝑑𝑑, 𝑞𝑞) with 𝑑𝑑 > 0, we would obtain an integrated
GARCH, or IGARCH model. This is a model where a shock to the conditional variance (and/or the
conditional expectation) never fades out.
We have not explored this yet, but theoretically even a simple unit root process does not have a
well-defined unconditional variance. This property carries over to the IGARCH process as well.
This is intuitive: since a unit root process, even without drift, may wander arbitrarily up or down
for arbitrarily long periods, there is no way to describe, with a single number, what variability we
may expect over all time. In formal terms, starting from some point in time, the process will have a
well-defined conditional variance.
If we attempt to take the limit of this to find the unconditional variance, we end up adding up an
infinite sequence of positive conditional variances, implying that the unconditional variance
“equals” infinity. This is what we mean by “not well defined”. In practice, it is hard to think of
examples of interest to the financial engineer that fall into this category, although Tsay (2010)
gives an application to value at risk.
α + β < 1 if the returns process is stationary, which is necessary for mean reverting processes.
If α + β = 1, the return process is a random walk, and the GARCH(1,1) process may be
expressed as:
When ω = 0, the IGARCH model becomes an EWMA (exponentially weighted moving average)
model.
Kalman filter
The Kalman filter is an extension to the maximum likelihood approach particularly built to provide
an estimation/filtering technique for what is known as state space models.
2 A state equation. This is a model equation that describes how the state of the system (in
this case, the conditional volatility) evolves over time.
The Kalman filter uses a probabilistic model of the evolution and disturbances of the observed
variables and the fundamentally unobserved state of the system to construct period by period
best estimates of the value of the state. That is, we use the best estimates of the parameters of the
system of the equation combined with our distributional assumption (e.g. that the innovations are
normal) to construct an estimate of the current state of the system.
Then, when a new observation arrives, we use the new information to update our best guess of the
parameters, and the previously estimated state variable, to provide our best estimates of the new
current level of the state variable, which in turn is used to improve the forecast of the observed
variables. Different versions of the Kalman filter exists, some employing only forward filtering,
others applying both forward and backward filtering (i.e. after going from first to last observation,
starting from the last observation and asking what would have been the best guess of the previous
state variable level). In this sense, the Kalman filter provides a smoothed estimate of the true state
variable process in any sample.
Tsay (2010) states that while the stochastic volatility models show some ability to improve in
sample fit, their out-of-sample forecasting ability relative to other models has yielded mixed
results.
Equations
Kalman Filter model assumes that the state of a system of time 𝑡𝑡 evolves from the prior state at
time 𝑡𝑡 − 1 according to the equation:
𝑥𝑥𝑡𝑡 – state vector containing the terms of interest for the system (volatility of returns) at time 𝑡𝑡
𝑢𝑢𝑡𝑡 – the vector containing any control inputs (steering angle for example at physics)
𝑓𝑓𝑡𝑡 – state transition matrix which applies the effect of each system state parameter at time 𝑡𝑡 − 1
on the system state at time t (example the position and velocity at time 𝑡𝑡 − 1 both affect the
position at time 𝑡𝑡)
𝑏𝑏𝑡𝑡 – the control-input matrix which applies the effect of each control input parameter in the vector
𝑢𝑢𝑡𝑡 on the state vector
𝑤𝑤𝑡𝑡 – the vector containing the process noise terms for each parameter in the state vector
An observation (or measurement) 𝑧𝑧𝑡𝑡 of the true state 𝑥𝑥𝑡𝑡 is made according to:
where 𝐻𝐻𝑡𝑡 is the observation model which maps the true state space into the observed space and 𝑣𝑣𝑡𝑡
is the observation noise which is assumed to be zero-mean Gaussian white noise with covariance
𝑅𝑅𝑡𝑡
Kalman filter can be used in pairs trading. Let's say two assets are cointegrated (there is a strong
connection between them). Example: a pair of Exchange Traded-Funds (ETFs).
You can perform a linear regression between the two ETFs. Determining the hedging ratio can be
a problem in this situation. The ratio shows how much of each asset has to be bought or sold and it
is not constant over time. Therefore, it is crucial to have a mechanism that automatically adjusts it.
The Kalman filter can be used in this situation. The “true” hedging ratio is treated as an
unobserved hidden variable in the state space model. The ratio is estimated considering the
“noisy” observations.
Bibliography
Alexander, C. (2008). Pricing, Hedging and Trading Financial Instruments. Wiley.
Chu J. et al. (2017). GARCH Modelling of Cryptocurrencies. Journal of Risk and Financial
Management.
Daroczi G. et al. Introduction to R for Quantitative Finance. Packt Publishing, Chapter 1 Time
Series Analysis, pages 7-26.
Ding J. (2018). Time Series Predictive Analysis of Bitcoin with ARMA-GARCH model in Python
and R.
Jeet P. and Vats P. Learning Quantitative Finance with R. Packt Publishing, Chapter 4: Time Series
Modeling, pages 96-125.
Jiang, W. (2012). Using the GARCH model to analyze and predict the different stock markets.
Letra, I. (2016). What Drives Cryptocurrency Value? A Volatility and Predictability Analysis.
Master Thesis.
Tusell, F. (2011). Kalman Filtering in R, Journal of Statistical Software, Volume 39, Issue 2.
1 Discuss what can be deduced from the time series properties (persistence and volatility)
of the series from the time series plot.
Next, consider the ACF and PACF of the process and its square:
Discuss what can be deduced from the time series properties of the series from these ACF and
PACF functions and discuss what this means for the appropriate modeling approach.
Consider the output of estimating 6 different ARMA-GARCH models for this process via maximum likelihood assuming normal errors:
Test
Test on: Distribution Statistic p-value Statistic p-value Statistic p-value Statistic p-value Statistic p-value Statistic p-value
Jarque-Bera residuals Chi^2 3.135 0.209 3.090 0.213 3.148 0.207 2.697 0.260 2.241 0.326 0.423 0.810
Shapiro-Wilk residuals W 0.996 0.183 0.996 0.183 0.995 0.082 0.995 0.129 0.996 0.234 0.995 0.132
Ljung-Box residuals Q(10) 3.348 0.972 3.379 0.971 4.617 0.915 6.424 0.778 166.590 0.000 177.066 0.000
Ljung-Box residuals Q(15) 5.473 0.987 5.515 0.987 6.774 0.964 8.633 0.896 182.690 0.000 194.670 0.000
Ljung-Box residuals Q(20) 8.571 0.987 8.611 0.987 9.693 0.973 11.450 0.934 193.579 0.000 207.810 0.000
Ljung-Box residuals^2 Q(10) 8.468 0.583 8.352 0.595 8.678 0.563 9.317 0.502 27.412 0.002 9.255 0.508
Ljung-Box residuals^2 Q(15) 9.250 0.864 9.062 0.874 9.479 0.851 10.422 0.792 32.123 0.006 12.362 0.651
Ljung-Box residuals^2 Q(20) 14.195 0.820 14.029 0.829 12.543 0.896 13.230 0.867 37.976 0.009 19.198 0.509
Analyze each of the six models and select the best one. Your analysis should touch on the values
and statistical significance of coefficients and tests and come to a conclusion on whether the
models are encompassing and/or parsimonious. All coefficients need not be discussed individually,
only to the extent necessary for giving a complete answer.