Professional Documents
Culture Documents
(Georg Bol, Svetlozar T. Rachev, Reinhold Würth) (BookFi)
(Georg Bol, Svetlozar T. Rachev, Reinhold Würth) (BookFi)
(Georg Bol, Svetlozar T. Rachev, Reinhold Würth) (BookFi)
Risk Assessment
Decisions in Banking and Finance
Physica-Verlag
A Springer Company
Editors
Prof. Dr. Georg Bol Prof. Dr. h.c. mult. Reinhold Würth
Prof. Dr. Svetlozar T. Rachev Reinhold-Wü rth-Str. 12-17
University of Karlsruhe (TH) 74653 Künzelsau-Gaisbach
Kollegium am Schloss, Geb. 20.12 Germany
76131 Karlsruhe
Germany
bol@statistik.uni-karlsruhe.de
rachev@statistik.uni-karlsruhe.de
DOI: 10.1007/978-3-7908-2050-8
This work is subject to copyright. All rights are reserved, whether the whole or part of the material
is concerned, specifically the rights of translation, reprinting, reuse of illustrations, recitation,
broadcasting, reproduction on microfilm or in any other way, and storage in data banks. Duplication
of this publication or parts thereof is permitted only under the provisions of the German Copyright
Law of September 9, 1965, in its current version, and permissions for use must always be obtained
from Physica-Verlag. Violations are liable for prosecution under the German Copyright Law.
The use of general descriptive names, registered names, trademarks, etc. in this publication does
not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
987654321
springer.com
Preface
On April 5–7, 2006, the 9th Econometric Workshop with the title “Risk
Assessment: Decisions in Banking and Finance” was held at the University
of Karlsruhe (TH), Germany. The workshop was organized by the Institute
for Statistics and Mathematical Economics and the Adolf Würth GmbH &
Co.KG, Künzelsau. More than 20 invited speakers and 70 participants at-
tended the workshop. The papers presented at the conference dealt with new
approaches and solutions in the field of risk assessment and management,
covering all types of risk (i.e., market risk, credit risk, and operational risk).
This volume includes 12 of the papers presented at the workshop. We are
delighted with the range of papers, especially from practitioners.
Many people have contributed to the success of the workshop: Sebastian
Kring and Sven Klussmeier did the major part of organizing the workshop.
The organizational skills of Markus Höchstötter, Wei Sun, Theda Schmidt,
Nadja Safronova, and Aksana Hurynovich proved indispensable. Jens Büchele
und Lyuben Atanasov were responsible for the technical infrastructure while
Thomas Plum prepared the design for this volume. All of their help is very
much appreciated.
The organization committee wishes also to thank the School of Economics
and Business Engineering, Vice-Dean Professor Dr. Christof Weinhardt, and
Professor Dr. Frank Fabozzi (Yale University’s School of Management) for
their cooperation. Last but certainly not least we thank Professor Dr. h.c.
Reinhold Würth and the Adolf Würth GmbH & Ko.KG for their generous
support of this conference.
Christian Diekmann
1 Introduction
2 Automotive Finance
The significant role automotive finance plays today should be seen against the
background that the market for motor vehicles has matured in most countries
over the past two decades while economic growth has slowed down. The face
of the industry serving this market has kept changing during this period more
rapidly than ever before. Three major trends can be identified:
• A global race for economies of scale, resulting in a series of mergers and
take-overs1 accompanied by an extreme expansion of production capacity.
1
Actually the number of independent manufacturers of motor vehicles has halved
since the mid 80s.
2 C. Diekmann
Up-stream activities
Manufacturer 16%
Systems & modules suppliers 7%
Component specialists 8%
Standard parts suppliers 2%
Raw material providers 5%
38%
Down-stream activities
New car retailing 5%
Leasing & financing 9%
Insurance business 15%
Used car retailing 12%
Car rental business 4%
Service & parts business 17%
62%
Automotive Finance 3
Captive Portfolio
GMAC 143.9
FMCC 123.4
DC FS 97.9
T FS 63.3
VW FS 51.9
BMW FS 43.5
RCI Banque 21.9
PSA Banque 21.5
Volvo FS 7.1
Although captives dominate the automotive finance market they are far
from having total control of it (Table 4). The situation in the US where com-
mercial banks and other types of financial institutions account for about half of
the relevant market may serve as an example. It should be noted that captives
and their competitors show very different characteristics. Captives pursue au-
tomotive finance as a core activity and benefit from the close relation to their
parent group. This gives them the advantage of a superior representation at
the point of sale and low costs of distribution. Another benefit is that the
manufacturers favour incorporating incentives into finance rates rather than
giving direct discounts. On the other hand, captives have a somewhat different
mission from that of other lending institutions. Their job is to support brand
sales, acquisition new and keeping existing customers on board. This means
the customer is viewed not only from the perspective of a lending institution
but also from that of an automotive manufacturer. Therefore, in spite of gen-
erating profits they do not follow unbiased financial objectives. In contrast,
4 C. Diekmann
Table 4. Outstanding finance receivables: US loans & leases automotive business [6]
banks and other financial institutions pursue pure profit targets. They rather
view automotive finance as another area of their credit and leasing business.
A potential advantage particularly banks have over captives is a stronger po-
sition in refinancing. Besides, captives and commercial banks there are two
other types of financial service providers in the market. The first category
consists of independent finance companies who often focus on sub-segments
of the automotive finance market, e.g. the sub-prime, used vehicle or fleet
business. The second category comprises internet and direct lenders whose
focus is to increase returns by saving distribution costs, regularly trying to
disintermediate the dealer network.
It should be seen however, that while automotive finance enjoyed a pe-
riod of exuberant growth in the past, the market is beginning to show signs
of saturation. Competition intensifies spilling over from the vehicle markets
and leaving less scope for conservative business policies. Generating profitable
growth in this environment calls for maintaining a careful balance between risk
and return in the future. This leads us to a more specific analysis of the risks
involved in the automotive finance business.
3 Risk Assessment
The theory of risk management has made significant advances over recent
years. However, most studies undertaken in this field have been concerned
with corporate default risk and risk assessment in the case of traditional bank
Automotive Finance 5
Leasing accounts for a substantial portion of the auto finance business. This
has important implications for risk management.
Essentially a lease is a contract by which the lessor conveys the right to
use an asset for a specific period of time to the lessee. In return he receives
a series of payments. A major characteristic of leasing being that while the
lessee takes full responsibility for the leased asset, the legal title of property
remains with the lessor. In addition to this, leases often contain various op-
tions and obligations regarding the use of the leased asset at the end of the
contract. Leasing owes much of its popularity to the favourable effect it has
on liquidity, the flexibility it affords for the lessee and the chance it offers to
realise off-balance sheet finance as well as a favourable tax treatment. For this
reason, leasing has experienced substantial growth over several decades with
6 C. Diekmann
For a considerable time, captives and leasing companies were the exclusive
providers of automotive leasing in the US. However, during the 90s automotive
leasing became more and more popular and developed into a highly profitable
business. Anxious not to miss the chance for profitable growth this business
offered, several large banks and finance companies entered the market. As a
result the competitive pressure in the leasing market began to increase. As all
competitors were concerned about losing profitable business, they reviewed
their residual value policies. Since the early 90s, residual values had steadily
increased. Facing the strong economic situation of the mid-90s, lessors as-
sumed this trend to continue. Therefore, contractual residual values were not
only adjusted to actual levels, rather their historic increase was extrapolated.
Considering that on average residual values had already increased from 42%
in 1990 to 62% in 1997, this extrapolation had a substantial effect. Several
market participants calculated average residuals up to 72% [9]. As a result
automotive leasing experienced an impressive boom.
However, in 1996 and 1997 the sales figures for SUVs, began to decline.
To counter this effect, manufacturers started incentive programs to further
new vehicle sales. Soon the trend of increasing new vehicle prices came to
a broad end at market level. Nevertheless, leasing continued to grow from
US $ 86 bn in 1997 to a new record level in 1999. In the same year the decline
in new vehicle prices spilt over to the used vehicle market. In response to the
apparent downturn in used vehicle values and the exceptionally good deals
available on new vehicles, a growing number of lessees chose to return their
leased vehicles at the end of the contract. Dealers showed little interest in
pre-empting the returned vehicles and passed them on to the lessors where
the inflated residual value estimates, the fall in used vehicle values and the
increase of turn-in rates multiplied to heavy losses (Table 5). Considering the
fact that lease terms had increased up to 60 months, these losses were to
affect their financial performance for several years to come. In 2000, residual
losses reached the limit of US $ 10 bn [3]. Until 2003 they had accumulated
to over US $ 20 bn [4] leading to a substantial consolidation in the market.
In 1999 Wachovia and First Union discontinued their auto lease operations.
GE Capital and National City followed in 2000 [9]. Key Corp was next and
quit leasing in 2001. Many smaller lessors followed their example or vanished
completely. Captives decided to keep up leasing which by then accounted for
8 C. Diekmann
Loans
Year 2000 2001 2002 2003 2004 2005E
Captive .35 .37 .41 .49 .44 .45
Bank .33 .33 .30 .28 .30 .30
Finance Co. .16 .15 .15 .11 .13 .13
Credit Union .14 .14 .13 .10 .10 .10
Other .02 .02 .01 .02 .03 .03
Leases
Year 2000 2001 2002 2003 2004 2005E
Captive .46 .55 .50 .48 .51 .55
Bank .33 .26 .25 .21 .23 .20
Finance Co. .15 .13 .19 .23 .18 .17
Credit Union .04 .04 .03 .03 .03 .02
Other .03 .03 .03 .03 .03 .05
about 30% of all passenger cars put on the road but reacted by adopting
a more conservative residual value policy. In the years to follow, auto lease
volumes declined and took until 2004 to show first signs of recovery.
4 Conclusion
The risk inherent in automotive finance is to a considerable extent driven
by industry and market dynamics. As a result, the broad economic perspec-
tive regularly taken when assessing default and recovery risk in diversified
bank portfolios proves to be inappropriate. The correct assessment and the
successful management of risk in this area of finance require considerable ex-
pertise. This is particularly true of the leasing business where residual values
are the central link between risk and return. Given the growing economic
importance of automotive finance as well as the increasing competitive pres-
sure in the market, there is substantial need for further research. This will
inevitably involve a great deal of empirical efforts as well as the development
of sophisticated models tailored to industry specifics. In this context, the com-
Automotive Finance 9
References
[1] Adesa Inc. (2005), Global Vehicle Remarketing 04-05.
[2] DaimlerChrysler Financial Services AG (2005), Company Slides 2005.
[3] Fahey J (2003), Residual Risk, Forbes Magazine, Vol. 171, Iss. 13.
[4] Manheim (2004), The Used Car Market Report, 2004 Edition.
[5] Manheim (2006), The Used Car Market Report, 2006 Edition.
[6] Miczeznikowski J, Hirsch E, Reppa P (2003), Rebundling of the Auto
Finance Industry. Booz Allen Hamilton.
[7] Städtler A (2005), Besseres Investitionsklima stärkt Leasingwachstum, ifo
Schnelldienst No. 23/2005.
[8] Volkswagen Financial Services AG (2004), Investor Relations Presenta-
tion.
[9] Wood D (2001), Auto Leasing Becomes Demolition Derby, ERisk
02.08.2001.
Evidence on Time-Varying Factor Models
for Equity Portfolio Construction
1 Introduction
return, have been widely used to evaluate systematic risk, i.e., the return risk
associated with market movements.
When estimating the CAPM it is common practice to assume stock betas
to be invariant over time. However, this stability assumption has been ques-
tioned and a considerable amount of empirical evidence reports important
beta variation over time (see among others, [2, 5, 8, 12, 16, 19], as well as [7]).
Beta variation over time goes hand in hand with unstable correlations among
stock returns and time-varying VCMs. This might have serious consequences
for the outcomes of portfolio optimization which are not widely recognized
by now.
In [6] we consider VCMs that are derived from time-varying beta estimates
for mean-variance optimization. When estimating time-varying betas we rely
on a time-varying market model given by
yi,t = αi,t + βi,t xt + ui,t , ui,t ∼ N (0, σu,i
2
), i = 1...N, t = 1...T,
with yi,t denoting the return of stock i at period t and xt the market return,
respectively. The error term ui,t captures specific risk of stock i measured by
the standard deviation σu,i , and the slope coefficient βi,t measures the stock’s
return sensitivity with respect to xt . The coefficient αi,t denotes the stock
specific return component at time t.
For estimating time-varying coefficients βi,t we employ three well es-
tablished estimation approaches, namely (i) Moving Window Least Squares
(MWLS); (ii) Flexible Least Squares (FLS) and (iii) the Random Walk Model
(RWM). See [11] and [15] for an illustration of the estimation methods. We
compare estimation results of these approaches with those, generated by the
time-invariant Recursive Least Squares-approach (RLS).
Our empirical findings for the U.S. suggest that betas, stock correlations
and, hence, VCMs are subject to significant variation in the short run as well as
in the long run. In fact, important benefits arise from time-varying estimation
of the market model when compared to time-invariant estimation via RLS.
Moreover, we examine the outcomes from mean-variance portfolio selec-
tion strategies based on variance-covariance matrices derived from these esti-
mates. We obtain improved ex-ante risk estimates as well as portfolios that
have superior risk and return characteristics while being well diversified. For
the estimation techniques considered in this paper, we find the same rank-
ing for nearly all investigated criteria. Due to our results, FLS is the best
method. It is followed by RWM, MWLS and RLS. The FLS procedure deliv-
ers the most precise beta estimates as well as the most precise portfolio risk
estimates. Moreover, efficient frontiers suggest higher returns for given volatil-
ities, trading strategies show the highest Sharpe Ratios and finally, portfolios
are the most diversified.
To summarize, the portfolio performances found in our empirical anal-
ysis indicate a strong need for the application of time-varying estimation
approaches for estimating correlations in risk analysis and portfolio construc-
tion. Due to our results, the FLS estimate is the favourable method to do so.
Evidence on Time-Varying Factor Models 13
References
[1] Best, M.J. and Grauer, R.R. (1991) On the Sensivity of Mean-Variance-
Efficient Portfolios to Changes in Asset Means: Some Analytical and
Computational Results. Journal of Financial Studies 4, 2, 315-342.
[2] Bos, T. and Newbold, P. (1984) An Empirical Investigation of the Pos-
sibility of Systematic Stochastic Risk in the Market Model. Journal of
Business 57, 35-41.
[3] Chan, L.K.C., Karceski, J. and Lakonishok, J. (1999) On Portfolio Op-
timization: Forecasting Covariances and Choosing the Risk Model. The
Review of Financial Studies 5, 937-974.
[4] Chopra, Kjay K. and William Z. Zimba (1993) The Effect of Errors
in Means, Variances and Covariances on Optimal Portfolio Choice. The
Journal of Portfolio Management, Winter 1993, 6-11.
[5] Collins, D.W., Ledolter, J. and Rayburn, J. (1987) Some further Evidence
on the Stochastic Properties of Systematic Risk. Journal of Business 60,
425-448.
[6] Ebner, Markus and Thorsten Neumann (2008) Time-Varying Factor
Models for Equity Portfolio Construction. The European Journal of Fi-
nance 14, 381-395.
[7] Ebner, Markus and Thorsten Neumann (2005) Time-Varying Betas of
German Stock Returns. Journal of Financial Markets and Portfolio Man-
agement. 19, 1, 29-46.
[8] Fabozzi, F.J. and Francis, J.C. (1978) Beta as a Random Coefficient.
Journal of Financial and Quantitative Analysis 13, 101-115.
[9] Jacquier, E. and Marcus, A.J. (2001) Asset Allocation Models and Mar-
ket Volatility. Financial Analysts Journal, 16-29.
[10] Jagannathan, R. and Ma, T. (2003) Risk Reduction in Large Portfolios:
Why Imposing the Wrong Constraints Helps. Journal of Finance 58,
1651-1683.
[11] Kalaba, Robert E. and L. Tesfatsion (1989) Time-Varying Linear Re-
gression via Flexible Least Squares. Computers and Mathematics with
Applications 17, 1215-1245.
[12] Kim, D. (1993) The Extent of Non-Stationarity of Beta. Review of Quan-
titative Finance and Accounting 3, 241-254.
[13] Ledoit, Ollivier and Michael Wolf (2002) Improved Estimation of the
Covariance Matrix of Stock Returns With an Application to Portfolio
Selection. Working paper. University of California, Los Angeles.
[14] Lintner (1965) The Valuation of Risk Assets and the Selection of Risky In-
vestments in Stock Portfolios and Capital Budgets. Review of Economics
and Statistics 47, 13-37.
[15] Neumann, T. (2003) Time-Varying Coefficient Models: A Comparison of
Alternative Estimation Strategies. Allgemeines Statistisches Archiv 87,
257-281.
14 M. Ebner and T. Neumann
[16] Schwert, G.W. and Seguin, P.J. (1990) Heterscedasticity in Stock Re-
turns. Journal of Finance 45, 1129-1155.
[17] Sharpe, C. (1964) Capital Asset Prices: A Theory of Market Equilibrium
under Conditions of Risk. The Journal of Finance 19, 3, 425-442.
[18] Shukla, R., Trzcinka, C. and Winston, K. (1995) Prediction Portfolio
Variance: Firm Specific and Macroeconomic Factors. Working Paper,
http://ssrn.com/abstract=6901.
[19] Sunder, S. (1980) Stationarity of Market Risk: Random Coefficients for
Individual Stocks. Journal of Finance 35, 4, 883-896.
Time Dependent Relative Risk Aversion
1 Introduction
Risk management has developed in the recent decades to be one of the most
fundamental issues in quantitative finance. Various models are being devel-
oped and applied by researchers as well as financial institutions. By modeling
price fluctuations of assets in a portfolio, the loss can be estimated using
statistical methods. Different measures of risk, such as standard deviation of
returns or confidence interval Value at Risk, have been suggested. These mea-
sures are based on the probability distributions of assets’ returns extracted
from the data-generating process of the asset.
However, an actual one dollar loss is not always valued in practice as a
one dollar loss. Purely statistical estimation of loss has the disadvantage of
ignoring the circumstances of the loss. Hence the notion of an investor’s utility
has been introduced. Arrow [2] and [10] were the first to introduce elementary
securities to formalize economics of uncertainty. The so-called Arrow-Debreu
securities are the starting point of all modern financial asset pricing theories.
Arrow–Debreu securities entitle their holder to a payoff of 1$ in one specific
state of the world, and 0 in all other states of the world. The price of such
a security is determined by the market, on which it is tradable, and is sub-
sequent to a supply and demand equilibrium. Moreover, these prices contain
information about investors’ preferences due to their dependence on the con-
ditional probabilities of the state of the world at maturity and due to the
imposition of market-clearing and general equilibrium conditions. The prices
reflect investors’ beliefs about the future, and the fact that they are priced
differently in different states of the world implies, that a one-dollar gain is not
always worth the same, in fact its value is exactly the price of the security.
A very simple security that demonstrates the concept of Arrow–Debreu
securities is a European option. The payoff function of a call option at maturity
T is
def
ψ(ST ) = (ST − K)+ = max(ST − K, 0) (1)
16 E. Giacomini et al.
The distinction between the actual data generating process of an asset and
the market valuations is the essence of macroeconomic dynamic equilibrium
asset-pricing models, in which market forces and investors’ beliefs are key
factors to value an asset with uncertain payoffs.
A standard dynamic exchange economy as discussed by [20], [29] and many
others, imposes that securities markets are complete, that they consist of one
consumption good and that the investors, which have no exogenous income
other than from trading the goods, seek to maximize their state-dependent
utility function. There is one risky stock St in the economy, corresponding to
the market portfolio in a total normalized supply. In addition, the economy is
endowed by a riskless bond with a continuously compounded rate of return r.
The stock price follows the stochastic process
dSt
= µdt + σdWt (3)
St
Time Dependent Relative Risk Aversion 17
St = e−(r+δ)t St
def
(4)
Since we are dealing with corrected data and in order to simplify the theoretic
explanations, we will consider δ = 0 from now on and omit the dividends from
the equations.
Taking the total differential yields
dSt = d(e−rt St )
= −re−rt St dt + e−rt dSt
= −re−rt St dt + e−rt [µSt dt + σSt dWt ]
= (µ − r)St dt + σ St dWt
= σ St dW t (5)
def
where W t = Wt + µ−r σ t can be perceived as a Brownian motion on the
probability space corresponding to the risk-neutral measure Q. The term µ−r σ
is called the market price of risk, it measures the excess return per unit of risk
borne by the investor and hence it vanishes under Q, justifying the name risk-
neutral pricing. Risk-neutral pricing can be understood as the pricing done
by a risk-neutral investor, an investor who is indifferent to risk and hence
not willing to pay the extra premium. The conditional risk-neutral density
of the stock price under Q, implied by (5) and denoted as qt (ST |St ), is the
state-price density which was described as the continuous counterpart of the
Arrow–Debreu prices from (2). The basic theorem of asset pricing states, that
absence of arbitrage implies the existence of a positive linear pricing rule ([8]),
and if the market is complete and indeed arbitrage-free, it can be shown that
the risk-neutral measure Q is unique.
In order to relate the subjective and risk-neutral densities to macroe-
conomic factors, we first need to review some of the basic concepts and
definitions of macroeconomic theory. Under some specific assumptions, it
is well known that a representative agent exists. The original representa-
tive agent model includes utility functions which are based on consumption
(see, for example, [21]). However, introducing labor income or intermediate
consumption do not affect the results significantly and hence, without loss of
generality, we review the concept of marginal rate of substitution with the help
of a simple consumption based asset pricing model. The fundamental desire
18 E. Giacomini et al.
where st denotes the state of the world at time t, ct denotes the consumption
at time t, cst+1 denotes consumption at the unknown state of the world at
time t + 1, pt (st+1 |st ) is the probability of the state of the world at time
t + 1 conditioned on information at time t, u(c) is the one-period utility of
consumption and β is a subjective discount factor. We further assume that
an agent can buy or sell as much as he wants from an asset with payoff ψst+1
at price Pt . If Yt is the agent’s wealth (endowment) at t and ξ is the amount
of asset he chooses to buy, then the optimization problem is
subjected to
ct = Yt − Pt · ξ
cst+1 = Yst+1 + ψst+1 · ξ
The first constraint is the budget constraint at time t, the agent’s endowment
at time t is divided between his consumption and the amount of asset he
chooses to buy. The budget constraint at time t + 1 sustains the Walrasian
property, i.e. the agent consumes all of his endowment and asset’s payoff at
the last period. The first order condition of this problem yields
u (cst+1 )
Pt = Et β ψst+1 (7)
u (ct )
u (c )
def s
We define MRSt = β Et u (ct+1 t)
as the Marginal Rate of Substitution at t,
meaning the rate at which the investor is willing to substitute consumption
at t + 1 for consumption at t. If consumption at t + 1 depends on the state of
the world (which is the case discussed here), the MRS is also referred to as a
stochastic discount factor.
Famous works like [20] or [24] address the asset pricing models in a more
general manner. The utility function depends on the agent’s wealth Yt at
time t and the payoff function depends on the underlying asset St . According
to [24], in equilibrium, the optimal solution is to invest in the risky stock at
every t < T and then consume the final value of the stock, i.e. Yt = St for
∀t < T and YT = ST = cT . This is a multi-period generalization of the model
introduced before (6), where period T corresponds to t + 1 in the previous
def
section. Defining time to maturity as τ = T − t, the date t price of an
Time Dependent Relative Risk Aversion 19
where qt (ST |St ) is the State Price Density and the expectation EQ t [ψ(ST )]
is taken with respect to the risk-neutral probability measure Q and not the
subjective probability measure, thus reflecting an objective belief about the
future states of the world.
Combining (8) and (9) we can define the pricing kernel Mt (ST ), which
relates to the state price density qt (ST |St ), the subjective probability and the
utility function as
and therefore MRSt = e−rτ Et [Mt (ST )]. Substituting out the qt (ST |St ) in (9)
and using (10) yields the Lucas asset pricing equation:
Pt = e−rτ EQ [ψ(ST )]
t∞
= e−rτ Mt (ST ) · ψ(ST )pt (ST |St )dST
0
= e−rτ Et [Mt (ST ) · ψ(ST )] (11)
The dependence of the pricing kernel on the investor’s utility function has
urged researchers to try and estimate distributions based on various utility
functions. Arrow [3] and [26] showed a connection between the pricing kernel
and the representative agent’s measure of risk aversion. The agent’s risk aver-
sion is a measure of the curvature of the agent’s utility function. The higher
the agent’s risk aversion is, the more curved his utility function becomes. If the
agent were risk-neutral, the utility function would be linear. In order to keep
a fixed scale in measuring the risk aversion, the curvature is multiplied by the
20 E. Giacomini et al.
level of the asset (the argument of the utility function), i.e. the representative
agent’s coefficient of Relative Risk Aversion (RRA) is defined as
U (ST )
Mt (ST ) = λ
U (St )
U (ST )
⇒ Mt (ST ) = λ (13)
U (St )
Substituting out the first and second derivatives of the utility function in (12)
using (13) yields
We now have a method of deriving the investor’s pricing kernel and his risk
aversion just by knowing, or being able to estimate, the subjective and the
risk-neutral densities. As an example, we consider the popular power utility
function 1 1−γ
c for 0 < γ = 1
u(ct ) = 1−nγ t (16)
log(ct ) for γ = 1
Rubinstein [29] showed, that for such a utility function, aggregate consump-
tion is proportional to aggregate wealth, corresponding to the utility of wealth
or asset prices discussed above. It can be seen, that as γ → 0 the utility is
reduced to a linear function. The logarithmic utility function when γ = 1 is
obtained by applying the L’Hospital rule.
The marginal rate of substitution of an investor with a power utility func-
tion is −γ
u (cT ) cT
MRSt = β Et = β E t (17)
u (ct ) ct
Time Dependent Relative Risk Aversion 21
Sustaining the assumptions of the [4] model and plugging (19) into (21)
yields
[ln(ST /St )−(r−0.5σ 2 )τ ]2
1
q BS (ST |St ) = √ · e− 2σ 2 τ (22)
ST 2πσ 2 τ
meaning that the underlying asset price follows the stochastic process
dSt
= r · dt + σ · dWt (23)
St
i.e. the stock price in a [4] world follows a geometric Brownian motion un-
der both probability measures, only with different drifts. Since the subjective
probability under the [4] is also log-normal but with drift µ, plugging the SPD
from (22) and the log-normal subjective density into (10) yields a closed-form
solution for the investor’s pricing kernel
− µ−r
ST σ2 (µ−r)(µ+r−σ 2 )τ
MtBS (ST ) = ·e 2σ 2 (24)
St
The only non constant term in this expression is SSTt , which corresponds to
consumption growth in a pure exchange economy. Since the pricing kernel in
(24) is also the ratio of the marginal utility functions (10), the investor’s utility
function can be derived by solving the differential equation. If we consider the
following constants
µ−r
γ=
σ2
(µ−r)(µ+r−σ 2 )τ
λ=e 2σ 2 (25)
The database used for this work consists of intraday DAX and options data
which has undergone a thorough preparation scheme. The data was obtained
from the MD*Base, maintained at the Center for Applied Statistics and Eco-
nomics (CASE) at the Humboldt-University of Berlin. The first trading day in
the database is January 4th 1999 and the last one is April 30th 2002, i.e. more
than three years of intraday data and 2,921,181 observations. The options
data contains tick statistics on the DAX index options and is provided by
the German–Swiss Futures Exchange EUREX. Each single contract is doc-
umented and contains the future value of the DAX (corresponding to the
24 E. Giacomini et al.
maturity and corrected for dividends according to (4)), the strike, the interest
rate (linearly interpolated to approximate a “riskless” interest rate for the
specific option’s time to maturity), the maturity of the contract, the closing
price, the type of the option, calculated future moneyness, calculated Black
and Scholes implied volatility, the exact time of the trade (in hundredths of
seconds after midnight), the number of contacts and the date.
In order to exclude outliers at the boundaries, only observations with a
maturity of more than one day, implied volatility of less than 0.7 and future
moneyness between 0.74 and 1.22 are considered, remaining with 2,719,640
observations on 843 trading days. For every single trading day starting April
1999, the static model described in the following section is run and the results
are collected. The daily estimation begins 3 months after the first trading day
in the database because part of the estimation process is conducted on his-
torical data, and the history “window” is chosen to be 3 months, as explained
in the next section.
P for Tau = 30,60,90,120 days on 19990416 P for Tau = 30,60,90,120 days on 20000403
4
3
3
Density
Density
2
2
1
1
0
0
0.7 0.8 0.9 1 1.1 1.2 1.3 0.7 0.8 0.9 1 1.1 1.2 1.3
Future Moneyness Future Moneyness
P for Tau = 30,60,90,120 days on 20010911 P for Tau = 30,60,90,120 days on 20020228
5
4
4
3
3
Density
Density
2
2
1
1
0
0.7 0.8 0.9 1 1.1 1.2 1.3 0.7 0.8 0.9 1 1.1 1.2 1.3
Future Moneyness Future Moneyness
Fig. 1. Subjective density for different maturities (30, 60, 90, 120 days) on different
trading days
26 E. Giacomini et al.
It can be seen in Fig. 1 that the distribution resembles a log normal distri-
bution, which is more spread the longer the maturity is. A well known feature
of financial data is that equity index return volatility is stochastic, mean-
reverting and responds asymmetrically to positive and negative returns, due
to the leverage effect. Therefore, this GARCH (1,1) model estimation, which
experiences a slight positive skewness, is an adequate measure for the index
returns, and it resembles the nonparametric subjective densities, which were
estimated by [1] and [6].
def
where κ = erτKSt is future moneyness, τ is time to maturity and σi is the im-
plied volatility. The kernel functions kκ and kτ together with the appropriate
bandwidths hκ and hτ are chosen such that the asymptotic properties of the
second derivative of the call price are optimized. The kernel function measures
the drop of likelihood, that the true density function goes through a certain
point, when it does not coincide with a certain observation. The price of the
call is then calculated using the [4] formula but with the estimated volatility,
and the SPD is estimated using (21).
A major advantage of such a method comparing to nonparametric ones is
that only the volatility needs to be estimated using a nonparametric regres-
sion. The other variables are parametric, thus reducing the size of the problem
significantly. Other important qualities of kernel estimators are a well devel-
oped and tractable statistical inference and the fact that kernel estimators
take advantage of past data, as well as future data, when estimating the cur-
rent distribution. The problem of kernel based SPDs is that they could, for
certain dates, yield a poor fit to the cross-section of option prices, although
for other dates the fit could be quite good.
The state-price density in this work is estimated using a local polynomial
regression as proposed by [27] and described thoroughly in [17]. The choice
of Nadaraya–Watson type smoothers, used by [1], is inferior to local poly-
nomial kernel smoothing. More accurately, the Nadaraya-Watson estimator
is actually a local polynomial kernel smoother of degree 0. If we use higher
order polynomial smoothing methods, we can obtain better estimates of the
functions. Local polynomial kernel smoothing also provides a convenient and
Time Dependent Relative Risk Aversion 27
SPD for Tau = 30,60,90,120 days on 19990416 SPD for Tau = 30,60,90,120 days on 20000403
5
5
4
4
Density
Density
3
3
2
2
1
1
0
6
4
4
Density
Density
3
2
2
1
0
Fig. 2. State-Price density for different maturities (30, 60, 90, 120 days) on different
trading days
The estimated risk neutral densities for the same dates and the same
maturities as in Fig. 1 are depicted in Fig. 2. The SPD is estimated on a
future moneyness scale, thus reducing the number of parameters that need to
be estimated.
One of the trading days plotted in Fig. 2 is September 11th 2001. It is
interesting to see that the options data on this trading day reflects some
increased investors’ beliefs, that the market will go down in the long run.
Similar behavior is found in the trading days following that particular day as
well as in other days of crisis. The highly volatile SPD for negative returns,
which could be explained, for example, by the leverage effect or the correlation
effect, could reflect a dynamic demand for insurance against a market crash.
This phenomenon is more apparent in days of crisis and was reported by [18]
as well.
Time Dependent Relative Risk Aversion 29
At this stage, we have the estimated subjective and state-price densities for the
same maturities and spread over the same grid. The next step is to calculate
the daily estimates for the pricing kernel and risk aversion.
The pricing kernel is calculated using (10), where the estimated subjective
density and the estimated SPD replace p(ST |St ) and q(ST |St ) in the equation,
respectively. Since the grid is a moneyness grid, and the p and q are estimated
on the moneyness grid, the estimated pricing kernel is actually Mt (κT ). The
coefficient of relative risk aversion is then computed by numerically estimating
the derivative of the estimated pricing kernel with respect to the moneyness
and then according to (14).
The estimated pricing kernels depicted in Fig. 3 for different trading days
and different maturities bear similar characteristics to those reported by [1],
EPK for Tau = 30,60,90,120 days on 19990416 EPK for Tau = 30,60,90,120 days on 20000403
1.5
2
1.5
EPK
EPK
1
1
0.5
0.5
EPK for Tau = 30,60,90,120 days on 20010911 EPK for Tau = 30,60,90,120 days on 20020228
10
6
4
EPK
EPK
5
2
0
Fig. 3. Estimated Pricing Kernel for different maturities (30, 60, 90, 120 days) on
different trading days
30 E. Giacomini et al.
[18], [28] and others, who conducted a similar process on the S&P500 index.
The pricing kernel is not a monotonically decreasing function, as suggested
in classic macroeconomic theory. It is more volatile and steeply upward slop-
ing for large negative return states, and moderately downward sloping for
large positive return states. Moreover, the pricing kernel contains a region of
increasing marginal utility at the money (around κ = 1), implying a nega-
tive risk aversion. This feature can clearly be seen in Fig. 4 which depicts the
coefficient of relative risk aversion and shows clearly, that the minimal risk
aversion is obtained around the ATM region and the relative risk aversion
is negative. The negative risk aversion around the ATM region implies the
possible existence of risk seeking investors, whose utility functions are locally
convex.
Jackwerth [18] named this phenomenon the pricing kernel puzzle and sug-
gested some possible explanations to it. One possible explanation is that,
RRA for Tau = 30,60,90,120 days on 19990416 RRA for Tau = 30,60,90,120 days on 20000403
40
20
30
15
20
RRA
RRA
10
10
5
0
0
−5
60
100
RRA
RRA
40
50
20
0
0
Fig. 4. Estimated relative risk aversion for different maturities (30, 60, 90, 120
days) on different trading days
Time Dependent Relative Risk Aversion 31
a broad index (DAX in this work, S&P500 in his work) might not be a good
proxy for the market portfolio and as such, the results are significantly differ-
ent than those implied in the standard macroeconomic theory. In addition to
the poor fit of the index, the assumptions for the existence of a representative
agent might not hold, meaning that markets are not complete or the utility
function is not strictly state-independent or time-separable.
Another possibility is that historically realized returns are not reliable
indicators for subjective probabilities, or that the subjective distribution is
not well approximated by the actual one. This deviation stems from the fact
that investors first observe historical returns without considering crash possi-
bilities, and only afterwards incorporate crash possibilities, which make their
subjective distribution look quite different than the one estimated here. The
historical estimation or the log-normal distribution assumptions ignore the
well known volatility clustering of financial data.
Looking from another interesting point of view, investors might make mis-
takes in deriving their own subjective distributions from the actual objective
one, thus leading to mispricing of options. Jackwerth [18] claims, that mispric-
ing of options in the market is the most plausible explanation to the negative
risk aversion and increasing marginal utility function.
This work does not aim, however, at finding a solution to the pricing
kernel puzzle. The implicit assumption in this work is that some frictions
in the market lead to the contradicting of standard macroeconomic theory,
resulting in a region of increasing marginal utility. In the following section, a
dynamic analysis of the pricing kernel and relative risk aversion is conducted
along the three-year time frame.
In order to explore the characteristics of the pricing kernel and the relative risk
aversion, their first four moments at any trading day have to be computed,
i.e. the mean (µt ), standard deviation (σt ), skewness (Skewt ) and kurtosis
(Kurtt ) of the functions across the moneyness grid. In addition, the daily
values of the estimated functions at the money (ATM) are calculated and
analyzed. Including this additional moment could prove essential as it was
32 E. Giacomini et al.
shown before that the functions behave quite differently at the money than in
other regions. Each of the estimates (pricing kernel and relative risk aversion)
is a function of moneyness and time to maturity, which was chosen to be a
vector of four predetermined maturities, and as in the previous section we
concentrate on τ = (30, 60, 90, 120) days.
Figures 5, 6, 7 and 8 depict the time-series of the ATM values and mean
values of the pricing kernel and the relative risk aversion, each estimated for
four different maturities on 589 trading days between April 1999 and April
2002. The trading days, on which the GARCH model does not fit the data, or
the local polynomial estimation experiences some negative volatilities, were
dropped. Time-series of the daily standard deviation, skewness and kurtosis,
as well as the differences time-series, were collected but not included in this
paper.
The plots show that the pricing kernel at the money (Fig. 5) behaves simi-
larly across different maturities and bears similar characteristics to its general
mean (Fig. 6). This result implies, that characterizing the pricing kernel using
the four first moments of its distribution is adequate. Contrary to the pricing
kernel, the relative risk aversion at the money (Fig. 7) looks quite different
than its general mean (Fig. 8). The ATM relative risk aversion is mostly neg-
ative, as detected already in the daily estimated relative risk aversion. The
mean relative risk aversion, however, is mostly positive. Another feature of the
relative risk aversion is that it becomes less volatile the longer the maturity is,
2.25 2.25
1.50 1.50
0.75 0.75
3.00 3.00
2.00 2.00
1.00 1.00
Fig. 5. ATM Pricing Kernel for different maturities (30, 60, 90, 120 days)
Time Dependent Relative Risk Aversion 33
3.00 2.25
2.00 1.50
1.00 0.75
2.25 2.25
1.50 1.50
0.75 0.75
Fig. 6. Mean of Pricing Kernel for different maturities (30, 60, 90, 120 days)
RRA (moneyness = 1), maturity = 30 days) RRA (moneyness = 1), maturity = 60 days)
14.00 8.00
8.25 3.50
2.50 −1.00
−3.25 −5.50
−9.00 −10.00
RRA (moneyness = 1), maturity = 90 days) RRA (moneyness = 1), maturity = 120 days)
40.00 107.00
27.75 77.25
15.50 47.50
3.25 17.75
−9.00 −12.00
Fig. 7. ATM Relative Risk Aversion for different maturities (30, 60, 90, 120 days)
34 E. Giacomini et al.
29.25 36.00
15.50 18.00
1.75
−12.00 −18.00
19.50 25.25
6.00 10.50
−7.50 −4.25
−21.00 −19.00
Fig. 8. Mean of Relative Risk Aversion for different maturities (30, 60, 90, 120
days)
implying the existence of more nervous investors for assets with short matu-
rities. The main conclusion we can draw from the relative risk aversion plots
is that the four first moments of the distribution do not necessarily represent
all the features of the relative risk aversion correctly, and the collection of
the extra details regarding the ATM behavior is justified, as it will be shown
by the principal component analysis. After describing the characteristics of
the different time-series, and before we concentrate on specific time-series for
further analysis, it is essential to determine which of the time-series are sta-
tionary. The test chosen to check for stationarity is the KPSS test, originally
suggested by [19].
Conducting stationarity tests for the various functions has shown, that the
moments of the time-series themselves are in most of the cases not stationary,
and the logarithmic differences of the moments are not always defined, due to
the existence of negative values. Contrary to that, the absolute differences of
all moments and across all maturities were found to be stationary. Therefore,
we concentrate from now on only on the absolute differences of the moments.
It can clearly be seen, that the dominant factors in the first principal com-
ponent are the changes in mean and standard deviation, whereas the dominant
factors in the second principal component are the changes in skewness and
standard deviation. The equations do not change much when other maturi-
ties are considered. As for the moments of the relative risk aversion, the first
principal component is dominated solely by the changes in standard deviation
and the second principal component is mainly dominated by the change in
relative risk aversion at the money.
We conclude therefore, that the variation of the pricing kernel and relative
risk aversion differences can be explained by two factors. The first factor of
pricing kernel differences explains 60% of the variability and can be perceived
as a central mass movement factor, consisting of the changes in expectation
and standard deviation. The second factor explains additional 20% of the
variability and can be perceived as a change of tendency factor, consisting
of changes in skewness and standard deviation. The principal components of
the relative risk aversion are a little different. The first one explains approx-
imately 35% of the variability and can be perceived as a dispersion change
factor, dominated by the change in standard deviation. The contribution of
the second principal component to the total variability is 35% as well and it
is dominated by the change in relative risk aversion of the investors at the
money. The mean of relative risk aversion differences seems to play no role in
examining the variability of the relative risk aversion.
The correlation between the ith moment and the j th principal component
is calculated as
lj
rXi ,Yj = gij (39)
sXi Xi
of skewness are not consistent across maturities, implying a bad fit. Since the
first principal component of the pricing kernel differences could explain ap-
proximately 60% of the variability, whereas the second factor can explain only
20%, the inconsistent behavior could be justified by the poor contribution of
the second principal component to the total variability.
The correlations of the first and second principal components of the rela-
tive risk aversion differences with their dominant factors (Table 2) are found
to be consistent across maturities. The first principal component is positively
correlated with its most dominant moment, the changes in the relative risk
aversion standard deviation. This correlation means essentially, that the less
homoscedastic the relative risk aversion is, i.e. the larger the changes in stan-
dard deviation are, the larger the first principal component of the relative risk
aversion differences becomes. The second principal component of the relative
risk aversion differences is positively correlated with its most dominant mo-
ment, the behavior at the money. The more volatile the relative risk aversion
at the money is, the higher the second principal component is. Both principal
38 E. Giacomini et al.
components of the relative risk aversion differences contribute more than 30%
of the variability and imply a good fit of the principal components to the
data. After constructing principal components, which explain the variability
of the time-series, it is essential to check the autocorrelation and the partial
autocorrelation functions of the time-dependent principal components. This
is illustrated in Fig. 9 for the pricing kernel differences. The same functions
for the principal components of the relative risk aversion differences have
similar characteristics and hence not reported here. Since the principal com-
ponents have similar autocorrelation and partial autocorrelation functions for
all different maturities, a maturity of 60 days was arbitrarily chosen to be pre-
sented. It can be seen, that the autocorrelation function drops abruptly after
0
0.5
-0.2
pacf
acf
0
-0.4
-0.5
0 5 10 15 20 25 30 0 5 10 15 20 25 30
lag lag
ACF of Second PC of PK PACF of Second PC of PK
1
0
0.5
-0.2
pacf
acf
0
-0.4
-0.5
0 5 10 15 20 25 30 0 5 10 15 20 25 30
lag lag
The last test conducted in this work is to detect a possible relation between
the principal components and easily observed data, such as changes in the
DAX level and in implied volatility at the money. It is well known, that the
simplest relation between an explanatory variable and a response variable can
be described and examined using a simple linear regression model
y = Xβ + (40)
for all t ∈ {1, . . . , n} with ut ∼ Nn (0, σu2 In ) as i.i.d. white noise and |ρ| < 1 for
stability. We could choose autoregressive processes of higher order, but since
most principal components were found to have an autocorrelated error term
of order 1, we concentrate here on AR(1) processes.
Iterating (43) from time 0 onwards yields
n ∞
t = lim (ρn+1 t−n−1 + ρs ut−s ) = ρs ut−s (44)
n→∞
s=0 s=0
40 E. Giacomini et al.
and hence E[t ] = 0 and the covariance matrix of the error term is
⎛ ⎞
1 ρ ρ2 . . . ρn−1
⎜ ρ 1 ρ . . . ρn−2 ⎟
2 σu2 ⎜⎜ ρ2 ρ
⎟
1 . . . ρn−3 ⎟
Cov() = σu Ω = ⎜ ⎟ (45)
1 − ρ2 ⎜ .. .. .. . . .. ⎟
⎝ . . . . . ⎠
ρn−1 ρn−2 ρ n−3
... 1
However, in a real application like the model discussed in this work, the
error-covariance matrix is not known and must be estimated from the data
along with the regression coefficients β. If the generating process is station-
ary, which is the case in the model discussed here, a commonly used algorithm
for estimating these errors is normally referred to as the [25] procedure. This
algorithm begins with running a standard OLS regression and examining the
residuals. The errors vector of the OLS regression is obtained simply by plug-
ging β in (40). Considering the residuals’ first order autocorrelations from
the preliminary OLS regression can suggest a reasonable form for the error-
generating process. These first order autocorrelations can be estimated as
n
t=2 t t−1
ρ = n 2 (46)
t=1 t
Replacing the ρ’s in (45) with the ρ ’s from (46) results in the estimated matrix
The best linear unbiased estimator in that case would be the estimated
Ω.
generalized least squares estimator
βGLS = (X Ω
−1 X)−1 X Ω
−1 y (47)
The [25] algorithm may seem to be a simple model, but it involves a compu-
tationally challenging estimation of Ω. Therefore, an alternative algorithm,
suggested by [31] is presented here. We define the following matrix as
⎛ ⎞
1 − ρ 2 0 0 . . . 0 0 0
⎜ − ρ 1 0 ... 0 0 0⎟
⎜ ⎟
⎜ 0 − ρ 1 ... 0 0 0⎟
⎜ ⎟
Ψ = ⎜ .. .. .. . . .. .. .. ⎟ (48)
⎜ . . . . . . .⎟
⎜ ⎟
⎝ 0 0 0 . . . − ρ 1 0⎠
0 0 0 . . . 0 − ρ1
It can be shown, that this matrix, multiplied by its transpose and the matrix
Ω (which is defined by (45)) is proportional to the unit matrix
1
Ψ Ψ Ω = In
1 − ρ 2
and hence the matrix Ψ has the following property
Ψ Ψ = (1 − ρ 2 )Ω
−1 (49)
Time Dependent Relative Risk Aversion 41
βGLS = (X Ψ ΨX)−1 X Ψ Ψy = [(ΨX) (ΨX)]−1 [(ΨX)] (Ψy) (50)
p
1 − ρ 2 y1 = 1 − ρ 2 βj x1j + 1 − ρ 2 1 (52)
j=0
As stated in the beginning of the current section, the changes in the DAX
level (St ) and the changes of ATM implied volatility (IVtAT M ) were chosen
to be tested as explanatory variables (X), whereas the first two principal
components of the pricing kernel and relative risk aversion differences for
different maturities were the dependent variables for the different models (y).
Since the dependency on the explanatory variable does not have to be linear,
different functions of the explanatory variables were tested. For each of the
explanatory variables the differences, the squared differences, the logarithmic
differences and the squared logarithmic differences were tested. The examined
models consisted of all possible combinations between the functions stated
above, as well as checking for interactions in each of the proposed models.
Since no interaction was ever found to be significant, they were dropped from
the model. The criterion for choosing the best model was a maximal value of
the F-statistic.
Table 3 describes the best fitted models for each of the principal com-
ponents (based on (51)). For this analysis, we consider a confidence level of
95%, i.e. any regression or regression coefficient yielding a Pvalue > 5% is re-
garded as non significant. The Pvalues for the regressions’ coefficients appear
in brackets.
The first principal component of the pricing kernel differences, which was
described before as a central mass movement factor, dominated by the changes
in the mean pricing kernel and the pricing kernel’s standard deviation, is
found to depend significantly on the logarithmic differences of ATM implied
volatility. This regression is only significant for short term maturities, and the
impact of the explanatory variables is positive and log-linear. The impact of
the DAX log return is not significant for a short term maturity, meaning the
first principal component of the pricing kernel differences is mainly influenced
by the logarithmic changes in the implied volatility at the money. Therefore,
42 E. Giacomini et al.
PC Maturity ρ 1
β xt,1 2
β xt,2 F
St IVtAT M
1 30 -0.43 -1.80 log St−1
1.76 log AT M
IVt−1
18.96
(0.289) (0.000) (0.000)
St IVtAT M
60 -0.47 2.71 log St−1
0.98 log AT M
IVt−1
10.78
(0.005) (0.001) (0.000)
90 Not Significant
St IVtAT M
2 30 -0.47 30.21 log St−1
12.77 log AT M
IVt−1
20.72
(0.001) (0.000) (0.000)
60 Not Significant
90 Not Significant
PC Maturity ρ 1
β xt,1 2
β xt,2 F
2 30 Not Significant
we can deduce the following: The larger the changes in ATM implied volatility
are and the higher the DAX log returns are (only for maturities of 60 days),
the more volatile the pricing kernel becomes, with bigger daily changes in its
mean and standard deviation.
We can not find a significant relationship between the second principal
component of the pricing kernel differences and the explanatory variables
(other than for very short maturities), a result that supports the second prin-
cipal component’s smaller contribution to the variability of pricing kernel
differences. The pricing kernel differences have one dominant factor which
explains approximately 60% of their variance and depends mainly on the log-
arithmic changes of the ATM implied volatility. The regression coefficients are
positive, as are the correlations of the first principal component with ∆µt (P K)
and ∆σt (P K) for the respective maturities.
The results regarding the principal components of the relative risk aver-
sion differences are quite different. These principal components are related to
the absolute changes in the DAX level and in ATM implied volatility. The
dependence is not log-linear, but strictly linear.
According to Table 2 in the previous section, the correlations of the first
principal components of the relative risk aversion differences with their dom-
inant moments are positive. The first principal component is a dispersion
factor, dominated by the change in the relative risk aversion standard devia-
tion. According to the regression, large changes in the DAX level and the ATM
implied volatility yield a larger principal component, which is associated with
a larger change in risk aversion standard deviation. This result implies the
existence of more uncertain investors with a more heteroscedastic risk aver-
sion, when the DAX level and ATM implied volatility are more time-varying.
This relation could be explained by the dispersion of information sets among
investors. Veldkamp [32] examines the impact of information markets on as-
sets prices. She basically claims, that information markets, not assets mar-
kets, are the source of frenzies and herds in assets prices. However, the price
fluctuations on the market affect these information sets and determine the
information prices, which are incorporated in the investors’ subjective beliefs.
More volatile markets lead necessarily to a higher risk and to less informa-
tion, which increases the demand for information in a competitive market.
Hence, more volatile markets cause more information to be provided at a
lower price. When less information is involved, individual agents are will-
ing to pay for information, and the information sets of the individual agents
become more dispersed. More dispersed information sets could increase het-
eroscedasticity of the aggregate relative risk aversion as a function of assets’
returns.
The results regarding the second principal component of the relative risk
aversion differences are slightly different. The second principal components
are positively correlated to the change of relative risk aversion at the money.
Nevertheless, the linear regression is not significant for a very short term
maturity of 30 days. For long term maturities the coefficients of the regression
44 E. Giacomini et al.
are positive, whereas for medium term maturities, they are negative. That
could be interpreted as follows: When the changes in DAX level and ATM
implied volatility are larger, the relative risk aversion at the money is more
volatile for long term maturities, but is less volatile for the medium term
maturities.
From this section we can conclude, that the principal components model
fits the relative risk aversion differences better than it fits the pricing kernel
differences. We were able to fit an autocorrelated regression model to the first
principal component of pricing kernel differences for short and medium term
maturities, and to both principal components of relative risk aversion differ-
ences. The autocorrelation is indeed found to be quite large (approximately
-0.5) for all of the above models, implying the existence of an autocorrelated
error as detected already.
5 Final Statements
This work focused on estimating the subjective density and the state-price
density of the stochastic process associated with the DAX. Based on the work
of [30], a good estimation of those two measures is sufficient for deriving the
investors’ preferences. However, this work did not include a direct approxima-
tion of the utility function based on empirical data, but rather an estimation
of the pricing kernel and the relative risk aversion as functions of the return
states. The utility function could be approximated numerically by solving the
differential equations discussed in Sect. 2, after the pricing kernel and relative
risk aversion function have been estimated. Nevertheless, this work aimed at
examining the dynamics of these two measures, characterizing the investors’
behavior, rather than deriving their implied utility function.
The daily estimated pricing kernel and relative risk aversion were found to
have similar characteristics to those reported by [18] and [1]. The pricing ker-
nel was shown not to be a strictly decreasing function as suggested by classical
macroeconomic theory, and the relative risk aversion experienced some nega-
tive values at the money. These findings were apparent throughout the three
year long database, implying existence of risk seeking investors with a locally
convex utility function, possibly due to some frictions in the representative
agent’s model.
The variability of the stationary daily changes in pricing kernel and rel-
ative risk aversion was found to be well explained by two factors. Since the
factors experienced some evident autocorrelation, the principal components
were tested as the response variable in a GLS regression model, which re-
gressed each of the principal components on the daily changes in the DAX
and in ATM implied volatility.
We found that large changes in ATM implied volatility lead to a more
volatile and time-varying pricing kernel. The absence of a significant fitted
regression model for the second principal component of the pricing kernel
Time Dependent Relative Risk Aversion 45
References
[1] Aı̈t Sahalia, Y. and Lo, A. W. [2000], ‘Nonparametric risk management
and implied risk aversion’, Journal of Econometrics 94, 9–51.
[2] Arrow, K. J. [1964], ‘The role of securities in the optimal allocation of
risk bearing’, Review of Economic Studies 31, 91–96.
[3] Arrow, K. J. [1965], ‘Aspects of the theory of risk-bearing’, Yrjö
Hahnsson Foundation, Helsinki.
[4] Black, F. and Scholes, M. [1973], ‘The pricing of options and corporate
liabililties’, Journal of Political Economy 81, 637–654.
[5] Breeden, D. and Litzenberger, R. [1978], ‘Prices of state-contingent
claims implicit in option prices’, Journal of Business 51, 621–651.
[6] Brown, D. P. and Jackwerth, J. C. [2004], ‘The pricing kernel puzzle:
Reconciling index option data and economic theory’, Working Paper,
University of Konstanz / University of Wisconsin.
[7] Campbell, J. and Cochrane, J. H. [1999], ‘A consumption-based explana-
tion of aggregate stock market behavior’, Journal of Political Economy
107.
[8] Cochrane, J. H. [2001], ‘Asset pricing’, Princeton University Press,
Princeton.
[9] Constantinides, G. [1982], ‘Intertemporal asset pricing with heteroge-
neous consumers and without demand aggregation’, Journal of Business
55, 253–268.
[10] Debreu, G. [1959], ‘The theory of value’, Wiley, New York.
[11] Derman, E. and Kani, I. [1994], ‘The volatility smile and its implied tree’,
Quantitative strategies research notes, Goldman Sachs.
[12] Fengler, M. R. [2005], ‘Semiparametric modelling of implied volatility’,
Springer, Berlin.
[13] Franke, J., Härdle, W. and Hafner, C. [2004], ‘Statistics of financial mar-
kets’, Springer, Heidelberg.
[14] Härdle, W. [1990], ‘Applied nonparametric regression’, Cambridge Uni-
versity Press, Cambridge.
[15] Härdle, W. and Hlávka, Z. [2005], ‘Dynamics of state price densities’, SFB
649 Discussion paper 2005-021, CASE, Humboldt University, Berlin.
[16] Härdle, W. and Zheng, J. [2002], ‘How precise are distributions predicted
by implied binomial trees?’, in: W. Härdle, T. Kleinow and G. Stahl
(eds.), Applied Quantitative Finance, Springer, Berlin, Ch. 7.
46 E. Giacomini et al.
1 Introduction
It is clear that these findings have important implications for financial de-
cisions. In general, if the return distribution is not Gaussian, standard mean-
variance analysis may not be reconcilable with expected utility theory, as
properties such as skewness and kurtosis will also affect investors’ decisions.
Moreover, as stressed in [11], the phenomenon of asymmetric dependencies,
with higher correlations in bear markets, is also of considerable relevance for
investment analysis, because it is in times of adverse market conditions that
the benefits from diversification are most urgently needed. However, models
not taking into account the state-dependent correlation structure will tend
to overestimate the benefits from diversification in bear markets, and, conse-
quently, they will underestimate the risk during such periods.
We attempt to tackle all these issues by assuming that returns are gen-
erated by multivariate normal mixture distributions. It is well-known that
normal mixture densities can capture the skewness and kurtosis observed in
empirical return distributions rather well. Moreover, regime-dependent corre-
lation structures are incorporated into the model in a natural and intuitively
appealing manner. By adopting the Markov-switching approach popularized
by Hamilton [24], we also allow for predictability of market regimes, which is
of great importance for portfolio selection.
It is clear, however, that the curse of dimensionality referred to at the
beginning of this section is even more burdensome in the mixture than in the
traditional framework, because we have as many covariance matrices to esti-
mate as we have mixture components. To effectively overcome this drawback,
we introduce a parsimonious parametrization of the regime-specific correlation
matrices by generalizing the common correlation model (CCM) of Elton and
Gruber [15] to the mixture of common correlation models (MCCM). Despite
its simplicity, the CCM has been shown in a number of previous studies to
deliver highly competitive correlation forecasts.
This article is organized as follows. Section 2 motivates the use of normal
mixture distributions to model asset returns. Section 3 introduces the MCCM
and develops parameter estimation via the EM algorithm. Section 4 presents
an application to international stock market returns, and Sect. 5 concludes
and identifies issues for further research.
k
f (rt ) = πjt φ(rt ; µj , Σj ), (3)
j=1
where φ(·; µj , Σj ) denotes the normal density with mean µj and covariance
matrix Σj , as given in (1). In (3), the πjt ’s are the (conditional) mixing
weights, and the φ(·; µj , Σj ) are the component densities, or mixture com-
ponents, with component means µj , and component covariance matrices Σj ,
j = 1, . . . , k. The normal mixture has finite moments of all orders, which are
easily found using the properties of the normal distribution. For example, the
mean and the covariance matrix are given by
k
µ := E(rt ) = πjt µj (4)
j=1
and
k
k
Var(rt ) = πjt Σj + πjt (µj − µ)(µj − µ) , (5)
j=1 j=1
respectively.
A finite mixture of a few normal distributions, say two or three, is capa-
ble of capturing the skewness and excess kurtosis detected in empirical asset
return distributions. While a general discussion of the moments of mixture
models may be found in [41, 54], let us briefly illustrate the skewness and
kurtosis properties in the univariate case, where N = 1. Then the centered
third moment of the mixture distribution is
k
k
E(rt − µ)3 = πjt (µj − µ)3 + 3 πjt σj2 (µj − µ) (6)
j=1 j=1
k
k
= πjt (µj − µ)3 + 3 πjt πit (σj2 − σi2 )(µj − µi ).
j=1 j=1 i<j
E(rt − µ)3 = π1t π2t [(π2t − π1t )(µ1 − µ2 )3 + 3(µ1 − µ2 )(σ12 − σ22 )]. (7)
In practice, we often find a bear market regime, say Regime 2, with a relatively
low regime probability (π2t < π1t ), a high variance (σ22 > σ12 ), and a low mean
return (µ2 < µ1 ). Then (7) implies that such a combination results in negative
skewness, which is often observed in the distribution of asset returns.
With respect to kurtosis, consider the case of equal component means, i.e.,
µ1 = · · · = µk = µ. Then the coefficient of excess kurtosis over the normal
distribution, κ, is
j πjt σj − (
4 2 2
E(rt − µ)4 j πjt σj ) Var(σ 2 )
κ := 2 − 3 = 3 = 3 2 2 > 0. (8)
E (rt − µ) 2 2
( j πjt σj )2
E (σ )
where µj = w µj , and σj = w Σj w.
There is considerable evidence for the presence of market regimes with
distinctly different stochastic properties of stock returns, e.g., [56, 28, 49, 57,
5, 6, 9, 20, 21, 22, 23, 3, 7, 48]. Often researchers identify a bull market regime
with high expected returns and relatively low variances, and a bear market
regime with lower returns and higher variances. Moreover, when mixture mod-
els are applied to multivariate return series, significantly different correlation
structures are usually found, where the correlations are higher in the bear
market regime [49, 13, 5, 6, 20, 23].
1
This can be seen, for example, by just writing down the moment generating
function of the portfolio return.
Portfolio Selection with Common Correlation Mixture Models 51
To complete the formulation of the mixture model given by (1) and (2), we
need to specify the stochastic process generating the market regimes, i.e., the
evolution of the mixing weights (2). While the independent mixture model,
where the regime probabilities are constant over time, has recently attracted
some interest in the context of normal mixture GARCH models [21, 3, 58], it
seems more likely that regimes are persistent. I.e., if we are in a bull market
currently, the probability of being in a bull market in the next period will be
larger than if the current regime were a bear market. If regimes are persistent,
it is clear that this persistency should be incorporated into the model, because
this implies that the regimes are predictable, and such predictability can be
exploited for asset allocation purposes.
To allow for predictability of regimes, we adopt the Markov-switching tech-
nique which has become very popular in econometrics and empirical finance
since the seminal work of Hamilton [24]. In this model, it is assumed that the
probability of being in regime j at time t depends on the regime at time t − 1
via the time-invariant transition probabilities pij , defined by
Cov(rt2 , rt−τ
2
) = π1,∞ (1 − π1,∞ )δ τ (σ12 − σ22 + µ21 − µ22 )2 ,
where π1,∞ = (1 − p22 )/(2 − p11 − p22 ) is the unconditional probability of the
first regime (to be defined more precisely in (15)), and δ = p11 + p22 − 1 may
52 M. Haas and S. Mittnik
As the market regimes are not directly observable, we can only use observed
returns to make probability statements about the market’s past, current, or
future states. Forecasts of future market regimes are needed for optimal out-
of-sample portfolio choices, and regime inferences are also an ingredient of the
EM algorithm for parameter estimation, which will be discussed in Sect. 3.1.
[24, 35] have developed algorithms to calculate such probabilities, and we
briefly summarize their results here (see also [26]).
To this end, we introduce, for each point of time, t, a new (unobserved)
k-dimensional random vector zt = (z1t , . . . , zkt ) , t = 1, . . . , T , with elements
zjt such that
%
1 if st = j
zjt = for j = 1, . . . , k, t = 1, . . . , T. (12)
0 if st = j
That is, zjt is one or zero according as whether the return vector at time t,
rt , has been generated by the jth component of the mixture.
Moreover, let Rτ be the return history up to time τ , i.e., Rτ = (r1 , . . . , rτ ),
τ = 1, . . . , T , and let θ be the vector of model parameters. Then our
probability inference of being in state j at time t, based on the return his-
tory up to time τ , Rτ , and the parameter vector, θ, will be denoted by
Pr(zjt = 1|Rτ , θ) = zjt|τ , and zt|τ = (z1t|τ , . . . , zkt|τ ) .
Using these definitions, [26] shows that zt|t and zt+1|t can be recursively
computed via
zt|t−1 ηt
zt|t = (13)
1k (zt|t−1 ηt )
zt+1|t = P zt|t , (14)
zt+τ |t = P τ zt|t .
T
T
log L(θ|RT ) = log f (rt |Rt−1 , θ) = log[1k (zt|t−1 ηt )]. (17)
t=1 t=1
We also need the so-called smoothed regime inferences, i.e., the regime
probabilities conditional on the entire return history, RT . [35] derived a con-
venient algorithm for this purpose, which works backwards through
1 + (N − 2)ρj
r̄j := (24)
1 + (N − 2)ρj − (N − 1)ρ2j
k
k
Cov(rt ) = πj,∞ Σj + πj,∞ (µj − µ)(µj − µ) , (27)
j=1 j=1
where the πj,∞ ’s are the unconditional regime probabilities defined in (15).
This shows that the pairwise correlations may differ due to different regime
means. In applications to financial returns, however, the differences between
the regime means tend to be small, relative to those between the vari-
ances, so that the overall covariance matrix (27) is approximately common
correlation-like.
The parameters of the standard, single-regime CCM are usually estimated
by simply equating (average) sample moments and theoretical quantities. That
is, estimate the individual means and standard deviations by their sample
analogies, and equate the common correlation coefficient to the average of the
2
This follows from (22) and the Sherman–Morrison formulas for the determinant
and the inverse, respectively, stating that, for an invertible matrix A and con-
formable vectors u and v, |A + uv | = |A|(1 + v A−1 u), and (A + uv )−1 =
A−1 − A−1 uv A−1 (1 + v A−1 u)−1 [53].
56 M. Haas and S. Mittnik
N (N − 1)/2 sample pairwise correlations [15, 4, 37]. This procedure is not fea-
sible for the mixture of CCMs, because we do not observe sample counterparts
of the regime-specific means, variances and correlation coefficients. However,
estimation can be carried out in a fast and stable manner by employing the
Expectation–Maximization (EM) algorithm of [14], which will be developed
in the next two subsections.
Note that the parameter vector, θ, for this model consists of the compo-
nent means, variances, and correlation coefficients, as well as of the k(k − 1)
independent elements of the transition matrix (11), i.e., it is of dimension
k(2N + k), which increases linearly with the number of assets under study.
that is, compute the expectation of the complete-data log likelihood with
respect to the missing data, given the observed data and the current fit, θ(n−1) .
Next, on the M-step, solve the complete-data likelihood equations
∂Q(θ; θ(n−1) )
= 0, (29)
∂θ
to find an update θ(n) for θ. The E-step (28) and the M-step (29) are alter-
nated repeatedly until convergence is achieved, that is, until the difference
3
[40] provides an exhaustive presentation of the EM algorithm, its theory and
numerous examples. A discussion of EM with special emphasis on mixture models
is given in [50]. The EM algorithm for a general class of Markov-switching models
was derived in [25].
Portfolio Selection with Common Correlation Mixture Models 57
θ(n) − θ(n−1) and/or log L(θ(n) |RT ) − log L(θ(n−1) |RT ) does not exceed a
prespecified (small) value, where log L(θ|RT ) = log f (RT ; θ) is the observed-
data log likelihood (17), i.e., the function to be maximized. A well-known
result of [14] is that log L(θ(n) |RT ) ≥ log L(θ(n−1) |RT ), i.e., the likelihood is
not decreased after an EM iteration.4
It turns out that, for the Markov-MCCM, we need an extension of the
EM algorithm, namely, the ECM algorithm of [44]. The ECM algorithm is an
extension of the EM algorithm, where the maximization on the M-step is bro-
ken into a number of conditional maximization (CM) steps. This procedure is
preferable if the complete-data maximization in the M-step of the EM algo-
rithm is complicated, and the CM-steps are simple. [44] shows that the ECM
algorithm shares the monotone convergence property of the EM algorithm.
To formulate the ECM algorithm for the Markov-MCCM, we let the un-
observed zt = (z1t , . . . , zkt ) , t = 1, . . . , T , be given by the quantities defined
in (12).
Under this missing-data formulation, the complete-data density is given by
(
k (
T (
k (
k
f c (XT ; θ) = (j φ1j )zj1 (pij φjt )zjt zi,t−1 , (30)
j=1 t=2 j=1 i=1
where φjt is short-hand notation for φ(rt ; µj , Σj ), denoting the normal density
with mean µj and covariance matrix Σj . Consequently, the complete-data log
likelihood is given by
k
T
k
k
log Lc (θ|XT ) = zj1 log j + zij,t log pij
j=1 t=2 j=1 i=1
T
k
+ zjt log φjt , (31)
t=1 j=1
where zij,t = zjt zi,t−1 . As can be seen from (31), the E-step of the nth iter-
ation requires the evaluation of the conditional expectation of the zjt ’s and
zij,t ’s, given the observed data, RT , and the current fit, θ(n−1) , which can be
accomplished using (18) and (20).
The updating formulas for the initial probabilities, j , j = 1, . . . , k,5 and
the elements of the transition matrix, pij , i, j = 1, . . . , k, have been derived
in [25]. They are given by
T (n)
(n) (n) (n) t=2 zij,t|T
j = z1j|T , and pij = T (n)
, i, j = 1, . . . , k, (32)
t=2 zi,t−1|T
4
Besides its theoretical appeal, this is a useful property in practice, as it may help
to detect programming errors by monitoring the change in log likelihood after
each iteration.
5
Recall the definition of j in (16).
58 M. Haas and S. Mittnik
(n) (n)
where zjt|T and zij,t|T are given by (18) and (20), respectively, with θ fixed at
θ(n−1) . Moreover, the updating formula for the component means is given by
!−1
(n)
T
(n)
T
(n)
µj = zjt|T zjt|T rt , j = 1, . . . , k. (33)
t=1 t=1
The E-step of the ECM algorithm is given by (28); and the M-step for the
initial probabilities and the elements of the transition matrix is given by (32).
Thus, to derive the M-step for the parameters of the component densities, we
only need to consider the term in the second line of (31). To this end, let
σj = (σ1j , . . . , σN j ) , j = 1, . . . , k, (34)
(n)
with respect to µj , σj , and ρj , separately for j = 1, . . . , k. As µj is given by
(33), (35) becomes
1 1 ) *
Qj2 (µj , σj , ρj ) = − log |Dj | − log |Rj | − tr Dj−1 Rj−1 Dj−1 Sj , (36)
2 2
where !−1
T
T + ,+ ,
(n) (n) (n) (n)
Sj = zjt|T zjt|T rt − µj rt − µj . (37)
t=1 t=1
N
r̄j sjii
N
rj sjim
N
log L(σj ) = − log σij − 2 − . (38)
i=1
2 i=1 σij 2 i=1 σij σmj
m=i
∂ log L(ψj ) 1 j
= − r̄j sj ψ j − rj si ψij = 0, = 1, . . . , N. (40)
∂ψ j ψj
i=
As Rj−1 and Sj are positive definite, their Hadamard product is positive defi-
nite by the Schur product theorem (cf. [30], Theorem 5.2.1); hence, the Hessian
of (39) is negative definite for all ψj , and the unique solution of (41) is the
global maximum of the log likelihood. To solve (41), we multiply (40) by ψ j
to observe that we require
j
r̄j sj ψ 2j + rj si ψij ψ j − 1 = 0, = 1, . . . , N. (43)
i=
= 1, . . . , N. (44)
60 M. Haas and S. Mittnik
N
sjii N
sjim
log L(ρj ) = − log |Rj | − r̄j − 2rj
σ2
i=1 ij
σ σ
i=1 m<i ij mj
where
aj1 = ψj diag(Sj )ψj , aj2 = ψj [Sj − diag(Sj )] ψj . (46)
Using (23), (24) and (25), we compute
drj 1 + (N − 1)ρ2j
=− , (48)
dρj [1 + (N − 2)ρj − (N − 1)ρ2j ]2
and
d log |Rj | −N (N − 1)ρj
= . (49)
dρj 1 + (N − 2)ρj − (N − 1)ρ2j
By differentiating (45), inserting (47), (48) and (49), and a few additional
(n)
algebraic manipulations, we find that ρj is a solution of the cubic equation
4
P(ρj ) = bi ρ4−i
j = 0, (50)
i=1
where
and
N
2 j
= −N (ψij 2
sii + ψmj sjmm − 2ψij ψmj sjim )
i=1 m<i
N
) * sjii −sjim ψij
= −N ψij ψmj < 0.
i=1 m<i
−sjim sjmm ψmj
Hence, (50) has at least one root in the admissible interval (−1/(N − 1), 1). It
may happen that (50) has three roots in this interval, but we did not observe
this in practice so far.7 If this occurs, one would choose the root with the
largest corresponding value of (45).
This completes the formulation of the ECM algorithm for the Markov-
MCCM.
t−1 k
∂f (rτ |sτ = j)
+ (zjτ |t − zjτ |t−1 ),
τ =1 j=1
∂θ1
t = 1, . . . , T,
where f (rt |Rt−1 , θ) is the conditional density of rt , given the history of returns
up to time t − 1; the quantities of the form zjτ |t can be computed using (18);
and the second line of (51) is set to zero for t = 1.
7
However, numerical examples can be constructed where this occurs.
62 M. Haas and S. Mittnik
1
t−1
− (zik,τ |t − zik,τ |t−1 ),
pik τ =2
i = 1, . . . , k, j = 1, . . . , k − 1, t = 2 . . . , T,
where the quantities of the form zij,τ |t can be obtained from (20); and the
second and third line of (52) are set to zero for t = 2.
An estimate of the information matrix can be constructed from the scores’
average outer product, i.e.,
! !
1 ∂ log f (rt |Rt−1 , θ)
T
∂ log f (rt |Rt−1 , θ)
I(θ) = ,
T t=1 ∂θ ∂θ
where δjt = rt − µj ,
In this subsection, we present the estimation results for the in-sample period,
covering the period from March 1991 to February 2001, with T = 520 obser-
vations. We denote the return vector at time t by rt = (r1t , r2t , r3t ) , where
r1t , r2t , and r3t are the time-t returns of the S&P500, the FTSE, and the
DAX, respectively. A few descriptive statistics of the three series, along with
the Jarque–Bera test for normality, are summarized in Table 1. While the
S&P500 has the largest mean and the smallest variance, it also has the great-
est kurtosis, while the FTSE has the lowest mean and lowest kurtosis value.
correlation/
mean covariance matrix skewness kurtosis JB
S&P500 FTSE DAX
S&P500 0.251 3.827 0.540 0.501 −0.100 4.321 38.675
(0.352) (0.000) (0.000)
FTSE 0.144 2.241 4.494 0.599 0.045 3.203 1.067
(0.676) (0.345) (0.587)
DAX 0.235 2.541 3.289 6.715 −0.094 4.075 25.797
(0.384) (0.000) (0.000)
8
All data have been obtained from Datastream.
64 M. Haas and S. Mittnik
The skewness coefficients are not significant for any of the series. Interestingly,
while the Jarque–Bera test strongly rejects normality for the S&P500 and the
DAX series due to their excess kurtosis, this is not the case for the FTSE. The
relatively mild deviations from normality reflected in the statistics reported
in Table 1 make the data a natural candidate to be modeled by a normal
mixture distribution.
We also note that the stock return series display a considerable degree
of comovement, with pairwise correlation coefficients ranging from approxi-
mately 0.5 to 0.6.
The Markov-MCCM developed in Sect. 3 will be specified with two and
three components. Parameter estimates for these models are reported in
Table 2,9 where the models are ordered with respect to a decreasing uncondi-
tional probability, i.e., π1,∞ > π2,∞ > π3,∞ .
The parameter estimates for the models have a rather striking pattern.
For example, for the two-component model, i.e., k = 2, the first regime is
characterized by a smaller correlation coefficient and, for all series, higher
Table 2. Parameter estimates for international stock returns. k refers to the number
of mixture components in the Markov-MCCM introduced in Sect. 3. Standard errors
are given in parentheses
k=2 k=3
µ1 (0.266 , 0.191 , 0.281 ) (0.196 , 0.136 , 0.233 )
(0.086) (0.108) (0.114) (0.103) (0.141) (0.150)
σ1 (1.469 , 1.844 , 1.947 ) (1.345 , 1.961 , 2.109 )
(0.053) (0.071) (0.074) (0.063) (0.093) (0.102)
ρ1 0.383 0.397
(0.040) (0.045)
π1,∞ 0.602 0.441
µ2 (0.228 , 0.074 , 0.165 ) (0.247 , 0.094 , 0.206 )
(0.188) (0.180) (0.237) (0.176) (0.166) (0.222)
σ2 (2.553 , 2.419 , 3.326 ) (2.515 , 2.395 , 3.273 )
(0.125) (0.131) (0.162) (0.115) (0.118) (0.155)
ρ2 0.649 0.643
(0.032) (0.030)
π2,∞ 0.398 0.425
µ3 – (0.442 , 0.333 , 0.328 )
(0.207) (0.172) (0.148)
σ3 – (1.691 , 1.346 , 1.153 )
(0.139) (0.132) (0.117)
ρ3 – 0.289
(0.087)
π3,∞ 0 ⎛ 0.134 ⎞
⎛ ⎞ 0.995 0.005 0.000
0.996 0.006 ⎜ (0.004) (0.006) (0.005) ⎟
⎝ (0.003) (0.007) ⎠ ⎜ 0.000 0.995 0.015 ⎟
P ⎜ (0.003) (0.006) (0.019) ⎟
0.004 0.994 ⎝ ⎠
(0.003) (0.007) 0.005 0.000 0.985
(0.008) (0.004) (0.018)
9
The ECM algorithm was terminated, i.e., convergence was considered to be
reached, when log L(θ
√
(n)
|RT ) − log L(θ(n−1) |RT ) ≤ 10−8 and θ(n) − θ(n−1) ≤
10−8 , where x = x x.
Portfolio Selection with Common Correlation Mixture Models 65
means and lower standard deviations than the second regime. Thus, we can
classify the first regime as a bull and the second as a bear market regime. It
should be noted, however, that the standard errors of the component means
are relatively large, so that their differences may not be statistically significant.
We also note that both the bull and the bear market regimes are highly
persistent, with both of the “staying probabilities” being larger than 99%.
In the long run, however, the market is more often in the bullish regime, as
reflected in its unconditional probability of approximately 60%.
The higher kurtosis of the S&P500 and the DAX, when compared to the
FTSE, as reported in Table 1, is reflected in Table 2 by the fact that
the regime-specific standard deviations are more different across regimes for
the former two indices. In fact, recall from (8) that, in normal mixture mod-
els, the coefficient of excess kurtosis can be interpreted as the coefficient of
variation of the component variances.
While some of the results for the two-component model carry over to the
specification with three components, there are also some differences, which
we discuss briefly. In particular, we still have a bull market regime with high
mean returns, low variances and a small correlation coefficient. However, this
regime now has the smallest unconditional probability and is, thus, classified
as Regime 3 in the three-component model. Instead, in this model, the regime
with the largest stationary probability is a “medium” regime with intermedi-
ate standard deviations (except for the S&P500) and correlation coefficient.
It seems that the bull market regime of the two-component model has been
split into a rather optimistic bull market of the new Regime 3 and the new
“business-as-usual” Regime 1, while the stochastic properties of the second
(bear) market component, as well as its unconditional probability, remain
essentially unchanged.10
Before we turn to out-of-sample portfolio selection, we shall also assess, for
the in-sample-period, the significance of both the presence of regimes as well
as regime-specific correlation coefficients. To do so, we estimate, in addition to
the models above, the standard single-regime CCM as well as Markov-MCCMs
with regime-independent correlation coefficients, i.e., with the restriction that
ρ1 = ρ2 (= ρ3 ).11 Table 3 reports likelihood-based goodness-of-fit measures for
the models. Specifically, we report the values of the maximized log likelihood
function, log L, and the AIC [1] and BIC [52] criteria, respectively.
The improvement in log likelihood when passing from the single-regime
to the two-regime model with regime-specific correlations is 52.46. Although
it is difficult to formally test for the number of regimes, this may be deemed
a significant improvement, in particular as the two-component model ranks
higher with respect to the rather conservative BIC. In this regard it may
10
Note, however, that the mean of the S&P500 is higher in Regime 2 than in Regime
1, although this may not be statistically significant.
11
The ECM algorithm developed in Sect. 3.2 can easily be modified to incorporate
such restrictions. Details are available from the authors upon request.
66 M. Haas and S. Mittnik
Table 3. Likelihood-based goodness-of-fit measures. The first row of the table indi-
cates the model, where k refers to the number of components in the Markov-MCCM.
The model with k = 1 is the standard single-regime CCM. The notation “ k = 2 ” is
(ρ1 =ρ2 )
used to denote the two-component Markov-MCCM with equal correlation structure
in both components, and an analogous notation is used for the three-component
model. K denotes the number of parameters of a model, log L is the log likelihood,
AIC = −2 log L + 2K, and BIC = −2 log L + K log T , where T is the number of
observations. Smaller values of AIC and BIC are preferred. Boldface entries indicate
the best model for the particular criterion
K 7 15 16 25 27
log L −3227.2 −3188.2 −3174.8 −3164.4 −3151.8
AIC 6468.5 6406.3 6381.6 6378.8 6357.7
BIC 6498.3 6470.1 6449.6 6485.1 6472.6
where c > 0 is the coefficient of risk aversion, and rtp is the portfolio return
at time t, i.e., rtp = wt rtp , with wt being the portfolio weight vector, satisfying
13 wt = 1 and wt ≥ 0. We will not make an attempt to argue that (54) is
a “realistic” assumption about real-world investor’s preferences. Since (54)
is characterized by a single parameter, c, we can investigate the impact of
increasing risk aversion on portfolio choice decisions. Moreover, in combina-
tion with mixed normally distributed asset returns, it allows a closed-form
computation and straightforward optimization of expected utility.
Using (54), and excluding short sales, a Gaussian investor will solve the
quadratic programming problem
Note that, both in (55) and (56), the (component) means and (component)
covariance matrices depend on t, because the parameter estimates are updated
every week. We also mention that the mixing weights, πjt , to be used in (56)
are the one-period-ahead predictive regime inferences given in (14).
Results
Table 4. Summary statistics for the out-of-sample portfolio returns, covering the
period from February 2001 to August 2005 (234 observations)
single-regime CCM
c = 0.025 0.05 0.1 0.25 0.5 0.75 1 1.25 1.5 2
mean 0.001 0.000 0.023 0.043 0.049 0.051 0.052 0.053 0.053 0.054
variance 5.386 5.247 4.928 4.768 4.739 4.732 4.729 4.728 4.726 4.725
skewness −0.379−0.366−0.342−0.305−0.291−0.287−0.284−0.283−0.282−0.281
kurtosis 5.293 5.457 5.753 5.785 5.772 5.765 5.761 5.758 5.756 5.754
two-regime Markov-MCCM
c = 0.025 0.05 0.1 0.25 0.5 0.75 1 1.25 1.5 2
mean 0.029 0.002 0.016 0.039 0.047 0.050 0.053 0.056 0.058 0.058
variance 6.175 5.504 4.972 4.750 4.704 4.695 4.697 4.702 4.704 4.707
skewness −0.329−0.316−0.278−0.258−0.256−0.254−0.255−0.255−0.255−0.255
kurtosis 4.353 4.879 5.323 5.592 5.691 5.717 5.719 5.715 5.715 5.713
three-regime Markov-MCCM
c = 0.025 0.05 0.1 0.25 0.5 0.75 1 1.25 1.5 2
mean 0.144 0.113 0.059 0.026 0.029 0.032 0.038 0.047 0.053 0.059
variance 6.954 6.413 5.347 4.815 4.677 4.648 4.650 4.670 4.689 4.692
skewness −0.303−0.310−0.289−0.261−0.260−0.257−0.255−0.255−0.257−0.260
kurtosis 3.922 4.161 4.807 5.456 5.723 5.796 5.803 5.772 5.741 5.743
Table 6. Summary statistics for the out-of-sample portfolio returns, covering the
period from March 2004 to August 2005 (75 observations)
single-regime CCM
c= 0.025 0.05 0.1 0.25 0.5 0.75 1 1.25 1.5 2
mean 0.169 0.185 0.213 0.230 0.236 0.238 0.238 0.239 0.239 0.240
variance 1.707 1.612 1.582 1.611 1.629 1.636 1.639 1.642 1.643 1.645
skewness 0.082 0.074 0.068 0.040 0.028 0.023 0.021 0.020 0.019 0.018
kurtosis 2.606 2.546 2.598 2.677 2.706 2.716 2.721 2.724 2.726 2.728
two-regime Markov-MCCM
c= 0.025 0.05 0.1 0.25 0.5 0.75 1 1.25 1.5 2
mean 0.299 0.248 0.225 0.232 0.233 0.234 0.237 0.241 0.243 0.243
variance 2.825 2.295 1.986 1.748 1.673 1.668 1.660 1.661 1.661 1.661
skewness −0.150 −0.044 0.057 0.073 0.069 0.048 0.019 0.007 0.003 0.001
kurtosis 2.581 2.363 2.366 2.581 2.677 2.718 2.762 2.801 2.815 2.820
three-regime Markov-MCCM
c= 0.025 0.05 0.1 0.25 0.5 0.75 1 1.25 1.5 2
mean 0.432 0.396 0.281 0.195 0.197 0.199 0.205 0.217 0.228 0.241
variance 3.755 3.216 2.362 1.867 1.690 1.664 1.663 1.670 1.677 1.658
skewness −0.064 −0.127 −0.016 0.025 0.053 0.061 0.056 0.036 0.014 0.010
kurtosis 2.732 2.548 2.345 2.330 2.415 2.500 2.587 2.685 2.736 2.771
riod from March 2004 to August 2005, the portfolio selection performance of
the models under study. While, just as in Table 4, the differences between the
models are negligible for the higher risk aversion coefficients, the differences
for the lower degrees of risk aversion are even more striking, in particular for
the three-regime model.
Figure 1 shows, for the entire out-of-sample period, the three return series
along with the one-step-ahead predictive regime probabilities, as given by (14),
for the bull and bear market regimes, as calculated from the two-component
S&P500
20
10
−10
−20
2002 2003 2004 2005
FTSE
20
10
−10
−20
2002 2003 2004 2005
DAX
20
10
−10
−20
2002 2003 2004 2005
0.9
0.8
0.7
0.6
0.3
0.2
0.1
0
2002 2003 2004 2005
Fig. 1. Shown are, in the top three panels, the three return series under study
over the out-of-sample period from February 2001 to August 2005. The bottom
panel displays the corresponding one-step-ahead predictive regime probabilities, as
given by (14), calculated from the two-regime Markov-MCCM. The solid line is the
probability of the bull market regime, and the dash-dot line is the probability of the
bear market regime
Portfolio Selection with Common Correlation Mixture Models 71
model. Recall that these are the mixing weights which enter the optimization
problem (56). Clearly the market was predicted to be in a bear state over the
first few years of the out-of-sample period, while it was predicted to be in the
bullish sate in the last period.
Thus it seems that most of the “excess returns” over the single component
model have been realized during the last 75 weeks of the out-of-sample period
under study. To confirm this, we plot, in Fig. 2, for each c-value considered
in Tables 4 and 6, and for each model, the average portfolio weights for the
three different stock indices over the last 75 weeks. Figure 2 reveals that there
are significant differences between the single-regime and the mixture models
only for the lower risk aversion coefficients. Namely, for small values of c,
the mixture investors (with low risk aversion) use their inferences about the
prevailing regime to put a large part of their wealth into the high-return
German market, which, on average, generated the highest returns during the
period under study (cf. Table 5). But as can be seen from Table 5, the DAX
was also a high-volatility stock, which prompts more risk averse investors to
abstain from allocating a large fraction of their wealth to the DAX. In this
regard, it is important to note that the conditional probabilities of the bear
market shown in Fig. 1 are not zero (or even close to zero) in the last period of
the sample, so that investors still assign a positive probability to the markets
being bearish. Clearly the importance of the bear market regime for asset
allocation decisions increases with the degree of risk aversion, c, and when
c becomes large, the portfolio choice will be entirely based on this regime.
Consequently, the (average) portfolio weights for larger values of c are very
similar for the single-regime and the multi-regime models.
5 Conclusions
Our results show that investors can benefit from accounting for regimes in
asset returns. The analysis should be extended in various ways. For exam-
ple, while we considered all-equity portfolios, investors could be allowed to
put part of their wealth into a risk-free asset, which would be relevant in
particular in times of falling stock markets. This would perhaps improve the
overall performance of the investment strategies, which, as seen in Table 4,
suffered from the bear market regime at the beginning of our out-of-sample
period. Presumably, the benefits from the opportunity to invest in a risk-free
asset will be especially significant for the multi-regime models, because the
fraction of wealth to be invested in the risk-free asset can be made dependent
on the prevailing regime, to hedge against low returns and high correlations
in the bear market. In addition, risk considerations may be taken into ac-
count. For example, Value-at-Risk or expected-shortfall restrictions could be
incorporated into the portfolio optimization. Then the mixture approach is
expected to further improve upon the traditional Gaussian approach, as it
better accommodates the excess kurtosis in the return distribution.
72 M. Haas and S. Mittnik
average weights for the single−regime CCM from March 2004 to August 2005
1
S&P500
0.8 FTSE
DAX
portfolio weights
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
risk aversion, c
average weights for the two−regime MCCM from March 2004 to August 2005
1
S&P500
0.8 FTSE
DAX
portfolio weights
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
risk aversion, c
average weights for the three−regime MCCM from March 2004 to August 2005
1
S&P500
0.8 FTSE
DAX
portfolio weights
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2
risk aversion, c
Fig. 2. Shown are, for the period from March 2004 to August 2005 (i.e., the last
75 observations of the out-of-sample period), and for values of the coefficient of
risk aversion, c, ranging from 0.025 to 2, the average portfolio weights for the three
different stock indices, as implied by the single-regime (top panel ) model, as well as
by the two- (center panel ) and three-regime (bottom panel ) MCCM
Portfolio Selection with Common Correlation Mixture Models 73
References
1 Introduction
series to be heavy-tailed and possibly skewed. Rachev and Mittnik [21] and
Rachev et al. [22] have developed financial models with α-stable distributions
and applied them to market and credit risk management, option pricing, and
portfolio selection. They also discuss the major attack on the α-stable models
in the 1970s and 1980s. That is, while the empirical evidence does not support
the normal distribution, it is also not consistent with an α-stable distribution.
The distribution of returns for assets has heavier tails relative to the normal
distribution and thinner tails than the α-stable distribution. Partly in re-
sponse to those empirical inconsistencies, various alternatives to the α-stable
distribution were proposed in the literature. Two examples are the “CGMY”
(or “KoBoL”) distribution (Carr et al. [7], Koponen [14], and Boyachenko
and Levendorskiĭ [6]) and the “Modified Tempered Stable” distribution (Kim
et al. [12]). These two distributions, sometimes called the tempered stable
distributions, have not only heavier tails than the normal distribution and
thinner than the α-stable distribution, but also have finite moments for all
orders. Recently, Rosiński [23] generalized the CGMY distributions and clas-
sified them using the “spectral” (or Rosiński) measure.
In this paper, we will introduce an extension of the CGMY distribution
named the “KR tempered stable” (or simply “KR”) distribution. The KR
distribution is characterized by a new spectral measure. We believe that the
simple form of the characteristic function, the exponential decayed tails, and
other desirable properties of the KR distribution will result in its use in the-
oretical and empirical finance, such as modeling asset return processes, port-
folio analysis, risk management, derivative pricing, and econometrics in the
presence of heavy-tailed innovations.
In the Black-Scholes model [5], the stock price process is described by the
exponential of Brownian motion with drift : St = S0 eXt where Xt = µt + σBt
and the process Bt is Brownian motion. Replacing the driving process Xt by
a Lévy process we obtain the class of exponential Lévy models. For example,
if Xt is replaced by the CGMY process then one can obtain the exponential
CGMY model (Carr et al. [7]). In the exponential Lévy model, the equiva-
lent martingale measure (EMM) of a given market measure is not unique in
general. For this reason, we have to find a method to select one of them.
One classical method to choose an EMM is the Esscher transform; another
reasonable method is finding the “minimal entropy martingale measure”, as
presented by Fujiwara and Miyahara [20]. However, while these methods are
mathematically elegant and have a financial meaning in a utility maximization
problem, the model prices obtained from the EMM did not match the market
prices observed for options. The other method for handling the problem is
to estimate the risk-neutral measure by using current option price data inde-
pendent of the historical underlying distribution. This method can fit model
prices to market prices directly, but it has a problem: the historical market
measure and the risk-neutral measure need not to be equivalent and it conflicts
with the the no-arbitrage property for option prices. To overcome these draw-
backs, one must estimate the market measure and the risk-neutral measure
A New Tempered Stable Distribution 79
d
The completely monotonicity of q(·, u) means that (−1)n dv q(v, u) > 0 for
all v > 0, u ∈ S d−1 , and n = 0, 1, 2, · · · . The tempering function q can be
represented as the Laplace transform
∞
q(v, u) = e−vs Q(ds|u) (3)
0
where
Γ (−α)((1 − iy)α − 1 + iαy), if α = 1
ψα (y) = (11)
(1 − iy) log(1 − iy) + iy, if α = 1
and b = E[X].
(ii) If α ∈ (0, 1) and
||x||R(dx) < ∞, (12)
||x||≤1
where
ψα0 (y) = Γ (−α)((1 − iy)α − 1) (14)
0
and b0 ∈ Rd is the drift vector (i.e. b0 = ||x||≤1 ||x||M (dx)).
Remark 2.4 Let X be a TαS distributed random vector with the spectral
measure R. By Proposition 2.7 in [23], we can say the following:
1. In the above definition, E[||X||] < ∞ if and only if α ∈ (1, 2) or
α = 1 and ||x|| log ||x||R(dx) < ∞, (15)
||x||>1
or
α ∈ (0, 1) and ||x||R(dx) < ∞. (16)
||x||>1
0
2. If α ∈ (0, 1) and Rd ||x||R(dx) < ∞, then both form (10) and (13) are
valid for X. Therefore
0 X ∼ T Sα0 (R, b0 ) and X ∼ T Sα0 (R, b), where b =
b0 + Γ (1 − α) Rd xR(dx).
The following Lemma shows some relations between the spectral measure
R of the TαS distribution and the Lévy measure of the α-stable distribution
given by (1).
Lemma 2.5 (Lemma 2.14. [23]) Let M be a Lévy measure of a proper TαS
distribution, as in (2), with the spectral measure R. Let M0 be the Lévy mea-
sure of α-stable distribution given by (1). Then
∞
M0 (A) = IA (tx)t−α−1 dtR(dx), A ∈ B(Rd ). (17)
Rd 0
82 Y.S. Kim et al.
Furthermore,
x
σ(B) = IB ||x||α R(dx), B ∈ B(S d−1 ). (18)
Rd ||x||
% ) * α ) *
Γ (−α) cos απ 2 |y| (1 − i tan απ
2 sgn(y)), if α = 1
ψ̄α (y) =
− 2 (|y| + i π y log(y)),
π 2
if α = 1
(See [24, Theorem 14.10]). In this case, we will write X ∼ Sα (σ, a).
Since TαS is infinitely divisible, there is a Lévy process (Xt )t≥0 in Rd such
that X1 has a TαS (proper TαS) distribution. The process (Xt )t≥0 will be
called a TαS (proper TαS) Lévy process.
Let Ω to be the set of all cadlag function on [0, ∞) into Rd , and (Xt )t≥0
is a canonical process on Ω (i.e, Xt (ω) = ω(t), t ≥ 0, ω ∈ Ω). Consider a
filtered probability space (Ω, F, (Ft )t≥0 ) where
F = σ{Xs ; s ≥ 0}
Ft = ∩s≥0 σ{Xu : u ≤ s}, t ≥ 0.
(Ft )t≥0 is the right continuous natural filtration. The canonical process
(Xt )t≥0 is characterized by a probability measure P on (Ω, F, (Ft )t≥0 ).
Theorem 2.6 (Theorem 4.1. [23]) In the above setting, consider two prob-
ability measures P0 and P on (Ω, F) such that the canonical process (Xt )t≥0
under P0 is an α-stable process while under P it is a proper TαS Lévy process.
Specifically, assume that under P0 , X1 ∼ Sα (σ, a), where σ is related to R
by (18) and α ∈ (0, 2), while under P, X1 ∼ T Sα0 (R, b) when α ∈ (0, 1) and
X1 ∼ T Sα (R, b) when α ∈ [1, 2). Let M , the Lévy measure corresponding to
R, be as in (2), where q(0+ , u) = 1 for all u ∈ S d−1 . Then P0 |Ft and P|Ft are
mutually absolutely continuous for every t > 0 if and only if
1
(1 − q(v, u))2 v −α−1 dv σ(du) < ∞ (19)
S d−1 0
and ⎧
⎨ 0,
0 if α ∈ (0, 1)
b−a= x(log ||x||
0 − 1)R(dx), if α = 1 (20)
⎩ R
d
with α ∈ (0, 2), k+ , k− , r+ , r− > 0 and p+ , p− > −α. Then the spectral
measure R corresponding to the Lévy measure M can be deduced as
−p+ −p−
R(dx) = (k+ r+ I(0,r+ ) (x)|x|p+ −1 + k− r− I(−r− ,0) (x)|x|p− −1 ) dx. (22)
Lemma 3.1 If M and R are given by (21) and (22), respectively, we have
(i) R({0}) = 0, |x|α R(dx) < ∞ and |x|R(dx) < ∞ for all α ∈ (0, 2).
R |x|>1
(ii) By Theorem 2.2, M can be written in the form
r+ ∞
−p+
M (A) = k+ r+ IA (tx)t−α−1 e−t dt xp+ −1 dx (23)
0
r− ∞
0
−p−
+ k− r− IA (−tx)t−α−1 e−t dt xp− −1 dx, A ∈ B(R0 ).
0 0
(iii) If α = 1 then
x log |x|R(dx) < ∞,
|x|>1
−1
If θ ≤ r+ then θx − 1 ≤ 0 where x ∈ (0, r+ ), and hence
r+ ∞
−p+
k+ r+ et(θx−1) t−α−1 dt xp+ −1 dx
0 1/x
r+ ∞
−p+
≤ k+ r+ t−α−1 dt xp+ −1 dx
0 1/x
−p
r+
xα+p+ −1 α
k+ r+
= k+ r+ + dx = ,
0 α α(α + p+ )
−1
Similarly if −r− ≤ θ then −θx − 1 ≤ 0 where x ∈ (0, r− ), and hence
r− ∞
−p
k− r− − et(−θx−1) t−α−1 dt xp− −1 dx
0 1/x
r− ∞
−p−
≤ k− r− t−α−1 dt xp− −1 dx
0 1/x
−p−
r−
xα+p− −1 α
k− r−
= k− r− dx = ,
0 α α(α + p− )
−1 −1
0
Thus, if −r−
≤ θ ≤ r+ then |x|>1 eθx M (dx) < ∞.
−1 −1
Conversely, if θ > r+ then θx − 1 > r+ x − 1 > 0 for all x ∈ (0, r+ ), so
−1
there is such that 0 < < r+ x − 1 for all h ∈ (0, r+ ). Hence
r+ ∞
−p
k+ r+ + et(θx−1) t−α−1 dt xp+ −1 dx
0 1/x
r+ ∞
−p+
> k+ r+ et t−α−1 dt xp+ −1 dx = ∞.
0 1/x
−1
Similarly, we can prove that, if θ < −r− then
r− ∞
−p
k− r− − et(−θx−1) t−α−1 dt xp− −1 dx = ∞.
0 1/x
A New Tempered Stable Distribution 85
Lemma 3.3 Let α ∈ (0, 2), p ∈ (−α, ∞) \ {−1, 0}, h > 0, and u ∈ R. Then
1,
we have, if α =
h
hp
xp−1 (1 − iux)α dx = F (p, −α; 1 + p; iuh) (24)
0 p
and, if α = 1,
h
((1 − iux) log(1 − iux) + iux) xp−1 dx (25)
0
+ ihu hu
= hp + (huF (2 + p, 1; 3 + p; ihu) − i(2 + p) log(1 − ihu))
1+p 2 + 3p + p2
(ihu)−p ) *,
+ (p − ihu)F3,2 (1, 1, 1 − p; 2, 2; 1−ihu)−(1 − (ihu)p ) log(1 − ihu) ,
p
where the hypergeometric function F (a, b; c; x) and the generalized hypergeo-
metric function Fp,q (a1 , · · · , ap ; b1 , · · · , bq ; x). (The hypergeometric function
and the generalized hypergeometric function are described in [3].)
Proof. Suppose |iux| < 1 and α = 1. Since
d ab
F (a, b; c; x) = F (a + 1, b + 1; c + 1; x),
du c
∞
(p)n (−α)n (iux)n ∞
p (iux)n
= (−α)n
n=0
(p + 1)n n! n=0
p+n n!
and
∞ ∞
(p + 1)n (−α)n+1 (iux)n+1 (p + 1)n−1 (−α)n (iux)n
=
n=0
(p + 1)n+1 n! n=1
(p + 1)n (n − 1)!
∞
n (iux)n
= (−α)n .
n=1
p+n n!
we have
d xp
F (p, −α, 1 + p; iux)
dx p
xp p(−α)
= xp−1 F (p, −α; 1 + p; iux) + F (p + 1, 1 − α; p + 2; iux)iu
p 1+p
∞ ∞
!
(p)n (−α)n (iux)n (p + 1)n (−α)n+1 (iux)n+1
= xp−1 + .
n=0
(p + 1)n n! n=0
(p + 1)n+1 n!
∞
!
(iux)n p n
= xp−1 1+ (−α)n +
n=1
n! p+n p+n
∞
!
(iux)n
= xp−1 1+ (−α)n
n=1
n!
= xp−1 (1 − iux)α .
86 Y.S. Kim et al.
Hence, (24) is proved if |iux| < 1 and this result can be extended analytically
if −1 < Re(iux) < 1, so (24) is true for all real u. Equation (25) can be proved
by the same method.
Theorem 3.4 Let X be a random variable with the proper TαS distribution
corresponding to the spectral measure R defined in (22) with conditions p = 0
and p = −1, and let m = E[X]. Then the characteristic function E[eiuX ],
u ∈ R, is given as follows:
(i) if α = 1,
iuX
E[e ] = exp Hα (u; k+ , r+ , p+ ) + Hα (−u; k− , r− , p− ) (26)
k+ r+ k− r−
+ iu m + αΓ (−α) − ,
p+ + 1 p− + 1
where
aΓ (−α)
Hα (u; a, h, p) = (F (p, −α; 1 + p; ihu) − 1) ,
p
(ii) if α = 1,
iuX
E[e ] = exp Gα (u; k+ , r+ , p+ ) + Gα (−u; k− , r− , p− ) (27)
k+ r+ k− r−
+ iu m + − ,
p+ + 1 p− + 1
where
ahu
Gα (u; a, h, p) = (huF (2 + p, 1; 3 + p; ihu) − i(2 + p) log(1 − ihu))
2 + 3p + p2
a(ihu)−p
+ ((p − ihu)F3,2 (1, 1, 1 − p; 2, 2; 1 − ihu) − (1 − (ihu)p ) log(1 − ihu).
p
Proof. By Lemma 3.1 (vi), m ≡ E[X] < ∞. By Definition 2.3, we have
⎧
⎪
⎪
⎨ Γ (−α)((1 − iux) − 1 + iαux)R(dx) + imu if α = 1
α
R
log E[eiuX
]=
⎪
⎪
⎩ ((1 − iux) log(1 − iux) + iux)R(dx) + imu if α = 1
R
In case α = 1, we have
Γ (−α)((1 − iux)α − 1 + iαux)R(dx) + imu
R
r+
−p
= k+ r+ + Γ (−α) ((1 − iux)α − 1 − iαux)xp+ −1 dx
0
r−
−p−
+ k− r− Γ (−α) ((1 + iux)α − 1 + iαux)xp− −1 dx + imu.
0
A New Tempered Stable Distribution 87
k± Γ (−α)(−α)n ik hk±
= F (p± + n, n − α; p± + n + 1; iuh± )
p± + n
k± Γ (n − α)in hn±
= F (p± + n, n − α; p± + n + 1; iuh± ).
p± + n
Thus, (28) can be shown.
Proposition 3.7 Let X ∼ KR(α, k+ , k− , r+ , r− , p+ , p− , m) with α = 1. Then
dk
the cumulants ck (X) ≡ i1k du k log E[e
iuX
] u=0 is given by c1 (X) = b and
k k
k+ r+ k− r−
ck (X) = Γ (k − α) + (−1)k
p+ + k p− + k
where k ≥ 2.
Remark 3.8 Let X ∼ KR(α, k+ , k− , r+ , r− , p+ , p− , m) with α = 1. By the
Corollary 3.7, we obtain the mean, variance, skewness and excess kurtosis of
X which are given as follows:
1. E[X] = c1 (X) = m
2 2
k+ r+ k− r−
2. Var(X) = c2 (X) = Γ (2 − α) +
p+++ 2 p− + 2,
3 3
k + r+ k − r−
c3 (X) Γ (3 − α) p+ +3 − p− +3
3. s(X) =
c2 (X)3/2
= + 2 2
,3/2
k r+ k − r−
Γ (2 − α)3/2 p+++2 + p− +2
+ 4 4 ,
k r+ k − r−
c4 (X) Γ (4 − α) p+++4 + p− +4
4. k(X) = = + ,2
c2 (X)2 k r+2 2
k − r−
Γ (2 − α)2 p+++2 + p− +2
Similarly, we have
k− Γ (−α)
lim (F (p− , −α; 1 + p− ; −ir− u) − 1)
p− →∞ p−
) −1 −α
*
= cΓ (−α) (r− + iu)α − r− .
Moreover, we have
k+ r+ k− r−
µ ≡ m + lim αΓ (−α) − lim αΓ (−α)
p+ →∞ p+ + 1 p− →∞ p− + 1
1−α 1−α
c(α + p+ )r+ c(α + p− )r−
= m + lim αΓ (−α) − lim αΓ (−α)
p+ →∞ p+ + 1 p− →∞ p− + 1
1−α
= m + cαΓ (−α)(r+ − r−
1−α
).
In all, we have
lim E[eiuX ]
p+ ,p− →∞
) )) −1 −α
* ) −1 −α
***
= exp iµu + cΓ (−α) (r+ − iu)α − r+ + (r− + iu)α − r− .
8
CGMY
KR p+=p−=−0.2
7
KR p =p =1
+ −
KR p+=p−=10
6
0
−0.2 −0.1 0 0.1 0.2 0.3
Proof. We have
r+ ∞
−p
|x|M (dx) = k+ r+ + txI(0,1) (tx)t−α−1 e−t dt xp+ −1 dx
|x|<1 0 0
r− ∞
−p
+ k− r− − (−tx)I(−1,0) (−tx)t−α−1 e−t dt xp− −1 dx
0 0
r+ 1/x
−p+
= k+ r+ t−α e−t dt xp+ dx
0 0
r− 1/x
−p−
− k− r− t−α e−t dt xp− dx.
0 0
Thus
< ∞ if α ∈ (0, 1)
|x|M (dx)
|x|<1 = ∞ if α ∈ [1, 2)
Proof. By Remark 3.8, X has a mean and variance, therefore we consider the
Chebyshev’s Inequality
1
P(|X − m| ≥ λ) ≤ Var(X).
λ2
We obtain
2 2
1 k+ r+ k− r−
P(|X − m| ≥ λ) ≤ Γ (2 − α) +
λ2 p+ + 2 p− + 2
as β → ∞.
and
∞
s−a−1 e−s ds
β
∞
−a−1 −β −a−2 −β
=β e − (a + 1)β e + (a + 1)(a + 2) s−a−3 e−s ds
β
Taking into account Proposition 3.14 and Lemma 3.15, we can prove the
following result.
e− r̄
2λ
P(|X − m| ≥ λ) ≥ C
λα+2
as λ → ∞, where C does not depend on λ and r̄ = max(r+ , r− ).
1 − exp(−z) ∼ z, z→0
λ−α−1
|x|α+1 e− |x| R(dx),
2λ
∼ α+3 (31)
2 R0
A New Tempered Stable Distribution 93
k+ − r2λ k− − r2λ
∼ (2λ)−1 α+p+ +2 e
+ + (2λ)−1 α+p− +2 e
−
r+ r−
∼ C̄(2λ)−1 e−
2λ
r̄
e− r̄
2λ
P(|X − m| ≥ λ) ≥ C .
λα+2
Let (Xt )t≥0 be a canonical process on Ω, the set of all cadlag function on
[0, ∞) into R, and consider a space (Ω, F, (Ft )t≥0 ), where
F = σ{Xs ; s ≥ 0}
Ft = ∩s>t σ{Xu : u ≤ s}, t ≥ 0.
Then P1 |Ft and P2 |Ft are equivalent for every t > 0 if and only if
α := α1 = α2 , (32)
α α α α
k1,+ r1,+ k2,+ r2,+ k1,− r1,− k2,− r2,−
= , = (33)
α + p1,+ α + p2,+ α + p1,− α + p2,−
and
94 Y.S. Kim et al.
m2 − m 1 (34)
⎧
⎪ kj,+ rj,+ pj,+ + 2
⎪
⎪ (−1)j log rj,+ −
⎪
⎪ pj,+ + 1 pj,+ + 1
⎪
⎪ j=1,2 if α = 1
⎨ kj,− rj,− pj,− + 2
= − log rj,− − .
⎪
⎪ pj,− + 1 pj,− + 1
⎪
⎪
⎪
⎪ kj,+ rj,+ kj,− rj,−
⎩ Γ (1 − α)
⎪ − if α = 1
j
(−1)
j=1,2
pj,+ + 1 pj,− + 1
Proof. In KR(αj , kj,+ , kj,− , rj,+ , rj,− , pj,+ , pj,− , mj ), the spectral measure Rj
is equal to
−p −p
Rj (dx) = (kj,+ rj,+j,+ Ix∈(0,rj,+ ) |x|pj,+ −1 + kj,− rj,−j,− Ix∈(0,rj,− ) |x|pj,− −1 )dx
and the polar coordinated Lévy measure Mj is equal to
Mj (dv, du) = v −αj −1 qj (v, u)dvσj (du)
where
α
j j α
kj,+ rj,+ kj,− rj,−
σj (A) = 11∈A + 1−1∈A , A ⊂ S0
αj + pj,+ αj + pj,−
and
rj,±
−α −pj,±
qj (v, ±1) = (αj + pj,± )rj,±j e−v/s sαj +pj,± −1 ds
0
0under Pj . Indeed, by Lemma 3.1 iii), EPj [|X1 |] < ∞ if αj ∈ (0, 2) and
|x|<1
|x|Rj (dx) < ∞ if αj ∈ (0, 1).
If pj,± > 12 − αj then we have
rj,±
d −αj −pj,±
qj (v, ±1) = −(αj + pj,± )rj,± e−v/s sαj +pj,± −2 ds
dv
∞
0
−αj −pj,±
= −(αj + pj,± )rj,± e−vt t−αj −pj,± dt
1/rj,±
∞
−α −p 1
≥ −(αj + pj,± )rj,±j j,± √ t−αj −pj,± dt
1/rj,± vt
αj + pj,± − 12
= −√ 1 v .
rj,± (αj + pj,± − 2 )
A New Tempered Stable Distribution 95
Let
⎧ & '
⎨ min − √ αj +pj,+
, − √
αj +pj,−
, αj ∈ (0, 1)
& rj,+ (αj +pj,+ −1/2) '−1/2)
rj,− (αj +pj,−
Kj =
⎩ min − α +p α +p
rj,+ (αj +pj,+ −1) , − rj,− (αj +pj,− −1) , αj ∈ [1, 2)
j j,+ j j,−
then
d Kj v −1/2 , αj ∈ (0, 1)
0> qj (v, ±1) ≥ .
dv Kj , αj ∈ [1, 2)
By the integration of the last inequality on the interval (0, v), we obtain
2Kj v 1/2 , αj ∈ (0, 1)
0 ≥ qj (v, ±1) − 1 = qj (v, ±1) − qj (0, ±1) ≥ .
Kj v, αj ∈ [1, 2)
Hence,
1
(1 − qj (v, u))2 v −αj −1 dv σ(du)
S0
% 00 01
4K 2 v −αj dv σ(du), αj ∈ (0, 1)
≤ 0S 0 001 2j −α +1
S0 0
Kj v j dv σ(du), αj ∈ [1, 2)
⎧
⎨ 4Kj 0 σ(du), α ∈ (0, 1)
2
1−αj S 0 j
⎩ Kj 0 0 σ(du), αj ∈ [1, 2)
= 2
2−αj S
< ∞.
By Theorem 2.6, there is a measure P0j such that P0j |Ft and Pj |Ft are equivalent
for every t > 0 and (Xt )t≥0 is an α-stable process with X1 ∼ Sαj (σj , aj ) under
P0j where
⎧
⎨ bj 0 if α ∈ (0, 1)
aj = bj − R x(log |x| 0 − 1)R j (dx) if α = 1
⎩
bj − Γ (1 − α) R xRj (dx) if α ∈ (1, 2)
0
mj − R x(log |x| 0 − 1)Rj (dx) if α = 1 .
=
mj − Γ (1 − α) R xRj (dx) if α = 1
96 Y.S. Kim et al.
and if α = 1, then
rj,+ rj,−
−pj,+ pj,+ −p
xRj (dx) = kj,+ rj,+ x dx + kj,− rj,−j,− xpj,− dx
R 0 0
kj,+ rj,+ kj,− rj,−
= −
pj,− + 1 pj,− + 1
Since P01 |Ft and P02 |Ft are equivalent for every t > 0 if and only if α1 = α2 ,
σ1 = σ2 , and a1 = a2 , we obtain the result that P1 |Ft and P2 |Ft are equivalent
for every t > 0 if and only if the parameters satisfy (32), (33) and (34).
In the remainder of this paper, let us denote a time horizon by T > 0 and the
risk-free rate by r > 0. Let Ω to be the set of all cadlag functions on [0, T ] into
R, and (Xt )t∈[0,T ] is a canonical process on Ω (i.e. Xt (ω) = ω(t), t ∈ [0, T ],
ω ∈ Ω). Consider a filtered probability space (Ω, FT , (Ft )t∈[0,T ] ) where
FT = σ{Xs ; s ∈ [0, T ]}
Ft = ∩s∈(t,T ] σ{Xu : u ≤ s}, t ∈ [0, T ].
(Ft )t∈[0,T ] is the right continuous natural filtration. The continuous-time mar-
ket is modeled by a probability space (Ω, FT , (Ft )t∈[0,T ] , P), for some measure
P named the market measure. In the market, the stock price is given by the
random variable St = S0 eXt , t ∈ [0, T ] for some initial value of the stock
price S0 > 0, and the discounted stock price S̃t of St is given by S̃t = e−rt St ,
t ∈ [0, T ]. The processes (St )t∈[0,T ] and (S̃t )t∈[0,T ] are called the stock price
process and the discounted (stock) price process, respectively. The process
(Xt )t∈[0,T ] is called the driving process of (St )t∈[0,T ] . The driving process
(Xt )t∈[0,T ] is completely described by the market measure P. If (Xt )t∈[0,T ]
A New Tempered Stable Distribution 97
is a Lévy process under the measure P, we say that the stock price process
follows the exponential Lévy model. Assume a stock buyer receives continu-
ous dividend yield d. A probability measure Q equivalent to P is called an
equivalent martingale measure (EMM) of P if the stock price process net of
the cost of carry (Lewis [15]) is a Q-martingale; that is EQ [St ] = e(r−d)t S0 or
EQ [eXt ] = 1.
Now, we will define the KR model. For convenience, we exclude the case
α = 1 and define a function
Remark 4.2
1. We have the condition r+ ∈ (0, 1) for ψα (−i; k+ , k− , r+ , r− p+ , p− , 0) and
E[eXt ] to be well defined.
p+ , p− ∈ (1/2 − α, ∞) \ {0}, if α ∈ (0, 1)
2. By the condition , we are able
p+ , p− ∈ (1 − α, ∞) \ {0}, if α ∈ (1, 2)
to use Theorem 3.17 for finding an equivalent measure.
3. Since m = µ − ψα (−i; k+ , k− , r+ , r− , p+ , p− , 0), we have
Theorem 4.3 Assume that (St )t∈[0,T ] is the the KR price process with pa-
rameters (α, k+ , k− , r+ , r− , p+ , p− , µ) under the market measure P, and with
parameters (α̃, ã+ , ã− , r̃+ , r̃− , p̃+ , p̃− , r − d) under a measure Q. Then Q is an
EMM of P if and only if
α = α̃, (35)
98 Y.S. Kim et al.
α α α α
k+ r+ k̃+ r̃+ k− r− k̃− r̃−
= , = (36)
α + p+ α + p̃+ α + p− α + p̃−
and
In this section, we will present the estimation results of the fit of our model
to the historical log-returns of the S&P 500 Index. In order to compare the
KR model with other well-known models, let us consider the normal, CGMY,
and KR density fit. The CGMY process is defined in the Appendix and in
[7]. In our empirical study, we focus on two sets of data. We estimated the
market parameters from time-series data on the S&P 500 Index over the period
January 1, 1992 to April 18, 2002, with ñ = 2573 closing prices (Data1), and
over the period January 1, 1984 to January 1, 1994, with n̄ = 2498 closing
prices (Data2). The estimation of market parameters based on Data1 will be
used to extract the risk-neutral density by using observed option prices, while
the historical series Data2 is selected to demonstrate the benefit of the KR
distribution in fitting historical log-returns containing extreme events (Black
Monday, October 19, 1987).
Our estimation procedure follows the classical maximum likelihood esti-
mation (MLE) method (see Table 1). The discrete Fourier transform (DFT) is
used to invert the characteristic function and evaluate the likelihood function
in the CGMY and KR cases.
In order to compare how the stock market process can be explained by
these different models, Figs. 2 and 3 show the results of density fits.
Let (Ω, A, P) be a probability space and {Xi }1≤i≤n a given set of inde-
pendent and identically distributed real random variables. In the following,
let us consider Xi (ω) = xi , for each i = 1, . . . , n. Let F be the distribution of
Xi , and x1 ≤ x2 ≤ . . . ≤ xn . The empirical cumulative distribution function
F̂n (x) is defined by
⎧
no. observations ≤ x ⎨ i
0, x < x1
F̂n (x) = = n , xi ≤ x ≤ xi+1 , i = 1, . . . , n − 1
n ⎩
1, xn ≤ x.
A statistic measuring the difference between F̂n (x) and F (x) is called the
empirical distribution function (EDF) statistic [10]. These statistics include
the Kolmogorov-Smirnov (KS) statistic [10, 18, 26] and Anderson-Darling
(AD) statistic [1, 2, 19]. Our goal is to test if the empirical distribution
A New Tempered Stable Distribution 99
60
Market data
Normal
CGMY
50
KR
40
30
20
10
0
−0.04 −0.03 −0.02 −0.01 0 0.01 0.02 0.03 0.04
Fig. 2. S&P 500 Index (from January 1, 1992 to April 18, 2002) MLE density fit.
Circles are the densities of the market data. The solid curve is the KR fit, the dotted
curve is the CGMY fit and the dashed curve is the normal fit
100 Y.S. Kim et al.
70
Market data
B−S
60 CGMY
KR
50
40
30
20
10
0
−0.04 −0.03 −0.02 −0.01 0 0.01 0.02 0.03 0.04
Fig. 3. S&P 500 Index (from January 1, 1984 to January 1, 1994) MLE density
fit. Circles are the densities of the market data. The solid curve is the KR fit, the
dotted curve is the CGMY fit and the dashed curve is the normal fit
where {xi }1≤i≤n is a given set of observations. Using the procedure of [18],
we can easily evaluate the distribution of Dn and find the p-value for our test.
It might be of interest to test the ability of the model to forecast ex-
treme events. To this end, we also provide the AD statistics. We consider two
different versions of the AD statistic. In its simplest version, it is a variance-
weighted KS statistic
A New Tempered Stable Distribution 101
|F (xi ) − F̂ (xi )|
ADn = sup (40)
xi F (xi )(1 − F (xi ))
Since the distribution of ADn is not known in closed form, p-values were
obtained via 1000 Monte Carlo simulations.
A more generally used version of this statistic belongs to the quadratic
class defined by the Cramér-von Mises family [10], i.e.
∞
(F̂n (x) − F (x))2
ADn2 = n dF (x) (41)
−∞ F (x)(1 − F (x))
1 1
n n
ADn2 = −n + (1 − 2i) log(zi ) − (1 + 2(n − i)) log(1 − zi )
n i=1 n i=1
p-value
Theoretical
Monte Carlo‡
Model χ2 KS AD2 χ2 KS AD AD2
Normal 0 0 0 0 0 0 0
CGMY 0.2045 0.9450 0.6356 0.43 0.908 0.098 0.656
KR 0.2216 0.9165 0.9082 0.53 0.875 0.242 0.916
Theoretical p-values were obtained from [18, 19] and χ2 distribution
‡
Monte Carlo p-values were obtained via 1,000 simulations
102 Y.S. Kim et al.
p-value
Theoretical Monte Carlo‡
Model χ2 KS AD2 χ2 KS AD AD2
Normal 0 0 0 0 0 0 0
CGMY 0.2451 0.3180 0.0865 0.893 0.305 0.696 0.086
KR 0.5055 0.9343 0.3723 0.974 0.875 0.872 0.361
Theoretical p-values were obtained from [18, 19] and χ2 distribution
‡
Monte Carlo p-values were obtained via 1000 simulations
πk = P (X ∈ Ak ) k = 1, . . . , m
If necessary, we collapse outer cells Ak , so that the expected value nπk of the
observations always becomes greater than 5 [25].
From the results reported in Tables 2 and 3, we conclude that H0normal is
rejected but H0CGM Y and H0KR are not rejected. QQ-plots (see Figs. 4 and 5)
show that the empirical density strongly deviated from the theoretical density
for the normal model, but this deviation almost disappears in both the CGMY
and KR cases.
0.04 0.04
0.03 0.03
0.02 0.02
0.01 0.01
0 0
−0.01 −0.01
−0.02 −0.02
−0.03 −0.03
−0.04 −0.04
−0.05 −0.05
−0.05 0 0.05 −0.05 0 0.05
0.05
0.04
0.03
0.02
0.01
−0.01
−0.02
−0.03
−0.04
−0.05
−0.05 0 0.05
Fig. 4. QQ-plots of S&P 500 Index (from January 1, 1992 to April 18, 2002) MLE
density fit. Normal model (left), CGMY model (right) and KR model (down)
0.05 0.05
0.04 0.04
0.03 0.03
0.02 0.02
0.01 0.01
0 0
−0.01 −0.01
−0.02 −0.02
−0.03 −0.03
−0.04 −0.04
−0.05 −0.05
−0.05 0 0.05 −0.05 0 0.05
0.05
0.04
0.03
0.02
0.01
−0.01
−0.02
−0.03
−0.04
−0.05
−0.05 0 0.05
Fig. 5. QQ-plots of S&P 500 Index (from January 1, 1984 to January 1, 1994) MLE
density fit. Normal model (left), CGMY model (right) and KR model (down)
104 Y.S. Kim et al.
density. Therefore, taking into account the estimation results of Section 4.1
under the market probability measure, we want to estimate parameters under
a risk-neutral measure.
Let us consider a given market model and observed prices Ĉi of call options
with maturities Ti and strikes Ki , i ∈ {1, . . . , N }, where N is the number of
options on a fixed day. The risk-neutral process is fitted by matching model
prices to market prices using nonlinear least squares. Hence, to obtain a prac-
tical solution to the calibration problem, our purpose is to find a parameter
set θ̃, such that the optimization problem
N
min (Ĉi − C θ̃ (Ti , Ki ))2 (42)
θ̃
i=1
α
k̃− r̃−
p̃− =
k− r− α (α − p− ) − α
and
µ − r = Hα (−i; k+ , r+ , p+ ) + Hα (i; k− , r− , p− )
− Hα (−i; k̃+ , r̃+ , p̃+ ) − Hα (i; k̃− , r̃− , p̃− ).
A New Tempered Stable Distribution 105
In the CGMY case we have only one free parameter but in the KR case we
have 3 free parameters to fit model prices to market prices; therefore, we can
obtain a better solution to the optimization problem. The KR distribution
is more flexible in order to find an equivalent change of measure and, at the
same time, takes into account the historical estimates.
The time-series data were for the period January 1, 1992–April 18, 2002,
while the option data were April 18, 2002.
Contrary to the classical Black-Scholes case, in the exponential-Lev́y mod-
els there is no explicit formula for call option prices, since the probability den-
sity of a Lev́y process is typically not known in closed form. Due to the easy
form of the characteristic functions of the CGMY and KR distributions, we
follow the generally used pricing method for standard vanilla options, which
can be applied in general when the characteristic function of the risk-neutral
stock-price process is known [8, 25]. Let ρ be a positive constant such that
the ρ-th moment of the price exists and φ be the characteristic function of the
random variable log ST . A value of ρ = 0.75 will typically do fine [25]. Carr
and Madan [8, 25] then showed that
exp (−ρ log K) ∞
C(K, T ) = exp(−iv log K)(v)dv,
π 0
where
exp(−rT )φ(v − (ρ + 1)i)
(v) =
ρ2 + ρ − v 2 + i(2ρ + 1)v
Furthermore, we need to guarantee the analyticity of the integrand function
in the horizontal strip of the complex plane, on which the line Lρ = {x + iρ ∈
C| − ∞ < x < ∞} lies [15, 16]. If we consider the exponential KR model, we
obtain the following additional inequality constraint,
−1
r+ ≥ 1 + ρ,
p+ , p− ∈ (1/2 − α, ∞),
by Remark 4.2.
Each maturity has been calibrated separately (see Table 4). Unfortunately,
due to the independence and stationarity of their increments, exponential Lévy
models perform poorly when calibrating several maturities at the same time
[9]. In Table 5, we resume the error estimator of our option price fits. If we
consider the exponential CGMY or KR models, we can estimate simultane-
ously market and risk-neutral parameters using historical prices and observed
option prices. The flexibility of the KR distribution allows one to obtain a
suitable solution to the calibration problem (see Table 5).
106 Y.S. Kim et al.
CGMY KR
T M̃ G̃ k̃+ k̃− r̃+ r̃−
0.0880 106.5827 96.1341 5325.8 33.727 0.0065 0.0330
0.1840 103.4463 93.3887 9126.3 33.024 0.0066 0.034
0.4360 92.4701 83.7430 4757.3 31.327 0.0074 0.0381
0.6920 89.4576 81.0851 3866.4 30.776 0.0076 0.0395
0.9360 90.0040 81.5675 6655.4 30.78 0.0075 0.03953
1.1920 82.6216 75.0354 9896.7 29.483 0.0079 0.0430
1.7080 77.3594 70.3609 10000 28.468 0.0084 0.046
5 Conclusion
References
[9] Cont, R. and Tankov P. (2004). Financial Modelling with Jump Processes,
Chapman & Hall/CRC, London.
[10] D’Agostino, R. B. and Stephens, M. A. (1986). Goodness of Fit Tech-
niques, Dekker, New York.
[11] Kawai, R. (2004). Contributions to Infinite Divisibility for Financial mod-
eling, Ph.D. thesis, http://hdl.handle.net/1853/4888.
[12] Kim, Y. S., Rachev, S. T., Chung, D. M., and Bianchi. M. L. The Modi-
fied Tempered Stable Distribution, GARCH-Models and Option Pricing,
Probability and Mathematical Statistics, to appear.
[13] Kim, Y. S. and Lee, J. H. (2007). The Relative Entropy in CGMY Pro-
cesses and Its Applications to Finance, to appear in Mathematical Meth-
ods of Operations Research.
[14] Koponen, I. (1995). Analytic Approach to the Problem of Convergence of
Truncated Lévy Flights towards the Gaussian Stochastic Process, Phys-
ical Review E, 52, 1197–1199.
[15] Lewis, A. L. (2001). A Simple Option Formula for General Jump-
Diffusion and Other Exponential Levy Processes, avaible from http://
www.optioncity.net.
[16] Lukacs, E. (1970). Characteristic Functions, 2nd Ed, Griffin, London.
[17] Mandelbrot, B. B. (1963). New Methods in Statistical Economics, Jour-
nal of Political Economy, 71, 421–440.
[18] Marsaglia, G., Tsang, W. W. and Wang, G. (2003). Evaluating Kol-
mogorov’s Distribution, Journal of Statistical Software, 8, 18.
[19] Marsaglia, G. and Marsaglia, J. (2004). Evaluating the Anderson-Darling
Distribution, Journal of Statistical Software, 9, 2.
[20] Fujiwara, T. and Miyahara, Y. (2003). The Minimal Entropy Martingale
Measures for Geometric Lévy Processes, Finance & Stochastics 7, 509–
531.
[21] Rachev, S. and Mitnik S. (2000). Stable Paretian Models in Finance,
Wiley, New York.
[22] Rachev, S., Menn C., and Fabozzi F. J. (2005). Fat-Tailed and Skewed
Asset Return Distributions: Implications for Risk Management, Portfolio
selection, and Option Pricing, Wiley, New York.
[23] Rosiński, J. (2006). Tempering Stable Processes, Working Paper,
http://www.math.utk.edu/˜rosinski/Manuscripts/tstableF.pdf.
[24] Sato, K. (1999). Lévy Processes and Infinitely Divisible Distributions,
Cambridge University Press, Cambridge.
[25] Schoutens, W. (2003). Lévy Processes in Finance: Pricing Financial
Derivatives, Wiley.
[26] Shao, J. (2003). Mathematical Statistics, 2nd Ed, Springer, Berlin
Heidelberg New York.
A New Tempered Stable Distribution 109
Appendix
Exponential CGMY Model
The CGMY process is a pure jump process, introduced by Carr et al. [7].
Definition 5.1 A Lévy process (Xt )t≥0 is called a CGMY process with pa-
rameters (C, G, M, Y, m) if the characteristic function of Xt is given by
φXt (u; C, G, M, Y, m)
= exp(iumt + tCΓ (−Y )((M − iu)Y − M Y + (G + iu)Y − GY )), u ∈ R.
1 Introduction
Classical models in financial risk management and portfolio optimization such
as the Markowitz portfolio optimization approach are based on the assumption
that risk factor returns and stock returns are normally distributed. Since the
seminal work of [9] and further investigations by [3], [5], [6], [10], [11], [13],
and [17] there has been overwhelming empirical evidences that the normal
distribution must be rejected. These investigations led to the conclusion that
112 S. Kring et al.
marginal distributions of risk factors and stock returns exhibit skewness and
leptokurtosis, i.e., phenomena that cannot be explained by the normal distri-
bution.
Stable or α-stable distributions have been suggested by the authors above
for modeling these pecularities of financial time series. Beside the fact that
α-stable distributions capture these phenomena very well, they have further
attractive features which allow them to generalize Gaussian-based financial
theory. First, they have the property of stability meaning, that a finite sum
of independent and identically distributed (i.i.d.) α-stable distributions is a
stable distribution. Second, this class of distribution allows for the generalized
Central Limit Theorem: A normalized sum of i.i.d. random variables converges
in distribution to an α-stable random vector.
A drawback of stable distributions is that, with a few exceptions, they
do not know any analytic expressions for their densities. In the univariate
case, this obstacle could be negotiated by numerical approximation based on
new computational possibilities. These new possibilities make the α-stable
distribution also accessible for practitioners in the financial sector, at least, in
the univariate case. The multivariate α-stable case is even much more complex,
allowing for a very rich dependence structure, which is represented by the
so-called spectral measure. In general, the spectral measure is very difficult
to estimate even in low dimensions. This is certainly one of the main reasons
why multivariate α-stable distributions have not been used in many financial
applications.
In financial risk management as well as in portfolio optimization, all the
models are inherently multivariate as stressed by [14]. The multivariate nor-
mal distribution is not appropriate to capture the complex dependence struc-
ture between assets, since it does not allow for modeling tail dependencies
between the assets and leptokurtosis as well as heavy tails of the marginal re-
turn distributions. In many models for market risk management multivariate
elliptical distributions, e.g. t-distribution or symmetric generalized hyperbolic
distributions, are applied. They model better than the multivariate normal
distributions (MNDs) the dependence structure of assets and offer an efficient
estimation procedure. In general, elliptical distributions (EDs) are an exten-
sion of MNDs since they are also elliptically contoured and characterized by
the so-called dispersion matrix. The dispersion matrix equals the variance
covariance matrix up to a scaling constants if second moments of the dis-
tributions exist, and has a similar interpretation as the variance-covariance
matrix for MNDs. In empirical studies1 it is shown that especially data of
multivariate asset returns are roughly elliptically contoured.
In this paper, we focus on multivariate α-stable sub-Gaussian distributions
(MSSDs). In two aspects they are a very natural extension of the MNDs. First,
they have the stability property and allow for the generalized Central Limit
Theorem, important features making them attractive for financial theory. Sec-
ond, they belong to the class of EDs implying that any linear combination
1
For further information, see [14]
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 113
holds for all n ∈ N with cn > 0 and dn ∈ R, then we call X stable or α-stable
distributed.
The definition justifies the term stable because the sum of i.i.d. random vari-
ables has the same distribution as X up to a scale and shift parameter. One
can show that the constant cn in Definition 1 equals n1/α .
The next definition represents univariate α-stable distributions in terms
of their characteristic functions and determines the parametric family which
describes univariate stable distributions.
114 S. Kring et al.
lim λα P (X > λ) → C+
λ→∞
lim λα P (X < λ) → C− .
λ→−∞
d
n
X= Xn,k .
k=1
Lemma 1. Let X ∼ Sα (σ, β, µ), 1 < α < 2 and β = 0. Then for any
0 < p < α there exists a constant cα,β (p) such that:
To get a first feeling for the sort of data we are dealing with, we display
in Fig. 1 the kernel density plots of the empirical returns, the Gaussian fit
and the α-stable fit of some representative stocks. We can clearly discern
the individual areas in the plot where the normal fit causes problems. It is
around the mode where the empirical peak is too high to be captured by
the Gaussian parameters. Moreover, in the mediocre parts of the tails, the
empirical distribution attributes less weight than the Gaussian distribution.
And finally, the tails are underestimated, again. In contrast to the Gaussian,
the stable distribution appears to account for all these features of the empirical
distribution quite well.
116 S. Kring et al.
35
Empirical Density
Stable Fit
Gaussian (Normal) Fit
30
25
20
15
10
0
−0.1 −0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 0.08 0.1
Fig. 1. Kernel density plots of Adidas AG: empirical, normal, and stable fits
2
In Fig. 2 we remove the two most extreme points in the upper and lower tails,
respectively.
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 117
0.08
0.06
0.04
0.02
empirical
−0.02
−0.04
−0.06
−0.08
−0.1
−0.04 −0.03 −0.02 −0.01 0 0.01 0.02 0.03 0.04 0.05
normal
0.1
0.08
0.06
0.04
0.02
empirical
−0.02
−0.04
−0.06
−0.08
−0.1
−0.1 −0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 0.08 0.1
stable
Fig. 2. Adidas AG quantile plots of empirical return percentiles vs. normal (top)
and stable (bottom) fits
of the stable random vector. In general, stable random vectors are difficult
to use for financial modeling, because the spectral measure is difficult to esti-
mate even in low dimensions. For stable financial model building, one has to
focus on certain subclasses of stable random vectors where the spectral mea-
sure has an easier representation. Such a subclass is the multivariate α-stable
sub-Gaussian law. They are obtained by multiplying a Gaussian vector by
W 1/2 where W is a stable random variable totally skewed to the right. Stable
118 S. Kring et al.
P (X ∈ A) = P (−X ∈ A)
(ii) If α = 1,
it X 2
E(e ) = exp{− |(t, s)|(1 + i sign((t, s)) ln |(t, s)|)Γ (ds) + i(t, µ)}
S d−1 π
In contrast to the univariate case, stable random vectors have not been
applied frequently in financial modeling. The reason is that the spectral mea-
sure, as a measure on the unit sphere S d−1 , is extremely difficult to estimate
even in low dimensions. (For further information see [17] and [15].)
Another way to describe stable random vectors is in terms of linear pro-
jections. We know from Theorem 1 that any linear combination
d
(b, X) = bi X i
i=1
The covariance between two normal random variables X and Y can be inter-
preted as the inner product of the space L2 (Ω, A, P). The covariation is the
analogue of two α-stable random variables X and Y in the space Lα (Ω, A, P).
Unfortunately, Lα (Ω, A, P) is not a Hilbert space and this is why it lacks some
of the desirable and strong properties of the covariance. It follows immediately
from the definition that the covariation is linear in the first argument. Unfor-
tunately, this statement is not true for the second argument. In the case of
α = 2, the covariation equals the covariance.
In general, as pointed out in the last section, α-stable random vectors have
a complex dependence structure defined by the spectral measure. Since this
measure is very difficult to estimate even in low dimensions, we have to retract
to certain subclasses, where the spectral measure becomes simpler. One of
these special classes is the multivariate α-stable sub-Gaussian distribution.
For α-stable sub-Gaussian random vectors, we do not need the spectral mea-
sure in the characteristic functions. This fact simplifies the calculation of the
projection functions.
2
If we choose σ(a) = ( 12 a Σa)1/2 , β(a) = 0 and µ(a) = a µ, then for all t ∈ R
we have
+ + πα , ,
E(exp(it(a X))) = exp −σ(a)α |t|α 1 − iβ(a) tan (sign t) + iµ(a)t .
2
σ 2 (ei + ej ) − σ 2 (ei − ej )
σij = .
2
Proposition 3. Let X = (X1 , . . . , Xn ) be a zero mean α-stable sub-Gaussian
random vector with dispersion matrix Σ. Then it follows
0.1 0.1
0.08 0.08
0.06 0.06
0.04 0.04
0.02 0.02
DCX
DBK
0 0
−0.02 −0.02
−0.04 −0.04
−0.06 −0.06
−0.08 −0.08
−0.1 −0.1
−0.1 −0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 0.08 0.1 −0.15 −0.1 −0.05 0 0.05 0.1 0.15 0.2
BMW CBK
(a) (b)
a X = ||a||X1
d
for any orthogonal matrix. The equation is√true, since Z is rotationally sym-
metric. Hence any linear combination of W Z is an elliptical random vec-
tor. The characteristic function of an α-stable sub-Gaussian random vector is
given by
1
E(eit X ) = eit µ e−( 2 t Σt)
α/2
X = µ + W 1/2 AZ
where
(i) Z ∼ Nd (0, Id );
(ii) W ≥ 0 is a non-negative, scalar-valued random variable which is inde-
pendent of G, and
(iii) A ∈ Rd×d and µ ∈ Rd are a matrix of constants, respectively.
We call a random variable X with NVMD a normal variance mixture
(NVM). We observe that Xw = (X|W = w) ∼ Nd (µ, wΣ), where Σ = AA .
126 S. Kring et al.
λu := λu (X1 , X2 ) := lim− P (X2 > F2← (q)|X1 > F1← (q)), (2)
q→1
provided a limit λu ∈ [0, 1] exists. If λu ∈ (0, 1], then X1 and X2 are said to
show upper tail dependence; if λu = 0, they are asymptotically independent in
the upper tail. Analogously, the coefficient of the lower tail dependence is
Equation (6) shows that beside the fact that a normal distribution is not
heavy tailed the components are asymptotically independent. This, again, is
a contradiction to empirical investigations of market behavior. Especially, in
extreme market situations, when a financial market declines in value, market
participants tend to behave homogeneously, i.e., they leave the market and sell
their assets. This behavior causes losses in many assets simultaneously. This
phenomenon can only be captured by distributions which are asymptotically
dependent.
[12] optimizes the risk and return behavior of a portfolio based on the
expected returns and the covariances of the returns in the considered asset
universe. The risk of a portfolio consisting of these assets is measured by the
variance of the portfolio return. In addition, he assumes that the asset returns
follow a multivariate normal distribution with mean µ and covariance Σ. This
approach leads to the following optimization problem
min w Σw,
w∈Rd
subject to
w µ = µp
w 1 = 1.
This approach can be extended in two ways. First, we can replace the
assumption of normally distributed asset returns by elliptically distributed
asset returns and second, instead of using the variance as the risk measure,
we can apply any positive-homogeneous, translation-invariant measure of risk
to rank risk or to determine the optimal risk-minimizing portfolio. In general,
due to the work of [1], a risk measure is a real-valued function : M → R,
where M ⊂ L0 (Ω, F, P ) is a convex cone. L0 (Ω, F, P ) is the set of all almost
surely finite random variables. The risk measure is translation invariant if
for all L ∈ M and every l ∈ R, we have (L + l) = (L) + l. It is positive-
homogeneous if for all λ > 0, we have (λL) = λ(L). Note, that value-at-risk
(VaR) as well as conditional value-at-risk (CVaR) fulfill these two properties.
The last theorem stresses that the dispersion matrix contains all the informa-
tion for the management of risk. In particular, the tail index of an elliptical
random vector has no influence on optimizing risk. Of course, the index has
an impact on the value of the particular risk measure like VaR or CVaR, but
not on the weights of the optimal portfolio, due to the Markowitz approach.
In risk management, we have very often to deal with portfolios consisting
of many different assets. In many of these cases it is important to reduce the
dimensionality of the problem in order to not only understand the portfolio’s
risk but also to forecast the risk. A classical method to reduce the dimension-
ality of a portfolio whose assets are highly correlated is principal component
analysis (PCA). PCA is based on the spectral decomposition theorem. Any
symmetric or positive definite matrix Σ can be decomposed in
Σ = P DP ,
More generally,
From equation (8), we can derive that the linear combination Y1 = P1 (X − µ)
has the highest dispersion of all linear combinations and Pi X has the highest
dispersion in the linear subspace {P1 , ..., Pi−1 }⊥ . If we interpret trace Σ =
d
j=1 σii as a measure of total variability in X and since we have
d
d
d
Pi ΣPi = λi = trace Σ = σii ,
i=1 i=1 i=1
Y = P (X − µ),
X = µ + P Y.
X = µ + P1 Y1 + P2 Y2 = µ + P1 Y1 + .
But one has to be careful. In contrast to the normal distribution case, the prin-
cipal components are only quasi-uncorrelated but not independent. Further-
more, we obtain for the coefficient of tail dependence between two principal
components, say Yi and Yj ,
0 √1/2 sα
√
0 2 ds
λ(Yi , Yj , 0, α) = 0 1 α1−s .
√ s
0 1−s2
ds
where p ∈ (1, α), cα,0 (p) = E(|Y |p )1/p > 0 and Y ∼ Sα (1, 0, 0).
(b) Let X1 , X2 , . . . , Xn be independent and identically distributed samples with
the same distribution as the random vector X. Let σ̂j be a consistent
estimator for σj , the scale parameter of the jth component of X, then, the
(2)
estimator σ̂ij (n, p), defined as
(2) 2 2−p 1
n
<p−1>
σ̂ij (n, p) = σ̂ Xti Xtj , (10)
cα,0 (p)p j n t=1
is a consistent estimator for σij , where Xti refers to the ith entries of the
observation Xt , t = 1, . . . , n, cα,0 (p) = E(|Y |p )1/p and Y ∼ Sα (1, 0, 0).
Proof. (a) Due to the Proposition 3 we have
Prop. 3 (2−p)/2
σij = 2α/2 σjj [Xi , Xj ]α
Prop.1 (2−α)/2
= 2α/2 σjj E(Xi Xj<p−1> )σjα /E(|Xj |p )
Lemma 1 (2−α)/2
= 2α/2 σjj E(Xi Xj<p−1> )σjα /(cα,0 (p)p σjp )
Corollary 1(i) (2−p)/2
= 2p/2 σjj E(Xi Xj<p−1> )/(cα,0 (p)p )
(b) The estimator σ̂j is consistent and f (x) = x2−p is continuous. Then, the
n
estimator σ̂j2−p is consistent for σj2−p . n1 k=1 Xki Xkj
<p−1>
is consistent
for E(Xi Xj<p−1> ) due to the law of large numbers. Since the product of
two consistent estimators is consistent, the estimator
1
n
(2) 2
σ̂ij = = p
σ̂j2−p <p−1>
Xti Xtj
cα,0 (p) n t=1
is consistent.
for all θ ∈ Θ.
We see from (11) that the covariance matrix of Z determines the dispersion
matrix of Xθ up to a scaling constant.
c(θ, p) > 0.
Since we have xp ≤ max{1, xα(θ) } for p ∈ (0, α(θ)) and x > 0, it follows
from Lebesque’s Theorem
p
lim c(θ, p) = lim E( Wθ ) lim E(|Z̃|p )
p→0 p→0 p→0
p
= E( lim Wθ )E( lim |Z̃|p )
p→0 p→0
= E(1)E(1)
= 1.
4
Note, if the random variable X has tail parameter α then E(|X|p ) < ∞ for all
p < α and E(|X|p ) = ∞ for all p ≥ α (see [19]).
134 S. Kring et al.
if p < α(θ)/2.
Proof. (i) follows directly from Lemma 2. For statement (ii), we have shown
that
The inequation (∗) holds because of the Chebyshev’s inequality and we have
E(|a (X − µ)|2p ) < ∞ because of the assumption p < α(θ)/2.
Note, that σ̂n (p, a)2/p , a ∈ Rd , is a biased, but consistent estimator for (aΣa ).
However, since we cannot determine c(θ, p) > 0 we have to use
1
n
p
σ̂n (p, a)c(θ, p) = |a (Xi − µ)|p (14)
n i=1
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 135
as the estimator. But then, Theorem 8 allows us the estimate the disper-
sion matrix only up to a scaling constant by using linear combinations
a X1 , . . . , a Xn , a ∈ Rd of the observations X1 , . . . , Xn . We can apply two
different approaches to do this.
The first approach is based on the fact that the following equation holds
= (a Σa)1/2 .
The last equation is true because of (ii) of the following theorem. The proof
of the equality (*) can be found in [21].
1 ( n
σ̂n (a) = |a (Xi − µ)|1/n
c(θ, 1/n) i=1
Proof. (i) follows directly from Lemma 2. For statement (ii), we have shown
that
The inequation (∗) holds because of the Chebyshev’s inequality. Then (ii)
follows from (12) in Lemma 2.
Note, that σ̂n2 (a), a ∈ Rd , is a biased but consistent estimator for (a Σa).
For the rest of this section we concentrate on α-stable sub-Gaussian ran-
dom vectors. In this case, the family of positive random variables (Wθ )θ∈Θ is
given by
πα
(Wα )α∈(0,2) and Wα ∼ Sα/2 (cos( ), 1, 0).
4
Furthermore, the scaling function c(., .) defined in Lemma 2 satisfies
2 )Γ (1 − p/α)
Γ ( p+1
c(α, p)p = 2p √
Γ (1 − p/2) π
2 + πp , + p,
= sin Γ (p)Γ 1 − , (17)
π 2 α
where Γ (.) is the Gamma-function. For the proof of (17), see [7] and [21].
With Theorems 8 and 9, we derive two estimators for the scale parameter
σ(a) of the linear projection a X for an α-stable sub-Gaussian random vector
X. The first one is
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 137
1 2 + πp , + p,
−1
n
σ̂n (p, a) = sin Γ (p) Γ 1 − |a Xi − µ(a)|p
n π 2 α i=1
1
n
α̂ = α̂k .
n
k=1
15
13
11
9
7
5
17
15
Values
13
11
16
15
14
13
Values
12
11
10
9
8
7
alpha=1.5 alpha=1.6 alpha=1.7 alpha=1.8 alpha=1.9
14
13
12
Values
11
10
Finally in Fig. 7, the interquartile range is equal to about 1.45 for all values of
α. The rate of decay is roughly n−1/2 . Extreme outliers can be observed for
small sample sizes larger than twice the median, regardless of the value of α.
For n = 1, 000, we have a maximal error around about 1.5 times the median.
Due to right-skewness, extreme values are observed mostly to the right of the
median.
(2)
5.2 Empirical Analysis of σ̂ij (n, p)
We examine the consistency behavior of the second estimator as defined in
(19) again using boxplots. In Fig. 5 through 12 we depict the statistical behav-
ior of the estimator. For generating independent samples of various lengths
for α = 1.5, 1.6, 1.7, 1.8, and 1.9, and two different values of p we use the al-
gorithm described in Sect. 4.4.5 For the values of p, we select 1.0001 and 1.3,
respectively. A value for p closer to one leads to improved properties of the
estimator as will be seen.
5
In most of these plots, extreme estimates had to be removed to provide for a clear
display of the boxplots.
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 141
30
25
20
Values
15
10
30
25
20
Values
15
10
−5
−10
alpha=1.5 alpha=1.6 alpha=1.7 alpha=1.8 alpha=1.9
980 estimations of 1000 are shown in each boxplot; p=1.3
In general, we can observe that the estimates are strongly skewed. This
is more pronounced for lower values of α while skewness vanishes slightly for
increasing α. All figures display a noticeable bias in the median towards low
(1) (2)
values. Finally, as will be seen, σ̂ij (n) seems more appealing than σ̂ij (n, p).
For a sample length of n = 100, Figs. 8 and 9 show that the bodies of
the boxplots which are represented by the innerquartile ranges are as high as
4.5 for a lower value of p and α. As α increases, this effect vanishes slightly.
However, results are worse for p = 1.3 as already indicated. For sample lengths
of n = 300, Figs. 10 and 11 show interquartile ranges between 1.9 and 2.4 for
lower values of p. Again, results are worse for p = 1.3. For n = 500, Figs. 12
and 13 reveal ranges between 1.3 and 2.3 as α increases. Again, this worsens
when p increases. And finally for samples of length n = 1, 000, Figs. 14 and
15 indicate that for p = 1.00001 the interquartile ranges extend between 1
for α = 1.9 and 1.5 for α = 1.5. Depending on α, the same pattern but on a
worse level is displayed for p = 1.3.
142 S. Kring et al.
18
16
14
Values
12
10
6
alpha=1.5 alpha=1.6 alpha=1.7 alpha=1.8 alpha=1.9
980 estimations of 1000 are shown in each boxplot; p=1.00001
25
20
Values
15
10
16
14
12
Values
10
22
20
18
16
Values
14
12
10
4
alpha=1.5 alpha=1.6 alpha=1.7 alpha=1.8 alpha=1.9
980 estimations of 1000 are shown in each boxplot; p=1.3
15
14
13
Values
12
11
10
7
alpha=1.5 alpha=1.6 alpha=1.7 alpha=1.8 alpha=1.9
980 estimations of 1000 are shown in each boxplot; p=1.00001
25
20
Values
15
10
5
alpha=1.5 alpha=1.6 alpha=1.7 alpha=1.8 alpha=1.9
980 estimations of 1000 are shown in each boxplot; p=1.3
0.2 0.2
0.15
0.15
0.1
0.1
0.05
MAN
LHF
0.05
0
0
−0.05
−0.05
−0.1
−0.1
−0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 0.08 0.1 0.12 −0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 0.08 0.1
BAS CON
(a) (b)
Fig. 16. Bivariate Scatterplots of BASF and Lufthansa in (a); and of Continental
and MAN in (b)
2 1
1.9 0.8
1.8 0.6
1.7 0.4
1.6 0.2
1.5 0
1.4 −0.2
1.3 −0.4
1.2 −0.6
1.1 −0.8
1 −1
0 20 40 60 80 100 0 20 40 60 80 100
Linear combinations
α β
Fig. 17. Scatterplot of the estimated α’s and β’s for 100 linear combinations
that we incorporate more information from the dataset and we can generate
more sample estimates α̂i and β̂i . In the former approach, we analyze only
the marginal distributions.
Figure 17 depicts the maximum likelihood estimates for 100 linear com-
binations due to (ii). We observe that the estimated α̂i , i = 1, . . . , n, range
from 1.5 to 1.84. The average, ᾱ, equals 1.69. Compared to the first approach,
the tail indices increase, meaning less leptokurtosis, but the range of the es-
timates decreases. The estimated β̂i ’s, i = 1, . . . , n, lie in a range of −0.4 and
0.4 and the average, β̄, is −0.0129. In contrast to the first approach, the vari-
ability in the β’s increases. It is certainly not to be expected that the DAX30
log-returns follow a pure i.i.d. α stable sub-Gaussian model, since we do not
account for time dependencies of the returns. The variability of the estimated
α̂’s might be explained with GARCH-effects such as clustering of volatility.
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 147
ADS
ALV
ALT
BAS
BMW
BAY
CBK
CON
DCX
DBK
DB1
DPW
DTE
EOA
FME
HEN3
IFX
LIN
LHA
MAN
MEO
MUV2
RWE
SAP
SCH
SIE
TKA
TUI
VOW
ADS
ALV
ALT
BAS
BMW
BAY
CBK
CON
DCX
DBK
DB1
DPW
DTE
EOA
FME
HEN3
IFX
LIN
LHA
MAN
MEO
MUV2
RWE
SAP
SCH
SIE
TKA
TUI
VOW
Fig. 18. Heat map of the sample dispersion matrix. Dark blue colors corresponds
to low values (min=0.0000278), to blue, to green, to yellow, to red for high values
(max=0,00051)8
7
The estimated β̂’s differ sometimes significantly from zero.
8
To obtain the heat map in color, please contact the authors.
148 S. Kring et al.
ADS
ALV
ALT
BAS
BMW
BAY
CBK
CON
DCX
DBK
DB1
DPW
DTE
EOA
FME
HEN3
IFX
LIN
LHA
MAN
MEO
MUV2
RWE
SAP
SCH
SIE
TKA
TUI
VOW
ADS
ALV
ALT
BAS
BMW
BAY
CBK
CON
DCX
DBK
DB1
DPW
DTE
EOA
FME
HEN3
IFX
LIN
LHA
MAN
MEO
MUV2
RWE
SAP
SCH
SIE
TKA
TUI
VOW
Fig. 19. Heat map of the sample covariance matrix. Dark blue colors corresponds
to low values (min=0.000053), to blue, to green, to yellow, to red for high values
(max=0,00097)9
−3 −3
x 10 x 10
3.5 7
3 6
2.5 5
Dispersion
Variance
2 4
1.5 3
1 2
0.5 1
0 0
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29
(a) (b)
Fig. 20. Barplots (a) and (b) depict the eigenvalues of the sample dispersion matrix
and the sample covariance matrix
9
To obtain the heat map in color, please contact the authors.
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 149
100 % 100%
90% 90%
80%
80%
70% 70%
60% 60%
50% 50%
40% 40%
30% 30%
20% 20%
10% 10%
0% 0%
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 1 2 3 4 5 6 7 8 9 10 11 1213 14 15 16 17 1819 20 21 22 23 2425 26 27 28 29
(a) (b)
Figure 21 (a) and (b) depict the cumulative proportion of the total vari-
ability explained by the first k principal components corresponding to the k
largest eigenvalues. In both figures, more than 50% is explained by the first
principal component. We observe that the first principal component in the sta-
ble case explains slightly more variability than in the ordinary case, e.g. 70%
of the total amount of dispersion is captured by the first six stable components
whereas in the normal case, only 65% is explained. In contrast to the normal
PCA the stable components are not independent but quasi-uncorrelated. Fur-
thermore, in the case of α = 1.69, the coefficient of tail dependence for two
principal components, say Yi and Yj , is
0 √1/2 s1.69
√ ds
0 1−s2
λ(Yi , Yj , 0, 1.69) = 0 1 1.69 ≈ 0.21
√s
0 1−s2
ds
ADS ADS
ALV ALV
ALT ALT
BAS BAS
BMW BMW
BAY BAY
CBK CBK
CON CON
DCX DCX
DBK DBK
DB1 DB1
DPW DPW
DTE DTE
EOA EOA
FME FME
HEN3 HEN3
IFX IFX
LIN LIN
LHA LHA
MAN MAN
MEO MEO
MUV2 MUV2
RWE RWE
SAP SAP
SCH SCH
SIE SIE
TKA TKA
TUI TUI
VOW VOW
0 0.05 0.1 0.15 0.2 0.25 0.3 −0.2 −0.1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8
(a) (b)
ADS ADS
ALV ALV
ALT ALT
BAS BAS
BMW BMW
BAY BAY
CBK CBK
CON CON
DCX DCX
DBK DBK
DB1 DB1
DPW DPW
DTE DTE
EOA EOA
FME FME
HEN3 HEN3
IFX IFX
LIN LIN
LHA LHA
MAN MAN
MEO MEO
MUV2 MUV2
RWE RWE
SAP SAP
SCH SCH
SIE SIE
TKA TKA
TUI TUI
VOW VOW
−0.5 −0.4 −0.3 −0.2 −0.1 0 0.1 0.2 0.3 −0.4 −0.3 −0.2 −0.1 0 0.1 0.2 0.3 0.4 0.5
(c) (d)
Fig. 22. Barplot summarizing the loadings vectors g1 , g2 , g3 and g4 defining the first
four principal components: (a) factor 1 loadings; (b) factor 2 loadings; (c) factor 3
loadings; and (d) factor 4 loadings
7 Conclusion
In this paper we present different estimators which allow one to estimate the
dispersion matrix of any normal variance mixture distribution. We analyze
the estimators theoretically and show their consistency. We find empirically
(1)
that the estimator σ̂ij (n) has better statistical properties than the estima-
(2)
tor σ̂ij (n, p) for i = j. We fit an α-stable sub-Gaussian distribution to the
DAX30 components for the first time. The sub-Gaussian model is certainly
more realistic than a normal model, since it captures tail dependencies. But
it has still the drawback that it cannot incorporate time dependencies.
Estimation of α-Stable Sub-Gaussian Distributions for Asset Returns 151
Acknowledgement
The authors would like to thank Stoyan Stoyanov and Borjana Racheva-Iotova
from FinAnalytica Inc for providing ML-estimators encoded in MATLAB. For
further information, see [22].
References
[1] Artzner, P., F. Delbaen, J.M. Eber and D. Heath. 1999. Coherent Measure
of Risk. Mathematical Finance 9, 203–228.
[2] Bingham, N.H., R. Kiesel and R. Schmidt. 2003. A Semi-parametric Ap-
proach to Risk Management. Quantitative Finance 3, 241–250.
[3] Cheng, B.N., S.T. Rachev. 1995. Multivariate Stable Securities in Finan-
cial Markets. Mathematical Finance 54, 133–153.
[4] Embrechts, P., A. McNeil and D. Straumann. 1999. Correlation: Pitfalls
and Alternatives. Risk 5, 69–71.
[5] Fama, E. 1965. The Behavior of Stock Market Prices. Journal of Business,
38, 34–105.
[6] Fama, E. 1965. Portfolio Analysis in a Stable Paretian market. Manage-
ment Science, 11, 404–419.
[7] Hardin Jr., C.D. 1984. Skewed Stable Variables and Processes. Technical
report 79, Center for Stochastics Processes at the University of North
Carolina, Chapel Hill.
[8] Höchstötter, M., F.J. Fabozzi and S.T. Rachev. 2005. Distributional Anal-
ysis of the Stocks Comprising the DAX 30. Probability and Mathematical
Statistics, 25, 363–383.
[9] Mandelbrot, B.B. 1963. New Methods in Statistical Economics. Journal
of Political Economy, 71, 421–440.
[10] Mandelbrot, B.B. 1963. The Variation of Certain Speculative Prices.
Journal of Business, 36, 394–419.
[11] Mandelbrot, B.B. 1963. New Methods in Statistical Economics. Journal
of Political Economy, 71, 421–440.
[12] Markowitz, H.M. 1952. Portfolio Selection. Journal of Finance 7, (1),
77–91.
[13] McCulloch, J.H. 1952. Financial Applications of Stable Distributions.
Hanbbook of Statistics-Statistical Methods in Finance, 14, 393–425.
Elsevier Science B.V, Amsterdam.
[14] McNeil, A.J., R. Frey and P. Embrechts. 2005. Quantitative Risk Man-
agement, Princeton University Press, Princeton.
[15] Nolan, J.P., A.K. Panorska, J.H. McCulloch. 2000. Estimation of Stable
Spectral Measure. Mathematical and Computer Modeling 34, 1113–1122.
[16] Nolan, J.P. 2005. Multivariate Stable Densities and Distribution Func-
tions: General and Elliptical Case. Deutsche Bundesbank’s 2005 Annual
Fall Conference.
152 S. Kring et al.
[17] Rachev, S.T., S. Mittnik. 2000. Stable Paretian Models in Finance. Wiley,
New York.
[18] Resnick, S.I. 1987. Extreme Values, Regular Variation, and Point Pro-
cesses. Springer, Berlin.
[19] Samorodnitsky, G., M. Taqqu. 1994. Stable Non-Gaussian Random Pro-
cesses, Chapmann & Hall, New York
[20] Schmidt, R. 2002. Tail Dependence for Elliptically Contoured Distribu-
tions. Mathematical Methods of Operations Research 55, 301–327.
[21] Stoyanov, S.V. 2005. Optimal Portfolio Management in Highly Volatile
Markets. Ph.D. thesis, University of Karlsruhe, Germany.
[22] Stoyanov, S.V., B. Racheva-Iotova 2004. Univariate Stable Laws in the
Fields of Finance-Approximations of Density and Distribution Functions.
Journal of Concrete and Applicable Mathematics, 2/1, 38–57.
Risk Measures for Portfolio Vectors
and Allocation of Risks
Ludger Rüschendorf
1 Introduction
In this paper we survey some recent developments on risk measures for port-
folio vectors and on the allocation of risk problem. The main purpose to study
risk measures for portfolio vectors X = (X1 , . . . , Xd ) is to measure not only
the risk of the marginals separately but to measure the joint risk of X caused
by the variation of the components and their possible dependence.
Thus an important property of risk measures for portfolio vectors is con-
sistency with respect to various classes of convex and dependence orderings.
It turns out that axiomatically defined convex risk measures are consistent
w.r.t. multivariate convex ordering. Two types of examples of risk measures
for portfolio measures are introduced and their consistency properties are in-
vestigated w.r.t. various types of convex resp. dependence orderings.
We introduce the general class of convex risk measures for portfolio vectors.
These have a representation result based on penalized scenario measures. It
turns out that maximal correlation risk measures play in the portfolio case the
same role that average value at risk measures have in one dimensional case.
The second part is concerned with applications of risk measures to the op-
timal risk allocation problem. The optimal risk allocation problem or, equiv-
alently, the problem of risk sharing is the problem to allocate a risk in an
optimal way to n traders endowed with risk measures 1 , . . . , n . This prob-
lem has a long history in mathematical economics and insurance. We show
that the optimal risk allocation problem is well defined only under an equi-
librium condition. This condition can be characterized by the existence of a
common scenario measure. A meaningful modification of the optimal risk allo-
cation problem can be given also for markets without assuming the equilibrium
condition. Optimal solutions are characterized by a suitable dual formulation.
The basic idea of this extension is to restrict the class of admissible allocations
in a proper way. We also discuss briefly some variants of the risk allocation
problem as the capital allocation problem.
154 L. Rüschendorf
where λ (X) = AV @Rλ (X) is the average value at risk 0 (also called expected
shortfall or conditional value at risk), β(µ) = supX∈A (0,1] AV @Rλ (X)µ(dλ)
is the penalty function, and A = {X ∈ L∞ (P ); (X) ≤ 0} is the acceptance
set of (see Kusuoka (2001) and Föllmer and Schied (2004)). Thus in di-
mension d = 1 the average value at risk measures λ are the basic building
blocks of the class of law invariant convex risk measures. For some recent
developments in the area of risk measures see [21].
For portfolio vectors X = (X1 , . . . , Xd ) ∈ L∞ d (P ) on (Ω, A, P ) a risk
measure : L∞ d (P ) → R 1
is called convex risk measure if
M1) X ≥ Y ⇒ (X) ≤ (Y )
M2) (X + mei ) = −m + (X), m ∈ R1
M3) (αX + (1 − α)Y ) ≤ α(X) + (1 − α)(Y ) for all α ∈ (0, 1);
thus is a monotone translation invariant, convex risk functional (see [8, 17]).
Like in d = 1 (X) denotes the smallest amount m to be added to the portfolio
vector X such that X +me1 is acceptable. ei denotes here the i-th unit vector.
A subset A ⊂ L∞ d (P ) with R not contained in A is called (convex)
d
acceptance set, if
(A1) A is closed (and convex)
(A2) Y ∈ A and Y ≤ X implies X ∈ A
(A3) X + mei ∈ A ⇔ X + mej ∈ A.
With
A (X) := inf{m ∈ R; X + me1 ∈ A}
risk measures are identified with their acceptance sets:
(a) If A is a convex acceptance set, then A is a convex risk measure
(b) If is a convex risk measure, then A = {X ∈ L∞d ; (X) ≤ 0} is a convex
acceptance set.
Risk Measures for Portfolio Vectors and Allocation of Risks 155
Let bad (P ) denote the set of finite additive, normed, positive measures
on L∞d (P ). Convex risk measures on portfolio vectors allow a representation
similar to d = 1.
= sup EQ (−X).
X∈A
P
For risk measures which are Fatou-continuous, i.e. Xn → X, (Xn ) uni-
formly bounded implies (X) ≤ lim inf (Xn ), bad (P ) can be replaced by
the class Md1 (P ) of P -continuous, σ-additive normed measures which can be
identified by the class of P -densities D = {(Y1 , . . . , Yd ); Yi ≥ 0, EP Yi = 1,
1 ≤ i ≤ d}.
For coherent risk measures, i.e. homogeneous, subadditive, monotone,
translation invariant risk measures the representation in (2) simplifies to
for some subset A ⊂ D. Thus the maximal correlation risk measures ΨY
are the basic building blocks of all law invariant convex risk measures on
portfolio vectors.
b) For d = 1 the representation in (5) can be shown to be equivalent to the
Kusuoka representation result in (1). For d ≥ 1 optimal couplings as in
the definition of the maximal correlation risk measure ΨY arise, have been
characterized in Rüschendorf and Rachev (1990). There are some examples
where ΨY can be calculated in explicit form but in general one does not have
explicit formulas. Therefore, it is useful to give more explicit constructions
of risk measures for portfolio vectors which generalize the known classes of
one dimensional risk measures. For some partial extensions of distortion
type risk measures see [8, 26].
X ≤F Y if Ef (X) ≤ Ef (Y ), ∀f ∈ F, (7)
i.e. smoothing by conditional expectation reduces the risk (for d = 1 see Schied
(2004) or Föllmer and Schied (2004)).
In insurance mathematics the monotonicity axiom M1) of a risk measure
has to be changed to monotonicity in the usual componentwise ordering. We
shall use the notation Ψ (X) for risk measures satisfying this kind of mono-
tonicity. The relation Ψ (X) = (−X) gives a one to one relation between
risk measures in the financial context and risk measures Ψ in the insurance
context.
A natural idea to construct risk measures for portfolio vectors X is to
measure the risk of some real aggregation of the risk vector like the joint
portfolio or the maximal risk, i.e. to consider
+
d ,
Ψ (X) = Ψ1 Xi or
i=1
(11)
Ψ (X) = Ψ1 (max Xi ),
i
As consequence of a), b) one gets that more positive dependent risk vectors
have higher risks. This extends some classical results on comparison of risk
vectors. Let Fi−1 denote the generalized inverse of the distribution function
Fi of Xi , then
d
d
Xi ≤icx Fi−1 (U ), (16)
i=1 i=1
(see Tchen (1980) and Rüschendorf (1980)). Thus as consequence of (15) and
(17) we conclude under the conditions of (14), (15)
the comonotonic risk vector leads to the highest possible risk under all risk
measures of type ΨM , ΨA . Extensions of (17) to compare risks also of two risk
vectors X, Y are given in [11, 24]. For a review of this type of comparison
results for risk vectors see the survey paper [25].
and Jewell (1979), Chevallier and Müller (1994), and many others). It was
later on extended to risk allocations in financial context (see e.g. Barrieu and
El Karoui (2005) and references therein.
An interesting point is that for translation invariant risk measures i ,
1 ≤ i ≤ n, the principle of Pareto optimal risk allocations is equivalent to
minimizing the total risk. This follows from the separating hyperplane theo-
rem and some simple arguments involving translation invariance. In particular
solutions are not unique and several additional (game theoretic) postulates
like fairness have been introduced to single out specific solutions of the risk
sharing problem. For example Chevallier and Müller (1994) single out condi-
tions which yield as possible solutions only portfolio insurance, tactical asset
allocation, and collar strategies. Classical results are the derivation of linear
quota sharing rules and of stop loss contracts as optimal sharing rules.
We discuss in the following some developments on the risk allocation
problem in the case where i are coherent risk measures with representa-
tion i (X) = supQ∈Pi EQ (−X) and scenario measures Pi . The more general
case of convex risk measures is discussed in [7, 9].
There is a naturally associated equilibrium condition coming from similar
equilibria conditions in game theory saying that in a balance of supply and
demand it is not possible to lower some risks without increasing others. In
formal terms this condition is formulated as:
n
(E) If Xi ∈ L∞ (P ) satisfy Xi = 0 and i (Xi ) ≤ 0, ∀i, then i (Xi ) = 0, ∀i.
i=1
which describes the optimal reachable total risk of an allocation. Both risk
measures have been considered in the literature (see [12, 2]).
It turns out (see [7]) that
X(ω) ≥ 0 ⇒ Xi (ω) ≥ 0
(25)
X(ω) ≤ 0 ⇒ Xi (ω) ≤ 0.
Considering the connection with multiple decision problems and using a non-
convex version of the minimax theorem we get the following dual represen-
tation of ∗ , which essentially simplifies the calculation (see Burgert and
Rüschendorf
; (2005)).
< Let X− , X+ denote the negative (positive) parts of
X and Pj , Pj denote the lattice supremum resp.; infimum <of Pj . Thus
in the case of P -continuous probability measure Pj , Pj and Pj are the
probability measures with the max resp. inf of the P -densities as their density
with respect to P .
Risk Measures for Portfolio Vectors and Allocation of Risks 161
Part b) says that our chosen restrictions on decompositions are not too
restrictive since as a result of them we get the largest possible coherent risk
measure below i . Several related classes of restrictions can be given which
lead to the same coherent risk measure. In particular we get a new useful
coherent risk measure describing the value of the total risk of the optimal
modified risk allocation problem.
A different new type of restrictions on the allocation problem has been
introduced in a recent paper by Filipovic
n and Kupper (2006) who consider
for a given risk allocation X = i=1 i as admissible risk transfers only
C
allocations of the form
n
X= Xi with Xi = Ci + xi · Z, (27)
i=1
N +
N ,
ki = Xi (29)
i=1 i=1
Using that is continuous from below Corollary 4.35 of Föllmer and Schied
(2004) implies the existence of some Q∗ ∈ P such that the supremum in (31)
is attained in Q∗ and with ki∗ = EQ∗ (−Xi ) holds
+ N , N
k = EQ∗ − Xi = ki∗ . (32)
i=1 i=1
Thus k1∗ , . . . , kN
∗
is a fair allocation of the risk capital.
Risk Measures for Portfolio Vectors and Allocation of Risks 163
References
[20] D. Müller and D. Stoyan. Comparison Methods for Stochastic Models and
Risks. Wiley, 2002.
[21] Risk Measures and Their Applications. Special volume, L. Rüschendorf
(ed.). Statistics & Decisions, vol. 24(1), 2006.
[22] L. Rüschendorf. Inequalities for the expectation of ∆-monotone func-
tions. Zeitschrift für Wahrscheinlichkeitstheorie und verwandte Gebiete,
54:341–349, 1980.
[23] L. Rüschendorf. Solution of statistical optimization problem by rear-
rangement methods. Metrika, 30:55–61, 1983.
[24] L. Rüschendorf. Comparison of multivariate risks and positive depen-
dence. J. Appl. Probab., 41:391–406, 2004.
[25] L. Rüschendorf. Stochastic ordering of risks, influence of dependence and
a.s. constructions. In N. Balakrishnan, I. G. Bairamov, and O. L. Gebi-
zlioglu, editors, Advances in Models, Characterizations and Applications,
volume 180 of Statistics: Textbooks and Monographs, pages 19–56. CRC
Press, 2005.
[26] L. Rüschendorf. Law invariant risk measures for portfolio vectors.
Statistics & Decisions, 24(1), 2006, 97–108.
[27] L. Rüschendorf and S. T. Rachev. A characterization of random variables
with minimum L2 -distance. Journal of Multivariate Analysis, 1:48–54,
1990.
[28] A. Schied. On the Neyman–Person problem for law invariant risk mea-
sures and robust utility functionals. Ann. Appl. Prob., 3:1398–1423, 2004.
[29] A. H. Tchen. Inequalities for distributions with given marginals. Ann.
Prob., 8:814–827, 1980.
The Road to Hedge Fund Replication:
The Very First Steps
Lars Jaeger
1 Introduction
The debate on sources of hedge fund returns is one of the subjects creating the
most heated discussion within the hedge fund industry. The industry thereby
appears to be split in two camps: Following results of substantial research, the
proponents on the one side claim that the essential part of hedge fund returns
come from the funds’ exposure to systematic risks, i.e. comes from their betas.
Conversely, the “alpha protagonists” argue that hedge fund returns depend
mostly on the specific skill of the hedge fund managers, a claim that they
express in characterising the hedge fund industry as an “absolute return” or
“alpha generation” industry. As usual, the truth is likely to fall within the
two extremes. Based on an increasing amount of empirical evidence, we can
identify hedge fund returns as a (time-varying) mixture of both, systematic
risk exposures (beta) and skill based absolute returns (alpha). However, the
fundamental question is: How much is beta, and how much is alpha?
There is no consensus definition of ‘alpha’, and correspondingly there is
no consensus model in the hedge fund industry for directly describing the
alpha part of hedge fund returns. We define alpha as the part of the return
that cannot be explained by the exposure to systematic risk factors in the
global capital markets and is thus the return part that stems from the unique
ability and skill set of the hedge fund manager. There is more agreement in
modeling the beta returns, i.e. the systematic risk exposures of hedge funds,
which will give us a starting point for decomposition of hedge fund returns
into ‘alpha’ and ‘beta’ components. We begin with stating the obvious: It is
generally not easy to isolate the alpha from the beta in any active investment
strategy. But for hedge funds it is not just difficult to separate the two, it is
already quite troublesome to distinguish them. We are simply not in a position
to give the precise breakdown yet. In other words, the current excitement
about hedge funds has not yet been subject to the necessary amount and
depth of academic scrutiny. However, we argue that the better part of the
confusion around hedge fund returns arises from the inability of conventional
166 L. Jaeger
risk measures and theories to properly measure the diverse risk factors of hedge
funds. This is why only recently progress in academic research has started to
provide us with a better idea about the different systematic risk exposures of
hedge funds and thus give us more precise insights into their return sources.1
Academic research and investors alike begin to realize that that the “search of
alpha” must begin with the “understanding of beta,” the latter constituting
an important – if not the most important - source of hedge fund returns.2
However, at the same time we are starting to realize that hedge fund beta
is different from traditional beta. While both are the result of exposures to
systematic risks in the global capital markets hedge fund beta is more com-
plex than traditional beta. Some investors can live with a rather simple but
illustrative scheme suggested by C. Asness3 : If the specific return is available
only to a handful investors and the scheme of extracting it cannot be simply
specified by a systematic process, then it is most likely real alpha. If it can
be specified in a systematic way, but it involves non-conventional techniques
such as short selling, leverage and the use of derivatives (techniques which are
often used to specifically characterize hedge funds), then it is possibly beta,
however in an alternative form, which we will refer to as “alternative beta.”
In the hedge fund industry “alternative beta” is often sold as alpha, but is
not real alpha as defined here (and elsewhere). If finally extracting the returns
does not require any of these special “hedge fund techniques” but rather “long
only investing,” then it is “traditional beta.”
But how do we model hedge fund returns explicitly and break them down
into alpha, alternative beta and traditional beta? Ultimately, what we are
looking for is a is a general equilibrium model, which relates hedge fund re-
turns to their systematic risk exposures represented by directly observable
market prices in the financial markets, similar to the Capital Asset Pricing
Model for the equity markets.4 This model does not exist yet in its entirety,
but there exists today a growing amount of academic literature on systematic
risk factors and hedge funds’ exposure to them (i.e. their factor loadings), in-
cluding a variety of “alternative beta factors.” We acknowledge that the qual-
ity of the offered model differs strongly for the different hedge fund strategy
1
See the recently published book by Jaeger (2005) and references therein.
2
Martin (2004) makes the pertinent point that measures of alpha inextricably de-
pend on the definition of benchmarks or beta components, going on to identify
ways in which techniques for measuring ‘alpha’ in a traditional asset management
environment are inappropriate or otherwise undermined by the specific charac-
teristics of hedge fund exposures. Moreover, most techniques for measuring hedge
fund alpha tend to reward fund managers for model and benchmark misspecifica-
tion, as imperfect specification of benchmark or ‘beta’ exposure tends to inflate
alpha.
3
Asness (2004).
4
While the CAPM is considered “dead” by most academics, there are extension of
it in various forms that continue to be subject of research. Further the CAPM is
still in extensive use by practioners.
The Road to Hedge Fund Replication 167
5
W. Fung, D. Hsieh, “Extracting Portable Alpha from Equity Long/Short Hedge
Funds” (2004),
6
See “Asset Allocation: Management Style and Performance Measurement” (1992)
by William Sharpe and the articles by Eugene Fama and Kenneth French “Mul-
tifactor explanations of Asset Pricing Anomalies” (1993) and “Common risk fac-
tors in the return of stocks and bonds” (1993). More information can also be
found at the websites of William Sharpe, www.wsharpe.com, and Ken French,
http://mba.tuck.dartmouth.edu/pages/faculty/ken.french/.
168 L. Jaeger
Fung and Hsieh were the first to extend Sharpe’s model to hedge funds
in 1997.7 They employed techniques similar to those Sharpe had applied to
mutual funds five years earlier, but introduced short selling, leverage and
derivatives – three important techniques employed by hedge funds - into their
model. The resulting factor equation would account for all hedge fund return
variation that derives from risk exposure to the risk factors of various asset
classes. Adding alpha to the equation, it allows us to decompose hedge fund
return as:
Hedge fund excess return = Manager’s alpha + Σ (βi * Factori ) + random
fluctuations
Fung and Hsieh performed multifactor regressions of hedge fund returns on
eight asset class indices: US equities, non-US equities, emerging market equi-
ties, US government bonds, non-US government bonds, one-month Eurodollar
deposit rate, gold, and the trade-weighted value of the US dollar. They
identified five risk factors (referred to as style factors), which they defined
as modelling Global Macro, Systematic Trend-Following, Systematic Oppor-
tunistic, Value, Distressed Securities. They further argued that hedge fund
strategies are highly dynamic and create option-like, non-linear, contingent
return profiles. These non-linear profiles, they argued, cannot be modelled
in simple asset class factor models. In their later research they explicitly in-
corporate assets with contingent payout profiles, e.g. options.8 Most of the
studies which have followed show results consistent with Fung and Hsieh.9
The recent literature offers an increasing number of studies around the ques-
tion of common style factor exposure and contingency in payoff profile for
hedge funds.10
As the formula above describes, we infer the hedge funds’ alphas by mea-
suring and subtracting out the betas times the beta factors. We can look at
alpha as the “dark matter” of the hedge fund universe. It can only be mea-
sured by separating everything else out and seeing what is left. In other words,
alpha is never directly observable, but is measured jointly with beta. It can
only be indirectly quantified by separating the beta components out. The ob-
tained value of alpha therefore depends on the chosen risk factors. If we leave
out a relevant factor in the model, the alpha will come out as fictively high.
To draw another analogy, we can equally say that alpha is the garbage bag
of the regression: We account for everything we can, and whatever is left gets
7
Fung, W., Hsieh, D., (1997).
8
The idea of option factors for the purpose of hedge fund modeling was already
introduced in the earliest work on hedge fund models by W. Fung and D. Hsieh,
(1997), and was since then discussed by many academic studies. See their recent
work: W. Fung, D. Hsieh. (2002).
9
See e.g. the article by S. Brown and W. Goetzmann, (2003). The authors identify
eight style factors, i.e. three more than Fung and Hsieh in their research.
10
See W. Fung, D. Hsieh, (2003); (2001); (2001); (2002); V. Agarwal, N. Naik,
(2000); D. Capocci, G. Hübner, (2004).
The Road to Hedge Fund Replication 169
put into alpha. As a consequence, some of the returns not accounted for by
these models are unaccounted beta rather than alpha. Surely, an incomplete
model of systematic risk factors doesn’t mean those additional risk factors do
not exist; only that we do not yet know how to model them. To draw another
image from astronomy, the outer planets of our solar system existed and ex-
erted their gravitational pull long before we had telescopes sensitive enough to
see them. Therefore the formula above on hedge fund returns should actually
read as follows:
Hedge fund return=Manager’s alpha +Σ (βi * Factori(modelled) ) + Σ (βi *
Factori(unmodelled ) + random fluctuations.
A simple example illustrates the problem: Consider a put writing strat-
egy on the S&P 500, or equivalently a covered call writing strategy, as e.g.
represented by the Chicago Board of Trade’s BXM index. To be precise, we
write monthly at-the-money call options on existing equity positions with one
month maturities. On regressing the BXM index against the S&P 500 over a
period of 11 years from 1994 to 2004 we obtain a statistically significant alpha
(i.e. a y-intercept of the regression) of around 0.4% per month, or almost 5%
p.a. There is surely not much true skill driven alpha in writing put options on
equities.11 All or most of the 0.4% is what we refer to as spurious or “phan-
tom” alpha, which results from the imperfect specification of the chosen model
(regression against the S&P 500). So we should not confuse pure manager skill
with an imperfect model. This is a common problem of multi-factor models
in the literature which claim to proof high alphas. We must therefore always
take any statistics of alpha with a grain of salt.
There is in fact little widely published data on historical hedge fund per-
formance, so industry analysis relies mostly on aggregated returns as provided
by a dozen of different index providers which differentiate hedge fund perfor-
mance across the various strategy sectors. Although these indices constitute an
important tool for comparison and possibly benchmarking within and outside
the hedge fund industry, measuring manager performance, classifying invest-
ment styles, and generally creating a higher degree of transparency in this still
rather opaque hedge industry, the results of these efforts vary significantly be-
tween providers and depend more on “committee decisions” regarding index
construction criteria - such as asset weighting, fund selection and chosen sta-
tistical adjustments - than on objectively determined rules. Although this is
also somewhat of a problem in traditional asset class indices, it is severely ex-
acerbated in the hedge fund space by the diverse, dynamic and opaque nature
of the hedge fund universe.
However, the built-in flaws of existing indices have as much to do with
the built-in complexities of hedge funds as with any fault of the index devel-
opers. It is simply more difficult to create unambiguous index construction
guidelines for the heterogeneous hedge fund universe. In particular, while the
construction of traditional asset class indices rests on the reasonably well
founded assumptions that the underlying assets are homogenous, and that
the investor follows a “buy and hold” strategy, hedge funds are diverse and
subject to dynamic change. In traditional asset classes, the average return
of the underlying securities in an index has a strong theoretical basis. It is
constructed to be the return of the “market portfolio,” which is the asset-
weighted combination of all investable assets in that class or a representative
proxy thereof. According to asset pricing theory – e.g. Sharpe’s Capital Asset
Pricing Models (CAPM) - this market portfolio represents exactly the combi-
nation of assets with the optimal risk-return trade-off in market equilibrium.
It is therefore not surprising that traditional equity indices became vehicles
for passive investment only after the development of a clear theoretical foun-
dation in the form of the CAPM.14 Traditional indices are designed to capture
directly a clearly defined risk premium available to investors willing to expose
themselves to the systematic risk of the asset class. So an investor in the S&P
14
It is worth noting here that equity indices remained almost solely performance
analysis tools rather than investment vehicles for many years. The first asset
weighted index tracker fund (on the S&P index) started in 1973, only about five
years after the CAPM became broadly accepted. The very first tracker fund was
launched in 1971 and was equally weighted (on the NYSE). The problem with
equally weighted indices is that they require constant rebalancing to maintain
those weightings, and in the pre-1975 period (i.e. prior to deregulation of stock
commissions) such rebalancing was extremely costly. Wells Fargo launched a cap-
weighted tracker fund in 1973 which enabled them to reduce transactions costs.
Some argue that the predominance of the S&P500 as a benchmark owes more to
the ease of replication than an inherent confidence in the theoretical jusutification
for cap-weighting, see Schoenfeld’s book “Active IUndex Investing” (2004)
The Road to Hedge Fund Replication 171
500 index knows exactly what he is getting; broad exposure to the risks and
risk premia of the US large cap equities market. In other words, there exists
a general equilibrium model in the case of the stock markets. However, such
a model is still missing for the asset class hedge funds.
The standard way to construct a hedge fund index has so far been to use
the average performance of a set of managers15 . However, indices constructed
from averaging single hedge funds inherit the errors and problems of the un-
derlying databases. Therefore they face several performance biases that limit
the usefulness of the result.16 These biases include (but are not limited to):
Survivorship. The survivorship bias is a result of unsuccessful managers
leaving the industry, thus removing unsuccessful funds ex post from the repre-
sentative index. Only their successful counterparts remain; creating a positive
bias. In the most extreme case this is like lining up a number of monkeys,
let them trade in the markets, take out all those that lost money, and then
checking the performance of the rest. The survivors may all be in good shape,
but they hardly represent the performance of the entire original group! Many
hedge fund databases only provide information on currently operating funds,
i.e. funds that have ceased operation are considered uninteresting for the in-
vestor and are purged from the database. This leads to an upwards bias in
the index performance, since the performance of the disappearing funds is
most likely worse than the performance of the surviving funds.17 Consensus
estimates about the size of the survivorship bias in hedge fund databases vary
from 2% to 4%. We note that hedge fund indices are only subject to this bias
to the extent that they are constructed after the fact/inception of the index.
Today index providers do not restate index returns on a going forward basis
as managers drop in and out of their database. Index users should only use
‘live’ index data rather than all historical pro forma data.
Backfilling. A variation of the survivorship bias can occur when a new fund
is included into the index and his past performance is added or “backfilled”
into the database. This induces another upward bias: New managers enter
the database only after a period of good performance, when entry seems most
15
Indices based on average performance of a set of managers have generally well
known pitfalls, already in traditional asset classes. See the article by Jeffrey Bailey
“Are Manager Universes Acceptable Performance Benchmarks,” Spring 1992.
16
Most of these issues are well known by practitioners and are discussed in details
in Chap. 9 of L. Jaeger “Through the Alpha Smoke Screens.” A good overview
of the problems can be found in A. Kohler, “Hedge Fund Indexing: A square
Peg in a round hole,” State Street Global Advisors (2003). See also “Hedge Fund
Indices” by G. Crowder and L. Hennessee, Journal of Alternative Investments,
(2001); “A Review of Alternative Hedge fund Indices.” by Schneeweis Partners
(2001); “Welcome to the Dark Side: Hedge Fund Attrition and Survivorship Bias
over the Period 1994-2001” by G. Amin et al. (2001).
17
The survivorship bias is also well known in the world of mutual funds, see for
example the paper by S. Brown et al., “Survivorship Bias in Performance Studies”
(1992).
172 L. Jaeger
attractive. Since fewer managers enter during periods of bad performance, bad
performance is rarely backfilled into the averages.18 Again, hedge fund indices
are only subject to this bias to the extent that they are constructed after the
fact/inception of the index.
Selection.Unlike public information used to compose equity and bond indices,
hedge fund index providers often rely on hedge fund managers to voluntarily
and correctly submit return data on their funds. Hedge fund managers are
private investment vehicles and are thus not required to make public disclosure
of their activities. Some bluntly refuse to submit data to any index providers.
This “self-selection bias” causes significant distortions in the construction of
the index and often skews the index towards a certain set of managers and
strategies on a going forward basis. Sampling differences produce much of the
performance deviation between the different fund indices. Hedge fund indices
draw their data from different provider, the largest of which are the TASS,
Hedge Fund Research (HFR) and CISDM (formerly MAR) database. These
databases have surprisingly few funds in common, as most hedge funds report
their data – if at all – only to a subset of the databases. Counting studies
have shown that less than one out of three hedge funds in any one database
contributes to the reported returns of all major hedge fund indices19 .
Autocorrelation. Time lags in the valuation of securities (especially for less
liquid strategies like Distressed Securities) held by hedge funds may induce
a smoothening of monthly returns which leads to volatility and correlations
being significantly underestimated. Statistically this effect expresses itself by
significant autocorrelation in hedge fund returns (as will be shown below).
Ironically, the theoretical and practical problems described above do not
disappear when the index is designed to be investable. Some problems are
actually exacerbated. A prerequisite for creating an investment vehicle is that
the underlying managers provide sufficient capacity for new investments. This
creates a severe selection bias, as hedge funds at full capacity (closed) are a
priori not considered in the index. In traditional assets, an investor in the
Dow Jones Industrial Average Index does not need to worry that IBM is
closed for further investment.20 But for hedge fund indices, capacity with top
18
R. Ibbotson estimates this bias to account for a total of up to 4% of reported hedge
fund performance (Presentation at GAIM conference 2004). See also: Brown, S,
Goetzmann, W., Ibbotson, R., “Offshore hedge funds: Survival and performance
1989–1995” (1999). A recent estimate of the backfilling bias is given by B. Malkiel
et. al in their paper “Hedge Funds: Risk and Return” (2004) where the backfilling
bias is estimated in the same region as by Ibbotson.
19
See the study by W. Fung and D. Hsieh, “Hedge Fund Benchmarks: A Risk Based
Approach” (2004)
20
To be more precise, IBM stocks are in fact “closed for further investments” as
there are only a finite number of shares available (assuming no capital increase).
In this way they actually resemble closed hedge funds. However, any investor who
desires can freely purchase IBM shares in the secondary markets (stock markets)
due to its high degree of liquidity (that is what stock markets are all about). In
this sense the comparison serves us well here.
M M
a a Ap Ap
92
94
96
98
100
102
104
106
108
110
100
105
110
115
120
125
130
100
105
110
115
120
125
130
135
140
145
150
100
105
110
115
120
125
130
135
140
Ap r-0
3 Ap r-0
3 M r-03
a
M r-03
a
M r-0 M r-0
a 3 a 3 Ju y-0 Ju y-0
Ju y-0 Ju y-0 n 3 n 3
n- 3 n- 3 Ju -03 Ju -03
Ju 03 Ju 0
3 Aul-03 Aul-03
Aul-03 Aul-03 g Se g-0
Se -03
Seg-0
3
Seg-0 p p 3
O -0
O p-0 p- 3 O -03
c c 3
c 3 O 03
c N t-0 N t-03
N t-0 N t-0 ov 3 o
o 3 o 3 D -0 D v-0
e 3 e 3
HFRI Macro
D v-0 D v-0
e 3 e 3 Ja c-0 Ja c-0
Ja c-0
3 Ja c-0
3 n 3 n 3
Fe -04 Fe -04
Fen-0
HFRX Macro
Ja c-0 Ja c-0 Ja c-0 Ja c-0
4 4 n 4 n 4
Fen-0 Fen-0 Fe -05 Fe -05
5 5
M b-0 M b-0
M M M M
a a a
100
105
110
115
120
100
105
110
115
120
125
130
135
140
a
95
100
105
110
115
100
110
120
130
140
150
160
Ap r-0
3 Ap r-0
3 Ap r-0
3 Ap r-0
3
M r-0 M r-0 M r-0 M r-0
a 3 a 3 a 3 a 3
Ju y-0 Ju y-0 Ju y-0 Ju y-0
n- 3
Ju 03
n- 3
Ju 0 n- 3 n 3
3 Ju 03 Ju -03
5 Ap -0
Ap r-0
5 5
Fig. 1. Comparison of cumulative performance for the HFR investable indices vs.
their non-investable counterparts since inception of the former. The last graph shows
173
formance of the HFRX, the investable counterpart of the HFRI index, to the
since inception of the former. The deviation is eye-catching: Let us just have
a specific look at the Equity Hedge indices. The average monthly underper-
representative and making it investable. Fig. 1 shows the divergence of various
174 L. Jaeger
21
The distinction between investible index providers and fund of funds is/should
be about systematic methodology and goals for manager selection. Most index
providers have virtually no selection methodology, and to that extent they are
just fund of funds. Those that do have well founded methodologies that are im-
plemented can, without demurring, be called indices. The biggest problem really
is that the index provider and the asset manager are in fact identical—this is
unlike the case for US Equity Indices, but not unlike the case for the most well
regarded bond indices (e.g. Lehman).
22
One important difference between the index provider and a fund of hedge funds re-
mains, though: The fund of funds manager is actively searching for alpha and trad-
ing talent, which justifies the comparably high feel level charged. He is not in the
business of “averaging the alpha,” an undertaking which almost by construction
will lead to lower results in the case of hedge funds. Note that alpha extraction
is on a global scale a “zero sum game.”
23
The reader is referred to the following article for another discussion on the prob-
lems and pitfalls of hedge fund indices: L. Jaeger, “Hedge Fund Indices – A new
way to invest in absolute returns strategies?,” (June 2004).
The Road to Hedge Fund Replication 175
construct derivatives from it and whether it can be sold short. The possibility
of short selling and constructing synthetic positions based on derivatives (in
a cost efficient way) creates the prospect of arbitrage opportunities using
the hedge fund indices. Ironically such arbitrage opportunities would most
likely be exercised by hedge funds, in a sort of Klein bottle of investments
that contain themselves. Whether or not such trades emerge will eventually
prove whether hedge fund indices can sustain market forces, which ultimately
enforce an arbitrage-free market equilibrium. Today, there is an active market
for structured products referencing hedge fund indices, including delta one
products that allow investors to synthetically short some of the investable
hedge fund indices.
400
350
300
250
200
150
100
94
95
95
96
96
97
97
98
98
99
99
00
00
01
01
02
02
03
03
04
04
05
05
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
1.
7.
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
.0
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
31
24
See L. Jaeger et al., “Case study: The sGFI Futures Index”, Journal of Alternative
Investments, (Summer 2002).
25
“Buy write” refers to holding long the underlying – in this case the S&P 500 index,
and simultaneously selling a call. This combination is economically identical to
selling a put on the S&P500 plus holding an equivalent amount of cash.
The Road to Hedge Fund Replication 177
8
6
Percent
Percent
6
4
4
2
2
0
−1 0 1 2 3 −2 −1 0 1 2
Convertible Equity Hedge Manager Small-Cap Spread Equity Hedge Manager
8
8
6
6
Percent
Percent
4
4
2
2
0
−1 −.5 0 .5 1 −.5 0 .5
CPPI Equity Hedge Manager AR(1) Equity Hedge Manager
28
A thorough discussion of the autoregressive factor can be found in Getmansky
M., Lo, A. W., Makarov, I., “An Econometric Model of Serial Correlation and
Illiquidity in Hedge Fund Returns” (2004). See also the paper by C. Asness et al,
“Do hedge funds hedge” (2001).
180 L. Jaeger
The CUSUM test considers the cumulated sum of the (normalized) recursive
residuals wr.
r=t
ωr
Wt = ,
σ̂
r=K+1
40
40
20
20
20
CUSUM
CUSUM
CUSUM
0
0
−20
−20
−20
−40
−40
−40
1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1
Time Time Time
40
40
20
20
20
CUSUM
CUSUM
CUSUM
0
0
−20
−20
−20
−40
−40
−40
1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1
Time Time Time
40
40
20
20
20
CUSUM
CUSUM
CUSUM
0
0
−20
−20
−20
−40
−40
−40
1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1
Time Time Time
Fig. 4. Results of a CUSUM stability test for the regression models in Table 1
182 L. Jaeger
to avoid the problem of data mining and in-sample over-fitting, the factors
chosen for the RFS were calculated on a rolling looking forward basis. To
be precise, the RFS returns in a given month were calculated using factors
obtained by a regression over data for the previous five years ending with
the previous month. The RFS are in spirit similar to what Jensen et al.35
describe as a generic replication of hedge fund strategy with the difference
however that the chosen factors/substrategies are explicitly modelled in the
regression set up.
The results for the most recent three years (since inception of the investable
indices) are rather astonishing: The cumulative replicating strategy’s returns
are often superior to the returns of the hedge fund indices, especially when
considering their investable versions. For the latter performance of the RFS
is better for every single strategy sector with the exception of the Distressed
strategy.
Interpreting our results leads us to a schematic illustration of where hedge
fund returns come from (Fig. 6). A long-only manager (represented by the left
bar) has two sources of returns: the market exposure and the manager excess
return, his “alpha” (which is negative for most managers in this domain).
The difference between long-only investing and hedge funds is largely that the
hedge fund will hedge away all or part of the broad market exposure. In order
to achieve this risk reduction, the hedge fund manager employs a variety of
Credit Risk
Equity Risk
Fig. 6. A schematic model for hedge fund return sources based on results in
Table 1
35
G. Jensen and J. Rotenberg “Hedge Funds Selling Beta as Alpha” (2003), updated
in 2004 and 2005.
184 L. Jaeger
techniques and instruments not typically used by the long-only fund manager
including short selling and the use of derivatives. This results in what appears
as a “pure alpha” product with low expected returns and low expected risk.
But in order to be attractive as a stand-alone investment, the hedge fund
manager has to conform to the market standard for return. This leads him
to scale the risk by using leverage, which provides the desired magnification
of return and risk. In this magnified configuration, systematic elements of
risk and return that before were hidden in the “Alpha” are suddenly large
enough to be analysed separately. In other words, we now have the necessary
magnifying glass to separate out the “beta in alpha’s clothing.” We estimate
that up to 80% of the returns from hedge funds originate as the result of beta
exposure (i.e. exposure to systematic risk factors) with the balance accounting
for manager skill based alpha (or not yet identified risk factors).
In the following we discuss our results for the individual strategy sec-
tors, the summary of which is presented in Table 2 in comparison with the
investable and non investable indices from Hedge Fund Research.
Most Long/Short Equity managers have exposure to both the broad equity
market and particularly to small cap stocks. Managers may find it easier to
find opportunities in a rising market, and it may also be easier to short sell
large cap and buy small cap stocks. Our risk factor model in Table 1 confirms
these results. The most significant factors are related to broad equity and small
cap equity markets. Fung and Hsieh obtain similar results in a specific study
on the Long/Short Equity strategy.36 They choose as independent variables
the S&P 500 index and the difference between the Wilshire 1750 index and
the Wilshire 750 index as a proxy for the small cap risk factor. We obtained
Table 2. Cumulated performance of the RFS and the HFRX strategy, data from
March 2003 to August 2005
36
See W. Fung, D. Hsieh, “The Risk in Long/Short Equity Hedge Funds” (2004).
The Road to Hedge Fund Replication 185
very similar results (having chosen the Russell 2000 and Russell 1000 for the
calculation of the small cap spread).
However, a closer look reveals that the exposure of Long/Short Equity
hedge funds has a strongly non-linear profile. This non-linear exposure is
reflected in the fact that the most explanatory independent variable is a con-
vertible bond index.37 Apparently, this profile models the Long/Short Equity
strategy well: Less participation on the upside, protection on the downside to
a certain point, but with more expressed losses in a severe downturn of the
equity markets (when convertible bonds loose their bond floor). The substi-
tution of an equity factor with a convertible bond factor thus yields a better
model than a simple equity factor.38 However there is another equity related
factor that comes into play: Hedge funds tend to decrease their exposure in
falling equity markets and increase it in rising markets, similar to a “Constant
Proportion Portfolio Insurance” strategy often employed in capital protected
structures. We simulate this behaviour by including such a CPPI factor based
on the rolling 12 month performance of the S&P 500. Figure 7 presents the
performance of the RFS next to the HFR non-investable (HFRI) and the in-
vestable versions of HFR (HFRX) and S&P indices since inception of the
HFRX (inception of the S&P index occurred later, and at its inception it was
taken to the same level as the HFRX in the graph). The chart confirms what
the numbers already indicated: We can very well replicate the performance
of the average Long/Short Equity manager in the index with a RFS model
0.05 140
RFS HFEH S&P Long/Short Equity HFRX RFS HFRI S&P Long / Short Equity HFRX
0.04 135
0.03 130
0.02 125
0.01 120
0 115
−0.01 110
−0.02 105
−0.03 100
J u -05
n 3
Ju y-04
Se -0 3
O -03
e 4
4
Se -04
O -04
ec 3
3
Ap r-04
3
05
Ju -03
Ju 05
ar 5
5
a 4
4
M r-05
a 4
Au l-0 4
o 4
M r-03
o 3
n 5
n 3
4
3
c 3
g 4
ec 4
4
ec 3
4
c 4
g 3
3
Au l-05
M r-0 3
05
Ju -03
Ju -05
a 5
5
a 4
4
a 5
a 4
o 4
a 3
o 3
Au l-04
Au l-03
g- 5
J u y-0
D v -0
Ja c -0
D v-0
Ja -0
Ap r-0
Ap -0
M b-0
M b-0
Fen-0
N t-0
N t-0
M r-0
Ju -0
Fen-0
J u y-0
J u y-0
Ju y-0
Se g-0
O p-0
D v -0
Ja -0
D v-0
Ja - 0
Seg-0
O p-0
Au l-0
Ap r-0
Ap r-0
Ap r-0
M b-0
M b-0
Fen-0
M r-0
N t-0
N t-0
M r-0
Jun-0
Fen-0
Au l-0
g-
n-
ay
p
p
n
a
c
a
a
M
M
with similar performance and volatility. The RFS performs along the HFRI
index despite some alpha displayed in Table 1. Figure 17 sheds some light
on this discrepancy: Table 1 displays the average alpha over the regression
period which as Fig. 17 indicates declines quite rapidly over time. Figure 7 in
contrast only matches the most recent performance since 2003. There is only
little alpha shown be Long/Short Equity managers in the most recent pe-
riod as Fig. 17 indicates. Finally, the RFS outperforms both investable indices
(HFRX and S&P) significantly.
Equity Market Neutral strategies aim at zero exposure to specific equity mar-
ket factors. Correspondingly, the model in Table 1 shows only a small (however
statistically significant) exposure to broad equity markets. However, the re-
sults indicate that the Equity Market Neutral style carries sensitivity to the
Fama-French momentum factor UMD and the value factor (the spread of the
MSCI value and growth indices). The R2 value of the regression for Equity
Market Neutral comes out lowest for all strategy sectors next to the Managed
Futures. In other words, simple linear models fall short of explaining a signif-
icant part of the variation of returns for this hedge fund style. However, to
mix the right combination of systematic risk exposures of Equity Market Neu-
tral strategies right, we must distinguish two distinctly different sub-styles of
this strategy. The one (often system based) approach buys undervalued stocks
and sells short overvalued stocks according to a value and momentum based
analysis. The second more short term oriented approach (also referred to as
“Statistical Arbitrage”) trades in pairs based on a statistical analysis of rel-
ative performance deviation of similar stocks. Both styles naturally have a
different exposure to the factors examined here.
Figure 8 confirms what the numbers in Table 1 indicate: The RFS un-
derperforms the HFRI index by some margin reflecting the positive alpha in
Table 1. However, it outperforms the HFRX investable index significantly.
0.02 115
RFS HFRI HFRX
0.015
0.01
110
RFS HFRI HFRX
0.005
0
105
−0.005
−0.01
100
−0.015
−0.02
−0.025 95
n- 5
n- 3
n 4
3
c 3
e 4
4
4
c 4
e 3
3
05
n 5
4
3
n 3
n- 4
3
ar 5
a 4
c 3
ec 4
n 4
Ju 03
Ju 05
4
c 4
e 3
n 3
4
5
05
Ju -03
Ju -05
a 5
4
5
ar 4
Ju -04
a 5
a 5
a 4
Ju 04
o 4
o 4
a 3
a 3
o 3
o 3
Au l-04
Au l-03
Au l-05
J u y-0
J u y-0
Ju y-0
Aul-04
Aul-03
Se g-0
O p-0
D v -0
Ja c -0
Aul-05
D v-0
Ja c - 0
Seg-0
O p-0
J u y-0
Ap r-0
Ap r-0
Ap -0
J u y-0
M b-0
Ju y-0
M b-0
Se g-0
O p-0
D v -0
Ja -0
D v-0
Ja c-0
Seg-0
O p-0
Ap r-0
Ap -0
Ap r-0
M b-0
Fen-0
M b-0
Fen-0
Fe -0
M r-0
M r-0
M r-0
N t-0
N t-0
M r-0
N t-0
N t-0
M r-0
M r-0
Fe -0
g-
g-
a
a
M
Fig. 8. Returns (monthly and cumulated) of the non-investable HFRI Equity Mar-
ket Neutral Index and the investable HFRX Equity Market Neutral Index (in light
color) vs. the RFS cumulative return (in dark color) based on the factor returns (see
text for details)
The Road to Hedge Fund Replication 187
Short Selling
The main exposure of the Short Selling strategy is, quite obviously, being short
the equity market. Interestingly, the exposure to the broad equity markets can
best be modeled with the same factor as for the Long/Short equity managers,
the Convertible Bond Index. This indicates the same type of non-linear expo-
sure as for the Long/Short Equity strategy, however with the signs inversed.
The strategy displays positive sensitivity to value stocks with as measured
by the spread between the MSCI value and growth indices. The alpha value
for Short Selling strategies stands at around 4-5% p.a. This indicates that
the short side does offer some profit opportunities, possibly explained in part
by most investors being restricted from selling short. However, the alpha of
this strategy must be high in order for the strategy to generate any profits
at all. This is because from the perspective of risk factor exposure, shorting
the equity markets starts off with an expected negative 4-7% return (long
term performance of the equity markets minus short rebate for the short posi-
tions). As a result Short Selling is the only hedge fund strategy with negative
past performance over the last 15 years. This is also reflected in Fig. 9 for the
more recent period. We observe that the Short Selling strategy can be well
replicated by the RFS model.
Event Driven
0.08 100
RFS HFRI
0.06
95
0.04 RFS HFRI
90
0.02
0 85
−0.02
80
−0.04
75
−0.06
−0.08 70
n 5
n 3
n 4
3
O p-03
e 4
n 4
p 4
O -04
e 3
n 3
Ap -04
Ap r-03
Ap -05
05
Ju -03
Ju -05
ar 5
Fe -0 5
ar 4
J u -03
n 4
Fe -04
p 3
a 5
a 4
Ju -04
N t-04
J u -05
a 3
N t-03
O -03
e 4
4
p 4
O -04
e 3
3
4
3
05
Au l-0 3
Au l-0 5
a 5
5
a 4
Au l-04
4
Au l-03
M r-05
a 4
Au l-05
Ju -04
D v -04
M r-03
o 3
4
3
g- 5
J u y-0
J u y- 0
Ju y-0
Se g-0
D v -0
Ja c -0
D v -0
Ja c - 0
Seg-0
M b-0
M b-0
Ju y-0
M r-0
M r-0
M r-0
Se g-0
Ja c -0
D v-0
Ja c-0
Seg-0
Ap r-0
Ap r-0
Ap r-0
M b-0
Ju -0
Jun-0
M b-0
Fen-0
N t-0
N -0
M r-0
Fen-0
Au l-0
g-
ay
ay
n
ct
c
a
o
o
M
M
0.01 130
0
120
−0.01
110
−0.02
−0.03 100
n 5
n 3
n 4
3
c 3
n 5
e 4
4
n 3
e 3
3
4
c 4
n 4
4
r 3
p 3
5
05
a 5
Ju -03
Ju -05
a 4
O -03
e 4
n 4
p 4
O -04
4
e 3
Ja c-03
Ap r-04
a 5
Ap r-05
05
a 4
Ju -04
o 4
M -03
o 3
a 5
Ju -03
Ju -05
a 4
Fe -0 5
Fe -04
Ju -04
a 5
a 4
N t-04
a 3
N t-03
Au l-04
3
Au l-05
Au l-04
Au l-03
Au l-05
Apr-03
J u y-0
J u y-0
Ju y-0
J u y-0
Se g-0
O p-0
D v -0
Ja c -0
J u y-0
D v -0
Ja c - 0
Seg-0
O p-0
Ju y-0
Ap r-0
Ap r-0
Ap r-0
M b-0
M b-0
D v -0
Se g-0
Ja c -0
D -0
Seg-0
Fen-0
M r-0
N t-0
N t-0
M r-0
Fen-0
M b-0
M b-0
M r-0
M r-0
M r-0
Au l-0
g-
g-
ov
n
a
c
o
a
M
M
Fig. 10. Returns (monthly and cumulated) of the non-investable HFRI Event
Driven Index and the investable HFRX Event Driven Index (in light color) vs.
the RFS cumulative return (in dark color) based on the factor returns (see text for
details)
the variation of Event Driven returns. Alpha is the highest for any strategy in
the hedge fund universe with roughly 5% p.a. over the analyzed period. This
is also reflected in Fig. 10, where we see that the RFS model yields roughly
about two thirds of the return of the Event Driven managers in the HFRI
index. However, again, the RFS outperforms the HFRX and S&P investable
index version significantly.
Distressed Securities
0.04
150
RFS HFRI HFRX
0.03
140
0.02
130
0.01
120
0
110
−0.01
−0.02 100
n 5
n 3
n 4
3
c 3
e 4
n 4
Se -04
O -04
e 3
n 3
05
Ap r-04
3
Ap -05
ar 5
M b-04
Ju -03
Ju -05
Fe -0 5
Fe -04
n 5
Ju -04
Ju n-0 3
Jun-0 4
M r-05
M r-04
ov 4
M r-03
N t-03
p 3
O -03
ec 4
4
e 3
3
4
c 4
05
4
3
5
a 5
a 4
J u -0
J u y-0
Aul-033
Ju -05
Ju y-0
5
4
Aul-04
Aul-03
a 5
a 4
D v -04
Se g-0
O p-0
D -0
Ja c -0
Aul-05
a 3
D v-03
D v-0
Ja c-0
Ap r-0
M -0
J u y-0
J u y-0
Ju y-0
N t-0
Au -04
Se g-0
Ja -0
Au -05
Ja c-0
Seg-0
O p-0
Ap r-0
Ap r-0
Ap r-0
M b-0
M b-0
Fen-0
Fen-0
M -0
N t-0
M r-0
N ct-0
M r-0
g-
ay
g-
g
p
b
a
c
o
r
a
o
l
M
M
Fig. 11. Returns (monthly and cumulated) of the non-investable HFRI Distressed
Index and the investable HFRX Distressed Index (in light color) vs. the RFS cumu-
lative return (in dark color) based on the factor returns (see text for details)
Merger Arbitrage
In their seminal paper on the Merger Arbitrage strategy, Mitchel and
Pulvino39 examine the conditional correlation properties of this strategy:
Merger Arbitrage strategies display rather high correlations to the equity
markets when the latter declines and comparably low correlations when
stocks trade up or sideways. This corresponds to a correlation profile similar
to that of a sold put on equities. As a matter of fact, the payout profile
of Merger Arbitrage strategies corresponds directly to a sold put option on
announced merger deals. This short put profile is reflected in the significance
of the BXM factor in Table 1. Shorting put options provides limited upside
but full participation on the downside (less the option premium). This argu-
ment extends beyond the immediate exposure to merger deals breaking up:
When the stock market falls sharply, merger deals are more likely to break.
In addition, a sharp stock market decline will reduce the likelihood of revised
(higher) bids and/or bidding competition for merger targets. Falling stock
markets also tend to reduce the overall number of mergers, which increases
the competition for investment opportunities and may thereby reduce the
expected risk premium. The strategy therefore has a slightly positive stock
market beta, however strongly non-linear. This overall exposure profile to
equity markets comes more from the correlation between the event risk and
the market than from the individual positions.
Mitchell and Pulvino calculated the historical track record of a simple
rule-based merger arbitrage strategy that at any time invests in each an-
nounced merger deal, both cash and stock-swap, with a pre-specified entry
and exit rule.40 They conducted this calculation for 4,750 merger transac-
tions from 1963 to 1998. The hedge fund manager Bridgewater performed a
very similar study but constrained themselves to the ten largest mergers at
any point in time. In both cases the resulting simulated returns came very
39
See M. Mitchel and T. Pulvino, “Characteristics of Risk in Risk Arbitrage” (2001).
40
See M. Mitchel and T. Pulvino, “Characteristics of Risk in Risk Arbitrage” (2001).
190 L. Jaeger
0.02 120
RFS HFRI HFRX
0.015 RFS HFRI HFRX
0.01 115
0.005
0 110
−0.005
−0.01 105
−0.015
−0.02 100
n 5
n 3
n 4
3
c 3
e 4
n 4
e 3
p 4
c 4
n 3
05
Ap -04
Ap r-03
Ap -05
ar 5
Ju -055
ar 4
Ju n-0 3
Jun-0 4
Ju -03
Ju -05
3
Fe -0 5
Fe -04
ct 3
Ju -04
Ja c -04
4
p 4
O -04
e 3
3
M r-05
M r-04
N t-04
05
M r-03
N t-03
4
Ap r-03
5
a 5
a 4
3
5
4
J u y-0
a 5
J u y-0
a 4
ov 4
a 3
o 3
Ju y-0
Aul-04
Aul-03
Se g-0
O p-0
D v -0
Ja c -0
Aul-05
D v-0
Ja c - 0
Seg-0
O -0
M b-0
J u y-0
J u y-0
M b-0
Ju y-0
Au 04
Aul-03
Se g-0
O p-0
D -0
Aul-05
D v-0
Ja c-0
Seg-0
Ap r-0
Ap r-0
M b-0
M b-0
Fen-0
Fen-0
M r-0
N t-0
M r-0
N -0
M r-0
g-
g-
n
a
-
a
o
o
a
e
l
M
M
Fig. 12. Returns (monthly and cumulated) of the non-investable HFRI Merger
Arbitrage Index and the investable HFRX Merger Arbitrage Index (in light color)
vs. the RFS cumulative return (in dark color) based on the factor returns (see text
for details)
close to the returns of the Merger Arbitrage hedge fund indices (HFR and
Tremont). We included a strategy which focuses on investing exactly along
the Mitchell/Pulvion study, the publicly available “Merger Fund.”41
Our regression shows what we expected, exposure to the equity markets,
in particular the small cap segment (furthermore the value sector), the BXM
index and the Merger Fund. However, the explanatory strength of the model
is not that high (considering that these factor should very well reflect what
the strategy is about). Just as with other Event Driven strategies the alpha
value is above average for this strategy with around 4% p.a. However, a com-
parison with the performance of the RFS in Fig. 12 shows that the skill based
component of returns has declined in recent years, as the RFS tracks the
performance of the HFRI Merger Arbitrage rather closely. Again, the RFS
outperforms the investable version of the HFR index by a safe margin.
skewed return distribution, where steady but small gains are countered with
rare but large losses. In other words, the managers are short some sort of
volatility, which makes the return profile resemble the payout profile of a
short option position.
0.005 110
108
0
106
−0.005
104
−0.01 102
−0.015 100
n 5
n- 3
4
p 3
O -03
e 4
4
e 3
3
p 4
O -04
05
n- 5
4
n- 3
3
n 4
a 5
a 4
3
3
Ju -05
5
c 3
e 4
4
e 3
n 3
4
c 4
4
05
4
4
3
5
M r-05
M r-04
N t-04
ar 5
M r-03
o 3
a 4
Ju 03
Ju 05
5
4
Ju -04
a 5
a 4
o 4
a 3
o 3
J u y-0
J u y-0
Ju y-0
Aul-04
Aul-03
Se g-0
D v -0
Ja c -0
Aul-05
D v-0
Ja c-0
Se -0
J u y-0
J u y-0
Ap r-0
Ap r-0
Ap r-0
Ju y-0
M b-0
M b-0
Aul-04
Ju 0
Aul-03
Se g-0
O p-0
D v -0
Ja c -0
Aul-05
D v-0
Ja c-0
Seg-0
O p-0
Fen-0
Jun-0
Fen-0
Ap r-0
Ap r-0
Ap -0
M b-0
N t-0
M b-0
Fe -0
Fen-0
M r-0
N t-0
M r-0
N t-0
M r-0
g-
g-
g
a
a
a
c
c
o
a
M
Fig. 13. Returns (monthly and cumulated) of the non-investable HFRI Fixed In-
come Index vs. the RFS cumulative return (in dark color) based on the factor returns
(see text for details). Note: An investable version of the HFR index does not exist
for Fixed Income Arbitrage hedge funds
42
See W. Fung, D. Hsieh, “The Risk in Fixed Income Hedge Fund Styles” (2002).
192 L. Jaeger
Convertible Arbitrage
0.02 105
0.01
0 100
−0.01
−0.02 95
−0.03
−0.04 90
n 4
n 4
n 3
ec 4
n 5
e 4
Ju -03
n- 5
3
c 3
05
4
ay 3
n- 3
p 3
O -03
05
4
Se -04
n 4
3
ar 4
3
M r-0 4
3
Apr-03
a 4
Apr-04
5
a 3
c 4
a 5
a 5
3
Ju -04
O -04
o 4
Ju -05
Ju -04
ov 4
Fe -05
Ju 05
a 5
a 5
a 4
M r-04
Ju -03
Ju -03
A u l-03
Au l-04
A u l-03
Au l-04
Au l-05
Au l-05
Ju y -0
Ju y -0
Ju y-0
D v -0
Ju y-0
D -0
Ju y-0
Seg-0
O p-0
Ja c -0
Seg-0
M r- 0
Ja c -0
Seg-0
Ja -0
Ja c-0
Dev-0
M b-0
A p r-0
A p -0
A p r-0
Dev-0
M b- 0
A p r-0
M r- 0
O p-0
M r- 0
M r- 0
No t-0
Fen-0
No t-0
N t- 0
Fen-0
Fe 0
N ct-0
M b-0
M b-0
g-
g-
p
g
n
c
a
a
M
M
Fig. 14. Returns (monthly and cumulated) of the non-investable HFRI Convertible
Arbitrage Index and the investable HFRX Convertible Arbitrage Index (in light
color) vs. the RFS cumulative return (in dark color) based on the factor returns (see
text for details)
Global Macro
0.04 125
0.02 120
0 115
−0.02 110
−0.04 105
−0.06 100
n 4
n 3
ec 4
n 5
n 4
3
ct 3
Fen-0 3
05
4
e 4
n 3
n- 5
3
ar 4
M r-03
M r-0 4
M r-05
3
c 3
n 3
05
4
a 3
c 4
a 5
n 4
3
o 4
a 4
3
Apr-04
5
a 5
a 3
a 5
No t-03
O p-04
4
Ju -04
o 4
a 4
Fe -05
Ju 5
A u l-0 3
a 5
3
Au l-04
a 4
Au l-05
Ju -03
A u l-03
Au l-04
Au l-05
Ju y -0
Ju y-0
D v -0
Ju y-0
Ju y - 0
Seg-0
O p-0
Ja c -0
Seg-0
Ja -0
D v -0
Ju y-0
Ju y-0
Dev-0
M b-0
A p r-0
A p -0
A p r-0
Seg-0
O p-0
Ja c -0
Seg-0
O p-0
Ja c-0
No -0
Ju -0
N t- 0
Fen-0
Ju -0
Dev-0
M b-0
A p r-0
A p r-0
M b-0
M r-0
M r- 0
Fe -0
N t- 0
0
Ju -0
M b-0
M r-0
g-
g-
a
c
a
M
M
Fig. 15. Returns (monthly and cumulated) of the non-investable HFRI Global
Macro Index and the investable HFRX Global Macro Index (in light color) vs. the
RFS cumulative return (in dark color) based on the factor returns (see text for
details)
Managed Futures
Managed Futures hedge funds are the main speculative agents in the global
futures markets, thus capturing what we referred to as the “commodity hedg-
ing demand premium.” A simple trend following trading rule (sGFII) applied
to the major global futures markets captures a large part of these returns and
shows up as the most dominant term in the regression in Table 1. Several
different studies have independently obtained this result.44 The sGFII index
is designed to model the return of trend following strategies with a simple rule
based momentum approach. It is a volatility weighted combination of trend
following strategies on 25 liquid futures contracts on commodities, bonds, and
currencies. This index shows a 48% correlation with the CISDM trend follow-
ing index, and equally a 48% correlation with the CSFB/Tremont index. Based
on the regression in Table 1 the average CTA in the CISDM Trendfollower
index displays negative alpha. Schneeweis/Spurgin and Jensen and Rotenberg
(Bridgewater) use similar trend following indicators on a much more restricted
set of contracts45 . They obtain an even higher correlation coefficient to the
CSFB-Tremont Managed Futures index (71% in the case of Bridgewater) or
the CISDM Managed Futures Indices (79% against the CISDM Trend follow-
ing index for Schneeweis/Spurgin). The lower correlation of the sGFII index
is possibly due to a comparably high exposure to commodity contracts com-
pared to Bridgewater’s and Schneeweiss/Spurgin’s model (which overweigh
the complex of financial futures contracts).
44
See L. Jaeger et al., “Case study: The sGFI Futures Index” (Summer 2002);
Jensen. G., Rotenberg, J., “Hedge Funds Selling Beta as Alpha” (2003);
R. Spurgin, “A Benchmark on Commodity Trading Advisor Performance” (1999).
45
T. Schneeweis and R. Spurgin, “Multifactor Analysis of Hedge Funds, Managed
Futures, and Mutual Fund Returns and Risk Characteristics” (1998); G. Jensen
and J. Rotenberg “Hedge Funds Selling Beta as Alpha” (2003).
The Road to Hedge Fund Replication 195
0.08 120
RFS CISDM S&P Managed Futures RFS CISDM S&P Managed Futures
0.06
115
0.04
0.02 110
0
105
−0.02
−0.04 100
−0.06
95
−0.08
−0.1 90
n 3
n 4
n 5
p 3
O -03
e 3
Ja c-03
p 4
c 4
e 4
Ja c-04
Ju y-03
n 4
Ju y-05
Apr-03
M -04
Ap -04
a 5
Apr-05
p 3
O -03
e 3
3
p 4
O -04
e 4
4
M r-03
N t-03
a 4
N t-04
M r-05
Ju -03
Fe -04
Ju -04
Fe -05
Ju -05
3
a 4
4
a 5
5
M r-03
Au l-0 3
o 3
a 4
Au l-0 4
o 4
M r-05
5
Ju -05
Au l-03
Au l-04
5
3
Ju -0
Ju y-0
Ju y-0
Seg-0
D v-0
Se -0
O -0
D v-0
Ju y-0
M b-0
Seg-0
D v-0
Ja c-0
Seg-0
D v-0
Ja c-0
M r-0
Ap r-0
M b-0
Ap r-0
M b-0
Ap r-0
Ju -0
N ct-0
Fen-0
M r-0
Ju -0
N ct-0
Fen-0
l-0
l-0
ay
g
b
ar
n
n
n
c
a
a
a
o
a
M
M
Fig. 16. Returns (monthly and cumulated) of the non-investable (!) CISDM Man-
aged Futures Qualified Universe Index (in grey color) vs. the RFS cumulative return
(in dark color) based on the factor returns (see text for details)
There is good reason to believe that generally the average alpha extracted by
hedge fund managers is destined to decline. As a matter of fact, we can al-
ready today observe that alpha has grown smaller in size over time, as Fig. 17
indicates for the most obvious strategy, Long/Short Equity, where we display
the alpha of a rolling regression over a 60 months time window. Independently
from our research, the attenuation of alpha has been observed elsewhere. Fung
et al. report in one of their latter research on the same phenomenon.46 One
possible explanation for this phenomenon comes quickly to mind: As more
46
W. Fung, D. Hsieh, N. Naik, T. Ramadorai, “Hedge Fund: Performance, Risk and
Capital Formation”, Preprint (2005)
196 L. Jaeger
1.20%
Rolling Alpha
1.00%
0.80%
0.40%
0.20%
0.00%
00 00 00 00 01 01 01 01 02 02 02 02 03 03 03 03 04 04 04 04 05 05 05
a n- pr- ul- ct- an- pr- ul- ct- an- pr- ul- ct- an- pr- ul- ct- an- pr- ul- ct- an- pr- ul-
J A J O J A J O J A J O J A J O J A J O J A J
Fig. 17. The development of alpha for Long/Short Equity funds (HFR sub-index)
based on a rolling regression over a 60 month time window. The risk factors were
chosen as in Table 1
our risk based models, such as trading volume, open short interest on stocks,
insider activity, leverage financing policies of prime brokers, etc. A direct de-
pendency of the hedge fund managers’ alpha creation from these variables will
lead us to a better understanding of their time variability that we empirically
observe in our models. This will ultimately lead us to an understanding of
the very alpha creation process of hedge funds, the part of hedge fund returns
which remains still in the dark for most investors. However, little effort has
been put into this task so far.
The main task of the investor will be to define what he wants from hedge
funds. Alpha is and will continue to be ultimately the most attractive sort of
return, as it comes with no systematic risk and no correlation to other asset
classes. But investors should realize both the scarcity of true alpha and the
power of alternative beta. It is the power of diversification into orthogonal
risk factors which will ensure that hedge funds remain broadly attractive for
investors. And when it come to the hedge funds’ beta there is surely a great
deal larger capacity available to investors than in the case of alpha. In fact, the
future growth prospects of the hedge fund industry become quite compelling
considering that we are far from any limit with respect to “beta capacity” in
the hedge fund industry. While the search of alpha surely remains compelling,
we believe it is investment in alternative betas which will be more and more
the key to successful hedge fund investing in the future.
the overall alpha in the global equity and bond market to be USD 220 billion.
We must further assume that hedge funds can participate from this “alpha
pie” only to a certain extent next to other professional players which are likely
to be “positive alpha players” and thus compete with hedge funds for alpha
(proprietary trading operations, large institutions, mutual funds – before their
fees, etc.). It seems realistic to assume that hedge funds can take one fourth
of that pie50 (a proportion which might grow larger over time, however, as
more players from the other “alpha parties” move into the hedge fund space).
This implies that there are USD 55 billion pure alpha available to hedge funds
each year. Further, assuming that hedge fund investors require a least a 15%
p.a. return gross of fees (before management, performance, trading fees, etc.),
which amounts into a net return of around 8%-10% and constitutes proba-
bly the minimum investors would require from hedge funds. This implies an
overall capacity of hedge funds based on alpha only of
USD 55 billion/0.15 = 366.6 billion USD,
about one third of the actual size of assets in the hedge fund industry. Even
with different, more beneficial assumptions on the overall investor tolerance
for inefficiencies and on how much hedge funds can participate in the total
“alpha pie”51 , we would not come up with a capacity significantly higher than
the current size of the industry. As a result, based on the assumption of ineffi-
ciencies in the global capital markets alone, we are not just lacking a satisfying
economic explanation of hedge fund return sources, we also find ourselves in
a position not being able to explain the current size of the industry!
But by now we understand that a large portion of hedge fund returns
is not related to pure alpha, but rather to “alternative beta.” The analysis
in our research suggests that a large part of the average hedge fund return
stems from alternative beta rather than alpha. We now consider our estimate
for that part to be as high as 80%. Well, this raises the bar for hedge fund
capacity significantly higher. Going along with our conclusion and estimating
that only 20% of the industry returns is related to pure alpha, we can calculate
the capacity of the industry to be
366.6 billion USD/0.2 = 1’833 billion USD,
about twice its current size. However, as large as this number seems, it
is exceeded by some of the estimates given by industry protagonists as to
what level the industry will grow within the following years. How can this
growth be managed considering our numbers? The answer is obvious: Only
by including a larger share of alternative beta in the overall return scheme of
hedge funds. Assuming that the ratio of alpha vs. alternative beta becomes
10%, the capacity reaches the number of 3670 billion USD (assuming that the
capacity of alternative beta is not limited at these levels, a fair assumption in
our view).
50
The reader is invited to perform the calculation with different numbers.
51
The reader may use his own set of assumptions.
The Road to Hedge Fund Replication 199
Summarizing, there is indeed plenty of room for the hedge fund industry
to grow, albeit only at the expense of becoming more and more beta driven.
This development will inevitably occur with the future growth of hedge funds.
As a matter of fact, recent performance suggests that this process has already
started.
At the end of this report we would like to point out a further direction
of research possibly not sufficiently covered in this research. Our analysis
suggests that the factor loads of hedge fund strategies are adequately modelled
as stationary. However, there is good reason to believe (and recent research
provides some evidence53 ) that there occur sudden and structural breaks in the
systematic risk exposures of hedge funds that cannot be modelled well enough
in a linear model context. Examples of such are easy to find: The blow up of
LTCM in the summer of 1998, the burst of the stock market bubble in the
spring of 2000, the turn in the equity market in March 2003. Upon a closer
look, a closer look at Fig. 4 reveals some evidence for such breaks, which our
analysis here does not account for. In order to model hedge fund exposure
during these breaks occurring in extreme market environment we need non-
linear exposure models. We will leave this topic for future research.
Generally, the progress recently on understanding the generic sources of
hedge fund returns leads us to the conclusion that investable benchmarks con-
structed by a joint venture of financial engineers and quant groups based on
risk factor analysis and replication has the potential to offer a valid, theoreti-
cally more sound, and cheaper alternative to the currently offered hedge fund
index products offered today. It is evident that once these indices become
more broadly recognized the hedge fund industry will be put upside down.
This will have some further important consequences on how hedge funds are
categorized by investors. So far, most consider them a separate asset class.
Realizing that hedge funds regarding their exposure to systematic risk fac-
tors are conceptually not that different from traditional types of investments
investors may find it conceptually easier to integrate them into their overall
asset allocation.
References
[1] Agarwal, V., Naik, N., “Performance Evaluation of Hedge Funds with
Option-Based and Buy-and-Hold Strategies,” Working paper (2001),
published under the title: “Risks and Portfolio Decisions involving Hedge
Funds”, Review of Financial Studies, 17, p. 63 (2004)
[2] Agarwal, V., Fung, W., Loon, Y., Naik, N., “Risks in Hedge Fund Strate-
gies: Case of Convertible Arbitrage,” Working Paper, London Business
School (2004)
[3] Amin, G., Kat, H., “Welcome to the Dark Side: Hedge Fund Attrition
and Survivorship Bias over the Period 1994-2001,” Journal of Alternative
Investments, (Summer 2003)
[4] Asness, C., Krail, R., Liew, J., “Do hedge funds hedge?”; Journal of
Portfolio Management, 28, 1 (Fall 2001)
53
W. Fung, D. Hsieh, N. Naik, T. Ramadorai, “Hedge Fund: Performance, Risk and
Capital Formation”, Preprint (2005)
202 L. Jaeger
[5] Asness, C., “An Alternative Future, I & II” Journal of Portfolio Manage-
ment, (October 2004)
[6] Bailey, J., “Are Manager Universes Acceptable Performance Bench-
marks,” The Journal of Portfolio Management, (Spring 1992).
[7] Brown, S, Goetzmann, W., Ibbotson, R., Ross, S., “Survivorship Bias in
Performance Studies”, Review in Financial Studies, 5, 4 (1992)
[8] Brown, S, Goetzmann, W., Ibbotson, R., “Offshore hedge funds: Survival
and performance 1989-1995”, Journal of Business, 92 (1999)
[9] Brown, S., Goetzmann, W. (2003), “Hedge Funds With Style,” Journal
of Portfolio Management, 29, 101-112
[10] Capocci, D., Hübner, G., “Analysis of hedge fund performance,” Journal
of Empirical Finance, 11 (2004)
[11] Crowder, G., “Hedge Fund Indices”, Journal of Alternative Investments,
(Summer 2001)
[12] Fama, E., French, K., “Common risk factors in the return of stocks and
bonds”, Journal of Financial Economics, 33 (1993)
[13] Fama, E., French, K., “Multifactor explanations of Asset Pricing Anoma-
lies”, Journal of Finance, 51, 55 (1996)
[14] Fung, W., Hsieh, D., Naik, N., Ramadorai, T., “Hedge Funds: Perfor-
mance, Risk and Capital Formation (July 19, 2006). AFA 2007 Chicago
Meetings Paper available at SSRN: http://ssrn.com/abstract=778124
[15] Fung, W., Hsieh, D., “Empirical Characteristics of Dynamic Trading
Strategies: The Case of Hedge Funds”, The Review of Financial Stud-
ies, 10, 2 (1997)
[16] Fung, W., Hsieh, D., “The Risk in Hedge Fund strategies: Theory and
Evidence from Trend-Followers,” The Review of Financial Studies, 14, 2,
p. 313 (Summer 2001) (2001).
[17] Fung, W., Hsieh, D., “Benchmarks of Hedge Fund Performance: Infor-
mation Content and Measurement Biases,” Financial Analyst Journal
(2001).
[18] Fung, W., Hsieh, D., “The Risk in Fixed Income Hedge Fund Styles”,
Journal of Fixed Income, 12, 2 (2002)
[19] Fung, W., Hsieh, D., “The Risk in Hedge Fund Strategies: Alternative
Alphas and Alternative Betas” in L. Jaeger (ed.), “The new genera-
tion of risk management for hedge funds and private equity investment”,
Euromoney (2003);
[20] Fung, W., Hsieh, D., “Hedge Fund Benchmarks: A Risk Based
Approach”, Working Paper (2004)
[21] Fung, W., Hsieh, D., “The Risk in Long/Short Equity Hedge Funds”,
Working Paper, London Business School, Duke University (2004)
[22] Fung, W., Hsieh, D., “Extracting Portable Alpha from Equity Long/
Short Hedge Funds” (2004), Journal of Investment Management, 2, 4,
1-19
The Road to Hedge Fund Replication 203
[23] Getmansky M., Lo, A. W., Makarov, I., “An Econometric Model of Serial
Correlation and Illiquidity in Hedge Fund Returns”, Journal of Financial
Economics, 74 (3), 529-610; Economics (2004)
[24] Jaeger, L , “Through the Alpha Smoke Screens: A Guide to Hedge Fund
Return Sources”, Euromoney Institutional Investors (2005)
[25] Jaeger, L., “Hedge Fund Indices – A new way to invest in absolute returns
strategies?”, AIMA Newsletter (June 2004)
[26] Jaeger, L. (ed.), “The new generation of risk management for hedge funds
and private equity investment”, Euromoney (2003)
[27] Jaeger, L., “Managing Risk in Alternative Investment Strategies”,
Financial Times/Prentice Hall (May 2002)
[28] Jaeger, Lars, “Sources of Return for Hedge funds and Managed Futures”,
The Capital Guide to Hedge Funds 2003, ISIPublications (Nov. 2002)
[29] Jaeger, L., Jacquemai, M., Cittadini, P., “Case study: The sGFI Futures
Index,” The Journal of Alternative Investment (Summer 2002).
[30] Jensen, J., Rotenberg, J., “Hedge Funds Selling Beta as Alpha”
Bridgewater (2003, updated 2004 and 2005)
[31] Kohler, A., “Hedge Fund Indexing: A square Peg in a round hole”, State
Street Global Advisors (2003).
[32] Malkiel, B., Saha, A., “Hedge Funds: Risk and Return”, Working Paper
(2004)
[33] Mitchel, M, Pulvino, T, “Characteristics of Risk in Risk Arbitrage”, Jour-
nal of Finance, 56, 6, 2135 (2001)
[34] Schneeweis, T., “A Review of Alternative Hedge Fund Indices.”
Schneeweis Partners (2001);
[35] Steven A. Schoenfeld, “Active Index Investing”, Wiley, New York (2004)
[36] Spurgin, R., “A Benchmark on Commodity Trading Advisor Perfor-
mance”, Journal of Alternative Investments (Fall 1999)
[37] Sharpe, W., “Asset Allocation: Management style and performance mea-
surement”, Journal of Portfolio Management, 2, 18 (Winter 1991)
[38] Till, H., “The capacity implications of the search of alpha”, AIMA
Newsletter (2004)
Asset Securitisation as a Profits Management
Instrument
Markus Schmidtchen
1 Introduction
The credit derivatives market has enjoyed a strong growth in liquidity for
loan products and credit derivatives for some years now. A growing number
of players populate both the supply and demand sides of the market, leading
to product diversification on the one hand and a broadening of demand on the
other. Rapid growth has been observed, in particular, in the single name credit
default swap market, through which institutions can hedge against credit de-
fault by large, well-known enterprises.
However, the instrument that can be used to hedge against credit default
by small and medium-sized enterprises (SMEs) is portfolio securitisation,
through which loans or loan default risk are transferred as a package. One
of the reasons for making a bundled transfer is the limited exposure to indi-
vidual SMEs.
In parallel with the development in the capital markets, many credit in-
stitutions are re-organising their credit risk management units. This is being
driven, on the one hand, by the regulatory demands in credit risk measure-
ment under Basel II. On the other hand, greater capital market liquidity is
making loan products more mobile. Banks are conducting increasingly active
credit portfolio management. For example, the capital market may be used to
deliberately increase credit exposure by means of an investment or to delib-
erately reduce risk by loan securitisation. This enables a bank to optimise its
credit portfolio, which in turn has a positive impact on its profits position.
Even if there are usually a number of different reasons for portfolio se-
curitisation, it is clear that economically a transaction is appropriate when
the economic capital released by risk reduction and reinvested in new lending
business generates enough profits to cover the cost of the securitisation.
†
This article presents the author’s opinion only and not an official statement of
KfW views.
206 M. Schmidtchen
20,0%
0,0%
0,0% 0,5% 1,0% 1,5% 2,0% 2,5% 3,0% 3,5% 4,0% 4,5%
Loss severity
Capital utilisation
Volume Expected loss 99,97% quantile Capital commitment
Before
securitisation 2.110.000.000 0,4% 9.114.585 3,9% 82.290.000 3,5% 73.175.415
1
The Monte Carlo model is based on a Gaussian Copula function and simulates
the loss distribution which occurs for the credit institution in the scenarios and
depicts the bank’s risk situation. To derive this loss distribution, it is assumed
that the bank expects an average asset correlation with the loan portfolio of
around 8%. This value is at the lower limit of the correlation assumptions used
by Basel II to derive the risk weightings for small and medium-sized enterprises.
Asset Securitisation 207
3 Securitisation Pool
In order to measure the effects of a securitisation, it is presumed that the credit
institution extracts a sub-portfolio worth 350 million Euros from its existing
loan portfolio with the intention of securitising this sub-portfolio synthetically
over a period of 5 years.
The randomly selected portfolio also has an average credit rating of “Ba1”
and 50% collateralisation. The loss distribution over the securitisation period
for this portfolio and the ensuing tranching are shown in Fig. 2.
The table in Fig. 2 shows that the pool has been subdivided into seven
tranches. The first loss piece (FLP) accounts for 2.35% of the volume. The
“Aaa” tranche accounts for 89% and is thus by far the largest securitisation
tranche.
In order to obtain as clear a distinction as possible between the placing
of expected and unexpected losses when analysing the securitisation effects,
the loan quality of the next tranche above the FLP has a very low “B3”
rating. This ensures that the FLP consists mainly of expected losses. This
can be seen, for example, from the relation between the expected losses of
the securitisation pool over 5 years (1.8%) and the size of the FLP, which is
around 76%.
208 M. Schmidtchen
5,0%
Probability of loss severity
4,5% Pool characteristics
4,0%
Volume 350 m
3,5%
Avg. rating Ba1
3,0%
Avg. recovery rate 50%
2,5%
2,0%
Exp. loss (5 years) 1,80%
1,5%
1,0%
0,5%
0,0%
0,0% 1,0% 2,0% 3,0% 4,0% 5,0% 6,0% 7,0% 8,0% 9,0%
Loss severity
Capital structure
Rating FLP B3 Ba2 Baa2 A2 Aa2 Aaa
Volume 2,35% 1,15% 1,50% 1,75% 2,00% 2,25% 89,00%
Spread p.a. 25,00% 8,00% 2,50% 0,75% 0,50% 0,32% 0,10%
In addition to the sizes of the individual tranches, the table also shows
the assumptions with regard to the spreads that the institution carrying out
the securitisation has to pay to the capital market investors for assuming the
risk. The spreads for categories “Aaa” to “Ba2” are based on observed SME
securitisations. The price for the FLP or the “B3” tranche has be selected in
such a way that, considering the expected losses of these tranches, the investor
achieves a return on the investment of around 12% for the FLP and around
5% for the “B3” tranche. Similar prices can currently also be observed in the
capital market.
It is also expected that the securitisation generates transaction costs to-
talling some 1.3 million Euros. These costs include payments to the arranger,
the rating agency, lawyers, etc, some of which have to be paid upfront while
some are running fees.
The impact on the bank’s portfolio loss distribution after placing the entire se-
curitisation pool is shown in Table 1. For the purpose of a comparative statical
analysis, the risk situation before securitisation is also shown in Table 1.
The table shows that after securitisation the bank’s total risk exposure is
reduced by the amount of the placed volume. In addition, the relative expected
loss increases marginally and the 99.97% quantile is now 4% of the remaining
volume. Overall, this is accompanied by a 0.08 percentage point increase in
the relative economic capital commitment.
The slight increase in the relative economic capital can be attributed to
the fact that, first, the quality of the loan portfolio retained by the bank has
Asset Securitisation 209
worsened slightly. This can be seen from the increase in the relative expected
loss and is due to the fact that the randomly selected securitised loans have
a credit rating that is slightly above average. Second, the placing reduces
the granularity of the bank’s loan portfolio, leading to an increase in the
probability of extreme losses. More economic capital must be retained to cover
this possibility. While the relative moments of the loss distribution increase
slightly, both the absolute expected loss and the absolute capital commitment
are obviously reduced.
The impact of placing the entire risk on the bank’s returns can be seen in
the simplified income statement presented below (see Table 2). The income
statement lists the total profit and expenditure at the securitising institution
over the entire securitisation transaction period. This approach is needed to
illustrate the profitability of a securitisation covering several periods in full.2
The expenses generated by this securitisation strategy comprise transac-
tion costs and capital market costs, which cover payments to the investors.
On the profits side, the institution can record the reduction in expected
loss. It was assumed that the expected loss or the standard risk costs is/are
part of the credit margin. Owing to the placing of the expected loss, this flow
of payments to the bank can now be considered entirely as profits.
Further profits from securitisation are derived from freeing up economic
capital. The amount of this variable is calculated by interest on the released
economic capital (roughly 10.7 Euros) being paid at the target return on
economic capital (11%) over the entire transaction period. If the income and
expenditure sides are added together, total profits are roughly – 3.6 million
Euros. Consequently, this securitisation strategy is not economical under the
given assumptions.
2
In order to quantify the intertemporal effects of securitisation, some simplifying
assumptions have been made. For example, it has been assumed that the risk
reduction effects that are presented in Table 1 and that are calculated for the first
year are also valid in the subsequent years. In addition, these future amounts are
not discounted.
210 M. Schmidtchen
Within the model, securitisation is shown to be optimal when only the mez-
zanine part of the portfolio is issued on the capital market. The FLP and the
“Aaa” part are retained by the credit institution. The risk situation arising
from this strategy is illustrated in Table 3.
The table shows that securitisation reduces the bank’s total risk exposure
by the amount of the placed volume only. In addition, there is a marginal
reduction in the expected loss. This can be attributed to the retention of the
FLP, which accounts for by far the largest portion of the expected losses on
the tranched portfolio.
In contrast to the first moment of the loss distribution, there is a clear
nominal and relative reduction in the 99.97% quantile, with the result that
the capital commitment also decreases markedly and falls to around 64 million
Euros. This value is only slightly above the nominal capital commitment when
the securitisation pool is fully placed (see Table 1).3
3
It can be shown that this somewhat higher capital commitment can be attributed
solely to the retention of the FLP. According to the model calculations, roughly
Asset Securitisation 211
Table 3. The bank’s risk situation when the FLP and the “Aaa” tranche are retained
Capital utilisation (1-year view)
Capital
Volume Expected loss 99.97% quantile commitment
Before
securitisation 2,110,000,000 0.43% 9,114,585 3.90% 82,290,000 3.47% 73,175,415
After
securitisation 2.079.725.000 0.44% 9,101,805 3.50% 72,790,375 3.06% 63,688,570
Table 4. Profits in case of the retention of the FLP and the “Aaa” tranche
Profits Income statement (5-year view) Expenditure
Reduction in
expected loss 63,902 1,298,463 Transaction costs
Profits Junior (FLP) Capital market costs
Economic
capital release 5,217,765 2,796,938 Mezzanine (Ba3 to Aa2)
Senior (Aaa)
Total 5,281,667 4,095,400
Total profits 1,186,267
Consequently, most of the economic risk has been placed with the mezza-
nine tranches and the retention of the senior tranches is not associated with
any significant risk for the bank. This result is also intuitively plausible as
the senior tranche has a higher credit rating than the bank VAR confidence is
and hence by the time the senior tranche defaults, chances are the bank will
already have defaulted.
The income statement effect, which derives from this securitisation strat-
egy, is presented in Table 4.
Compared with Table 2, the income statement expenditure in Table 4
is reduced by the capital market costs that are saved by not placing the
junior and senior parts of the portfolio. In addition to expenditure, profits also
decline. However, the profits do not decline as strongly as the expenditure.
The profits are reduced by roughly 6 million Euros by retaining the FLP
or through the expected losses not placed. By contrast, costs amounting to
some 9.9 million Euros are saved by not placing the FLP. This imbalance
can be explained, inter alia, by the fact that an FLP investor – in contrast
to the securitising institution – is invariably entitled to interest for assuming
the expected losses. In this case, this is 12% per annum. This leads to the
66% of the FLP or 5.4 million Euros are expected losses, which from the bank’s
perspective are not an economic risk. By contrast, the remainder of the FLP
amounting to roughly 1.6 million Euros is a high risk position, almost all of
which must be deducted from the economic capital.
212 M. Schmidtchen
conclusion that in the model used here it is not economically sound to place
the FLP.
The second central result is that placing the senior tranche is not eco-
nomically sound. Retaining this tranche saves around 1.6 million Euros in
expenditure. By contrast, the profits from released economic capital fall by
comparison with full securitisation by roughly 0.7 million Euros only.
Overall, the securitisation strategy of retaining the junior and senior risks
yields positive profits of around 1.2 million Euros.
The economic success of the securitisation strategy presented here is es-
sentially driven by the bank’s targeted return on economic capital and the
targeted solvency level. Whereas the targeted return on economic capital de-
termines the opportunity costs of the use of capital, the solvency level affects
the absolute amount of capital commitment for a specific credit risk. If the
solvency level is set at the 99.97% quantile as in the calculations in the exam-
ple, the total profits from the securitisation strategy rise the more the return
on capital increases. This situation is shown by the solid line in Fig. 3.
However, if the bank’s targeted solvency level falls from 99.97 to 99.90%,
this corresponds roughly to an “A” rating the line in Fig. 3 shifts to the right.
In this case, the break-even return on economic capital is roughly 11%, as
opposed to roughly 8.5% in the previous case.
This result can be attributed to the fact that the lower target credit rating
leads to lower capital commitment before securitisation and hence to overall
less economic capital being released through the securitisation. Accordingly,
the return on economic capital must be higher to ensure that securitisation
makes sense in terms of profits.4
3
Earnings (m)
0
7% 8% 9% 10% 11% 12% 13% 14% 15% 16% 17%
−1
−2
Return on Equity
Retention of FLP & Senior (99,97%) Retention of FLP & Senior (99,90%)
4
The same qualitative effect is achieved if the securitising bank anticipates an
average asset correlation of less than 8% (see footnote 2). In this case, too, less
economic capital would need to be retained before securitisation.
Asset Securitisation 213
6 Conclusion
The analysis has shown that a credit institution which specialises in SMEs
can significantly affect both the risk situation and the profits situation by
using the instrument of portfolio securitisation. In the model presented above
a portfolio strategy in which only the mezzanine part of a portfolio, which
contains most of the economic risk, is securitised is shown to be particularly
effective.
Net profits from this strategy depend, inter alia, on the target credit rat-
ing of the securitising institution, which affects the absolute amount of the
economic capitalisation. The better the target credit rating, the higher the
capitalisation before securitisation and the higher the amount of economic
capital released.
In addition to the target credit rating, the return on economic capital re-
quired is a further important factor behind the profitability of a transaction
of this kind. The return on economic capital required determines the oppor-
tunity costs of the capital utilisation. Together with the absolute amount of
economic capital released, it thus determines the profits from the optimal
securitisation strategy.
The fact that the economic success of the securitisation strategy depends
on the target credit rating and return on economic capital leads, in particular,
to institutions that set themselves a high solvency level and which require a
substantial return on economic capital can use the instrument of loan securi-
tisation to optimise profits.
Recent Advances in Credit Risk Management
1 Introduction
In the last decade, the market for credit related products as well as tech-
niques for credit risk management have undergone several changes. Financial
crises and a high number of defaults during the late 1990s have stimulated
not only public interest in credit risk management, but also their awareness of
its importance in today’s investment environment. Also the market for credit
derivatives has exhibited impressive growth rates. Active trading of credit
derivatives only started in the mid 1990s, but since then has become one of
the most dynamic financial markets. The dynamic expansion of the market
requires new techniques and advances in credit derivative and especially de-
pendence modelling among drivers for credit risk. Finally, the upcoming new
capital accord (Basel II) encourages banks to base their capital requirement
for credit risk on internal or external rating systems [4]. This regulatory body
under the Bank of International Settlements (BIS) becoming effective in 2007
aims to strengthen risk management systems of international financial institu-
tions. As a result, the majority of international operating banks sets focus on
an internal-rating based approach to determine capital requirements for their
loan or bond portfolios. Another consequence is that due to new regulatory
requirements there is an increasing demand by holders of securitisable assets
to sell or to transfer risks of their assets.
Recent research suggests that while a variety of advances have been made,
there are still several fallacies both in banks’ internal credit risk management
systems and industry wide used solutions. As [15] point out, the use of the
normal distribution for modelling the returns of assets or risk factors is not
adequate since they generally exhibit heavy tails, excess kurtosis and skewness.
All these features cannot be captured by the normal distribution. Also the
notion of correlation as the only measure of dependence between risk factors
or asset returns has recently been examined in empirical studies, for example
[7]. Using the wrong dependence structure may lead to severe underestimation
of the risk for a credit portfolio. The concept of copulas [13] allowing for more
216 F. Cowell et al.
1 week
140
Kernel Estimate
Stable Fit
Gaussian Fit
120
100
80
60
40
20
0
−0.06 −0.04 −0.02 0 0.02 0.04 0.06
the basis of ETL leads to higher risk-adjusted returns. ETL compared to VaR
possesses a number of advantages, among others ETL is a smooth function
which can be readily optimized. Moreover, ETL reflects only the downside and
does not penalize for upside potential of the portfolio/asset returns, which is
not true for the standard deviation.
Recently, a variety of alternative risk measures has been introduced in the
literature. One may also like to consider individual assets and to ascertain
how much risk each asset contributes to the portfolio. Hence, also marginal
(incremental) statistics should be considered. For an overview on desirable
properties of a risk measure see for example [3], [19], [20] or [16].
Yj,t = βj,0 + βj,1 X1,t + βj,2 X2,t + ..... + βj,m Xm,t + vj,t . (6)
224 F. Cowell et al.
Hereby, Xk,t−1 and Xk,t−2 denote the lagged values of the variable Xk ,
while ek,t denotes an error term that is assumed to be i.i.d, i.e. ek,t ∼ N (0, σe ).
Obviously, based on parameter estimates for equations (6)-(7) a macro-
economic index also for future periods can be estimated. This index can be
used to determine conditional default probabilities pj,t for rating class j in
period t. The author suggests a logit model of the form:
1
pj,t = . (8)
1 + e−Yj,t
while other models could be applied.
Finally, for estimation of the conditional migration matrix a shifting pro-
cedure is used that redistributes the probability mass within each row of the
unconditional migration matrix [24]. The shift operator is written in terms of
a matrix S = {Sij } and the shift procedure is accomplished by
associated with the bins defined above best approximate the given year’s ob-
served transition rates. The estimation problem then results in determining
ρ such that the distance between the forecasted conditional transition matrix
and empirically observed migrations is minimized, see e.g. [10]. Note that [23]
extends the one-factor model representation by a multi-factor, Markov chain
model for rating migrations and credit spreads.
The different approaches point out the importance to incorporate busi-
ness cycle effects into the estimation of credit migration matrices. [10], [12]
and more recently [21] show that the conditional approach outperforms a
naive approach of simply taking historical average or previous year’s transi-
tion matrices.
There are four key steps in the Monte Carlo approach to credit risk modelling
in the asset value model:
Step 1. Modelling the dependence structure between market risk factors
and the credit risk drivers.
Step 2. Scenario Generation - each scenario corresponds to a possible “state
of the world” at the end of the risk horizon which the portfolio risk is estimated
for. For purposes of this article, the “state of the world” is just the credit rating
of each of the obligors in the portfolio and the corresponding values of the
market risk factors affecting the portfolio.
Step 3. Portfolio valuation - for each scenario, the software evaluates the
portfolio to reflect the new credit ratings and the values of the market risk
factors. This step creates a large number of possible future portfolio values.
Recent Advances in Credit Risk Management 227
Under the asset value models, the general assumption is that the driver of
credit events is the asset value of a company. The dependence structure be-
tween the asset values of two firms can be approximated by the dependence
structure between the stock prices of those firms. In case there is no stock price
information for a given obligor we employ the idea of segmentation described
in CreditMetrics. The essence of this approach is that the user determines the
percentage obligor volatility allocation among the volatilities of certain mar-
ket indices and explains the dependence between obligors by the dependence
of the market indices that drive obligors’ volatilities.
As discussed in Sect. 3 modelling the dependence structure requires a
greater flexibility than the one offered by the correlation concept. Hence, Cog-
nity Credit Risk Module supplies flexible dependence structure models:
• A copula approach
• A subordinated model approach
• A simplified approach using correlations as a measure for the dependency
for comparison purposes
A copula suitable for modelling dependencies between financial variables
and credit drivers in particular, should be flexible enough to capture the
dependence of extreme events and also asymmetries in dependence. There
are few flexible multivariate copula functions which can be applied to large-
dimensional problems. Examples include the Gaussian copula (the one behind
the multivariate Gaussian distribution), the multivariate Student’s t-copula,
etc. Cognity utilizes a flexible copula model which contains the multivariate
Student’s t-copula as a special case and allows for asymmetry in the depen-
dence model as well as for dependence in the extreme events. The copula model
is based on an asymmetric version of the multivariate Student’s t-distribution
and is flexible enough for all market conditions including severe crises in which
the asymmetric dependence is most pronounced.
the subordinated approach arises from the so-called subordinated distri-
butions. The symmetric stable distribution discussed in Section 2 is one rep-
resentative of this class. In particular, a random variable X is said to be
subordinated if its distribution allows the following stochastic representation:
228 F. Cowell et al.
Scenario Generation
X = µ + γY + g(Y )Z (11)
Two models should be defined and estimated under the SDR approach: the
first model provides an econometric approach for default probabilities of a
segment based on explanatory variables like macro-factors, indices, etc. The
second model deals with a time series approach for the explanatory variables.
Default probability models are evaluated for each user-defined segment.
The segment definitions can be flexible based on criteria like the credit rating,
the industry, the region and the size of the company, provided that the time
series of default rates are available for each of the segments. The explanatory
variables that might be appropriate to represent the systematic risk of the
default rates in the chosen country/industry/segment depend on the nature
of the portfolio and might comprise industry indices, macro variables (GDP,
unemployment rate) as well as long-term interest rates or exchange rates,
etc. When defining the model for the default probability of a segment based
on explanatory variables (macro-factors, indices, etc.) we use historical data
Recent Advances in Credit Risk Management 231
for default frequencies in a given segment and historical time series for the
explanatory variables. The idea is similar to the CreditPortfolioView described
in Sect. 5. Hereby, a function f is chosen and estimated such that
where DFs,t is the default frequency in the segment s for the time period t; Xi,t
is the value of the i-th explanatory variable at time t, i = 1, . . . , N . It should
be mentioned that in general explanatory variables can be observable factors
but also factors estimated by the means of fundamental factor analysis based
on stock returns in a given segment or latent variables coming from statistical
factor models.
The second model is a time-series model for the explanatory variables.
The usual way to model dependent variables (as suggested also in CreditPort-
folioView) is to employ some kind of ARM A(p, q) model. That is the same
as assuming that
p
q
Xt = a0 + ai Xt−i + bj εt−i + εt , (13)
i=1 j=1
Xt = A1 Xt−1 + Et (14)
There are five key steps in the Monte Carlo approach to credit risk modelling
based on stochastic modelling of the default rate:
Step 1. Build econometric models for default rates and explanatory risk
variables. Based on the explanatory risk variables (macro-factors, indices, etc.)
232 F. Cowell et al.
an econometric model for the default probability for each segment is fitted
using historical data for default probabilities in a given segment and historical
time series data for the explanatory variables.
Step 2. Generate scenarios – each scenario corresponds to a possible “state
of the world” at the end of the risk horizon. Here, the “state of the world” is
a set of values for the market and explanatory risk variables defined.
Step 3. Estimate default probabilities under each scenario for each segment
using the scenario (simulation) values of the explanatory variables and the
model estimated in step 1. Sample a new default rate for each obligor and
adjust the respective migration probabilities based on the new default rate.
Determine the credit rating status of each obligor based on the new migration
and default probabilities. Technically this is accomplished by making use of
a uniform (0,1) random variable, which is drawn for each counterparty and
each simulation of the default rate.
Step 4. Portfolio valuation – for each scenario, revalue the portfolio to
reflect the new credit status of the obligor and the values of the market risk
factors. This step generates a large number of possible future portfolio values.
Step 5. Summarize results – once the scenarios in the previous steps are
generated, we come up with an estimate for the distribution of portfolio values.
We may then choose to report any descriptive statistics for this distribution.
7 Conclusion
Acknowledgement
The authors are grateful to Georgi Mitov (FinAnalytica) and Dobrin Penchev
(FinAnalytica) for the fruitful comments, suggestions and computational as-
sistance. We also thank Zari Rachev (University of Karlsruhe, UCSB and
FinAnalytica) and Stoyan Stoyanov (FinAnalytica) for helpful discussions.
References
[1] Alessandrini, F., 1999. Credit Risk, Interest Rate Risk, and the Business
Cycle. Journal of Fixed Income 9 (2), 42–53.
[2] Allen, L., Saunders, A., 2003. A Survey of Cyclical Effects in Credit Risk
Measurement Model. BIS Working Paper 126.
[3] Artzner, P., Delbaen, F., Eber, J.-M., Heath, D., 1999. Coherent measures
of risk. Mathematical Finance 9 (3), 203–228.
[4] Basel Committee on Banking Supervision, 2001. The new Basel Capital
Accord, Second Consultative Document.
[5] Belkin, B., Forest, L., Suchower, S., 1998. The Effect of Systematic Credit
Risk on Loan Portfolio Value-at-Risk and Loan Pricing. CreditMetrics
Monitor.
[6] Belkin, B., Forest, L., Suchower, S., 1998. A one-parameter Representa-
tion of Credit Risk and Transition Matrices. CreditMetrics Monitor.
[7] Embrechts, P., McNeil, A., Straumann, D., 1999. Correlation and Depen-
dence in Risk Management: Properties and Pitfalls. In: Risk management:
value at risk and beyond, ed. Dempster, M.
[8] Fama, E., 1965. The Behaviour of Stock Market Prices. Journal of Busi-
ness 38, 34–105.
234 F. Cowell et al.
[9] Helwege, J., Kleiman, P., 1997. Understanding aggregate default rates of
high-yield bonds. Journal of Fixed Income 7(1), 55–61.
[10] Kim, J., 1999. Conditioning the Transition Matrix. Risk Credit Risk
Special Report, 37–40.
[11] Mandelbrot, B., 1963. The Variation of certain speculative Prices. Jour-
nal of Business 36, 394–419.
[12] Nickell, P., Perraudin, W., Varotto, S., 2000. Stability of Rating Transi-
tions. Journal of Banking and Finance 1-2, 203–227.
[13] Picone, D., 1959. Fonctions de répartition à n dimensions et leurs marges.
Working Paper, Cass Business School 8, 229–231.
[14] Rachev, S., Martin, R., Racheva, B., Stoyanov, S., 2006. Stable ETL
Portfolios and Extreme Risk Management. Working Paper.
[15] Rachev, S., Mittnik, S., 2000. Stable Paretian Models in Finance. Wiley,
New York.
[16] Rachev, S., Ortobelli, S., Stoyanov, S., Fabozzi, F., Biglova, A., 2006. De-
sirable Properties of an Ideal Risk Measure in Portfolio Theory. Working
Paper.
[17] Samorodnitsky, G., Taqqu, M., 1994. Stable Non-Gaussian Random Pro-
cesses. Chapman & Hall, New York.
[18] Schweizer, B., Sklar, A., 1983. Probabilistic Metric Spaces. North Holland
Elsevier, New York.
[19] Szegö, G., 2002. Measures of Risk. Journal of Banking and Finance 26(7),
1253–1272.
[20] Szegö, G., 2004. Risk Measures for the 21st Century. Wiley, Chichester.
[21] Trück, S., 2008. Forecasting Credit Migration Matrices with Business
Cycle Effects - A Model Comparison. European Journal of Finance 14(5),
359–379.
[22] Trück, S., Rachev, S., 2005. Credit Portfolio Risk and PD Confidence
Sets through the Business Cycle. Journal of Credit Risk 1(4).
[23] Wei, J., 2003. A Multi-Factor, Credit Migration Model for Sovereign
and Corporate Debts. Journal of International Money and Finance 22,
709–735.
[24] Wilson, T., 1997. Measuring and Managing Credit Portfolio Risk.
McKinsey & Company.
[25] Wilson, T., 1997. Portfolio Credit Risk I/II. Risk 10.
Stable ETL Optimal Portfolios and Extreme
Risk Management
1 Introduction
We introduce a practical alternative to Gaussian risk factor distributions
based on Svetlozar Rachev’s work on Stable Paretian Models in Finance (see
[4]) and called the Stable Distribution Framework. In contrast to normal dis-
tributions, stable distributions capture the fat tails and the asymmetries of
real-world risk factor distributions. In addition, we make use of copulas, a gen-
eralization of overly restrictive linear correlation models, to account for the
dependencies between risk factors during extreme events, and multivariate
ARCH-type processes with stable innovations to account for joint volatility
clustering. We demonstrate that the application of these techniques results in
more accurate modeling of extreme risk event probabilities, and consequently
delivers more accurate risk measures for both trading and risk management.
Using these superior models, VaR becomes a much more accurate measure of
downside risk. More importantly Stable Expected Tail Loss (SETL) can be
accurately calculated and used as a more informative risk measure for both
market and credit portfolios. Along with being a superior risk measure, SETL
enables an elegant approach to portfolio optimization via convex optimiza-
tion that can be solved using standard scalable linear programming software.
We show that SETL portfolio optimization yields superior risk adjusted re-
turns relative to Markowitz portfolios. Finally, we introduce an alternative
investment performance measurement tools: the Stable Tail Adjusted Return
Ratio (STARR), which is a generalization of the Sharpe ratio in the Stable
Distribution Framework.
“When anyone asks me how I can describe my experience of nearly 40
years at sea, I merely say uneventful. Of course there have been winter gales
and storms and fog and the like, but in all my experience, I have never been
in an accident of any sort worth speaking about. I have seen but one vessel in
236 S.T. Rachev et al.
distress in all my years at sea (...) I never saw a wreck and have never been
wrecked, nor was I ever in any predicament that threatened to end in disaster
of any sort.”
E.J. Smith, Captain, 1907, RMS Titanic
“It has often been argued that the true distributions returns (even
after standardizing by the volatility) imply a larger probability of
extreme returns than that implied from the normal distribution. Al-
though we could try to specify a distribution that fits returns better, it
would be a daunting task, especially if we consider that the new distri-
bution would have to provide a good fit across all asset classes.” (Tech-
nical Manual, RMG, 2001, http://www.riskmetrics.com/publications/
index.html).
Stable ETL Optimal Portfolios and Extreme Risk Management 237
In spite of wide-spread awareness that most risk factor distributions are heavy-
tailed, to date, risk management systems have essentially relied either on his-
torical, or on univariate and multivariate normal (or Gaussian) distributions
for Monte Carlo scenario generation. Unfortunately, historical scenarios only
capture conditions actually observed in the past, and in effect use empirical
probabilities that are zero outside the range of the observed data, a clearly
undesirable feature. On the other hand Gaussian Monte Carlo scenarios have
probability densities that converge to zero too quickly (exponentially fast) to
accurately model real-world risk factor distributions that generate extreme
losses. When such large returns occur separately from the bulk of the data
they are often called outliers.
Figure 1 below shows quantile–quantile (qq)-plots of daily returns versus
the best-fit normal distribution of nine randomly selected microcap stocks for
the two-year period 2000–2001. If the returns were normally distributed, the
quantile points in the qq-plots would all fall close to a straight line. Instead
they all deviate significantly from a straight line (particularly in the tails),
reflecting a higher probability of occurrence of extreme values than predicted
by the normal distribution, and showing several outliers.
Such behavior occurs in many asset and risk factor classes, including well-
known indices such as the S&P 500, and corporate bond prices. The latter
are well known to have quite non-Gaussian distributions that have substantial
negative skews to reflect down-grading and default events. For such returns,
non-normal distribution models are required to accurately model the tail be-
havior and compute probabilities of extreme returns.
Various non-normal distributions have been proposed for modeling ex-
treme events, including:
• Mixtures of two or more normal distributions
• t-distributions, hyperbolic distributions, and other scale mixtures of nor-
mal distributions
• Gamma distributions
• Extreme value distributions
• Stable non-Gaussian distributions (also known as Lévy-stable and Pareto-
stable distributions)
238 S.T. Rachev et al.
CMED.returns
AER.returns
0.1
0.0
−0.1
−0.2
−0.3
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Quantiles of Standard Normal Quantiles of Standard Normal Quantiles of Standard Normal
CVST.returns
0.6
0.00
0.2
−0.10
−0.2
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Quantiles of Standard Normal Quantiles of Standard Normal Quantiles of Standard Normal
0.10
−0.1 0.0 0.1 0.2
−0.4 −0.2 0.0 0.2 0.4
WFHC.returns
ALLE.returns
IQW.returns
0.00
−0.10
−3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3 −3 −2 −1 0 1 2 3
Quantiles of Standard Normal Quantiles of Standard Normal Quantiles of Standard Normal
Among the above, only stable distributions have attractive enough math-
ematical properties to be a viable alternative to normal distributions in trad-
ing, optimization and risk management systems. A major drawback of all
alternative models is their lack of stability. Benoit Mandelbrot [3] demon-
strated that the stability property is highly desirable for asset returns. These
advantages are particularly evident in the context of portfolio analysis and
risk management.
An attractive feature of stable models, not shared by other distribution
models, is that they allow generation of Gaussian-based financial theories
and, thus allow construction of a coherent and general framework for financial
modeling. These generalizations are possible only because of specific proba-
bilistic properties that are unique to (Gaussian and non-Gaussian) stable laws,
namely: the stability property, the central limit theorem, and the invariance
principle for stable processes.
Benoit Mandelbrot [3], then Eugene Fama [2], provided seminal evidence
that stable distributions are good models for capturing the heavy-tailed
(leptokurtic) returns of securities. Many follow-on studies came to the same
conclusion, and the overall stable distributions theory for finance is provided
in the definitive work of Rachev and Mittnik [4], see also [5, 6, 9].
Stable ETL Optimal Portfolios and Extreme Risk Management 239
But in spite the convincing evidence, stable distributions have seen vir-
tually no use in capital markets. There have been several barriers to the
application of stable models, both conceptual and technical:
• Except for three special cases, described below, stable distributions have
no closed form expressions for their probability densities.
• Except for normal distributions, which are a limiting case of stable distri-
butions (with α=2 and β = 0, stable distributions have infinite variance
and only a mean value for α > 1.
• Without a general expression for stable probability densities, one cannot
directly implement maximum likelihood methods for fitting these densities,
even in the case of a single (univariate) set of returns.
The availability of practical techniques for fitting univariate and mul-
tivariate stable distributions to asset and risk factor returns has been the
barrier to the progress of stable distributions in finance. Only the recent de-
velopment of advanced numerical methods has removed this obstacle. These
patent-protected methods are at the foundation of the CognityTM risk man-
agement and portfolio optimization software system (see further comments in
Sect. 5.6).
where
1 x−µ
fµ,σ (x) = f
σ σ
is any probability density function in a location-scale family for X:
α) ) **
−σ α |t| ) 1 − iβsgn(t) tan πα 2* + iµt, α = 1
log F (t) =
−σ |t| 1 − iβ π2 sgn(t) log |t| + iµt, α = 1
Symmetric PDFs
0,7
0,6
0,5
α=0.5
0,4
α=1
0,3 α=1.5
α=2
0,2
0,1
0
−5 −4 −3 −2 −1 0 1 2 3 4 5
values of α decrease the distribution exhibits fatter tails and more peakedness
at the origin.
The
√ case of α = 2 and β = 0 and with the reparameterization in scale,
= 2σ, yields the Gaussian distribution, whose density is given by:
σ
1 (x−µ)2
fµ,σ (x) = √ e− 2σ 2 .
2π
σ
The case α=1 and β = 0 yields the Cauchy distribution with much fatter
tails than the Gaussian, and is given by:
!−1
2
1 x−µ
fµ,σ (x) = 1+
π·σ σ
Aside from the Gaussian, Cauchy, and one other special case of stable distri-
bution for a positive random variable with α = 0.5, there is no closed form
expression for the probability density of a stable random variable.
Thus one is not able to directly estimate the parameters of a stable distri-
bution by the method of maximum likelihood. To estimate the four parameters
Stable ETL Optimal Portfolios and Extreme Risk Management 241
0,3
0,25
β=0
0,2 β=0.25
β=0.5
0,15 β=0.75
β=1
0,1
0,05
0
−5 −4 −3 −2 −1 0 1 2 3 4 5
NORMAL DENSITY
EMPIRICAL DENSITY
1.5
1.0
0.5
0.0
of the stable laws, the CognityTM system uses a special patent-pending ver-
sion of the FFT (Fast Fourier Transform) approach to numerically calculate
the densities with high accuracy, and then applies maximum likelihood esti-
mation (MLE) to estimate the parameters.
The results from applying the CognityTM stable distribution modeling to
the DJIA daily returns from January 1, 1990 to February 14, 2003 is displayed
in Fig. 4. The figure shows the left-hand tail detail of the resulting stable den-
sity, along with that of a normal density fitted using the sample mean and
sample standard deviation, and that of a non-parametric kernel density esti-
mate (labeled “Empirical” in the plot legend). The parameter estimates are:
242 S.T. Rachev et al.
A daily return smaller than −0.04 with the stable distribution occurs with
probability 0.0066, or roughly seven times every four years, whereas such a
return with the normal fit occurs on the order of once every four years.
Similarly, a return smaller than −0.05 with the stable occurs about once
per year and with the normal fit about once every 40 years. Clearly the nor-
mal distribution fit is an exceedingly optimistic predictor of DJIA tail return
values.
Figure 5 below displays the central portion of the fitted densities as well
as the tails, and shows that the normal fit is not nearly peaked enough near
DJIA
Empirical Density
Stable Fit
50 Gaussian (Normal) Fit
40
30
20
10
0
−0.08 −0.06 −0.04 −0.02 0 0.02 0.04 0.06 0.08
Fig. 5. The fitted stable and normal densities together with the empirical density
Stable ETL Optimal Portfolios and Extreme Risk Management 243
the origin as compared with the empirical density estimate (even though the
GARCH model was applied), while the stable distribution matches the em-
pirical estimate quite well in the center as well as in the tails.
by a strictly positive stable random variable with index αi /2, i=1,2,. . . ,K.
The vector of stable random variable scale multipliers is usually independent
of the normal scenario vectors, but it can also be dependent. See for example
Rachev and Mittnik [4], and [5, 6, 9].
Another very promising approach to building the cross-sectional depen-
dence model is through the use of copulas, an approach that is quite attrac-
tive because it allows for modeling higher correlations during extreme market
movements, thereby accurately reflecting lower portfolio diversification at such
times. The next section briefly discussion copulas.
Why Copulas?
where
C(ui ) = ui for i = 1, 2, · · · , n.
It is known that for any multivariate cumulative distribution function:
F (x1 , x2 , · · · , xn ) = P (X1 ≤ x1 , X2 ≤ x2 , · · · Xn ≤ xn )
−2 0 2
RISK FACTOR 1
infrequent occurrence of very large returns. This can be seen by noting that
the density contours of points in the scatter plot are somewhat elliptical near
the origin, but are nowhere close to elliptical for more extreme events. This
situation is in contrast to a Gaussian linear dependency relationship where
the density contours are expected to be elliptical.
It is well known that asset returns and risk factors returns exhibit volatility
clustering, and that even after adjusting for such clustering the returns will
still be non-normal and contain extreme values. There may also be some serial
dependency effects to account for. In order to adequately model these collec-
tive behaviors we recommend using ARIMA models with an ARCH/GARCH
“time-varying” volatility input, where the latter has non-normal stable inno-
vations. This approach is more flexible and accurate than the commonly used
simple exponentially weighted moving average (EWMA) volatility model, and
provides accurate time-varying estimates of VaR and expected tail loss (ETL)
risk measures. See Sect. 4 for discussion of ETL vs. VaR that emphasizes the
advantages of ETL. However, we stress that those who must use VaR to sat-
isfy regulatory requirements will get much more accurate results with stable
VaR than with normal VaR, as the following example vividly shows.
Consider the following portfolio of Brady bonds:
• Brazil C 04/14
• Brazil EIB 04/06
• Venezuela DCB Floater 12/07
• Samsung KRW Ord Shares
• Thai Farmers Bank THB
We have run normal, historical and stable 99% (1% tail probability) VaR
calculations for one-year of daily data from January 9, 2001 to January 9, 2002.
We used a moving window with 250 historical observations for the normal
VaR model, 500 for the historical VaR model and 700 for the stable VaR
model. For each of these cases we used a GARCH(1,1) model for volatility
clustering of the risk factors, with stable innovations. We back-tested these
VaR calculations by using the VaR values as one-step ahead predictors, and
got the results shown in Fig. 8.
The figure shows: the returns of the Brady bond portfolio (top curve); the
normal+EWMA (a la RiskMetrics) VaR (curve with jumpy behavior, just
below the returns); the historical VaR (the smoother curve mostly below but
sometimes crossing the normal+EWMA VaR); the stable+GARCH VaR (the
bottom curve). The results with regard to exceedances of the 99% VaR, and
keeping in mind Basel II guidelines, may be summarized as follows:
• Normal 99% VaR produced 12 exceedances (red zone)
• Historical 99% VaR produced 9 exceedances (on upper edge of yellow zone)
Stable ETL Optimal Portfolios and Extreme Risk Management 247
0.05
Observed Portf. Returns
Stable 99% VaR
0.04 Normal 99% VaR
Historical 99% VaR
0.03
0.02
0.01
−0.01
−0.02
−0.03
−0.04
−0.05
0 50 100 150 200 250 300
• Stable 99% VaR produced 1 exceedence and nearly two (well in the green
zone)
Clearly stable (+GARCH) 99% VaR produces much better results with
regard to Basel II compliance. This comes at the price of higher initial cap-
ital reserves, but results in a much safer level of capital reserves and a very
clean bill of health with regard to compliance. Note that organizations in the
red zone will have to increase their capital reserves by 33%, which at some
times for some portfolios will result in larger capital reserves than when using
the stable VaR, this in addition to being viewed as having inadequate risk
measures relative to the organization using stable VaR.
• VaR does not give any indication of the risk beyond the quantile, and so
provides very weak information on downside risk.
• VaR portfolio optimization is a non-convex, non-smooth problem with
multiple local minima that can result in portfolio composition disconti-
nuities. Furthermore it requires complex calculation techniques such as
integer programming.
• VaR is not sub-additive; i.e. the VaR of the aggregated portfolio can be
larger than the sum of the VaR’s of the sub-portfolios.
• Historical VaR limits the range of the scenarios to data values that have
actually been observed, while normal Monte Carlo tends to seriously un-
derestimate the probability of extreme returns. In either case, the prob-
ability functions beyond the sample range are either zero or excessively
close to zero.
Expected Tail Loss (ETL) is simply the average (or expected value) loss con-
ditioned on the loss being larger than VaR. ETL is also known as Conditional
Value-at-Risk (CVaR), or Expected Shortfall (ES). (We assume that the un-
derlying return distributions are absolutely continuous, and therefore, ETL
is equal to CVaR). As such ETL is intuitively much more informative than
VaR. We note however that ETL offers little benefit to investors who use a
normal distribution to calculate VaR at the usual 99% confidence limit (1%
tail probability). The reason is that the resulting VaR and ETL values differ
by very little, specifically:
• For CI = 1%, VaR = 2.336 and ETL = 2.667
ETL really comes into its own when coupled with stable distribution mod-
els that capture leptokurtic tails (“fat tails”). In this case ETL and VaR values
will be quite different, with the resulting ETL often being much larger than
the VaR.
As in the graph in Fig. 9, consider the time series of daily returns for the
stock OXM from January 2000 to December 2001. Observe the occurrences
of extreme values.
While this series also displays obvious volatility clustering that deserves
to be modeled as described in Sect. 4.3, we shall ignore this aspect for the
moment. Rather, here we provide a compelling example of the difference be-
tween ETL and VaR based on a well-fitting stable distribution, as compared
with a poor fitting normal distribution.
Figure 10 shows a histogram of the OXM returns with a normal density
fitted using the sample mean and sample standard deviation, and a stable
density fitted using maximum-likelihood estimates of the stable distribution
parameters. The stable density is shown by the solid line and the normal
density is shown by the dashed line. The former is obviously a better fit than
the latter, when using the histogram of the data values as a reference. The
Stable ETL Optimal Portfolios and Extreme Risk Management 249
OXM Returns
0.10
0.05
0.0
−0.05
−0.10
−0.15
Fig. 10. The stable and normal 99% VaR for OXM
estimated stable tail thickness index is α̂ = 1.62. The 1% VaR values for the
normal and stable fitted densities are 0.047 and 0.059 respectively, a ratio of
1.26 which reflects the heavier-tailed nature of the stable fit.
Figure 11 displays the same histogram and fitted densities with 1% ETL
values instead of the 1% VaR values. The 1% ETL values for the normal
and stable fitted densities are 0.054 and 0.174 respectively, a ratio of a little
over three-to-one. This larger ratio is due to the stable density’s heavy tail
contribution to ETL relative to the normal density fit.
250 S.T. Rachev et al.
30
25
20
15
Fig. 11. The stable and normal 99% ETL for OXM
1
n
+
F (w, γ) = γ + [w ri − γ] ,
ε · n i=1
4
RISK ADJUSTED RETURN (MU / VAR)
Stable VaR
VaR
3
2
1
0
SETL
VaR
3
2
1
0
A SETL optimal portfolio is one that minimizes portfolio expected tail loss
subject to a constraint of achieving expected portfolio returns at least as large
as an investor defined level, along with other typical constraints on weights,
where both quantities are evaluated in the SETL framework. Alternatively,
a SETL optimal portfolio solves the dual problem of maximizing portfolio
expected return subject to a constraint that portfolio expected tail loss is
not greater than an investor defined level, where again both quantities are
evaluated in the SETL framework. In order to define the above ETL precisely
we use the following quantities:
Rp : the random return of portfolio p
SERp : the stable distribution expected return of portfolio p
Lp = −Rp + SERp : the loss of portfolio p relative to its expected return
ε: a tail probability of the SETL distribution Lp
SV aRp (ε): the stable distribution Value-at-Risk for portfolio p
The latter is defined by the equation
where the probability is calculated in the SETL framework, that is SV aRp (ε)
is the ε-quantile of the stable distribution of Lp . In the value-at-risk literature
(1 − ε) × 100% is called the confidence level. Here we prefer to use the simpler,
unambiguous term tail probability. Now we define SETL of a portfolio p as
subject to
SERq ≥ µ.
Here we use ωα to mean either the resulting portfolio weights or the label
for the portfolio itself, depending upon the context. The subscript α to remind
us that we are using a GMstable distribution modeling approach (which entails
different stable distribution parameters for each asset and risk factor). In other
words the SETL optimum portfolio ωα minimizes the expected tail loss among
all portfolios with mean return at least µ , for fixed tail probability ε and asset
allocation constraints λ. Alternatively, the SETL optimum portfolio ωα solves
the dual problem
ωα (η|ε) = arg max SERq
q∈Q
subject to
SET Lq (ε) ≤ η.
The SETL efficient frontier is given by ωα (µ|ε) as a function of µ for fixed
ε, as indicated in Fig. 15. If the portfolio includes cash account with risk free
rate rf , then the SETL efficient frontier will be the SETL capital market line
(CM Lα ) that connects the risk-free rate on the vertical axis with the SETL
tangency portfolio (Tα ), as indicated in the figure.
We now have a SETL separation principal analogous to the classical sep-
aration principal: The tangency portfolio Tα can be computed without refer-
ence to the risk-return preferences of any investor. Then an investor chooses
a portfolio along the SETL capital market line CM Lα according to his/her
risk-return preference.
SER CMLα
SETL efficient frontier
Tα
r1
SETL
Fig. 15. The SETL efficient frontier and the capital market line
Stable ETL Optimal Portfolios and Extreme Risk Management 257
Keep in mind that in practice when a finite sample of returns one ends
up with a SETL efficient frontier, tangency portfolio and capital market line
that are estimates of true values for these quantities.
While the SETL investor has optimal portfolios described above, the
Markowitz investor is not aware of the SETL framework and constructs
a mean-variance optimal portfolio. We assume that the Markowitz investor
operates under the same assumptions A1-A4 as the SETL investor. Let ERq
be the expected return and σq the standard deviation of the returns of a
portfolio q. The Markowitz investor’s optimal portfolio is
ω2 (µ) = min σq
q∈Q
subject to
ERq ≥ µ
along with the other constraints λ.
The Markowitz optimal portfolio can also be constructed by solving the
obvious dual optimization problem.
The subscript 2 is used in ω2 as a reminder that α = 2 you have the
limiting Gaussian distribution member of the stable distribution family, and
in that case the Markowitz portfolio is optimal. Alternatively you can think
of the subscript 2 as a reminder that the Markowitz optimal portfolio is a
second-order optimal portfolio, i.e., an optimal portfolio based on only first
and second moments.
The Markowitz investor ends up with a different portfolio, i.e., a differ-
ent set of portfolio weights with different risk versus return characteristics,
than the SETL investor. It is important to note that the performance of the
Markowitz portfolio, like that of the SETL portfolio, is evaluated under a
GMstable distributional model. If in fact the distribution of the returns were
exactly multivariate normal (which they never are) then the SETL investor
and the Markowitz investor would end up with one and the same optimal
portfolio. However, when the returns are non-Gaussian SETL returns, the
Markowitz portfolio is sub-optimal. This is because the SETL investor con-
structs his/her optimal portfolio using the correct distribution model, while
the Markowitz investor does not. Thus the Markowitz investors frontier lies
below and to the right of the SETL efficient frontier, as shown in Fig. 16, along
with the Markowitz tangency portfolio T2 and Markowitz capital market line
CM L2 .
As an example of the performance improvement achievable with the SETL
optimal portfolio approach, we computed the SETL efficient frontier and the
Markowitz frontier for a portfolio of 47 micro-cap stocks with the smallest
alphas from the random selection of 182 micro-caps in Sect. 3.1. The results
258 S.T. Rachev et al.
SER CMLα
SETL efficient frontier
CML2
Markowitz frontier
Tα
T2
rf
xe
SETL
Fig. 16. The SETL and the Markowitz efficient frontiers
SETL
EXPECTED RETURN (Basis Points per Day)
Markowitz
50
40
30
20
10
TAIL PROBABILITY = 1%
0
are displayed in Fig. 17. The results are based on 3,000 scenarios from the
fitted GMstable distribution model based on two years of daily data during
years 2000 and 2001. We note that, as is generally the case, each of the 47
stock returns has its own estimate stable tail index α̂i , i = 1, 2, . . . , 47.
Here we have plotted values of T ailRisk = ε · SET L(ε), for ε = 0.01,
as a natural decision theoretic risk measure, rather than SET L(ε) itself. We
note that over a considerable range of tail risk the SETL efficient frontier
dominates the Markowitz frontier by 14–20 bp’s daily!
Stable ETL Optimal Portfolios and Extreme Risk Management 259
We note that the 47 micro-caps with the smallest alphas used for this
example have quite heavy tails as indicated by the box plot of their estimated
alphas shown below.
Here the median of the estimated alphas is 1.38, while the upper and lower
quartiles are 1.43 and 1.28 respectively. Evidently there is a fair amount of
information in the non-Gaussian tails of such micro-caps that can be exploited
by the SETL approach.
We conclude that the risk adjusted return of the SETL optimal portfolio
ωα is generally superior to the risk adjusted return of the Markowitz mean
variance optimal portfolio ω2 . The SETL framework results in improved in-
vestment performance.
Rachev Ratio
The Rachev Ratio (R-ratio) is the ratio between the expected excess tail-
return at a given confidence level and the expected excess tail loss at another
confidence level:
ET Lγ1 (x (rf − r))
ρ(r) =
ET Lγ2 (x (r − rf ))
Here the levels γ1 and γ2 are in [0,1], x is the vector of asset allocations
and r − rf is the vector of asset excess returns. Recall that if r is the portfolio
return, and L = −r is the portfolio loss, we define the expected tail loss as
ET Lα% (r) = E(L/L > V aRα% ), whereP (L > V aRα% ) = α, and α is in (0,1).
The R-Ratio is a generalization of the STARR. Choosing appropriate levels γ1
and γ2 in optimizing the R-Ratio the investor can seek the best risk/return
profile of her portfolio. For example, an investor with portfolio allocation
maximizing the R-Ratio with γ1 = γ2 =0.01 is seeking exceptionally high
returns and protection against high losses.
We mentioned earlier that when using SET Lp (ε) rather than V aRp (ε), risk
managers and portfolio optimizers may wish to use other values of ε than the
conventional VaR values of .01 or .05, for example values such as 0.1, 0.15,
0.2, 0.25 and 0.5 may be of interest. The choice of a particular ε amounts to
a choice of particular risk measure in the SETL family of measures, and such
a choice is equivalent to the choice of a utility function. The tail probability
parameter ε is at the asset manager’s disposal to choose according to his/her
asset management and risk control objectives.
Note that choosing a tail probability ε is not the same as choosing a risk
aversion parameter. Maximizing
for various choices of risk aversion parameter c for a fixed value of ε merely
corresponds to choosing different points along the SETL efficient frontier. On
the other hand changing ε results in different shapes and locations of the SETL
efficient frontier, and corresponding different SETL excess profits relative to
a Markowitz portfolio.
It is intuitively clear that increasing ε will decrease the degree to which a
SETL optimal portfolio depends on extreme tail losses. In the limit of ε = 0.5,
which may well be of interest to some managers since it uses the average
Stable ETL Optimal Portfolios and Extreme Risk Management 261
loss below zero of Lp as its penalty function, small to moderate losses are
mixed in with extreme losses in determining the optimal portfolio. There is
some concern that some of the excess profit advantage relative to Markowitz
portfolios will be given up as ε increases. Our studies to date indicate, not
surprisingly, that this effect is most noticeable for portfolios with smaller
stable tail index values.
It will be interesting to see going forward what values of ε will be used by
fund managers of various types and styles.
A generalization of the SETL efficient frontier is the R-efficient fron-
tier, obtained by replacing the stable portfolio expected return SERp in
SERp − c · SET Lp (ε) by the excess tail return , the numerator in the R- ratio.
R-efficient frontier allows for fine tuning of the tradeoff between high excess
means returns and protection against large loss.
The SETL framework described in this paper has been implemented in the
CognityTM Risk Management and Portfolio Optimization product. This
product contains solution modules for Market Risk, Credit Risk (with inte-
grated Market and Credit Risk), Portfolio Optimization, and Fund-of-Funds
portfolio management, with integrated factor models. CognityTM is imple-
mented in a modern Java based server architecture to support both desktop
and Web delivery. For further details see www.finanalytica.com.
Acknowledgements
References
[1] Bradley, B. O. and Taqqu, M. S. (2003). “Financial Risk and Heavy
Tails”, in Handbook of Heavy Tailed Distributions in Finance, edited by
S. T. Rachev, Elsevier/North-Holland, Amsterdam
[2] Fama, E. (1963). “Mandelbrot and the Stable Paretian Hypothesis”,
Journal of Business, 36, 420–429
[3] Mandelbrot, B. B. (1963). “The Variation in Certain Speculative Prices”,
Journal of Business, 36, 394–419
[4] Rachev, S. and Mittnik, S. (2000). Stable Paretian Models in Finance.
Wiley, New York
262 S.T. Rachev et al.
[5] Racheva-Iotova B., Stoyanov, S., and Rachev S. (2003). Stable Non-
Gaussian Credit Risk Model; The Cognity Approach, in Credit Risk
(Measurement, Evaluations and Management), edited by G. Bol,
G. Nakhaheizadeh, S. Rachev, T. Rieder, K-H. Vollmer, Physica-Verlag
Series: Contributions to Economics, Springer, Heidelberg, NY, 179–198
[6] Rachev, S., Menn, C., and Fabozzi, F.J. (2005). Fat Tailed and
Skewed Asset Return Distributions: Implications for Risk, Wiley-Finance,
Hoboken
[7] Rockafellar, R. T. and Uryasev, S. (2000). “Optimization of Conditional
Value-at-Risk”, Journal of Risk, 3, 21–41
[8] Sklar, A. (1996). “Random Variables, Distribution Functions, and Copu-
las – a Personal Look Backward and Forward”, in Distributions with Fixed
Marginals and Related Topics, edited by Ruschendorff et. al., Institute of
Mathematical Sciences, Hayward, CA
[9] Stoyanov, S., Racheva-Iotova, B. (2004) “Univariate Stable Laws in the
Field of Finance-Parameter Estimation, Journal of Concrete and Applied
Mathematics, 2, 369–396
Pricing Tranches of a CDO and a CDS Index:
Recent Advances and Future Research
1 Introduction
In recent years, the market for credit derivatives has developed rapidly with
the introduction of new contracts and the standardization of trade documen-
tation. These include credit default swaps, basket default swaps, credit default
swap indexes, collateralized debt obligations, and credit default swap index
tranches. Along with the introduction of new products comes the issue of how
to price them. For single-name credit default swaps, there are several factor
models (one-factor and two-factor models) proposed in the literature. How-
ever, for credit portfolios, much work has to be done in formulating models
that fit market data. The difficulty in modeling lies in estimating the correla-
tion risk for a portfolio of credits. In an April 16, 2004 article in the Financial
Times [5], Darrell Duffie made the following comment on modeling portfo-
lio credit risk: “Banks, insurance companies and other financial institutions
managing portfolios of credit risk need an integrated model, one that reflects
correlations in default and changes in market spreads. Yet no such model
exists.” Almost a year later, a March 2005 publication by the Bank for Inter-
national Settlements noted that while a few models have been proposed, the
modeling of these correlations is “complex and not yet fully developed.” [1].
In this paper, first we review three methodologies for pricing CDO
tranches. They are the one-factor copula model, the structural model, and
the loss process model. Then we propose how the models can be improved.
The paper is structured as follows. In the next section we review credit de-
fault swaps and in Sect. 3 we review collateralized debt obligations and credit
default swap index tranches. The three pricing models are reviewed in Sects. 4
(one-factor copula model), 5 (structural model), and 6 (loss process model).
Our proposed models are provided in Sect. 7 and a summary is provided in
the final section, Sect. 8.
264 D. Wang et al.
The major risk-transferring instrument developed in the past few years has
been the credit default swap. This derivative contract permits market partici-
pants to transfer credit risk for individual credits and credit portfolios. Credit
default swaps are classified as follows: single-name swaps, basket swaps, and
credit default index swaps.
a first-to-default swap provides insurance for only the first default, a second-to-
default swap provides insurance for only the second default, an nth-to-default
swap provides insurance for only the nth default. For example, in an nth-to-
default swap, the protection seller does not make a payment to the protection
buyer for the first n − 1 defaulted reference entities, and makes a payment for
the nth defaulted reference entity. Once there is a payment upon the default
of the nth defaulted reference entity, the swap terminates. Unlike a single-
name CDS, the preferred settlement method for a basket default swap is cash
settlement.
months. The composition of each version of a CDX remains static in its lifetime
if no default occurs to the underlying reference entities, and the defaulted
reference entities are eliminated from the index.
There are two kinds of contracts on CDXs: unfunded and funded. An
unfunded contract is a CDS on a portfolio of names. This kind of contract is
traded on all the Dow Jones CDX and the iTraxx indexes. For some CDXs
such as the Dow Jones CDX NA HY index and its sub-indexes7 and the
iTraxx Europe index, the funded contract is traded. A funded contract is a
credit-linked note (CLN), allowing investors who because of client imposed or
regulatory restrictions are not permitted to invest in derivatives to gain risk
exposure to the CDX market. The funded contract works like a corporate bond
with some slight differences. A corporate bond ceases when a default occurs
to the reference entity. If a default occurs to a reference entity in an index,
the reference entity is removed from the index (and also from the funded
contract). The funded contract continues with a reduced notional principal
for the surviving reference entities in the index. Unlike the unfunded contract
which uses physical settlement, the settlement method for the funded contract
is cash settlement.
The index swap premium of a new version index is determined before the
roll day and unchanged over its life time, which is referred to as the coupon
or the deal spread. The price difference between the prevailing market spread
and the deal spread is paid upfront. If the prevailing market spread is higher
than the deal spread, the protection buyer pays the price difference to the
protection seller. If the prevailing market spread is less than the deal spread,
the protection seller pays the price difference to the protection buyer. The
index premium payments are standardized quarterly in arrears on the 20th of
March, June, September, and December of each calendar year.
The CDXs have many attractive properties for investors. Compared with
the single-name swaps, the CDXs have the advantages of diversification and
efficiency. Compared with basket default swaps and collateralized debt obli-
gations, the CDXs have the advantages of standardization and transparency.
The CDXs are traded more actively than the single-name CDSs, with low
bid–ask spreads.
7
The Dow Jones CDX NA HY index includes 100 equal-weighted North America
High Yield reference entities. Its sub-indexes include the CDX NA HY B
(B-rated), CDX NA HY BB (BB-rated), and CDX NA HY HB (High Beta)
indexes.
Pricing Tranches of a CDO and a CDS Index 267
for example, the first default, the second default, and the nth default. Cor-
respondingly, there are the first layer protection, the second layer protection,
and the nth layer protection. These protection layers work like basket default
swaps with some differences. The main difference is that the n basket default
swap protects the nth default in a portfolio and the nth protection layer pro-
tects the nth layer of the principal of a portfolio, which is specified by a range
of percentage, for example 15–20%. The layer protection derivative products
include collateralized debt obligations and CDS index tranches.
Tranche 4
30 − 70%
Coupons Spreads
Principal Principal Tranche 3
Bond 1 15 − 30%
Bond 2 SPV
Bond 3
... ... Tranche 2
Bond n 5 − 15%
Proceeding Proceeding
Tranche 1
0 − 5%
Collateral Pool
CDX NA IG (5 year)
Index 0–3% 3–7% 7–10% 10–15% 15–30%
63.25 48.1% 347 135.5 47.5 14.5
basis points premium per annum. The market quote is the upfront percent-
age payment. For example, the market quote of 27.8% for the iTraxx equity
tranche means that the protection buyer pays the protection seller 27.8% of the
notional principal upfront. In addition to the upfront payment, the protection
buyer also pays the protection seller the fixed 500 basis points premium per
annum on the outstanding notional principal. For all the non-equity tranches,
the market quotes are the premium in basis points, paid quarterly in arrears.
Just like the indexes, the premium payments for the tranches (with the ex-
ception of the upfront percentage payment of the equity tranche) are made
on the 20th of March, June, September, and December of each calendar year.
Following the commonly accepted definition for a synthetic CDO, CDX
tranches are not part of a synthetic CDO because they are not backed by a
portfolio of bonds or CDSs [6]. In addition, CDX tranches are unfunded and
they are insurance contracts, while synthetic CDO tranches are funded and
they are CLNs. However, the net cash flows of index tranches are the same as
synthetic CDO tranches and these tranches can be priced the same way as a
synthetic CDO.
The critical input for pricing synthetic CDO and CDS index tranches is an es-
timate of the default dependence (default correlation) between the underlying
assets. One popular method for estimating the dependence structure is using
copula functions, a method first applied in actuarial science. While there are
several types of copula function models, Li [10, 11] introduces the one-factor
Gaussian copula model for the case of two companies and Laurent and Gre-
gory [9] extend the model to the case of N companies. Several extensions to
the one-factor Gaussian copula model were subsequently introduced into the
literature. In this section, we provide a general description of the one-factor
copula function, introduce the market standard model, and review both the
270 D. Wang et al.
one-factor double t copula model [6] and the one-factor normal inverse Gaus-
sian copula model [8].
Suppose that a CDO includes n assets i = 1, 2, . . . , n and the default time
τi of the ith asset follows a Poisson process with a parameter λi . The λi is the
default intensity of the ith asset. Then the probability of a default occurring
before time t is
P (τi < t) = 1 − exp(−λi t). (1)
In a one-factor copula model, it is assumed that the default time τi for the
ith company is related to a random variable Xi with a zero mean and a unit
variance. For any given time t, there is a corresponding value x such that
Moreover, the one-factor copula model assumes that each random variable Xi
is the sum of two components
4
Xi = ai M + 1 − a2i Zi , i = 1, 2, . . . , n, (3)
total defaults that have occurred by time t conditional on the market condition
M = m, then Nt|m follows a binomial distribution Bin(n, Dt|m ), and
n!
P (Nt|m = j) = Dj (1 − Dt|m )n−j , j = 0, 1, 2, . . . , n. (6)
j!(n − j)! t|m
The probability that there will be exactly j defaults by time t is
∞
M
P (Nt = j) = E P (Nt|m ) = P (Nt|m = j)fM (m)dm, (7)
−∞
where fM (m) is the probability density function (pdf) of the random vari-
able M .
Li [10, 11] was the first to suggest that the Gaussian copula can be employed
in credit risk modeling to estimate the default correlation. In a one-factor
Gaussian copula model, the distributions of the common market component
M and the individual component Zi ’s in (3) are standard normal Gaussian
distributions. Because the sum of two independent Gaussian distributions is
still a Gaussian distribution, the Xi ’s in (3) have a closed form. It can be
verified that the Xi ’s have a standard normal distribution.
The one-factor copula Gaussian copula model is the market standard
model when implemented under the following assumptions:
• A fixed recovery rate of 40%
• The same CDS spreads for all of the underlying reference entities
• The same pairwise correlations
• The same default intensities for all the underlying reference entities
The market standard model does not appear to fit market data well (see
[6, 8]. In practice, market practitioners use implied correlations and base cor-
relations.
The implied correlation for a CDO tranche is the correlation that makes
the value of a contract on the CDO tranche zero when pricing the CDO with
the market standard model. For a CDO tranche, when inputting its implied
correlation into the market standard model, the simulated price of the tranche
should be its market price.
McGinty, Beinstein, Ahluwalia, and Watts [14] introduced base correla-
tions in CDO pricing. To understand base correlations, let’s use an example.
Recalling the CDX NA IG tranches 0–3%, 3–7%,7–10%, 10–15%, and 15–30%,
and assuming there exists a sequence of equity tranches 0–3%, 0–7%, 0–10%,
0–15%, and 0–30%, the premium payment on an equity tranche is a combina-
tion of the premium payment of the CDX NA IG tranches that are included
in the corresponding equity tranche. For example, the equity tranche 0–10%
includes three CDX NA IG tranches: 0–3%, 3–7%, and 7–10%. The premium
272 D. Wang et al.
payment on the equity tranche 0–10% includes three parts. The part of 0–3%
is paid the same way as the CDX NA IG tranche 0–3%, the part of 3–7% is
paid the same way as the CDX NA IG tranche 3–7%, and the part of 7–10%
is paid the same way as the CDX NA IG tranche 7–10%. Then the definition of
base correlation is the correlation input that make the prices of the contracts
on these series of equity tranches zero. For example, the base correlation for
the CDX NA IG tranche 7–10% is the implied correlation that makes the price
of a contract on the equity tranche 0-10% zero.
Kalemanova, Schmid, and Werner [8] propose utilizing normal inverse Gaus-
sian distributions in a one-factor copula model. A normal inverse Gaussian
distribution is a mixture of normal and inverse Gaussian distributions.
An inverse Gaussian distribution has the following density function
%
(ζ−ηx)2
√ ζ x−3/2 exp(−
fIG (x; ζ, η) = 2πη 2ηx ), if x > 0,
(9)
0, if x ≤ 0,
where ζ > 0 and η > 0 are two parameters. We denote the inverse Gaussian
distribution as IG(ζ, η).
Suppose Y is an inverse Gaussian distribution. A normal Gaussian distri-
bution X ∼ N (υ, σ 2 ) is a normal inverse Gaussian (NIG) distribution when
its mean υ and variance σ 2 are random variables as given below
υ = µ + βY, σ 2 = Y,
(10)
Y ∼ IG(δγ, γ 2 ),
Pricing Tranches of a CDO and a CDS Index 273
where δ > 0, 0 ≤ |β| < α, and γ := α2 − β 2 . The distribution of the random
variable X is denoted by X ∼ (α, β, µ, δ). The density of X is
δα exp(δγ + β(x − u))
f (x; α, β, µ, δ) = K(α δ 2 + (x − µ)2 ), (11)
π δ 2 + (x − µ)2
where K(.) is the modified Bessel function of the third kind as defined below
1 ∞ 1
K(ω) := exp(− ω(t − t−1 ))dt. (12)
2 0 2
The mean and variance of the NIG distribution X are respectively
δβ δα2
E(X) = µ + , V ar(X) = . (13)
γ γ3
The family of NIG distributions has two main properties. One is the closure
under the scale transition
α β
X ∼ N IG(α, β, µ, δ) ⇒ cX ∼ N IG( , , cµ, cδ). (14)
c c
The other is that if two independent NIG random variables X and Y have
the same α and β parameters, then the sum of these two variables is still an
NIG variable as shown below
X ∼ N IG(α, β, µ1 , δ1 ), Y ∼ N IG(α, β, µ2 , δ)
(15)
⇒ X + Y ∼ N IG(α, β, µ1 + µ2 δ1 + δ2 ).
When using NIG distributions in a one-factor copula model, the model
is referred to as a one-factor normal inverse Gaussian copula model. The
distributions for M and Zi ’s in (3) are given below
!
αβ
M ∼ N IG α, β, − ,α ,
α2 − β 2
! (16)
α 1 − a2i β 1 − a2i αβ 1 − a2i α 1 − a2i
Zi ∼ N IG , ,− , .
ai ai ai α2 − β 2 ai
5 Structural Model
Hull, Predescu, and White [7] propose the structural model to price the default
correlation in tranches of a CDO or an index. The idea is based on Merton’s
model [17] and its extension by Black and Cox [3]. It is assumed that the value
of a company follows a stochastic process, and if the value of the company
goes below a minimum value (barrier), the company defaults.
In the model, N different companies are assumed and the value of company
i (1 ≤ i ≤ N ) at time t is denoted by Vi . The value of the company follows a
stochastic process as shown below
where µi is the expected growth rate of the value of company i, σi is the volatil-
ity of the value of company i, and Xi (t) is a variable following a continuous-
time Gaussian stochastic process (Wiener process). The barrier for company
i is denoted by Bi . Whenever the value of company i goes below the barrier
Bi , it defaults.
Without the loss of generality, it is assumed that Xi (0) = 0. Applying
Ito’s formula to ln Vi , it is easy to show that
ln Hi − ln Vi (0) µi − σi2 /2
βi = γi = − , (21)
σi σi
then Bi∗ = βi + γi t.
To model the default correlation, it is assumed that each Wiener process Xi
follows a two-component process which includes a common Wiener process
M and an idiosyncratic Wiener process Zi . It is expressed as
4
dXi (t) = ai (t)dM (t) + 1 − a2i (t)dZi (t), (22)
Loss process models for pricing correlation risk have been developed by
Schönbucher [20], Sidenius et al. [21], Di Graziano and Rogers [4], and Bennani
[2]. Here we introduce the basic idea of the loss process model as discussed by
Schönbucher. We omit the mathematical details.
The model is set up in the probability space (Ω, (Ft )0≤t≤T , Q), where Q is
a spot martingale measure, (Ft )0≤t≤T is the filtration satisfying the common
definitions, and Ω is the sample space. Assume that there are N company
names in a portfolio. Each name has the same notional principal in the port-
folio. Under the assumption of a homogenous recovery rate for all the com-
panies, all companies have identical losses in default which is normalized to
one. The cumulative default loss process is defined by
N
L(t) = 1{τk ≤t} , (23)
k
where τk is the default time of company k, and the default indicator 1{τk ≤t}
is 1 when τk ≤ t and 0 when τk > t. The loss process is an N -bounded,
integer-valued, non-decreasing Markov chain. Under Q-measure, the proba-
bility distribution of L(T ) at time t < T is denoted by the vector p(t, T ) :=
(p0 (t, T ), . . . , pN (t, T )) , where the pi ’s are conditional probabilities
d P (t, T ) = −λ
N −1 (T )Pi,N −1 (t, T ),
dT i,N
for all i, j = 0, 1, . . . , N and 0 ≤ t ≤ T . The initial conditions are Pi,j (t, t) =
1{i=j} . The solution of the Kolmogorov equations in (26) is given below
⎧
⎨0 0T
for i > j,
Pi,j (t, T ) = exp{− t λi (t, s)ds} 0 for i = j, (27)
⎩0T − tT λj (t,u)du
t
Pi,j−1 (t, s)λj−1 e ds for i < j.
In the dynamics version of the loss process model, the loss process follows
a Poisson process with time- and state-dependent inhomogeneous default in-
tensities λL(t) (t), L(t) = 0, . . . , N − 1, which are the transition rates in the
Pricing Tranches of a CDO and a CDS Index 277
generator matrix in (25). The aggregate default intensity λL(t) (t) can be ex-
pressed in terms of the individual intensities λk (t)
λL(t) (t) = λk (t), (30)
k∈S(t)
where S(t) := {1 ≤ k ≤ N |τk > t} is the set of companies that have not
defaulted by time t.
The loss process is assumed to follow a Poisson process with stochastic
intensity, a process referred to as a Cox process.
dλi (t, T ) = µi (t, T )dT + σi (t, T )dB(t), i = 0, . . . , N − 1, (31)
where B(t) is a d-dimension Q-Brownian motion, the µi (t, T )’s are the drifts
of the stochastic processes, and the σi (t, T )’s are the d-dimension volatilities
of the stochastic processes. To keep the stochastic processes consistent with
the loss process L(t), the following conditions must be satisfied
PL(t),i (t, T )µi (t, T ) = σi (t, T )υL(t),i (t, T ), 0 ≤ i ≤ N − 1, t ≤ T, (32)
where, υn,m (t, T )’s are given by
⎧
⎪
⎨0 for i>j
0T
υi,j = Pnm (t,0 T ){− t σi (t, s)ds} for i = j , (33)
⎩ 0 T e− sT λj (t,u)du [σ P a (t, s) − P (t, s)σ (t, s)]ds for
⎪
i<j
t i,j−1 ij j
with
Pa
σi,j−1 (t, T ) = Pi,j−1 (t, T ) + λm−1 (t, T )υn,m−1 (t, T ). (34)
The model can be implemented by a Monte Carlo method. For pricing a CDO
with a maturity T , the procedure is as follows:
1. Initial condition: t = 0, L(0) = 0 (p0 (0, 0) = 1), and specify λi (0, 0)’s and
σi (0, .)’s.
2. Simulate a Brownian motion trial.
3. s → s + ∆s: (until s = T )
• Calculate P0,m (0, s) from (27), and υ0,j (0, s) from (33), and use them
and σi (0, s) to calculate µi (0, s) from (32).
• Calculate λi (0, s+∆s) using the Euler scheme and µi (0,s) and σi (0, s).
• In a Euler scheme, calculate the loss distribution pi (0, s + ∆s) from
(27) and using the representation of the loss distribution in (29).
• The loss distribution pi (0, .) on the time period of (0, T ) is then
calculated.
4. Repeat steps 2–4 until the average loss distributions pi (0, .) of all the trials
converge.
5. Using the average loss distributions pi (0, .) to price a CDO.
The loss process model can also be used to price other portfolio credit
derivatives such as basket default swaps, options on CDS indexes, and options
on CDS indexes tranches.
Hull and White [6] first use heavy-tailed distributions (Student’s t distribu-
tions) in a one-factor copula model. In their so-called one-factor double t
copula model, Hull and White use the t distribution with ν degrees of free-
dom for the market component M and the individual components Zi ’s in
equation (3). The degrees of freedom parameter ν of the t distribution can
be 3, 4, 5, . . . . When the degrees of freedom parameter of ν is equal to 3, the
copula function has the maximum tail-fatness. When the degrees of freedom
parameter of ν increases, the tail-fatness of the copula function decreases.
As mentioned before, Hull and White find that the double t copula model
fits market data well when the degrees of freedom parameter ν is equal to 4.
But the simulation by Kalemanova et al. [8] shows a different result. When
Pricing Tranches of a CDO and a CDS Index 279
Kalemanova et al. compare their model with the double t copula model, in
addition to the simulation results by their own model, they also give the
simulation results by the double t copula model for both the cases of the
degrees of freedom parameter ν equal to 3 and 4. These simulation results
show that the double t copula model fits market data better when ν = 3 than
ν = 4. One difference in these two works is that different market data are
used in the simulation. Hull and White use market data for the 5-year iTraxx
Europe tranches on August 4, 2004, while Kalemanova et al. use market data
on April 12, 2006. Therefore, the difference, related to how many degrees of
freedom make the double t copula fit market data well, may suggest that
for market data in different times, the double t copula model with different
tail-fatnesses works well.
The drawbacks of the double t copula are that its tail-fatness cannot be
changed continuously and the maximum tail-fatness occurs when the degrees
of freedom parameter ν is equal to 3. In order to fit market data well over
time, it is necessary that the tail-fatness of a one-factor copula model can be
adjusted continuously and can be much larger than the maximum tail-fatness
of the one-factor double t copula model.
In the following, we suggest four one-factor heavy-tailed copula models.
Each model has (1) a tail-fatness parameter that can be changed continuously
and (2) a maximum tail-fatness much larger than that of the one-factor double
t copula model.
1
Y = 2 Y. (40)
σ1 + σ22
The pdf of Y is
√ + 2 2 ,
p pσ12 +(1−p)σ22 y (pσ1 +(1−p)σ22 )
fY (y) = √
2πσ1
exp − 2σ12
√ + 2 2 , (41)
(1−p) pσ12 +(1−p)σ22 y (pσ1 +(1−p)σ22 )
+ √
2πσ
exp − 2σ 2 .
2 2
Γ ( ν+1
2 )
fT (t|ν) = ν √ (1 + t2 /ν)−(ν+1)/2 , 0 < x < ∞, ν > 0. (44)
Γ ( 2 ) νπ
Stable Distribution
A non-trivial distribution g is a stable distribution if and only if for a sequence
of independent, identical random variables Xi, , i = 1, 2, 3, . . . , n with a distri-
bution g, the constants cn > 0 and dn can always be found for any n > 1 such
that
d
cn (X1 + X2 + · · · + Xn ) + dn = X1 .
In general, a stable distribution cannot be expressed in a closed form except
for three special cases: Gaussian, Gauchy, and Lévy distributions. However,
the characteristic function always exists and can be expressed in a closed
form. For a random variable X with a stable distribution g, the characteristic
function of the X can be expressed in the following form
exp(−γ α |t|α [1 − iβsign(t) tan( πα2 )] + iδt), α = 1 ,
ϕX (t) = E exp(itX) =
exp(−γ|t|[1 + iβ π2 sign(t) ln(|t|)] + iδt), α=1
(50)
where 0 < α ≤ 2, γ ≥ 0, −1 ≤ β ≤ 1, and −∞ ≤ δ ≤ ∞, and the function of
sign(t) is 1 when t > 0, 0 when t = 0, and −1 when t < 0.
There are four characteristic parameters to describe a stable distribution.
They are: (1) the index of stability or the shape parameter α, (2) the scale
parameter γ, (3) the skewness parameter β, and (4) the location parameter
δ. A stable distribution g is called the α stable distribution and is denoted
Sα (δ, β, σ) = S(α, σ, β, δ).
The family of α stable distributions has three attractive properties:
• The sum of independent α stable distributions is still an α stable distri-
bution, a property referred to as stability.
• α stable distributions can be skewed.
• Compared with the normal distribution, α stable distributions can have a
fatter tail and a high peak around its center, a property which is referred
to as leptokurtosis.
Real world financial market data indicate that assets returns tend to be
fat-tailed, skewed, and peaked around center. For this reason α stable distri-
butions have been a popular choice in modeling asset returns (see [19]).
where hi (x), i = 1, 2 are the pdf of two normal distributions with means µi
and standard deviations σi , and gθ (x) is the pdf of an α stable distribution
with its parameter vector θ = (α, γ, β, δ). To secure a well-defined smooth
probability distribution, the following regularities are imposed:
where ψ and ϕ denote the density and distribution functions of the standard
normal distribution, respectively. A smoothly truncated α stable distribution
[a,b]
is referred to as an STS-distribution and denoted by Sα (γ, β, δ). The proba-
bilities p1 and p2 are referred to as the cut-off probabilities. The real numbers
a and b are referred to as the cut-off points.
The family of STS-distributions has two important properties. The first is
that it is closed under the scale and location transitions. This means that if
the distribution X is an STS-distribution, then for c, d ∈ R, the distribution
[a,b]
Y := cX + d is an STS-distribution. If X follows Sα (γ, β, δ), then Y follows
S
[
a,b]
( δ)
γ , β, with
α
a = ca + d, b = cb + d,
α =α,
cδ + d α = 1, (53)
= |c|γ,
γ β = sign(c)β, δ =
cδ − π2 c log |c|σβ + d α = 1,
The base-case structural model suggested by Hull et. al [7] can be an alterna-
tive method to the one-factor Gaussian copula model. The results of the two
models are close. Consider the fact that the one-factor double t copula model
fits market data much better than the one-factor Gaussian copula model ac-
cording to Hull and White [6]. A natural way to enhance the structural model
is by applying heavy-tailed distributions.
Unlike the one-factor copula model, where any continuous distribution
with a zero mean and a unit variance can be used, in the structural model there
is a strong constraint imposed on the distribution of the underlying stochastic
processes. The distribution for the common driving process M (t) and the
individual driving process Zi ’s in (22) must satisfy a property of closure under
summation. This means that if two independent random variables follow a
given distribution, then the sum of these two variables still follow the same
distribution. As explained earlier, the α stable distribution has this property
and has been used in financial modeling (see [18]). We suggest using the α
stable distribution in the structural model.
The non-Gaussian α stable distribution has a drawback. Its variance does
not exist. The STS distribution is a good candidate to overcome this problem.
For a STS distribution, if the two cut-off points a and b are far away from the
peak, the STS distribution is approximately closed under summation. Based
on this, employing the STS distribution in the structural model should be the
subject of future research.
In the dynamic loss process model, the default intensities λi ’s follow
stochastic processes as shown in (31). It is also a possible research direc-
tion to use the α stable distribution and the STS distribution for the driving
processes.
8 Summary
In this paper, we review three models for pricing portfolio risk: the one-factor
copula model, the structural model, and the loss process model. We then
propose how to improve these models by using heavy-tailed functions. For
the one-factor copula model, we suggest using (1) a double mixture Gaussian
copula, (2) a double t distribution with fractional copula, (3) a double mixture
distribution of t and Gaussian distributions copula, and (4) a double smoothly
truncated α stable copula. In each of these four new extensions to the one-
factor Gaussian copula model, one parameter is introduced to control the
tail-fatness of the copula function. To improve the structural and loss process
models, we suggest using the stable distribution and the smoothly truncated
stable distribution for the underlying stochastic driving processes.
Pricing Tranches of a CDO and a CDS Index 285
References
[1] Amato J, Gyntelberg J (2005) CDS index tranches and the pricing of
credit risk correlations. BIS Quarterly Review, March 2005, pp 73–87
[2] Bennani N (2005) The forward loss model: a dynamic term structure
approach for the pricing of portfolio credit derivatives. Working paper,
available at http://www.defaultrisk.com/pp crdrv 95.htm
[3] Black F, Cox J (1976) Valuing corporate securities: some effects of bond
indenture provision. The Journal of Finance, vol 31, pp 351–367
[4] Di Graziano G, Rogers C (2005) A new approach to the modeling and
pricing of correlation credit derivatives. Working paper, available at
www.defaultrisk.com/pp crdrv 88.htm
[5] Duffie D (2004) Comments: irresistible reasons for better models of credit
risk. Financial Times, April 16, 2004
[6] Hull J, White A (2004) Valuation of a CDO and nth to default CDS
without Monte Carlo simulation. The Journal of Derivatives, vol 2, pp
8–23
[7] Hull J, Predescu M, White A (2005) The valuation of correlation-
dependent credit derivatives using a structural model. Working paper,
Joseph L. Rotman School of Management, University of Toronto, avail-
able at http://www.defaultrisk.com/pp crdrv 68.htm
[8] Kalemanova A, Schmid B, Werner R (2005) The normal inverse Gaus-
sian distribution for synthetic CDO pricing. Working paper, available at
http://www.defaultrisk.com/pp crdrv 91.htm
[9] Laurent JP, Gregory J (2003) Basket default swaps, CDOs and factor cop-
ulas. Working paper, ISFA Actuarial School, University of Lyon, available
at http://www.defaultrisk.com/pp crdrv 26.htm
[10] Li DX (1999) The valuation of basket credit derivatives. CreditMetrics
Monitor, April 1999, pp 34–50
[11] Li DX (2000) On default correlation: a copula function approach. The
Journal of Fixed Income, vol 9, pp 43–54
[12] Lucas DJ, Goodman LS, Fabozzi FJ (2006) Collateralized debt obliga-
tions: structures and analysis, 2nd edn. Wiley Finance, Hoboken
[13] Mardia K, Zemroch P (1978) Tables of the F- and related distributions
with algorithms. Academic, New York
[14] McGinty L, Beinstein E, Ahluwalia R, Watts M (2004) Credit correlation:
a guide. Credit Derivatives Strategy, JP Morgan, London, March 12, 2004
[15] Menn C, Rachev S (2005) A GARCH option pricing model with alpha-
stable innovations. European Journal of Operational Research, vol 163,
pp 201–209
[16] Menn C, Rachev S (2005) Smoothly truncated stable distribu-
tions, GARCH-models, and option pricing. Working paper, University
of Karlsruhe and UCSB, available at http://www.statistik.
uni-karlsruhe.de/download/tr smoothly truncated.pdf
286 D. Wang et al.
[17] Merton R (1974) On the pricing of corporate debt: the risk structure of
interest rates. The Journal of Finance, vol 29, pp 449–470
[18] Rachev S, Mittnik S (2000) Stable Paretian models in finance. John Wi-
ley, Series in Financial Economics and Quantitative Analysis, Chichester
[19] Rachev S, Menn C, Fabozzi FJ (2005) Fat-tailed and skewed asset return
distributions: implications for risk management, portfolio selection, and
option pricing. Wiley Finance, Hoboken
[20] Schönbucher P (2005) Portfolio losses and the term structure of loss
transition rates: a new methodology for the pricing of portfolio credit
derivatives. Working paper, available at http://www.defaultrisk.
com/pp model 74.htm
[21] Sidenius J, Piterbarg V, Andersen L (2005) A new framework for
dynamic credit portfolio loss modeling. Working paper, available at
http://www.defaultrisk.com/pp model 83.htm