Professional Documents
Culture Documents
(Research in Finance) Andrew H. Chen-Research in Finance, Volume 21 - Emerald Group Publishing Limited (2005) PDF
(Research in Finance) Andrew H. Chen-Research in Finance, Volume 21 - Emerald Group Publishing Limited (2005) PDF
INTRODUCTION ix
v
vi
vii
viii
ABSTRACT
We use stochastic dominance to test whether investors should prefer riskier
securities as the investment horizon lengthens. Simulated return distributions
for stocks, bonds, and U.S. Treasury bills are generated for holding periods
of one to 20 years and stochastic dominance tests are run to establish
preferences among the alternative portfolios. With independent returns, we
find no evidence that high-risk securities (stocks) dominate low-risk securities
(bonds) as the investment horizon lengthens. Under the assumption that
security returns are correlated across time, we find that common stocks
dominate corporate bonds and U.S. Treasury bills for sufficiently long
investment horizons.
1. INTRODUCTION
The issue of time diversification has generated considerable controversy.
Theoreticians, most notably Merton and Samuelson, reason that, if markets are
efficient and security returns are independent and identically distributed, then
lengthening the investment horizon should not reduce risk. Thus, an investor’s
optimal mix of securities should be independent of the planned holding period.1
Many market professionals, however, recommend that the proportion of an
investor’s holdings of high-risk securities such as equities should increase with
Research in Finance
Research in Finance, Volume 21, 1–15
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21001-2
1
2 CHARLES W. HODGES ET AL.
the length of the desired investment horizon. Lee (1990) and Reichenstein and
Dorsett (1995) offer support for this position by arguing that time diversification
will hold true if stock returns are mean reverting across time. In this case, returns
have negative autocorrelation so that volatility is reduced, because a positive or
negative price movement tends to be followed by a price movement in the
opposite direction. While numerous empirical studies employing a mean-variance
framework have been conducted on the efficacy of time diversification, they have
yielded conflicting results and have failed to resolve the issue.2
We use stochastic dominance (SD) to analyze the issue of time diversification.
Data from Ibbotson Associates 2001 Yearbook is used to generate simulated return
distributions for portfolios of small stocks, common stocks, long-term corporate
bonds, and U.S. Treasury bills for various holding periods. For each holding
period, we apply stochastic dominance tests to determine whether preferences
can be established among the equity and fixed-income portfolios. Our analysis is
first conducted under the assumption that returns are independent across time and
then under the assumption that security returns are correlated across time so that
potential mean reversion in equity returns is captured.
We employ stochastic dominance because it is a more general, analytical
technique than mean-variance analysis. Stochastic dominance examines the
entire distribution of returns and, thus, considers all the moments of the
return distribution. Furthermore, SD does not require any specific distributional
assumption such as normality. Empirical studies of time diversification based on
mean-variance analysis ignore higher moments, such as skewness and kurtosis, and
are not consistent with maximizing expected utility unless the return distributions
are normal. Albrecht (1998) demonstrates the pitfalls of applying mean-variance-
based performance measures to non-normal security return distributions.
In this paper, the stochastic dominance results clearly show that the issue of time
diversification is ultimately a question of autocorrelation in security returns. When
returns are independent across time, no dominance exists among the stock and bond
portfolios, even for long holding periods. This is consistent with the assumptions
and analyses by Merton and Samuelson. When autocorrelation across security
returns is captured, however, stock portfolios dominate the bond portfolios for suffi-
ciently long holding periods. This is consistent with the practitioner view that stock
returns are mean reverting so that lengthening the investment horizon reduces risk.
2. STOCHASTIC DOMINANCE
Stochastic dominance has important advantages over the mean-variance frame-
work that underlies other empirical time diversification studies. It is theoretically
Time Diversification and Stochastic Dominance 3
In the case where F(r) lies entirely to the right of G(r), preference is readily
apparent. When two cumulative distributions cross, other factors than return
(e.g. risk aversion) must be considered in order to establish dominance. Further
assuming that investors are risk averse (second derivative of the utility function is
negative), second degree stochastic dominance (SSD) can be applied. Formally, F
dominates G for all risk-averters if and only if:
r
[G(t) − F(t)] dt ≥ 0 (1)
−∞
for all values of r, and the inequality is strict for at least one value of r. G(t) is
the cumulative distribution associated with G. F(t) is the cumulative distribution
associated with F. When F(t) lies to the right of G(t), the integral of G(t) − F(t) is
positive.
Figure 3 shows that, when the conditions of Eq. (1) are met, F(t) lies far enough
to the right of G(t) so that investment F would be preferred to investment G. This
is because the expected utility gained from the positive area to the left of R0 is
more than the decrease in expected utility lost between R0 and R1 .
Analysis of third degree stochastic dominance (TSD) is similar except that
it assumes that investors’ absolute risk aversion decreases, which implies that
investors prefer positive skewness. Formally, the TSD rule asserts that F dominates
G if, and only if:
r v
[G(t) − F(t)] dt dv ≥ 0 (2)
−∞ −∞
Time Diversification and Stochastic Dominance 5
through 2000 are collected from the Ibbotson Associates data. For a specified
holding period of n years and a given portfolio, we randomly select (with
replacement) n returns out of the 75 sample returns. We then compute the holding-
period return (HPR) by:
n
HPRn = (1 + R i ) − 1 (3)
i=1
where: HPRn = return for holding period of n years. Ri = ith return observation.
n = number of years in holding period.
This procedure is repeated 250 times. Thus, we generate sample return
distributions for each portfolio for annual holding periods of one to 20 years.
Note that this procedure is consistent with an efficient market because it generates
independent returns.
To illustrate, consider the case of a five-year holding period and the small-
stock portfolio. Five annual returns are selected at random from the 1926 through
2000 historical returns and the five-year holding-period return is computed. This
is repeated until a sample distribution of 250 five-year holding-period returns is
obtained.
Descriptive statistics are computed for each sample return distribution. Tables 1
through 3 list the portfolio means, standard deviations, and skewnesses,
respectively. Table 1 shows that the mean return for all portfolios increases with the
length of the holding period. The mean return for Treasury bills, for example, grows
from 3.84% for a one-year holding period to 115.84% for a 20-year holding period.
The corresponding mean returns for small stocks are 18.656% and 2234.42%.
One might be tempted to conclude that a long-term (20-year investment horizon)
investor should invest in small stocks rather than Treasury bills since the expected
return is larger.
Risk, however, also increases with the length of the holding period. Table 2
shows that the standard deviation of returns for Treasury bills increases from
3.14% for a one-year holding period to 28.84% for a 20-year holding period. The
corresponding standard deviations for small stocks are 34.151% and 3,712.34%.
Risk, as measured by the standard deviation, grows much more rapidly with the
length of the holding period for small stocks than for Treasury bills.
Skewness coefficients given in Table 3 are generally positive and are inconsistent
with normally distributed random variables.5 Indeed, the Kolomogorov-D statistics
8 CHARLES W. HODGES ET AL.
reject normality at the 1% significance level for all the stock and bond portfolios
for all holding periods. Thus, preferences established strictly on the basis of a
mean-variance analysis would not be valid.
Stochastic dominance tests are run to determine whether preferences can be
established among the portfolios for each holding period. The SD algorithms
for discrete distributions used here are discussed in Levy (1992).6 The results
of the 20 tests are summarized in Table 4 which shows membership in the TSD
efficient set.7 As the results are the same for all holding periods, we only show the
results in five-year increments. The SSD and FSD results are identical to the TSD
results.
Table 4 shows no evidence that time diversification results in preferences among
the portfolios for risk-averse investors who prefer positive skewness. The efficient
set for each holding period includes all four portfolios. This means that an investor
with a 20-year investment horizon could rationally select any of the stock, bond,
or T-bill portfolios. Thus, with independent returns, time diversification fails and
Merton and Samuelson are correct.
Time Diversification and Stochastic Dominance 9
results.
We now repeat the analysis allowing for autocorrelation across time in the security
returns. Mean reversion (negative autocorrelation) in long-run stock returns and
mean aversion (positive autocorrelation) in fixed-income securities have been
Descriptive statistics are computed for each sample return distribution. Tables 5,
6, and 7 give the portfolio means, standard deviations, and skewnesses, respec-
tively. Note that the tables start with a two-year holding period since autocorrelation
across time requires at least two holding periods. Tables 5 and 6 show that expected
returns and standard deviations increase with the holding period for all portfolios.
Table 7 shows that the skewness coefficients remain positive, indicating
that the return distributions are not normal. Mean-variance analysis remains
inappropriate.
A comparison of corresponding standard deviations in Table 2 with those in
Table 6 indicates that the portfolio returns are autocorrelated across time. The
two equity portfolios exhibit negative autocorrealtion or mean-reverting behavior.
Under mean-reversion, a positive return tends to be followed by a negative return
so that volatility is dampened. The standard deviations for the small-stock and
common-stock portfolios are all less than those for corresponding independent
returns for holding periods of five years or longer. For example, the standard
deviation for the common-stock portfolio for a 20-year investment horizon is only
576.70% in Table 6 (autocorrelated returns) compared to 1018.19% in Table 2
(independent returns).
12 CHARLES W. HODGES ET AL.
The corporate bond and Treasury bill portfolios, however, exhibit positive
autocorrelation or mean aversion. Under mean aversion, a positive return tends to
be followed by another positive return so that volatility is enhanced. The standard
deviation of the corporate bond and Treasury bill portfolios in Table 6 are greater
than those for corresponding independent returns given in Table 2 for all holding
periods. For example, the standard deviation for the corporate bond portfolio for
a 20-year investment horizon is 212.65% in Table 6 (autocorrelated returns) but
only 122.54% in Table 2 (independent returns).
These results suggest that time diversification may hold with autocorrelated
returns. Mean reversion in the stock portfolios combined with mean aversion in
the fixed-income portfolios imply that the relative risk of the equity portfolios
compared to the fixed-income portfolios should decline as the holding period
lengthens. Thus, equities would be preferred to fixed-income securities for
sufficiently long holding periods as investment professionals claim.
Stochastic dominance tests are run to determine if preferences can be established
when autocorrelation in the return generating process is captured. The results given
in Table 8 show membership in the TSD efficient set. As predicted, preferences
for the equity portfolios are now established. Common stocks and small stocks
dominate Treasury bills for holding periods of eight years or longer and corporate
bonds for holding periods of 16 years or longer. Common stocks are preferred
over small stocks for holding periods of 18 years or longer. The SSD results are
identical to the TSD results.
4. CONCLUDING COMMENTS
The results in this paper reconcile the views of theoreticians and market
professionals. With independent security returns, no dominance exists between
the equity and fixed-income portfolios. All portfolios are in the efficient set for all
holding periods. This is consistent with the arguments of Merton and Samuelson
that time diversification does not reduce risk and that asset allocation should not
be influenced by the intended holding period.
However, when autocorrelation in security returns is considered, we find that
the stock portfolios dominate the fixed-income portfolios given sufficiently long
holding periods. Mean reversion in stock returns combined with mean aversion in
bond returns makes equities more attractive as the holding period lengthens. This is
consistent with the practitioner view that the proportion of equity holdings should
increase with the intended holding period. Thus, the issue of time diversification
is ultimately a debate about the degree to which security returns are autocorrelated
across time.
NOTES
8. See, for example, Fama and French (1988), Poterba and Summers (1988), Reichenstein
and Dorsett (1995), Lewellen (2001), and Strong and Taylor (2001).
REFERENCES
Albrecht, T. (1998). The mean-variance framework and long horizons. Financial Analysts Journal, 54,
44–49.
Asness, C. S. (1996). Why not 100% equities? Journal of Portfolio Management, 22, 29–34.
Butler, K. C., & Domian, D. L. (1991). Risk, diversification, and the investment horizon. Journal of
Portfolio Management, 17, 41–47.
Efron, B. (1979). Bootstrap methods: Another look at the jack-knife. Annals of Statistics, 7, 1–12.
Fama, E., & French, K. (1988). Permanent and temporary components of stock prices. Journal of
Political Economy, 96, 246–273.
Ferguson, R., & Simaan, Y. (1996). Portfolio composition and the investment horizon revisted. Journal
of Portfolio Management, 22, 62–67.
Gunthorpe, D., & Levy, H. (1994). Portfolio composition and the investment horizon. Financial Analysts
Journal, 50, 51–55.
Hansson, B., & Persson, M. (2000). Time diversification and estimation risk. Financial Analysts
Journal, 56, 55–62.
Hodges, C., Taylor, W., & Yoder, J. A. (1997). Stocks, bonds, the Sharpe ratio, and the investment
horizon. Financial Analysts Journal, 53, 74–80.
Hodges, C., & Yoder, J. A. (1996). Time diversification and security preferences: A stochastic
dominance approach. Review of Quantitative Finance and Accounting, 7, 289–298.
Ibbotson, R. (2001). Stocks, bonds, bills, and inflation: 2001 yearbook. Chicago, IL: Ibbotson.
Lee, W. (1990). Diversification and time: Do investment horizons matter? Journal of Portfolio
Management, 16, 21–26.
Levy, H. (1992). Stochastic dominance and expected utility: Survey and analysis. Management Science,
38, 555–593.
Levy, H., & Gunthorpe, D. (1993). Optimal investment proportions in senior securities and equities
under alternative holding periods. Journal of Portfolio Management, 19, 30–36.
Levy, H., & Sarnat, M. (1984). Portfolio investment selection: Theory and practice. Englewood Cliffs,
NJ: Prentice-Hall.
Lewellen, J. (2001). Temporary movements in stock prices. Working Paper, MIT.
Lloyd, W. P., & Haney, R. L., Jr. (1980). Time diversification: Surest route to lower risk. Journal of
Portfolio Management, 6, 5–9.
Lloyd, W. P., & Modani, N. K. (1983). Stocks, bonds, bills, and time diversification. Journal of Portfolio
Management, 9, 7–11.
McEnally, R. W. (1985). Time diversification: Surest route to lower risk? Journal of Portfolio
Management, 11, 24–26.
Merton, R. C., & Samuelson, P. A. (1974). Fallacy of the Log-Normal approximation to
optimal portfolio decision-making over many periods. Journal of Financial Economics, 1,
67–95.
Olsen, R. A., & Khaki, M. (1998). Risk, rationality, and time diversification. Financial Analysts Journal,
54, 58–63.
Poterba, J., & Summers, L. (1988). Mean reversion in stock prices: Evidence and implications. Journal
of Financial Economics, 22, 27–59.
Time Diversification and Stochastic Dominance 15
Reichenstein, W., & Dorsett, D. (1995). Time diversification revisted. Charlottesville, VA: Research
Foundation of the Institute of Chartered Financial Analysts.
Samuelson, P. A. (1990). Asset allocation could be dangerous to your health. Journal of Portfolio
Management, 16, 5–8.
Samuelson, P. A. (1994). The long-term case for equities. Journal of Portfolio Management, 21, 15–26.
Siegel, J. J. (2002). Stocks for the long run. New York, NY: McGraw-Hill.
Strong, N., & Taylor, N. (2001). Time diversification: Empirical tests. Journal of Business Finance and
Accounting, 28, 263–302.
Thaler, R. H., & Williamson, J. P. (1994). College and university endowment funds: Why not 100%
equities? Journal of Portfolio Management, 23, 27–37.
Thorley, S. R. (1995). The time-diversification controversy. Financial Analysts Journal, 51, 68–76.
MINORITY EQUITY INVESTMENTS
AND INTER-FIRM COLLABORATIONS
ABSTRACT
1. INTRODUCTION
Research in Finance
Research in Finance, Volume 21, 17–44
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21002-4
17
18 SU HAN CHAN ET AL.
alliances, joint ventures, and mergers. We focus on strategic alliances which are an
intermediate form of collaboration. Specifically, we study strategic alliances that
involve a minority equity investment by one of the partnering firms (the purchaser)
in the other (the seller).1 These agreements are particularly interesting because the
minority equity stake not only provides an infusion of cash to the seller but gives
the purchaser partial control over the seller’s resources without creating a new
entity as in a joint venture. In addition, the partial ownership interest provides a
mechanism for bonding the partnering firms to the alliance agreement.
We examine investor reactions to the announcement of corporate purchases
of minority equity stakes associated with the formation of strategic alliances.
Despite the prevalence of minority equity investments among partners to interfirm
collaborations, especially in high-tech industries, there is little empirical evidence
as to their valuation consequences. Previous studies of interfirm equity investments
(see Choi, 1991; Mikkelson & Ruback, 1985) or block shareholdings (see Bethel,
Liebeskind & Opler, 1998; Hertzel & Smith, 1993; Wruck, 1989) have not focused
on the use of minority equity stakes as a bonding mechanism for a strategic
alliance.2
Prior research on interfirm collaboration has examined joint ventures (Johnson
& Houston, 2000; McConnell & Nantell, 1985), non-equity strategic alliances
(Chan, Kensinger, Keown & Martin, 1997), and equity alliances (Allen & Phillips,
2000). The results of the joint venture and alliance studies suggest that interfirm
collaborations, in general, benefit the partnering firms.
Our study is most closely related to a recent study by Allen and Phillips (2000)
who investigate long-term block ownership by corporations and performance
changes in firms with corporate block owners. Using a sample of 150 corporate
equity purchases that accompany explicit product market relationships and 252
equity stakes that are not associated with business relationships during the
1980–1991 period, Allen and Phillips examine three potential reasons for block
ownership: accruing the benefits of product market relationships, the alleviation of
financing constraints, and improved board monitoring by corporate owners. They
find that the largest increases in the sellers’ stock prices, investment, and operating
profitability occur when the equity ownership is accompanied by alliances, joint
ventures, and other product market relationships between the purchasing and
selling firms. Furthermore, their findings were strongest in industries with high
R&D expenditures. The study concludes that block ownership has significant
benefits in product market relationships.
We provide additional evidence as to the sources of value creation in inter-firm
collaborations supported by equity participation. Our study focuses exclusively on
minority equity alliances and adds to the findings of the Allen and Phillips (2000)
study in the following ways.
Minority Equity Investments and Inter-Firm Collaborations 19
Wruck (1989) shows that a private sale of equity results in a shift to more
concentrated holdings by non-managers and this generally increases firm value
if the blockholder uses his votes to see that corporate resources are managed more
efficiently or if the existence of the block increases the probability of a value-
increasing takeover. In the case of the equity alliances we study, the purchasing
partner generally does not intend to take over control or restructure the selling firm.
Therefore, any shift in assessment of the value of the seller is more likely to come
from a more efficient use of corporate resources. Also, the purchase of a minority
equity stake to form an alliance by an informed purchaser is an endorsement of the
selling firm’s management and its future prospects (e.g. the value of the seller’s
option to develop and market its new technology). Therefore, a reassessment of
the seller’s value could also be a result of the revelation of new information to the
market concerning the seller’s future prospects (which the market is incapable of
evaluating without this signal).
Following the above reasoning, we expect that the formation of an equity
alliance will create value for both the purchasing and selling firms. This value
arises from a lowering of transaction costs through pooling of resources while
outsourcing non-core functions and taking advantage of the capabilities of the
allied firms. It also arises from the mitigation of adverse selection and information
asymmetry problems regarding the seller’s future prospects. However, we expect
that the value creation will vary across the partnering firms in predictable ways. For
example, Mody (1993) suggests that the flexibility inherent in alliances facilitates
experimentation with new combinations of participants in the pursuit of new
technologies or marketing strategies. Therefore, we expect firms that are highly
R&D intensive or that have high growth opportunities will gain more value from
experimentation as well as from the pooling of resources. Also we expect firms
facing higher risks in their operations to benefit more from the pooling of resources
with a partner firm in an alliance. Given information asymmetry problems, we
expect more value added for alliances involving firms in related industries (see
also Balakrishnan & Koza, 1993).
On the downside, alliances carry risks not incurred in the fully integrated
corporation. These include a variety of uncertainties arising out of dependence
on another party for resources or services. (See Klein, Crawford and Alchian
(1978), for a discussion of this general class of problems.) Furthermore, such
Minority Equity Investments and Inter-Firm Collaborations 23
partnerships are inherently less stable than fully integrated (merged) firms because
the contracts used to form them cover a limited scope of activities, and residual
rights to profits are sometimes purposefully left ambiguous. According to Murray
and Siehl (1989) these agreements constitute “complex contracts” for which it is
difficult to identify a priori the potential outcomes of the relationship, the factors
causing those outcomes, the appropriate responses to the outcomes, or which party
is responsible for each response. Due to this ambiguity, a relationship of mutual
interest (bonding) replaces legalistic enforcement mechanisms. Typically, the bond
takes one of two basic forms: an equity investment by one of the partnering firms
(a minority equity stake) or hostage assets.6
For example, Mody (1993) points out that alliances have weak incentives to
curb opportunistic behavior on the part of the partnering firms when compared
to the fully integrated firm that arises out of a merger. Equity ownership in an
equity alliance helps create an ownership structure that safeguards investments
(and other claims to future income) while aligning the incentives of the partners
for cooperative behavior (Pisano, 1989). Opportunism by an equity partner
is penalized through reductions in the value of its equity investment. Thus a
higher minority equity stake ensures greater bonding of the firm’s efforts and
reduces the possibility of opportunstic behavior as it helps align the incentives of
the partners.
Furthermore, in the event of termination, equity ownership protects the acquiring
partner by providing an enforceable mechanism for dissolving the alliance. That
is, the proportionate ownership interest of the acquiring partner provides a ready
basis for determining exactly what is due to each partner to the agreement should
the alliance be terminated. We hypothesize, therefore, that the larger the size of
the minority equity stake relative to the total shares the seller has outstanding the
lower is the risk of opportunism and the greater the potential value created by the
alliance.
Pisano (1989) argues that R&D is inherently more costly to govern through
contracts and is subject to incentive losses when internalized completely. For R&D
activities partial equity ownership is preferred to pure contractual governance.
Based on this argument, we would expect equity alliances involving R&D to be
more prevalent than those involving non-R&D activities.
In summary, because minority equity investments benefit the allied firms from
a lowering of transaction costs through the pooling of resources, we expect stock
market gains for both the purchasing and the selling firm. However, the wealth
gains will vary across the partnering firms in predictable ways. We expect more
gains to accrue to firms that are highly R&D intensive, that have high growth
opportunities, and that face high risks in their operations. Also given information
asymmetry problems experienced by selling firms, we expect more value added
24 SU HAN CHAN ET AL.
for selling firms that form alliances with purchasing firms from related industries.
We also hypothesize that the larger the size of the minority equity stake relative
to the total shares the seller has outstanding, the lower is the risk of opportunism
and the greater the potential value created by the alliance.
To obtain the sample of firms entering into alliances involving the purchase of
minority equity stakes, we searched both the Lexis/Nexis database (including
the Business Wire, PR Newswire, Southwest Newswire Reuters and United Press
International) and the Dow Jones News Retrieval Service database (including the
Dow Jones News Wire and the Wall Street Journal) for the 1981–1992 period. We
combine various keywords including “alliance,” “minority-equity” and “minority
stake” with the different agreement types (licensing, marketing, distribution,
supply, production, manufacturing, development, research, and technology).
Unlike some previous studies of interfirm equity investments (see Choi, 1991;
Madden, 1981; Mikkelson & Ruback, 1985), we did not use the SEC Schedule
13D filings to identify our initial set of firms. This was done for the following
reasons: (1) equity alliances often involve small, non-publicly traded seller firms
(13D filings are required only when 5% or more of a publicly traded firm’s stock
is acquired); (2) 13D filings do not give information as to the purpose of the
investment; and (3) previous studies have noted that the 13D filings are often
preceded by press releases or WSJ reports that reveal plans to acquire a minority
stake in another firm.7
We found 93 announcements of minority-equity investments that meet the
following sample selection criteria: (1) the purpose of the buy-in is to form an
alliance to pool knowledge and resources to achieve a common objective; (2)
the equity stake creates a minority position (does not exceed 50% of the selling
firm’s outstanding shares); (3) at least one of the partners in the alliance has
stock return data available in the CRSP daily returns file for NYSE, AMEX, or
NASDAQ firms during the period of analysis;8 and (4) at the time of the buy-in, the
purchasing firm gave no indication of an intention to acquire the seller firm. The
following announcement by Hewlett-Packard is an example of an announcement
of a minority-equity investment:
Hewlett-Packard Co. announced that it had entered into a letter of intent with Santa Barbara
Laboratories Inc. providing for a multi-faceted relationship between the two companies, as part
of which Hewlett-Packard will acquire a minority-equity position in Santa Barbara Laboratories
Minority Equity Investments and Inter-Firm Collaborations 25
for an undisclosed amount of cash. . . . The company said the proposed relationship includes
arrangements for technical collaboration (Dow Jones News Wire, 9/24/84).
Of the firms involved in the minority-equity alliances, stock return data was
available in the CRSP file for 85 purchasing firms and 29 selling firms (two of
these announcements involve multiple firms purchasing the same seller firm).9 The
minority equity stake (as reported by the purchasing/selling firm in the news article)
averages 15%, ranging from 3% up to 32% of the selling firm’s outstanding shares
(59 of the announcements reported precise figures on the size of the minority-
equity position, while the others simply reported that a minority-equity position
was taken). Mikkelson and Ruback (1985) in their study of the value effects
of accumulation of large blocks report an average equity position of 37% for
13D filings associated with an outstanding takeover proposal and 9.8% where the
equity purchase is for investment purposes. In Wruck’s (1989) study of private
sales of equity securities she reports that after taking into account the purchasers’
stockholdings prior to the sale, the stake averages 26%, and ranges from about 2
to 84%. The higher stakes reported by Wruck may reflect the fact that her sample
includes cases where the purchaser intends to gain control of the selling firm,
obtain the right to elect or nominate directors, or help impede a takeover of the
selling firm by another firm. Allen and Phillips (2000) report that the average size
of the equity stakes in their sample was 20% (median 14%) which is very similar
to our sample. Thus, it appears that the average equity stake in minority equity
alliances lies somewhere between that associated with a takeover proposal and a
passive investment.
Panel A of Table 1 presents the sample observations classified by the year of
their announcement. While the sample spans twelve years, approximately 45% of
the sample is concentrated in 1989 and 1990. The concentration of minority equity
alliance announcements in these two years is consistent with a report indicating
that acquisitions and minority equity investments rose in 1989 as the U.S. high-
technology industry restructured (Electronic Business, July 9, 1990, pp. 42–44).
According to the same report the number of foreign firms acquiring equity interests
in U.S. electronics firms also rose in that year. In our sample we find four incidences
of buy-ins by foreign firms, one in 1988, two in 1989 and one in 1990. For the
foreign purchasing firms, the primary motive is to tap an existing infrastructure
and to establish an international working partnership.10
Panel B of Table 1 classifies the alliances by the purpose of the agreement. We
find that 26% of the announcements (24 cases) involve development or research
alliances followed by 23 cases of marketing alliances (25%). There are 25 cases
(27%) that involve combinations of the four types of agreement and 9 cases
(10%) that did not specify the purpose for forming the equity alliance. The above
26 SU HAN CHAN ET AL.
Year of Number of % of
Announcement Announcements Total
1981 2 2.2
1982 4 4.3
1983 5 5.4
1984 4 4.3
1985 3 3.2
1986 10 10.8
1987 6 6.5
1988 8 8.6
1989 22 23.7
1990 20 21.5
1991 2 2.2
1992 7 7.5
Total 93 100.0
Panel B: Distribution of Announcements by Type of Cooperative Agreement
I. Licensing 4 3 1
II. Marketing 23 19 7
III. Development or Research 24 24 7
IV. Technology 8 8 2
V. Combination of II and III 14 12 9
VI. Other combinations of I–IV 11 11 1
VII. Not Specified 9 8 2
Total 93 85 29
Note: The table reports the distribution by year and by type of cooperative agreement for the 93 alliance
announcements involving the purchase/sale of minority equity stakes by NYSE, AMEX, and
NASDAQ firms in the 1981–1992 period. The sample of announcements (identified by searching
the Lexis-Nexis database) reported below include only cases where at least one of the partnering
firms in the alliance is publicly traded with return data available on the CRSP tape. Panel
A reports the number of announcements in each year during the 1981–1992 period. Panel B
classifies the announcements by the type of cooperative agreement as stated in the news article
containing the announcement of the alliance. It also reports the number of announcements under
each type made by purchasing firms and selling firms.
Minority Equity Investments and Inter-Firm Collaborations 27
Note: The table classifies the purchasing and selling firms in our sample into a high-technology group or
a low-technology group. We classify a firm as being in the high-technology group if its operations
are primarily hitech as described in CorpTech, a database containing information about high-
tech manufacturing, development and service companies in the U.S. Both the Purchasing and
selling firms reported below are publicly-traded firms with return data available on the CRSP
tape.
28 SU HAN CHAN ET AL.
(5 purchasing firms and 1 seller) comes from the motor vehicle and parts industry.
We also find that of the 93 announcements of minority-equity purchase, 38 (41%)
involve combinations between firms in related industries (the same three-digit SIC
industry).11
Most of the buy-ins involve larger firms acquiring the shares of a smaller firm.
Furthermore, roughly 69% of the purchasing firms bought into private firms. There
are only 23 cases where both purchasing and selling firms are publicly-traded with
data available on the CRSP tapes. (Note that there are 29 cases if we include
6 cases in which one of the partnering firms is a foreign-listed firm.) Note that
those sellers that are publicly-traded are relatively young in terms of the number
of years since listing. The average as well as the median age since listing is about
6 years.
Table 3 presents sample statistics to show the basic characteristics of both
the purchasing and selling firms that collaborate using a minority equity stake.
The purchasing firms are much larger, about 85 times, than the selling firms (as
measured by the market value of equity 21 days prior to the announcement). It
Purchasing firms
Mean 21,369 0.08 0.58 0.49 0.07%
Standard deviation 0.04 0.37 0.16 0.19%
Median 0.07 0.55 0.46 0.03%
Sample size 85 76 83 84 84
Sellers firms
Mean 252 0.95 0.40 0.50 0.16%
Standard deviation 2.32 0.28 0.26 0.11%
Median 0.17 0.38 0.48 0.12%
Sample size 29 23 27 28 29
Note: The table provides summary statistics for the Purchasing and selling firms in our sample. The
market value of equity is the market value of the firm’s common stock 21 trading days before the
initial announcement of the equity alliance. RD/Sales is the ratio of research and development
expenditures divided by firm sales in the year prior to the announcement of the minority equity
investment. BV/MV is the ratio of the book value of the firm’s assets to their market value
(measured as the sum of the market value of the firm’s equity plus the book value of its liabilities).
Leverage is the ratio of the total firm liabilities to total assets. Variance is the variance in daily
stock returns over the two years preceding the announcement of the minority equity investment.
Except for the stock returns which are obtained from CRSP tapes, all the other variables are taken
from the Compustat tapes in the year prior to the announcement. The number of observations
for each variable varies depending on the availability of data from the Compustat tapes.
Minority Equity Investments and Inter-Firm Collaborations 29
is noteworthy that the average size of the selling firm sample ($252 million) is
similar to that of Wruck’s (1989) private equity sales sample ($234 million).
Purchasing firms’ R&D/sales ratio averages 8% with a median of 7% while
selling firms’ R&D/sales ratio is much higher, averaging 95% with a median of
17%. The t-statistic for the difference in the mean R&D/sales between the two
samples is significant (t-statistic = 1.80). The high average R&D/sales ratio for
the seller sample is driven mainly by four high-tech selling firms for which the
R&D/sales ratios exceed 100%. Even when comparing the purchasing and selling
samples using the median R&D/sales ratio, the selling firms are still at least twice
as R&D intensive as the purchasing firms.
Although both purchasing and selling firms exhibit growth opportunities
measured using the book-to-market value (BV/MV) ratio, the selling firms exhibit
a significantly lower mean BV/MV ratio than the purchasing firms (t-statistic of
difference in mean BV/MV = 2.67). (BV/MV is the ratio of the book value of
the firm’s assets to their market value, where market value is equal to the sum
of the market value of the firm’s equity plus the book value of its liabilities.) This
suggests that the selling firms in general exhibit higher growth opportunities than
the purchasing firms. The lower BV/MV ratio for the selling firms could also
suggest that the fraction of market value of the selling firms’ assets attributable to
tangible assets is lower than that for the purchasing firms.
The above observations about the differences in size, R&D intensity and growth
opportunities between the purchasing and selling firms are consistent with the
argument that the larger and relatively less R&D intensive firms with lower
growth prospects use minority equity alliances to gain access to the capabilities
of smaller, knowledge-intensive firms with attractive growth prospects. Although
both purchasing and selling firms exhibit similar levels of financial leverage,
the selling firms exhibit significantly higher stock return volatility (as measured
by the variance in daily stock returns) than their purchasing firm counterparts
(t-statistic of difference in mean variance = 3.09). The higher variance in the
selling firms’ stock returns could be largely due to the higher risk of their
operations.
We also examine the ownership concentration in our sample firms using the
percentage of insider ownership information from CDA/Spectrum. CDA/Spectrum
defines an insider as an officer, director or beneficial owner (holder of 10% or
more) of a company’s stock. We measure the total direct insider ownership as a
percentage of the number of shares outstanding for the quarter just prior to the
announcement of the equity alliance. The mean (median) insider holdings for the
purchasing firms and sellers are 14.6% (5%) and 17.5% (17.5%), respectively.
These data are consistent with evidence reported in other studies (Morck, Shleifer
& Vishny, 1988; Wruck, 1989) that ownership is more concentrated in smaller
30 SU HAN CHAN ET AL.
The purchasing firms in our sample use a minority-equity buy-in as a means to enter
into a collaborative relationship with a selling firm. Hence we expect the outcomes
of this form of interfirm investment to differ from those buy-in situations where
the purchasing firm is considering a takeover or making a passive investment. In
the study of interfirm investment by Mikkelson and Ruback (1985) they report (see
their Table 1) that for those investments where the purchasing firm is considering
a takeover, about 49% of the selling firms were taken over by the purchasing firm
within three years of the initial buy-in while for passive investments only about
8% of the selling firms were taken over by the purchasing firm. Also for passive
investments, about 44% ended in a targeted repurchase or a sale of shares to a third
party.
It is also interesting to see how these relationships evolve when compared to
non-equity strategic alliances and joint ventures. In Chan et al.’s (1997) study of
non-equity strategic alliances, they find that in the eight year period following
the announcement of the alliances, only a small percentage (around 6%) of the
alliances resulted in early termination of the agreement. The majority of them
(about 83%) involve the signing of new agreements to strengthen or expand the
relationship. Only a small percentage of the alliances (around 6%) evolved into a
more permanent form of relationship, such as a joint venture or a merger. This is
in contrast to joint ventures in which roughly 80% end in a takeover by one of the
partners (see Bleeke & Ernst, 1995). Since minority equity alliances lie somewhere
in between non-equity alliances and joint ventures in terms of the integration of the
allied firms, we would expect to find the outcomes of the minority equity alliances
falling somewhere between the other two modes of interfirm relationships.
To track the evolution of the minority-equity alliances we examine the Wall Street
Journal (WSJ) index and news articles from the wire services for any related news
reports related to these investments. We track these news reports starting from the
date of the announcement of the minority-equity relationship and terminating three
full years after the announcement. Table 4 summarizes the results of this analysis.
We find a total of 41 outcome events reported for our sample firms in the WSJ
and newswire services within three years. More than 70% of them relates to the
strengthening or extension of the initial collorative agreement (categories A–E).12
There are nine instances where the purchasing firm increases stake in the sellers,
fifteen incidences of formation of new pact, two incidences where the partnering
Minority Equity Investments and Inter-Firm Collaborations 31
Note: This table reports the outcomes of the interfirm alliances in our sample. The table only reports
the outcomes involving both purchasing and selling firms that are announced in either the
Dow Jones News Retrieval Service or the Lexis/Nexis database. The outcomes span the period
between the initial announcement of the alliance and the end of the third full year following the
initial announcement made during the 1981–1992 period.
firms decide to form a joint venture, and four instances (about 10% of the total
announced outcomes) where a purchasing firm agrees to acquire the selling firm.
(In two of the acquisition cases the purchasing firm agreed to acquire the selling
firm a year after purchasing a minority equity stake.) The later finding is in contrast
to that for interfirm investments with a takeover intention in which about 49% of
the selling firms were taken over (see Mikkelson & Ruback, 1985) or to that for
joint ventures in which roughly 80% end in a takeover by one of the partners (see
Bleeke & Ernst, 1995).
Table 4 also shows a total of eleven outcomes (about 27% of the total announced
outcomes) in which the alliance agreement was reduced in scope or terminated
(categories F–I). There are two cases in which the partnering firms were involved
in legal disputes over violation by one of the partners of some key elements of the
partnering agreement. There are two cases where the purchasing firm reduces its
stake in the seller and three cases (about 7% of total announced outcomes) where
a purchasing firm sells its entire equity interest in the selling firm either back to
the selling firm or to a third party.13 This contrasts with passive investments in
32 SU HAN CHAN ET AL.
which about 44% ended in a targeted repurchase or a sale of shares to a third party
as reported by Mikkelson and Ruback (1985). Finally, there are five cases in the
sample where the alliance was dissolved because one of the partnering firms was
acquired by a third party.
In summary, the evidence in Table 4 indicates that for many firms the formation
of an alliance represents the beginning of a dynamic relationship between the
partnering firms. This evidence provides some limited support for Mody’s (1993)
conjecture that alliances can be used as a learning experiment. Changes in the
relationship may also be the result of adaptation to evolving market conditions as
well.
4. RESULTS
4.1. Gains to Purchasing and Selling Firms in a Minority-Equity Alliance
Table 5 presents the abnormal returns and the distribution of wealth gains between
firms participating in alliances involving a minority-equity commitment. For the
sample of 85 acquiring firms, 55% experience positive abnormal returns on the
announcement day; however, the average abnormal return is an insignificant
0.07%. The two-day results are similar. This evidence indicates that on average
the purchasing firms neither pay too much nor too little.14
For the sample of 29 selling firms, 76% of these firms experience positive
abnormal returns on the announcement day and the average abnormal return
Minority Equity Investments and Inter-Firm Collaborations 33
Note: The table reports the announcement day as well as pre- and post-announcement returns for both
the purchasing and selling firms forming an alliance involving the purchase/sale of a minority
equity stake. The market value of equity as well as the dollar change in equity value is also
reported for both purchasing and selling firms. The market value of equity is the market value of
the firm’s common stock 21 trading days before the initial announcement of the equity alliance.
The mean dollar change in wealth is computed by first multiplying each participating firm’s
market value of equity by its two-day (0, +1) cumulative abnormal return and then averaging
the product across firms in the Purchasing firm or selling firm sample.
∗ Significant at the 5% level for a two-tailed test.
is 5.42% (Z-statistic = 7.84). The average abnormal return over the two-day
announcement period is a statistically significant 8.15% with a Z-statistic of 8.24.
This evidence indicates that significant value is created for the stockholders of the
selling firms when a minority equity buy-in to form an alliance is announced.
For both the purchasing and selling firms, we find that the cumulative
average abnormal return (CAAR) in the pre-announcement period (−20, −1) is
insignificant. In the post-announcement period (2, 10) the CAAR is insignificant
for the purchasing firms, but is a significant −2.87% for the selling firms. This
suggests that the market may have overreacted to the news of the minority equity
alliance for the selling firms when it was announced. However, considering that
the CAAR on days 0 and +1 is a significant 8.15%, the selling firms still garner a
net gain.
The finding that the gains accrue to selling firms but not to purchasing firms
is consistent with prior studies of mergers and acquisitions. The result is also in
line with findings for joint-ventures and non-equity strategic alliances in which the
34 SU HAN CHAN ET AL.
smaller strategic partner (analogous to the selling firm) reaps a higher abnormal
return compared to the larger partner (analogous to the purchasing firm). Similarly,
in the subset of their sample that involved the announced formation of alliances
and joint ventures, Allen and Phillips (2000) find that the stock prices of selling
firms respond positively while purchasing firms do not. For their entire sample
of block equity purchases they find that the cumulative average excess stock
returns to sellers over a 21-day (−10, +10) interval is a significant 6.9%. The
response is greater (9.1%) for the subset of their sample where the equity purchase
is accompanied by the announced formation of a product market relationship in
the form of joint ventures or alliances. Similarly, we find that for the same 21-
day period the selling firms involved in minority equity alliances the cumulative
average excess stock returns is a significant 8.35%.
In their study of accumulations of 5% or more of another company’s shares
Mikkelson and Ruback (1985) find that, in general, the share prices of both the
acquiring and selling firms increase in response to the initial disclosure of the
investment position. In contrast, we find that where the motive of the equity buy-in
is to form an alliance, only the stock prices of the selling firms respond favorably.
It is possible that our inability to detect a significant abnormal return for the acquirer
is due to the large size differential between the purchasing and the selling firms in
our sample. Table 5 shows that purchasing firms are significantly larger (85 times
the market capitalization of) than the selling firms. To investigate this possibility
we investigate dollar returns corresponding to the announcement period abnormal
returns. To obtain a measure of the dollar change in wealth for the purchasing and
the selling firms, we first multiply each participating firm’s market value of equity
by its two-day cumulative abnormal return and then average the product across
firms in the sub-samples. We find that the mean dollar change in wealth (based on
two-day abnormal returns) for the sellers is $20.04 million while that experienced
by the purchasing firms is $119.36 million. These dollar returns suggest that on
average the formation of a minority equity alliance benefits both the purchasing
and selling firms.
The average excess stock returns for selling firms in our sample are higher
than that obtained by Mikkelson and Ruback (1985) in their study of interfirm
investment and by Wruck (1989) who examines private equity sales. Mikkelson
and Ruback report a two-day excess return of 3.24% while Wruck reports an excess
return of 4.5%.15 One possible explanation for this result is that the alliances
we study are expected to generate improvements in the utilization of corporate
resources or to mitigate information asymmetry problems such that they create
greater value than the sale of equity to a passive investor.
The purchasing firm results are different from those of Mikkelson and Ruback
(1985). They find statistically significant two-day abnormal returns (1.27%,
Minority Equity Investments and Inter-Firm Collaborations 35
see their Table 4, p. 535) for purchasing firms while our sample produces an
insignificant two-day abnormal return (0.33%).16 This result may be due to
differences in sample composition for the two studies. Interfirm equity purchases
can also be made as a part of a passive investment strategy wherein the acquiring
firm does not intend to become involved in the business of the selling firm and only
hopes to reap its proportionate share of the selling firm’s returns. The Mikkelson
and Ruback sample may include alliances formed with both active and passive
investment motives whereas we study only those cases in which the acquirer
plans to engage in an active investment strategy involving the selling firm. It is
also noteworthy that the majority of selling firms in equity alliances are small
entreprenuerial firms, and consequently the relative size of the purchasing firm
compared to the seller may be much larger in our sample than in Mikkelson and
Ruback’s sample. Thus the lack of significance in the returns to the purchasing
firms in our study may simply be due to the relative size of the event’s valuation
impact relative to the size of the firm.
We test for the possibility of a wealth transfer between the purchasing and
selling firms by estimating the correlation between the two-day announcement
period abnormal returns for the subsample of 23 equity alliances for which both
the purchasing and selling firms are in our sample. If sellers were benefiting at the
expense of purchasing firms or vice versa, then we would expect this correlation
to be negative and significant. The resulting coefficient was −0.03 and was not
statistically different from zero. Consequently, we find no evidence of wealth
transfers between purchasing and selling firms.
We include the logarithm of the market value of firm equity (on event day −21) in
the regression to control for the influence of firm size (Firm Size). We also construct
an indicator variable that equals 1 if the alliance agreement involves research
and/or development to test whether there is a difference in share price response
between alliances where the stated purpose of the agreement involves research
and/or development versus non-R&D types of agreements (such as marketing
agreements). Since we do not find the type of agreement to be a differentiating
factor we do not report this regression result.
Panels A and B of Table 6 present the estimated coefficients from the cross-
sectional regressions involving the two-day cumulative abnormal return for the
purchasing and selling samples, respectively. Panel A shows that the valuation
effect for the purchasing firms is positively correlated with their R&D intensity.
In other words, purchasing firms that are more R&D intensive seem to reap higher
returns from forming a minority equity alliance than their less R&D intensive
counterparts.
Panel B shows that the abnormal returns for the selling firms have significant
positive relationship with the Variance, Stake and Related Industry variables. A
positive coefficient for the Variance variable indicates that selling firms with more
volatile stock returns enjoy higher abnormal returns from forming the alliance. The
positive and significant coefficient for the Stake variable suggests that the larger the
minority stake purchased by the purchasing firm, the higher the abnormal returns
for the selling firms. This result supports the hypothesis that the larger the size of
the minority equity stake relative to the seller’s outstanding shares, the lower is the
risk of opportunism and consequently the greater the potential value that is created
by the alliance. Finally, selling firms that tie-up with purchasing firms that are in
a related industry are associated with significantly higher abnormal returns. This
result conforms to the hypothesis that purchasing firms from a related industry
are better able to evaluate the selling firm’s assets-in-place and its growth options
thus mitigating the information asymmetry problems regarding the selling firms.
Although the BV/MV variable has a negative coefficient (suggesting that firms
with higher (lower) growth opportunities tend to get a higher (lower) payback
from forming an alliance), it is not significant.
Note: This table reports the cross-sectional regression results obtained by regressing the two-day (0, +1) cumulative abnormal return for investing (and
sellers) firms on selected variables representing firm size, technology status of firm (HITECH), R&D intensity, growth opportunities (BV/MV),
riskiness of firm (variance), percentage of equity stake in the selling firm (STAKE), and industry relatedness (RELATED INDUSTRY). Firm
size is measured by the logarithm of the market value of equity 21 days prior to the announcement of the alliance. Hitech is a dummy variable
that takes a value of one if the firm is classified as a high tech firm. R&D intensity is the ratio of the firm’s R&D expenditures to its net
sales. BV/MV measures a firm’s growth options and is computed as the ratio of book value of the firm’s assets to the market value of the
firm’s assets (measured as the sum of the market value of the firm’s equity plus the book value of its liabilities). Variance, a measure of the
riskiness of the firm, is the variance in daily stock returns (from CRSP tapes) over the two years preceding the announcement of the minority
equity investment. Stake is the percentage of equity of the selling firm reported in the news article that the investing firm says it will purchase.
RELATED INDUSTRY is a dummy variable that takes a value of one if the investing and selling firm are from the same 3-digit SIC industry.
R&D intensity and BV/MV are computed using data from the Compustat tapes the year prior to the announcement of the formation of the
equity alliance. There are 46 observations from the investing firm sample and 21 observations from the seller sample that have complete data
on all the variables used in the regression.
∗ Significant at the 5% level for a two-tailed test.
∗∗ Significant at the 10% level for a two-tailed test.
37
38 SU HAN CHAN ET AL.
gain insight into the circumstances under which managers choose to engage in
such alliances. There are two issues addressed in this analysis. First, do firms that
enter into alliances do so because they are performing poorly or simply because
the purchasing firms have the cash flow to invest? Mohanram and Nanda (1996)
observe that firms entering into joint ventures tend to be experiencing deteriorating
performance. Second, does the operating performance of alliance partners improve
or deteriorate following the equity tie-up?
Data on operating performance spanning the six-year period surrounding (two
years before to three years after) the announcement of the minority equity alliance
is obtained from the COMPUSTAT. We measure operating performance using
operating cash flow return on assets (OCF-ROA) which is computed by dividing the
earnings before interest, taxes, depreciation, amortization, and extraordinary items
(COMPUSTAT mnemonic EBITDA) by the book value of total assets. To adjust for
industry influences we calculate industry-adjusted performance measures as the
difference between the announcing firm value in a given year and the median value
for all firms on COMPUSTAT with the same four-digit SIC code.17 The number
of observations in each year varies depending on the availability of COMPUSTAT
data.
Panels A and B of Table 7 report the operating performance results for the
set of purchasing and selling firms. These results indicate that firms that acquire
equity shares in an alliance partner, on average, have higher operating cash flow
to assets than their industry peers over the entire six year period beginning two
years before the announcement of the formation of an alliance. This suggests that
purchasing firms that engage in minority equity alliances tend to be cash rich
firms and therefore their participation in such alliances is not motivated by poor
performance. Furthermore, Panel A shows that when firm performance relative
to its industry peers is analyzed for three years following the formation of the
alliance, the purchasing firm’s relative cash flow position is reduced significantly
(although it is still above the industry) when compared to its cash flow position in
the year preceding the equity participation. This evidence suggests that purchase
of a minority equity stake is a way for the cash-rich firms to spend their cash.
Selling firms, on the other hand, show negative operating cash flow-to-asset
position in most of the years surrounding the announcement of the minority equity
alliance. Furthermore, the selling firms perform significantly worse than their
industry peers in the two years leading up to the equity tie-up announcement.
Unlike the purchasing firms, however, the operating cash flow-to-asset position of
the sellers show significant improvement in the third year following the equity tie-
up when compared to the year preceding the alliance announcement. The above
results suggest that poor performance may be a motivation for selling firms to
enter into interfirm collaborations involving minority equity participation by the
Minority Equity Investments and Inter-Firm Collaborations
Table 7. Operating Performance of Purchasing and Selling Firms in Minority Equity Alliances Surrounding Minority
Equity Purchase Announcements.
Year Relative to Minority Equity Purchase Announcement
−2 −1 0 1 2 3
Note: Performance is measured using operating cash flow return on asset (OCF-ROA). OCF-ROA is computed using the earnings before interest, taxes,
depreciation, amortization, and extraordinary items (EBITDA) divided by total assets. The industry-adjusted performance measure equals the
difference between the announcing firm’s measure in a given year and the median value for all firms on COMPUSTAT with the same four-digit
SIC code. The number of observations in each year varies depending on the availability of COMPUSTAT data.
∗ Significant at the 5% level for a two-tailed test.
∗∗ Significant at the 10% level for a two-tailed test.
39
40 SU HAN CHAN ET AL.
partnering firm. In addition, the alliance and equity stake seem to help improve
the operating performance of the selling firms.
In summary, the above results suggest that, unlike joint venture agreements,
poor performance is not a motivation for purchasing firms to enter into interfirm
collaborative agreements involving a minority equity stake. For selling firms,
however, poor operating performance may be a motivation for them to enter
into minority equity alliances. In addition, the post-performance results for both
the purchasing firms and selling firms seem to be in line with the evidence we
present on the market’s response to the announcement of minority equity tie-ups
by purchasing and target firms.
5. CONCLUSIONS
Minority-equity investment in a strategic partner’s common stock is frequently
used to finance an alliance and to bond the partnering firms to the agreement.
The partner that sells the equity stake gets an immediate cash infusion, while the
purchasing partner gets a percentage of the seller’s profits plus access to the seller’s
skills and technology as prescribed in the alliance agreement.
Based on 93 announcements of companies making minority-equity
commitments in strategic partners who indicate no intention to seek control of
the partnering firm, we find that the average stock price response is neutral for
the acquiring firms, but positive and significant for the selling firms. Further,
we find no evidence that the gains to the selling firms are a result of a wealth
transfer from the purchasing firms. The finding of significant positive wealth gain
for the selling firms is consistent with Mikkelson and Ruback’s (1985) study of
interfirm equity investments, Wruck’s (1989) study of private equity sales, and
Allen and Phillips’ (2000) study of corporate equity ownership associated with
product market relationships. Our cross-sectional analysis reveals that the two-
day announcement abnormal returns for the selling firms are positively related to
the amount of equity invested by the purchasing firms, their riskiness as reflected in
the variability of their stock returns, and the relatedness of the industry to which the
purchasing firm belongs. The analysis also reveals that purchasing firms that are
more R&D intensive seem to reap higher returns from a minority equity alliance
than their less R&D intensive counterparts.
The evidence from this study suggests that the pooling of resources in a minority
equity alliance benefits the selling firms. The change in the selling firm’s value
could be derived from enhanced financing flexibility, a lowering of transactions
costs, and a more efficient use of corporate resources under the minority equity
arrangement. In addition, the purchase of minority equity stake by a sophisticated
Minority Equity Investments and Inter-Firm Collaborations 41
purchasing firm solely for the purpose of forming a strategic alliance can constitute
an endorsement of the selling firm’s management and future prospects. Therefore,
the positive valuation observed for the selling firm’s stock could also be due to
this revelation of new information to the market concerning the seller’s future
prospects.
Within three years of the announced formation of an alliance the majority of the
agreements continue unchanged. Of those that do change, the largest number are
expanded or strengthened. The result is an increase in the equity stake, formation
of a joint venture or merger of the partnering firms. A very small percentage of
the alliances terminate within three years of their formation. These results provide
evidence that firms engaging in minority equity alliances often take advantage of
the flexibility afforded in alliances to alter their relationships, possibly to adapt to
evolving market conditions.
Finally, firms that acquire minority stakes in alliance partner firms tend to
have better operating performance than their peers during the period two years
before through three years following the announcement of the agreement. The
cash flow position of the purchasing firms, however, deteriorates somewhat in
the three years following the announcement, compared with their own prior cash
flow position (although still better than industry peers). Selling firms, on the other
hand, tend to perform worse than their industry peers in the two years before
the announcement of their engaging in the formation of an equity alliance. Their
performance in the three years following the announcement, however, improves
significantly when compared to their pre-alliance performance. This evidence on
the selling firms differs from that observed for non-equity alliances and suggests
that poor performance may be a motivator for selling firms to participate in alliances
that involves equity participation by a cash rich partner.
NOTES
3. Allen and Phillips (2000) examine a sub-sample of block equity purchases that are
accompanied by alliances and joint ventures formed within a two-year interval either prior
to or subsequent to the purchase.
4. Morck et al. (1988) find that smaller firms find information asymmetry problems to
be particularly acute, which would make minority equity alliances particularly attractive
for these firms. For more about public versus private equity offerings, see Wu (2000).
5. For more discussion on this issue, see Parkhe (1993) and Chan et al. (1997).
6. See Pisano (1989) for a discussion of the role of equity financing in reducing the
incentives for opportunism by partners in an equity alliance. Williamson (1985, p. 191)
describes a double hostage system for bonding an alliance “. . . reciprocity involves the sale
of specialized product to B conditioned on the procurement of specialized product from
B. The argument here is that reciprocity can serve to equalize the exposure of the parties,
thereby reducing the incentive of the purchaser or seller to defect from the exchange –
leaving the supplier to redeploy specialized assets at greatly reduced value.”
7. Madden (1981) found that the Wall Street Journal date preceded the Insider’s
Chronicle date (that reports Schedule 13D filings) by an average of 37 days. The average
delay between the filing date of the Schedule 13-D and the date of the Insider’s Chronicle
was 23 days.
8. Minority equity purchases have been used extensively to forge alliances between large
publicly-held firms and smaller privately-held firms and also between domestic U.S. and
foreign firms. Therefore, we include publicly-traded firms (with returns data available on
the CRSP) regardless of whether they form minority equity alliances with a publicly traded
firm, a private firm or a foreign firm.
9. The number of purchasing (selling) firms in the final sample excludes four (two)
publicly-traded foreign firms that do not have stock return data in the CRSP file.
10. For example, in 1989, the Ascom Group of Switzerland (the largest communications
and services automation equipment company in Switzerland) made a $6 million direct
investment in New York based Comverse Technology Inc. (one of the leading international
suppliers of computerized message management systems) for 17 million shares, giving it
an ownership position of 13.7% (Business Wire, August 17, 1989). The investment was
made as part of a multi-faceted strategic alliance formed by the two companies to distribute
message management systems to organizations that provide subscriber services in selected
markets of Western Europe. Under the pact, Ascom will also market Comverse’s defense
and related communications processing systems in Switzerland, West Germany and other
markets.
11. For those firms that are not publicly-traded, we read the news articles describing the
minority equity alliance to find out the industry in which those firms are operating.
12. Since we rely on press releases for information we recognize the possibility that
firms may simply not announce changes in the status of their alliances following the initial
announcement.
13. For example, in October, 1993 IBM sold its entire equity interest in Hogan Systems
Inc. back to Hogan in a private market transaction. IBM has bought approximately 5% of
Hogan’s system’s stock since September 1990 to strengthen a strategic partnership that has
existed between the two companies since 1986. The sale was profitable for IBM due to the
strong increase in price of Hogan’s stock. For Hogan, the sale was potentially profitable
because it hopes to make a better return from the repurchase of those shares than from
holding cash. A Hogan senior vice president said it is not clear yet if Hogan and IBM will
Minority Equity Investments and Inter-Firm Collaborations 43
extend the current exclusive marketing agreement after it expires in 1995 (Business Wire,
10/4/93).
14. We also compare the excess returns for firms that purchase equity stakes in private
firms with those that purchase equity stakes in public firms and find no significant difference.
15. The results for the Mikkelson and Ruback (1985) study reflect the abnormal returns
for a sample of 106 firms that sold shares to other firms who stated that their purchase
objective was for investment purposes. Similar results for selling firms that were the target
of a takeover attempt were much higher (7.74%).
16. As in Mikkelson and Ruback, we also evaluated the abnormal returns of the
purchasing firms based on whether they are frequent (appearing in our sample six or more
times) or infrequent purchasers of minority equity. As for the full sample, the abnormal
returns for both sub-samples are insignificant.
17. Allen and Phillips (2000) use a similar technique to examine operating income
changes surrounding corporate equity purchases. See also Barber and Lyon (1996).
ACKNOWLEDGMENTS
We thank Marshall Bloom, Andrew Chen, David Denis, Diane Denis, Ning Gong,
Mike Impson, Paul Laux, Ray Miles, Ko Wang as well as participants at the
Texas Finance Symposium and the Financial Management Association Meeting
for comments on previous versions of this paper. We are also grateful to Yajat
Bindal and Jaafar Salwani for their research assistance.
REFERENCES
Allen, J. W., & Phillips, G. M. (2000). Corporate equity ownership and product market relationships.
Journal of Finance, 25, 2791–2815.
Balakrishnan, S., & Koza, M. P. (1993). Information asymmetry, adverse selection and joint ventures.
Journal of Economic Behavior and Organization, 20, 99–117.
Barber, B., & Lyon, J. (1996). Detecting abnormal operating performance: The empirical power and
specification of test-statistics. Journal of Financial Economics, 3, 359–400.
Bethel, J. E., Liebeskind, J. P., & Opler, T. (1998). Block share purchases and corporate performance.
Journal of Finance, 53, 605–634.
Chan, S. H., Keown, A. J., Kensinger, J. W., & Martin, J. D. (1997). Do strategic alliances create value?
Journal of Financial Economics, 46, 199–221.
Chan, S. H., Martin, J. D., & Kensinger, J. W. (1990). Corporate research and development expenditures
and share value. Journal of Financial Economics, 26, 255–276.
Choi, D. (1991). Toehold acquisitions, shareholder wealth, and the market for corporate control. Journal
of Financial and Quantitative Analysis, 26, 391–407.
Dodd, P., & Warner, J. B. (1983). On corporate governance. Journal of Financial Economics, 11,
401–438.
Harrigan, K. R. (1985). Strategies for joint ventures. Lexington, MA: Lexington Books.
44 SU HAN CHAN ET AL.
Hertzel, M., & Smith, R. L. (1993). Market discounts and shareholder gains for placing equity privately.
Journal of Finance, 48, 459–485.
Jensen, M. C., & Meckling, W. (1991). Specific and general knowledge, and organizational structure.
In: L. Werin & H. Wijkander (Eds), Main Currents in Contract Economics. Oxford: Blackwell.
Johnson, S., & Houston, M. B. (2000). A reexamination of the motives and gains in joint ventures.
Journal of Financial and Quantitative Analysis, 35, 67–85.
Kensinger, J. W., & Martin, J. D. (1991). Financing network organizations. Journal of Applied
Corporate Finance, 4, 66–76.
Klein, B., Crawford, R., & Alchian, A. (1978). Vertical integration, appropriable rents, and the
competitive contracting process. Journal of Law and Economics, 21, 297–326.
Madden, G. P. (1981). Potential corporate takeovers and market efficiency: A note. Journal of Finance,
36, 1191–1197.
McConnell, J., & Nantell, T. (1985). Common stock returns and corporate combinations: The case of
joint ventures. Journal of Finance, 40, 519–536.
Mikkelson, W. H., & Ruback, R. S. (1985). An empirical analysis of the interfirm equity investment
process. Journal of Financial Economics, 523–553.
Mody, A. (1993). Learning through alliances. Journal of Economic Behavior and Organization, 20,
151–170.
Mohanram, P., & Nanda, A. (1996). When do joint ventures create value? Academy of Management
Proceedings, 36–40.
Morck, R., Shleifer, A., & Vishney, R. (1988). Management ownership and market valuation: An
empirical analysis. Journal of Financial Economics, 20, 293–315.
Murray, A. I., & Siehl, C. (1989). Joint venture and other alliances: Creating a successful cooperative
linkage. Morristown, NJ: Financial Executives Institute.
Parkhe, A. (1993). Strategic alliance structuring: A game theoretic and transaction cost examination
of interfirm cooperation. Academy of Management Journal, 36, 794–829.
Pisano, G. (1989). Using equity participation to support exchange: Evidence from the biotechnology
industry. Journal of Law, Economics, and Organization, 5, 109–126.
Williamson, O. (1985). The economic institutions of capitalism. New York, NY: Free Press.
Wruck, K. H. (1989). Equity ownership concentration and firm value: Evidence from private equity
financings. Journal of Financial Economics, 23, 3–28.
Wu, Y. (2000). The choice between public and private equity offerings. Unpublished Working Paper,
University of Chicago.
SIZE AND BOOK-TO-MARKET
EFFECTS IN THE RETURNS
ON INFORMATION TECHNOLOGY
STOCKS
ABSTRACT
This paper explores the relationship between size, book-to-market, beta, and
expected stock returns in the U.S. Information Technology sector over the
July 1990–June 2001 period. Two models, the multivariate model and the
three-factor model, are employed to test these relationships. The risk-return
tests confirm the relationship between size, book-to-market, beta and stock
returns in IT stocks is different from that in other non-financial stocks.
However, the sub-period results (the periods before and after the technology
crash in April 2000) show that the nature of the relationship between stock
returns, size, book-to-market, and market factors, or the magnitude of the
size, book-to-market, and market premiums, is on average unchanged for
both sub-periods. This result suggests the technology stock crash in April
2000 was not a correction of stock prices.
Research in Finance
Research in Finance, Volume 21, 45–91
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21003-6
45
46 QUANG-NGOC NGUYEN ET AL.
1. INTRODUCTION
One of the most important debates among financial economists, as well as financial
practitioners, in recent years concerns the valuation of information technology
(IT) stocks. The valuation of IT stocks is not an easy task as many of these firms
do not have, or have only a short history of stock performance. Even for those
few firms that do have a long history of performance, the prediction of future
earnings is extremely difficult due to unique features of the sector. Future earnings
of firms in the IT sector depend on the opportunity, the company, the management,
or the future outcome of the research and development activity, which are as
difficult to estimate as predicting future earnings themselves (Trueman, Wong
& Zhang, 2001). Empirical studies investigating IT stock valuation have used
options (Franklin, 2000) or web traffic (for internet stocks, Rajgopal, Kotha &
Venkatachalam, 2000), while Sadorsky (2003) finds that the conditional volatilities
of oil prices, the term premium, and the consumer price index each have a signifi-
cant impact on the conditional volatility of technology stock prices. An implicit fea-
ture of these studies is the assumption that IT stocks are “special” and consequently
traditional valuation models – such as the discounted cash flow model – may argued
to be inappropriate. However, no single paper has empirically explained why IT
stocks are special, except for those that cite the abnormal price-earnings ratios
(P/E) associated with these stocks as evidence for their “specialty” (Shiller, 2000).
The objective of this paper is to investigate why IT stocks are special by utilizing
a comparative perspective which tests the relation between size, book-to-market
equity, beta, and stock returns and then compares the magnitude of this relationship
with that found in earlier studies (Fama & French, 1992, 1993). The paper also
tests the difference in the risk-return relationship for IT stocks before and after
the technology stock crash in April 2000. Investigating IT stocks in a comparative
perspective can lead to evidence that would help financial academics as well as
practitioners differentiate IT stocks from other stocks, as well as providing further
insight into questions concerning valuation.
There is one obvious benefit arising from the distinction between IT stocks and
“old economy” stocks.1 It is logical that a new valuation model for IT stocks should
be developed if IT stocks are radically different from “old economy’ stocks. It is
equally reasonable a priori that the extant discounted cash flows model should
be sufficient to produce accurate measurements of IT stock prices, if there is
no distinction between IT stocks and “old economy” stocks. In the case of no
distinction, looking for a new model for the valuation of IT stocks would be
pointless.
The evidence on the relation between size, book-to-market equity, beta, and
stock returns in the IT sector is investigated in this paper using two alternative
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 47
risk-return testing models. In the first model, or the multivariate model, direct
measurements of firm size, book-to-market ratio, and market betas are applied. The
model is then used in the Fama and MacBeth (1973) cross-sectional regressions
to test the significance of the coefficients on the size, book-to-market, and
beta variables. This is the simplest approach and was widely used prior to the
introduction of the three-factor model by Fama and French (1993). In the three-
factor model, firm size, book-to-market ratio, and market beta are not applied
directly. In contrast, these variables are represented by the returns on the three
portfolios formed in a way that closely mimics firm size, book-to-market ratio,
and market beta. The time-series regressions are applied to the three-factor model.
To address any biases introduced by the impact of the technology crash in April
2000 on the behavior of stock prices, this analysis is carefully divided into two
sub-periods: one before the crash and one after the crash. Thus the paper provides
an insight into how IT stocks behave in the lead-up to the crash, after the crash as
well as how they behave over the whole sample period, from July 1990 to June
2001.
The paper improves the regression approaches used by Fama and French (1992,
1993). This paper employs the Fama and MacBeth (1973) procedure to estimate
the coefficient slopes on the size, book-to-market, and beta variables. Another
improvement, is that the paper uses the Seemingly Unrelated Regressions (SUR)
methodology to estimate the coefficients for the three-factor model. This procedure
runs time-series regressions for all portfolios as a whole, and gives estimators of the
coefficients after adjusting for cross-correlation between the portfolios’ residuals.
Individually the results are robust. However the two models, the multivariate
model and the three-factor model, produce different results. In tests using the
multivariate model, the book-to-market ratio is the only significant factor that
explains variation in stock returns. However, the book-to-market effect is only
significant at the 10% level, in contrast to Fama and French (1992) who find that
the effect is very robust (at the 5% level) for all non-financial stocks listed on the
NYSE, AMEX, and NASDAQ over the 1963–1991 period. Another difference is
that this paper finds no relationship between size and stock returns. Nevertheless,
the paper does discover that beta has weak explanatory power, which is similar to
their earlier finding.
In tests using the three-factor model, all the returns on the portfolios that are
designed to mimic the size, book-to-market, and market (beta) factors are strongly
related to returns on size – book-to-market sorted portfolios. This evidence is in
favor of the three-factor model. The different results produced by the multivariate
model and the three-factor model suggest that the two models might not be
comparable. The excess returns on the size – book-to-market portfolios and the
market portfolios before the technology crash in April 2000 (the pre-crash excess
48 QUANG-NGOC NGUYEN ET AL.
returns) are negative while the excess returns on those portfolios after the crash (the
post-crash excess returns) are indistinguishable from zero. However, this paper
finds that the relation between size, book-to-market, beta, and portfolio returns
(using the three-factor model) is similar for the periods before and after the crash.
The remainder of this paper is organized as follows. Section 2 describes the
differing empirical evidence on the risk-return relation as well as summarizes
alternative explanations for the size and book-to-market anomaly. Specific
hypotheses are then proposed. The data, its source, and the measurements of
variables for the risk-return tests are presented in Section 3. Section 4 discusses
the multivariate model and the three-factor model as well as the reasons for their
uses. Section 5 provides the regression results. Section 6 concludes the paper and
suggests the direction for future research.
2. LITERATURE REVIEW
2.1. Background
Sharpe (1964) and Lintner (1965) suggest a positive linear relation between
individual security risk and its expected return. This relationship is expressed
by the well-known Capital Asset Pricing Model (CAPM) where:
E(R i ) = E(R f ) + [E(R m ) − E(R f )]i (1)
The symbols in Eq. (1) are defined as follows: E(R i ) is the expected return on
security i for the period and is equal to the change in the price of the security, plus
any dividends, interest, or other distributions, divided by the price of the security
at the start of the period; E(R m ) is the expected return on the market portfolio of all
securities taken together; E(R f ) is the return on the riskless security for the period;
i is the contribution of security i to the risk of the market portfolio and is often
called the “systematic risk” of security i. is defined algebraically by:
Cov(R i , R m )
i = (2)
Var(R m )
Nevertheless, the applicability of the CAPM has been questioned with a
number of studies in the asset-pricing literature demonstrating a weak relationship
between beta and stock returns. Many findings show that market s of stocks
are not sufficient to explain stocks returns (e.g. Banz, 1981), or more bluntly
the relationship between market s and average stock returns is flat (e.g. Fama
& French, 1992). Additional factors, such as firm size, book-to-market equity,
have been introduced to accommodate actual stock return behavior. The failure
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 49
of the CAPM to explain empirical evidence, or the deviation from the CAPM is
considered as an “anomaly” in the risk-return relationship and is attributed to the
following two anomalies.
The apparent violations of the CAPM have inspired financial academics to find
plausible explanations for these violations. The reasons for the failure of the CAPM
50 QUANG-NGOC NGUYEN ET AL.
Controlling for BE/ME, small stocks tend to have lower earnings on book equity
than do big stocks.
A third risk based explanation concerns the presence of “data-snooping
statistics” as used by Aldous (1989, p. 252) to describe the situation “where you
have a family of test statistics T(a) whose null distribution is known for fixed a,
but where you use the test statistic T = T(a) for some a chosen using the data.”
Lo and MacKinlay (1990b) argue that if the choice of a is based on the data, then
the sampling distribution of the resulting test statistic is generally not the same as
the null distribution with a fixed a; hence, the actual size of the test may differ
substantially from its nominal value under the null.
Other possible biases related to the data-snooping bias are also suggested
by many financial academics. Black (1993) and MacKinlay (1995) say that the
deviations from the CAPM are sample-specific results that are unlikely to be
observed in various sets of data. However, other studies do not support the data-
snooping bias, sample-specific bias, and the selection bias explanations for the
deviations from the CAPM. Chan, Hamao and Lakonishok (1991), Capaul, Rowley
and Sharpe (1993), and Fama and French (1998) document strong relationships
between average return and BE/ME in markets outside the U.S. Davis (1994) finds
that the relationship between average return and BE/ME observed in recent U.S.
returns extends back to 1941. Even the relationship exists for data back to 1926
with the pre-1963 value premium is close to that observed for the subsequent period
in earlier work.
the multivariate model and the time-series regressions on the three-factor model.
These regressions techniques will be described in the Method section.
3. DATA
This study investigates Information Technology (IT) stocks that are listed on the
New York Stock Exchange (NYSE), the American Stock Exchange (AMEX) and
the NASDAQ from July 1990 to June 2001 as provided by the COMPUSTAT
Database. Not all stocks are chosen for the risk-return tests in the paper; only those
who have the market equity and book-to-market data as described in Section 3.2.2.
are included. In addition, only stocks that have monthly beta for every month during
a testing year (12 months) are selected for the tests conducted during that year. Due
to the fact that new firms enter the IT sector and existing firms exit the sector each
year, there are circumstances that a stock might be selected in a particular period
but might not be selected in a later period. Thus, the number of stocks satisfying
the testing conditions varies year by year.
Table 1 shows the total number of stocks available for the risk-return tests in
each year from July 1990 to 2000. The number of stocks in each portfolio during a
testing year (from July of year t to June of year t + 1) is also reported. For simplicity
the formation of portfolios is described in the next section. From Table 1, it can be
seen that the total number of stocks that satisfy the required conditions increases
nearly four fold over the 1990–2001 period, starting from 379 stocks for the testing
year July 1990–June 1991 to 1474 stocks for the testing year July 2000–June 2001.
This reflects the “technology boom” during the 1990s.
According to Kothari, Shanken and Sloan (1995), there are at least two aspects
of COMPUSTAT selection procedures that appear to impart a survivorship bias.
First, prior to 1978 COMPUSTAT routinely included historical financial statement
information for as many years as available going back to 1946 on firms added to
their database in a given year. In 1978 COMPUSTAT launched a major database
expansion project from about 2700 NYSE-AMEX and high-profile NASDAQ
companies to about 6,000 companies. Five years of annual data from 1973
to 1978 was added for most of these firms. Consider a firm in 1973, with
substantial assets but relatively poor earnings prospects, considerable uncertainty,
and correspondingly low market value. Suppose this high BM (book-to-market
ratio) firm performed poorly over the next five years, with earnings even lower than
expected and negative stock returns. If this company was not on COMPUSTAT
to begin with, it might not be added to the database in 1978, either because of
54 QUANG-NGOC NGUYEN ET AL.
Table 1. Year-by-Year Numbers of Stocks Available for Testing from July 1990
to June 2001.
Portfolios July July July July July July July July July July July
1990 1991 1992 1993 1994 1995 1996 1997 1998 1999 2000
Big-High 12 8 3 16 17 18 21 22 24 12 24
Big-BM2 25 25 41 32 44 39 50 68 59 46 77
Big-ME3 44 37 37 47 53 59 64 87 106 98 122
Big-Low 28 47 42 43 50 60 93 103 113 142 170
ME2-High 23 20 20 15 28 24 29 50 48 47 51
ME2-BM2 37 32 24 25 41 42 65 81 84 100 100
ME2-BM3 22 33 34 35 38 56 72 69 81 88 108
ME2-Low 28 24 37 40 39 42 50 54 73 73 101
ME3-High 23 35 39 35 38 50 64 76 95 107 107
ME3-BM2 25 22 26 34 39 45 51 65 81 89 119
ME3-BM3 21 18 17 26 26 29 44 56 51 70 94
ME3-Low 20 17 25 29 38 41 41 59 60 54 59
Small-High 33 32 43 58 62 71 97 107 111 133 183
Small-BM2 14 23 19 27 29 37 42 38 54 62 75
Small-BM3 11 14 18 17 29 24 29 41 40 39 48
Small-Low 13 7 5 8 19 22 26 36 29 31 36
Total 379 394 430 487 590 659 838 1012 1109 1191 1474
The table reports the number of stocks in each portfolio at the beginning of July of each year t from
July 1990 to June 2001.
At the end of June of each year t, all Information Technology stocks on the NYSE, AMEX, and
NASDAQ are grouped into 16 portfolios based on their market capitalization (ln(ME)) and book-to-
market ratio (ln(BM)). The size breakpoints are the first (25%), second (50%) and third (75%) quartiles
of ln(ME) of all information technology stocks on the NYSE, AMEX, and NASDAQ. Similarly, the
book-to-market breakpoints are the first, second, and third quartiles of ln(BM) of all information
technology stocks on these exchanges. Stocks with negative BM are excluded so that ln(BM) is
calculable. A portfolio’s ln(ME), ln(BM), and beta are the value-weighted averages of stocks’ ln(ME),
ln(BM), and beta, respectively. The weights are determined by stocks’ ln(ME).
Big, ME2, ME3, and Small represent the biggest size quartile, second largest size quartile, third
largest size quartile and smallest size quartile, respectively. High, BM2, BM3, and Low represent the
highest book-to-market quartile, second highest book-to-market quartile, third highest book-to-market
quartile, and the lowest high book-to-market quartile, respectively. Big-High means the size – book-
to-market portfolio whose stocks belong to the biggest size quartile and the highest book-to-market
quartile.
This section briefly describes how various variables such as firm size, book-to-
market ratio, market beta, and stock returns are measured.
ME, making it more normally distributed and facilitating the portfolio selection
procedures. Most of previous authors also use ln(ME) in their tests of capital asset
pricing model. (e.g. Chan & Chen, 1998; Fama & French, 1992; Jegadeesh, 1992).
The table reports the descriptive statistics for the 16 size – book-to-market portfolios from July 1990
to June 2001.
At the end of June of each year t, all Information Technology stocks on the NYSE, AMEX, and
NASDAQ are grouped into 16 portfolios based on their market capitalization (ln(ME)) and book-to-
market ratio (ln(BM)). ME is common equity, which is denominated in millions of U.S. dollars and
is measured at the end of June of year t. BM is the ratio of book equity to market equity, with book
equity measured at fiscal yearend falling in year t − 1, and market equity measured at the end of
58 QUANG-NGOC NGUYEN ET AL.
Table 2. (continued )
December of year t − 1. ln(ME) and ln(BM) are the natural logarithms of ME and BM, respectively.
The ME breakpoints are the first (25%), second (50%), and third quartiles (75%) of the stocks’ ln(ME).
Similarly, the BM breakpoints are the first, second, and third quartiles of the stocks’ ln(BM). Big,
ME2, ME3, and Small represent the biggest largest size quartile, second largest size quartile, third
largest size quartile and smallest size quartile, respectively. High, BM2, BM3, and Low represent the
highest book-to-market quartile, second highest book-to-market quartile, third highest book-to-market
quartile, and the lowest high book-to-market quartile, respectively.
Panel A shows the time-series averages of book-to-market ratios (ln(BM)) and the associated statistics
for the 16 size – book-to-market portfolios. A portfolio’s book-to-market ratio is the value-weighted
average of the ln(BM) of all stocks in the portfolio, with the weights calculated based on the stocks’
ln(ME).
Panel B shows the time-series averages of market equity (ln(ME)) and the associated statistics for
the 16 size – book-to-market portfolios. A portfolio’s market equity is the value-weighted average of
the ln(ME) of all stocks in the portfolio, with the weights calculated based on the stocks’ ln(ME).
annual beta. Annual beta from July of year t to June of year t + 1 is the arithmetic
average of monthly betas during the same period. That is:
beta(Jult ) + beta(Augt ) + · · · + beta(Jant+1 ) + · · · + beta(Junt+1 )
Betai,t =
12 (4)
where betai,t is the annual stock beta for the testing year beginning in July of year
t and ending in June of year t + 1; beta(Jult ), . . ., beta(Junt+1 ) are the monthly
stock betas for months from July of year t to June of year t + 1.
Monthly stock betas are extracted from the COMPUSTAT database.
COMPUSTAT calculates monthly stock beta by regressing monthly stock
returns of the previous 60 months since the calculation month (inclusive) on
monthly S&P 500 index returns of the same period. If less history is available,
COMPUSTAT uses a minimum of 24-month data to measure monthly beta. Thus,
firms that do not have at least 24 consecutive monthly returns will not have beta
available on the COMPUSTAT database and will not included in this paper’s
tests.2
Due to the fact that firms in the information technology sector do not usually
have a long history of stock prices, calculating annual beta using annual returns of
the past five years (at least two years) would leave very few firms with available
beta. This would result in too few observations for the tests of the risk-return
relationship and could give rise to spurious inferences. The use of monthly returns
in the calculation of annual beta though not producing very precise estimates of
annual beta, would work better than the use of annual returns given the short history
of stock prices for firms in the IT sector.
The procedure that COMPUSTAT uses to measure monthly beta is different
from that employed by Fama and French (1992). In this last paper (monthly) beta
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 59
is measured using the Scholes and Williams (1977) procedure, which accounts
for the degree of stock trading infrequency. Scholes-Williams beta estimates are
defined as
+1
B ik
i = (5)
(1 + 2r)
k=−1
3.2.4. Returns
Annual returns from July of year t to June of year t + 1 are calculated in the
COMPUSTAT database; they are the percentage increase (or decrease) in the
closing stock prices in June of year t and June of year t + 1, adjusted for dividends
payments and the compounding effect of reinvested dividends. Similarly, monthly
returns are the percentage increase (or decrease) in the closing stock prices on
the last days of two consecutive months. Monthly returns are also adjusted for
the effect of dividend payments. One-year Treasury bill rate and three-month
Treasury bill rate will respectively be used to calculate annual returns and monthly
returns.
Table 3 reports the time-series averages of the portfolios’ excess returns (i.e.
the spreads of portfolio’s annual returns over the one-year Treasury Bill rates),
betas, book-to-market ratios (ln(BM)), and market equity (ln(ME)) for the July
1997–June 2001 period. The summary statistics for these variables are used in the
multivariate model. There seems to be no relationship between portfolios’ excess
returns and size (or ln(ME)). Across each book-to-market quartile, the excess
returns show no particular pattern when moving from big- to small-stock portfolios.
The relationship between the excess returns and ln(BM) is somewhat stronger, with
the excess returns increasing from high- to low-BM portfolios within the biggest-
size quartile. Within the second biggest-size quartile (ME2) the excess returns
decrease from high-BM to low-BM portfolios. Beta is highly positively correlated
60 QUANG-NGOC NGUYEN ET AL.
The table reports the time-series averages of portfolios’ excess returns, beta, book-to-market ratio
(ln(BM)), and market equity (ln(ME)) for the July 1997–June 2001 period.
At the end of June of each year t, all Information Technology stocks on the NYSE, AMEX, and
NASDAQ are grouped into 16 portfolios based on their market capitalization (ln(ME)) and book-to-
market ratio (ln(BM)). The size breakpoints are the first (25%), second (50%) and third (75%) quartiles
of ln(ME) of all information technology stocks on the NYSE, AMEX, and NASDAQ. Similarly, the
book-to-market breakpoints are the first, second, and third quartiles of ln(BM) of all information
technology stocks on these exchanges. Stocks with negative BM are excluded so that ln(BM) is
calculable. A portfolio’s ln(ME), ln(BM), and beta are the value-weighted averages of stocks’ ln(ME),
ln(BM), and beta, respectively. The weights are determined by stocks’ ln(ME).
Big, ME2, ME3, and Small represent the biggest size quartile, second largest size quartile, third
largest size quartile and smallest size quartile, respectively. High, BM2, BM3, and Low represent the
highest book-to-market quartile, second highest book-to-market quartile, third highest book-to-market
quartile, and the lowest high book-to-market quartile, respectively. Big-High means the size – book-
to-market portfolio whose stocks belong to the biggest size quartile and the highest book-to-market
quartile.
Annual stock returns for the period July of year t to June of year t + 1 are adjusted for monthly stock
price appreciation plus reinvestment of monthly dividends and the compounding effect of dividends paid
on reinvested dividends. A portfolio’ return is the value-weighted average of the returns on individual
stocks in the portfolio. A portfolio’s excess return is the difference between the portfolio’ return and
the 1-year U.S. Treasury bill rate in the same testing year. A testing year is from July of year t to June
of year t + 1.
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 61
4. METHOD
This section consists of two parts. Section 4.1. describes the multivariate model that
is traditionally used in tests of the size and book-to-market effects. Section 4.2.
describes the three-factor model, proposed by Fama and French (1993).
Detailed portfolio formation procedures for each model are also discussed in
each part.
pair of the three variables would lead to spurious results and render any inference
from the results meaningless. To overcome this problem, various combinations of
these three factors will be used in regressions. Fama and French (1992) use the
Fama and MacBeth (1973) cross-sectional approach to estimate the coefficients
and residuals in the model. This paper improves the Fama and MacBeth procedure,
which use Ordinary Least Squares (OLS) method, by adopting the Newey-West
(1987) technique. The OLS methodology will give best, linear, unbiased estimates
of the coefficients of a model only when the residuals have constant variance, (i.e.
homoskedasticity), and are independent of each other (i.e. not autocorrelated).
If the residuals show patterns of heteroskedasticity (non-constant variance), the
variances, and therefore the standard deviations of the coefficients, will be
biased. Because the Newey-West (1987) procedure accounts for the problems
of hetoroskedasticity and autocorrelation, it is more efficient and produces more
accurate t-statistics than the OLS.
where R p is the return on portfolio p; |ln(MEi )| is the absolute value of the natural
logarithm of stock i’s market equity; n is the number of stocks in portfolio p. We
obtain the value-weighted market equity, value-weighted book-to-market ratio,
and value-weighted market beta for each portfolio in a similar way: A portfolio’s
market equity is calculated as:
n
|ln(MEi )| × ln(MEi )
ln(MEp ) = i=1 (9)
|ln(MEi )|
A portfolio’s book-to-market ratio is calculated as:
n
|ln(MEi )| × ln(BMi )
ln(BMp ) = i=1n (10)
i=1 |ln(MEi )|
where RM(t) is the value-weighted monthly returns on the market portfolio that
consists of all IT stocks; RF(t) is the risk free rate; SMB(t) is the difference between
returns on small stocks and returns on big stocks; HML(t) is the difference between
returns on high book-to-market stocks and low book-to-market stocks.
cover the full July 1990–June 2001 period as described in the Data section. Whereas
annual data is used for the multivariate model, we use monthly data for the three-
factor model. In Section 4.1. we argue that annual data is better than monthly data
because the effect of seasonal change in returns will be minimized when annual
data are used. This is true if the multivariate model is used. However, for the three-
factor model, it is not necessary to eliminate the seasonal effect. The reason is
that the explanatory variables, SMB, HML, and RM–RF are themselves returns,
in which the seasonal effect is inherent. The dependent variable is also return and
as such contains in itself the seasonal effect. When the time-series regressions of
the dependent variable on the explanatory variables are run, the seasonal effects in
two sides of the three-factor model will be canceled out. Thus, the period (month
versus annual) used to measure returns is appropriate and also has the advantage
of providing more observations, which would lead to more precise estimates of
the coefficients for the model.
The 16 size-BE/ME portfolios are formed much like the six size-BE/ME portfolios
discussed earlier. In June of each year t all Information Technology stocks are sorted
by size (ln(ME) and (independently) by book-to-market equity (ln(BM)). For the
size sort, ME is measured at the end of June. For the book-to-market sort, ME is
market equity at the end of December of t − 1, and BE is book common equity for
the fiscal year ending in calendar year t − 1. The first (25%), second (50%), and
third (75%) quartiles of stocks’ ln(ME) and ln(BM) are used as the breakpoints for
the allocation of stocks into four size quartiles and four book-to-market quartiles.
Sixteen portfolios are formed from the intersections of the size and BM quartiles.
Value-weighted monthly returns on the portfolios from July of year t to June of
year t + 1 are used as dependent variables in the time-series regressions.
Table 4 shows the descriptive statistics for SMB, HML, RM–RF, and the excess
returns on sixteen size – BM portfolios from July 1990 to June 2001. The time-
series mean of SMB is 1.12%, with a t-statistic of 1.63 (less than 1.98). This
suggests that over the sample period, small stocks do not outperform big stocks.
However, the returns on high book-to-market stocks do exceed those on low
book-to-market stocks over the same period. The time-series mean for HML is
a positive 1.48%, with a t-statistic of 2.37. The excess returns on the market
portfolio (RM–RF) are significant in the negative territory (a mean of −2.71%,
with a t-statistic of −3.04). Except for the portfolios in the smallest-size quartile
and the ME3-BM2 portfolio, other portfolios have mean excess returns that are
distinguishable from zero. These portfolios have negative mean excess returns
with −2.53% being the highest (the ME3-High portfolio). This fact, coupled with
a negative mean excess return on the market portfolio (RM–RF), indicates that IT
stocks on average produce negative returns over the July 1990–June 2001 period.
It is also clear that the mean excess return on each of the size – book-to-market
portfolio is very low in practical terms. For instance, the ME3-High portfolio,
which has the highest mean excess return compared with other size – BM portfolios,
produces a mean excess return of −2.53% per month, which is translated into
approximately −30.36% per year. That is not to mention the ME2-Low portfolio,
which loses nearly 51.12% (−4.26 × 12) per year. The lowest mean excess return
on a size – book-to-market portfolio reported by Fama and French (1993) was
0.32% per month or 3.84% per year. Even after accounting for slight differences
in research design, the positive 3.84% is still significantly different to the negative
−30.36%. Table 4 also reports low correlations between SMB, HML, and RM–RF.
The correlation coefficients between SMB and HML, SMB and RM–RF, and HML
68 QUANG-NGOC NGUYEN ET AL.
Table 4. Descriptive Statistics for the Dependent and Explanatory Variables for
the Three-Factor Model: July 1990–June 2001.
Panel A. Descriptive Statistics for SMB, HML, and RM–RF
The table reports the descriptive statistics for the returns that mimic size (SMB), book-to-market
(HML), and market (RM–RF) factors (panel A), and the excess returns on the 16 size – book-to-market
portfolios (panel B) for the July 1990–June 2001 period. SMB, HML, and RM–RF are used as the
explanatory variables, and the portfolios’ excess returns are used as the dependent variable for the
three-factor model.
Panel A shows the descriptive statistics for SMB, HML, and RM–RF. SMB (small minus big) is the
difference between the returns on small-stock and big-stock portfolios with about the same weighted
average book-to-market equity. HML (high minus low) is the difference between the returns on high
and low book-to-market equity portfolios with about the same weighted average size. RM is the value-
weighted monthly% return on the stocks in the 16 size-BM portfolios. RF is the three-month Treasury
bill rate, observed at the beginning of each month.
Panel B shows the descriptive statistics for the excess returns on the 16 size – book-to-market
portfolios. Three-month Treasury bill is used to calculate the excess returns. The portfolios are formed
as follows. At the end of June of each year t, all Information Technology stocks on the NYSE,
AMEX, and NASDAQ are grouped into 16 portfolios based on their market capitalization (ln(ME)) and
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 69
Table 4. (Continued )
book-to-market ratio (ln(BM)). ME is common equity, which is denominated in millions of U.S.
dollars and is measured at the end of June of year t. BM is the ratio of book equity to market
equity, with book equity measured at fiscal yearend falling in year t − 1, and market equity measured
at the end of December of year t − 1.The ME breakpoints are the first (25%), second (50%), and
third (75%) quartiles of the stocks’ ln(ME). Similarly, the BM breakpoints are the first, second,
and third quartiles of the stocks’ ln(BM). Stocks with negative BM are excluded so that ln(BM) is
calculable.
The reported means (in percentage) are the time-series averages of the monthly values of the
corresponding variables. The sample period ranges from July 1990 to June 2001.
A value with a t-statistic being less than –1.98 or greater than 1.98 is considered significant at the
5% level.
and RM–RF are 0.34, 0.36, and −0.13, respectively. All these figures are well
below 0.50.
5. RESULTS
This section contains three parts. Section 5.1. presents the empirical results of the
Fama and MacBeth (1973) regressions on the multivariate model. Section 5.2.
presents the empirical results of the time-series regressions on the three-factor
model. Section 5.3. compares the behavior of information technology stocks’ price
before and after the technology stock crash in April 2000.
This section consists of two sub-sections. Section 5.1.1. discusses the results
from the FM (Fama & MacBeth) regressions on individual stocks. Section 5.1.2.
discusses the results from the FM regressions on the size – book-to-market
portfolios.
Table 5. Results from the (Improved) Fama and MacBeth Regressions with
Stock Data: July 1997–June 2001.
Models Variables Intercept (%) Coefficients (%)
a1t ln(ME)it a2t ln(BM)it a3t it
The table reports the time-series average coefficient slopes and the associated t-statistics (in brackets)
from the (improved) Fama and MacBeth regressions of stock returns on size, book-to-market ratio, and
beta for the July 1997–June 2001 period. The reported intercepts and slopes are in percentage.
The multivariate model above is a general model. There is a total of seven models where the explanatory
variables are combined in various ways. For instance, model 1 includes all three variables, ln(ME),
ln(BM), and beta while model 2 consists of only two variables, ln(BM) and beta. The aim is to separate
the effect of the multicollinearity problem.
ME is common equity, which is denominated in millions of U.S. dollars and is measured at the end
of June of year t. BM is the ratio of book equity to market equity, with book equity measured at fiscal
yearend falling in year t − 1, and market equity measured at the end of December of year t − 1. ln(ME)
and ln(BM) are the natural logarithms of ME and BM, respectively. Beta is the average of monthly
stock betas during the period July of year t to June of year t + 1. Monthly stock beta is calculated for
a 5-year (60-month) time period, ending in the current month. If less history is available, beta will be
calculated for as few as 24 months. Month-end closing prices (including dividends) are used in the
calculation. A testing year (for each yearly cross-sectional regression) is from July of year t to June of
year t + 1. There are four yearly cross-sectional regressions for the July 1997–June 2001 period.
∗ Significant at the 0.10 level.
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 71
the regressions, the average slope on ln(BM) is 7.51%, with a t-statistic of 2.68
(Model 7). The magnitude of the book-to-market coefficient increases to 8.42%
(t-statistic of 3.08) when beta is added (Model 2). This change might be caused
by a slight negative correlation coefficient of −0.08 (Table 7) between book-to-
market ratio and beta. In the regressions where ln(ME), ln(BM), and beta are
included (Model 1), the slope on ln(BM) decreases to 5.03%, with a t-statistic
of 2.56.
The slope coefficient on ln(BM) is indistinguishable from zero in Model 4, where
stock returns are regressed on the size and book-to-market variables. This lack of
explanatory power from ln(BM) is due to the fact that ln(BM) is highly negatively
correlated with ln(ME). The time-series average of the correlation coefficient
between ln(BM) and ln(ME) is reported at −0.37 (Table 7). Nevertheless, Model 7
provides strong evidence that the book-to-market ratio is statistically related to
variation in stock returns. The strong explanatory power of book-to-market ratio
sharply contrasts the weak relationship between size, ln(ME), beta, and stock
returns. In univariate regressions where either beta or size is included, the t-statistics
for the slopes on the size and beta variables are just −1.45 (Model 6) and 1.27
(Model 5), respectively. Models 1, 2, 3, and 4 say that size and beta have no power
to explain average returns in FM regressions that use various combinations of beta
with size and book-to-market ratio.
It should be noticed that our interpretation of the results in Table 5 might be
affected by an extremely low adjusted R-squared for each model; the values of
the adjusted-R2 never exceeds 0.05 across models and across years. Thus, the
book-to-market variable is significant in explaining stock returns but it is not the
only significant variable; some unknown variable(s) are more significant and more
powerful in explaining variation in stock returns.
Table 6. Results from the (Improved) Fama and MacBeth Regressions with
Portfolio Data: July 1997–June 2001.
Models Variables Intercept (%) Coefficients (%)
a1t ln(ME)pt a2t ln(BM)pt a 3t b pt
The table reports the time-series average coefficient slopes and the associated t-statistics (in brackets)
from the (improved) Fama and MacBeth regressions of stock returns on size, book-to-market ratio, and
beta for the July 1997–June 2001 period. The reported intercepts and slopes are in percentage.
R pt − R ft = a 0t + a 1t ln(ME)pt + a 2t ln(BM)pt + a 3t pt + pt
The multivariate model above is a general model. There is a total of seven models where the explanatory
variables are combined in various ways. For instance, model 1 includes all three variables, ln(ME),
ln(BM), and beta while model 2 consists of only two variables, ln(BM) and beta. The aim is to separate
the effect of the multicollinearity problem.
ME is common equity, which is denominated in millions of U.S. dollars and is measured at the end
of June of year t. BM is the ratio of book equity to market equity, with book equity measured at fiscal
yearend falling in year t − 1, and market equity measured at the end of December of year t − 1. ln(ME)
and ln(BM) are the natural logarithms of ME and BM, respectively. Beta is the average of monthly
stock betas during the period July of year t to June of year t + 1. Monthly stock beta is calculated for
a 5-year (60-month) time period, ending in the current month. If less history is available, beta will be
calculated for as few as 24 months. Month-end closing prices (including dividends) are used in the
calculation. A testing year (for each yearly cross-sectional regression) is from July of year t to June of
year t + 1. There are four yearly cross-sectional regressions for the July 1997–June 2001 period.
Portfolios are reformed based on stocks’ market equity, ln(ME), and book-to-market ratio, ln(BM)
at the end of June of each year t. Portfolios’ ln(ME), ln(BM), and beta are the value-weighted averages
of individual stocks’ ln(ME), ln(BM), and beta, respectively. The weights are determined by stocks’
ln(ME).
∗ Significant at the 0.10 level.
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 73
slope on ln(ME) is only −1.04 (Model 6). With the presence of other explanatory
variables, such as ln(BM) and beta, the average slopes on the firm size variable
are even smaller and have smaller t-statistics. The risk premiums on  in Models
1, 2, 3, and 5 are interestingly also negative. This could indicate that portfolios
with lower  earn higher returns than portfolios with higher . Nevertheless, the
evidence on the negative relationship between portfolios’  and returns is not
conclusive because all the premiums on the beta variable are indistinguishable
from 0. The largest t-statistic (in absolute value) for the premium on  is −2.14,
which is insignificant at the 0.10 level.
5.1.3. Comparison between the FM Regressions on Individual Stocks and the FM
Regressions on the Size – Book-to-Market Portfolios
The book-to-market effect exists regardless of whether portfolio data or stock data
are used. In addition, the magnitudes of this effect are similar in both cases. When
portfolio data are used, the slope on ln(BM) in Model 7 is 7.42%, with a t-statistic
of 2.77 (Table 6). When stock data are used, the slope on ln(BM) in Model 7 is
7.51%, with a t-statistic of 2.68 (Table 5). These results confirm the significance of
the book-to-market ratio in explaining stock returns. Although the slope on ln(ME)
and beta in Table 5 is different from the slope on ln(ME) and beta in Table 6, both
tables report that size and beta are not related to stock returns. Instead, both tables
suggest that ln(BM) is the only variable that explains variation in stock returns.
Table 7 shows that when stocks are formed into size – book-to-market portfolios,
the correlation between the size variable and the book-to-market variable is
reduced to −0.08 (−0.37 if stock data are used). However, the achieved low
correlation between size and book-to-market ratio comes at the expense of the
correlation between size and beta, which is increased to 0.75. The change in
the correlation coefficients between ln(ME), ln(BM), and beta due to alternative
uses of stock data and portfolio data is one of the reasons for the differing
slopes on ln(ME) and beta in Table 5 and Table 6. Nevertheless, the advantage of
sorting stocks into portfolios is that the return-generating process is more correctly
specified by the multivariate model, which assumes portfolio size, book-to-market
ratio, and beta are the three most important factors that explain variation in portfolio
returns. This is indicated by the fact that the values of R2 in regressions using
portfolio data are higher than the values of R2 obtained in regressions using stock
data, with the (average) values of adjusted R2 in the cross-sectional regressions
using portfolio data are as high as 0.20.
5.1.4. Comparison with Fama and French’s (1992) results
Similar to Fama and French’s (1992) paper, this paper finds no relationship between
beta and stock returns. Whether beta is used alone or in various combinations with
size and book-to-market ratio, the slopes on beta is indistinguishable from 0. This
74 QUANG-NGOC NGUYEN ET AL.
Stocks data
ln(ME) 1
ln(BM) −0.37 1
Beta 0.13 −0.08 1
Portfolios data
ln(ME) 1
ln(BM) −0.08 1
Beta 0.75 −0.24 1
The table reports the time-series averages of the correlation coefficients between ln(ME), ln(BM), and
beta for individual stocks as well as for the size – book-to-market portfolios. ME is common equity,
which is denominated in millions of U.S. dollars and is measured at the end of June of year t. BM is
the ratio of book equity to market equity, with book equity measured at fiscal yearend falling in year
t − 1, and market equity measured at the end of December of year t − 1. ln(ME) and ln(BM) are the
natural logarithms of ME and BM, respectively. Beta is the average of monthly stock betas during the
period July of year t to June of year t + 1. Monthly stock beta is calculated for a 5-year (60-month)
time period, ending in the current month. If less history is available, beta will be calculated for as few
as 24 months. Month-end closing prices (including dividends) are used in the calculation.
Portfolios are reformed at the end of June of each year t, using size and book-to-market quartile
breakpoints. The size breakpoints are the first (25%), second (50%) and third (75%) quartiles of ln(ME)
of all information technology stocks on the NYSE, AMEX, and NASDAQ. Similarly, the book-to-
market breakpoints are the first, second, and third quartiles of ln(BM) of all information technology
stocks on these exchanges. Stocks with negative BM are excluded so that ln(BM) is calculable. A
portfolio’s ln(ME), ln(BM), and beta are the value-weighted averages of stocks’ ln(ME), ln(BM), and
beta, respectively. The weights are determined by stocks’ ln(ME).
result confirms the weak explanatory power of beta reported by many financial
economists (e.g. Barber & Lyon, 1997b; Davis, 1994; Jegadeesh, 1992). However,
the results in this paper provide a different picture to those found by Fama and
French (1992). First, Fama and French (1992) find that the book-to-market effect
is robust at the 5% level of significance while the book-to-market effect in this
study is only significant at the 10% level. Secondly, the combination of size and
book-to-market ratio captures variation in stock returns, whereas this paper finds
no relationship between size and stock returns regardless of whether size is alone
or in combination with book-to-market ratio.
This section contains six sub-Sections 5.2.1., 5.2.2. and 5.2.3. reports the results
on the one-factor model, with HML, SMB, and RM–RF in turn plays the role of the
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 75
explanatory variable. 5.2.4. discusses the results on the two-factor model, in which
SMB and HML are the two explanatory variables. 5.2.5. discusses the results on the
three-factor model, in which SMB, HML, and RM–RF are the three explanatory
variables. And finally, 5.2.6. considers the use of the three-month Treasury bill
rate as an alternative to the one-month Treasury bill rate.
77
78
Table 8. (Continued )
BM 3 −0.40 −0.63 −0.89 −0.81
(t-statistic) (−3.35) (−4.83) (−6.35) (−5.32)
Low −0.52 −0.89 −1.23 −1.20
(t-statistic) (−4.10) (−6.37) (−9.06) (−7.06)
Adjusted R2 = 0.56
Adjusted R2 = 0.44
The table reports the Seemingly Unrelated Regressions (SUR) results for the following model:
At the end of June of each year t, all Information Technology stocks on the NYSE, AMEX, and NASDAQ are grouped into 16 portfolios based on their
5.2.5. Regressions with Three Explanatory Variables: SMB, HML, and RM–RF
Table 9 shows the SUR estimates of the intercepts and coefficients on SMB, HML,
and RM for the 16 size – book-to-market portfolios. With the exception of three
portfolios, the Small-High, Small-Low, and ME2-Low portfolios, the intercepts for
the remaining 13 size – book-to-market portfolios are indistinguishable from zero
and have statistically insignificant t-statistics ranging from −1.40 (ME3–BM2)
to 1.10 (Small–BM3). Nearly all the slope coefficients on SMB are significant
and wit only one exception have t-statistics that lie outside the [−1.98, 1.98]
interval. The one exception is the ME2-BM3 portfolio, which returns a t-statistic
of −1.84. This quantity, nevertheless, is very close to −1.98. More interestingly,
the slope on SMB appears related to size. In three of the four book-to-market
quartiles (the exception is the second highest book-to-market quartile (BM2))
the slope on SMB increases monotonically from bigger- to smaller-size quartiles.
The explanatory power of HML in stock returns is somewhat weaker than that
of SMB. Six out of the 16 size – book-to-market portfolios return t-statistics
are within the [−1.98, 1.98] interval. The weak explanatory power, which is in
contrast with a strong relationship between stock returns and HML where HML
is the only explanatory variable, is due to the interaction between SMB, HML,
and RM–RF. Table 4 reports that the correlation coefficients between HML and
SMB, and HML and RM–RF are 0.34 and −0.13, respectively. Therefore, the
apparently weak power of HML in the three-factor model does not invalidate
itself as a significant variable that captures stock returns. This is evidenced in the
one-factor model where HML is the only explanatory variable.
As further evidence for the role of HML in mimicking book-to-market ratio, the
slope on HML tends to decrease from higher- to smaller- BM portfolios. The two
portfolios that do not follow this tendency are ME2-BM2 and ME3-BM3. The
slope on RM–RF (or the market factor) are all more than six standard errors from
zero. The market factor, either in combination with the size and book-to-market
factors or standing alone, strongly captures variation in portfolio returns. The
adjusted R2 for the three-factor model scores an impressive 0.83, much higher
than those reported for the one-factor models (0.16 for HML, 0.30 for SMB, and
0.56 for RM–RF, Table 9). This is a strong indication that the three-factor model
is a good model that describes the relationship between size, book-to-market
ratio, beta, and stock returns well.
5.2.6. Does the Use of the 3-Month Treasury Bill Rate Matter?
In Section 4.2., it was argued that the use of the 3-month Treasury bill rate
(USTBL3M) instead of the 1-month Treasury bill rate (USTBL1M) would lead
to the same inference as if the 1-month rate was used. This section empirically
improves that statement. As the data for the 1-month Treasury bill rate on
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 81
Intercepts
High −0.44 0.18 −0.34 1.72
(t-statistic) (−0.88) (0.50) (−0.95) (2.91)
BM 2 0.22 0.10 −1.50 0.98
(t-statistic) (0.80) (0.38) (−1.40) (1.75)
BM 3 0.18 0.14 −0.03 0.93
(t-statistic) (0.90) (0.48) (−0.06) (1.10)
Low −0.01 −0.65 −0.14 2.49
(t-statistic) (−0.05) (−2.20) (−0.37) (2.36)
SMB coefficients
High −0.51 −0.11 0.18 0.46
(t-statistic) (−7.32) (−2.09) (3.60) (5.54)
BM 2 −0.38 −0.30 1.59 0.65
(t-statistic) (−9.91) (−7.67) (10.53) (8.20)
BM 3 −0.45 −0.08 0.39 0.74
(t-statistic) (−15.66) (−1.84) (6.82) (6.13)
Low −0.44 0.09 0.72 1.44
(t-statistic) (−12.03) (2.07) (13.57) (9.71)
HML coefficients
High 0.22 −0.03 −0.08 −0.08
(t-statistic) (3.09) (−0.50) (−1.56) (−0.96)
BM 2 0.03 0.00 2.57 −0.37
(t-statistic) (0.72) (0.03) (16.43) (−4.49)
BM 3 0.00 −0.20 −0.45 −0.48
(t-statistic) (0.02) (−4.65) (−7.51) (−3.88)
Low −0.10 −0.43 −0.80 −0.91
(t-statistic) (−2.64) (−9.98) (−14.45) (−5.89)
RM–RF coefficients
High 0.89 0.99 0.84 0.80
(t-statistic) (17.39) (26.72) (22.86) (13.14)
BM 2 0.88 0.96 1.48 0.90
(t-statistic) (31.85) (34.15) (13.43) (15.82)
BM 3 0.97 1.04 1.07 0.81
(t-statistic) (46.26) (33.87) (25.76) (9.26)
Low 1.02 1.13 1.05 0.70
(t-statistic) (38.28) (37.30) (27.14) (6.53)
Adjusted R2 0.83
82 QUANG-NGOC NGUYEN ET AL.
Table 9. (Continued )
The table reports the Seemingly Unrelated Regressions (SUR) results for the three-factor model:
At the end of June of each year t, all Information Technology stocks on the NYSE, AMEX, and
NASDAQ are grouped into 16 portfolios based on their market capitalization (ln(ME)) and book-to-
market ratio (ln(BM)). ME is common equity, which is denominated in millions of U.S. dollars and
is measured at the end of June of year t. BM is the ratio of book equity to market equity, with book
equity measured at fiscal yearend falling in year t − 1, and market equity measured at the end of
December of year t − 1. ln(ME) and ln(BM) are the natural logarithms of ME and BM, respectively.
The ME breakpoints are the first, second, and third quartiles of the stocks’ ln(ME). Similarly, the BM
breakpoints are the first, second, and third quartiles of the stocks’ ln(BM). Big, ME2, ME3, and Small
represent the biggest size quartile, second largest size quartile, third largest size quartile and smallest
size quartile, respectively. High, BM2, BM3, and Low represent the highest book-to-market quartile,
second highest book-to-market quartile, third highest book-to-market quartile, and the lowest high
book-to-market quartile, respectively.
SMB (small minus big) is the difference between the returns on small-stock and big-stock portfolios
with about the same weighted average book-to-market equity. HML (high minus low) is the difference
between the returns on high and low book-to-market equity portfolios with about the same weighted
average size. RM is the value-weighted monthly% return on the stocks in the 16 size-BM portfolios.
RF is the three-month Treasury bill rate, observed at the beginning of each month.
DataStream is not available after May 1996, the sample period used for the
comparison between the effects of these two rates is from July 1990 to May 1996.
Employing the SUR procedure, two time-series regressions of the size – book-
to-market portfolios’ excess returns on the size (SMB), book-to-market (HML),
and market factor (RM–RF) are undertaken. The size factor, SMB, and the book-to-
market factor (HML) are the same in these two regressions. Only the excess returns
on the size – book-to-market portfolios and the market portfolio are different: in
one regression, the 1-month rate is used; in the other, the 3-month rate is used.
Table 10 reports the results of the two regressions.
It is clear that the coefficients on SMB, HML, and RM–RF produced by the use
of the 1-month rate are nearly identical to those produced by the use of the 3-month
rate. If there is any difference, the difference is only 0.01%. The t-test for the mean
difference between the 1-month rate SMB coefficients and the 3-month rate SMB
coefficients returns a t-statistic of 0.42, which is not significant at the 5% level to
reject the null hypothesis of no difference. Similarly, the results of the t-tests for
the mean difference between the 1-month rate HML coefficients and the 3-month
rate HML coefficients, and between the 1-month rate RM–RF coefficients and
the 3-month rate RM–RF coefficients show no sign of difference (in a statistical
sense). Thus, the use of the 3-month Treasury bill rate in the preceding sections
would not invalidate the conclusion that the three-factor model is well specified
for information technology stocks on the NYSE, AMEX, and NASDAQ.
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 83
Table 10. SUR Results Using 1-Month and 3-Month Treasury Bill Rates:
July 1990–May 1996.
Portfolios SMB Coefficients HML Coefficients RM–RF Coefficients
USTBL1M USTBL3M USTBL1M USTBL3M USTBL1M USTBL3M
The table reports the results for the Seemingly Unrelated Regressions of the size – book-to-market
portfolios’ excess returns on SMB, HML, and RM–RF, with the excess returns are measured in one of
the two ways: (1) using the 1-month treasury bill rate (USTBL1M), and (2) using the 3-month treasury
bill rate (USTBL3M).
SMB (small minus big) is the difference between the returns on small-stock and big-stock portfolios with
about the same weighted average book-to-market equity. HML (high minus low) is the difference be-
tween the returns on high and low book-to-market equity portfolios with about the same weighted aver-
age size. RM is the value-weighted monthly% return on the stocks in the 16 size-BM portfolios. RF is the
three-month Treasury bill rate or the one-month Treasury bill rate, depending on each of the approach.
The portfolios are formed as follows. At the end of June of each year t, all Information
Technology stocks on the NYSE, AMEX, and NASDAQ are grouped into 16 portfolios based on
their market capitalization (ln(ME)) and book-to-market ratio (ln(BM)). ME is common equity, which
is denominated in millions of U.S. dollars and is measured at the end of June of year t. BM is the ratio
of book equity to market equity, with book equity measured at fiscal yearend falling in year t − 1, and
market equity measured at the end of December of year t − 1.The ME breakpoints are the first (25%),
second (50%), and third (75%) quartiles of the stocks’ ln(ME). Similarly, the BM breakpoints are the
first, second, and third quartiles of the stocks’ ln(BM). Stocks with negative BM are excluded so that
ln(BM) is calculable.
∗ Tests for mean difference between two-matched samples.
∗∗ At the 5% level of significance.
84 QUANG-NGOC NGUYEN ET AL.
This section investigates how well the three-factor model describes the size and
book-to-market effects in two sub-periods, the period before the technology stock
crash in April 2000 (the pre-crash period), and the period after the crash (the
post-crash period). Section 5.3.1. reports the descriptive statistics for the pre-crash
and after-crash variables. Section 5.3.2. presents the regressions results for each
of the period.
5.3.1. Inputs for the Three-Factor Model in the Periods Before (July 1990–March
2000) and After the Technology Crash (May 2000–June 2001)
Table 11 presents the descriptive statistics for the SMB, HML, and RM–RF
variables (Panel A) as well as for the size – book-to-market portfolios’ excess
returns (Panel B) for the July 1990–March 2000 period (pre-crash period) and
the May 2000–June 2001 period (post-crash period). The month when the
crash occurs, April, is excluded to separate the immediate effect of the crash.
86 QUANG-NGOC NGUYEN ET AL.
Table 11 reports that most of the (dependent and independent) variables have
the post-crash means (i.e. the means for the May 2000–June 2001 period) that
are indistinguishable from zero. The post-crash means for SMB, HML, and
RM–RF are 0.57%, 2.92%, and –8.09%, with t-statistics of 0.21, 1.31, and –1.59,
respectively (Panel A). The post-crash means of the excess returns on the size –
book-to-market portfolios also have t-statistics that are less than 2 standard errors
from 0, except for the Big-Low portfolio (Panel B).
In contrast, (in Panel B) 11 out of the 16 size – book-to-market portfolios
have pre-crash mean excess returns (i.e. the mean excess returns for the July
1990–March 2000) that are statistically different from zero. The five portfolios
whose pre-crash mean excess returns are indistinguishable from zero are the
ME3-BM2 portfolio and those from the smallest – size quartile (the bottom four
portfolios). The pre-crash mean for the market excess returns (RM–RF) is also
different from zero (t-statistic of −2.44) (Panel A). Although SMB and HML
have the pre-crash means that are not different from zero at the 5% level of
significance, they are on average significant at the 10% level. The t-statistics for
the pre-crash means of these two variables, SMB and HML, are 1.88 and 1.92,
respectively.
The table reports the results for the Cross-Section Weighting regressions of the size – book-to-market
portfolios’ excess returns on SMB, HML, and RM–RF, using pre-crash data and post-crash data. Pre-
crash data are the data for the July 1990–March 2000 period. Post-crash data are the data for the May
2000–June 2001 period. The term “crash” implies the technology crash, which occurred in April 2000.
The model for the regressions is:
R(t) − RF(t) = a + b[RM(t) − RF(t)] + sSMB(t) + hHML(t) + e(t)
SMB (small minus big) is the difference between the returns on small-stock and big-stock portfolios
with about the same weighted average book-to-market equity. HML (high minus low) is the difference
between the returns on high and low book-to-market equity portfolios with about the same weighted
average size. RM is the value-weighted monthly% return on the stocks in the 16 size-BM portfolios.
RF is the three-month Treasury bill rate, observed at the beginning of each month.
The portfolios are formed as follows. At the end of June of each year t, all Information
Technology stocks on the NYSE, AMEX, and NASDAQ are grouped into 16 portfolios based on
their market capitalization (ln(ME)) and book-to-market ratio (ln(BM)). ME is common equity, which
is denominated in millions of U.S. dollars and is measured at the end of June of year t. BM is the ratio
of book equity to market equity, with book equity measured at fiscal yearend falling in year t − 1, and
market equity measured at the end of December of year t − 1.The ME breakpoints are the first (25%),
second (50%), and third (75%) quartiles of the stocks’ ln(ME). Similarly, the BM breakpoints are the
first, second, and third quartiles of the stocks’ ln(BM). Stocks with negative BM are excluded so that
ln(BM) is calculable.
∗ Tests for mean difference between two-matched samples.
∗∗ At the 5% level of significance.
88 QUANG-NGOC NGUYEN ET AL.
6. CONCLUSION
This paper explores the relationship between size, book-to-market, beta, and
expected stock returns in the Information Technology sector. In particular, the
paper uses two models to test this relationship. The first model assumes stock
returns are related to the direct measure of size, book-to-market, and beta (the
multivariate model). The second model assumes stock returns are related to returns
on the three portfolios that are designed to mimic the size, book-to-market, and
beta factors (the three-factor model).
The risk-return tests using the multivariate model indicate that the book-to-
market effect is present in stock returns. However, this effect only exists at
the 10% level of significance in contrast to the more significant relationship
documented by Fama and French (1992) for all non-financial stocks on the NYSE,
AMEX, and NASDAQ over the 1963–1990 period. Thus, the book-to-market
effect for IT stocks, though present, is not robust over the period 1997–2001.
Except for discovering a weak book-to-market effect, the risk-return tests using
the multivariate model find no relationship between size, beta and stock returns.
Although Fama and French (1992) also document a weak explanatory power of
beta, they find that size is significant and that size and book-to-market ratio in
combination absorb variation in stock returns. The different results documented in
Fama and French’s (1992) paper and this paper might be due to different research
designs employed by the two papers. However, research design seems to be playing
a trivial part since the results from the FM regressions on individual stocks and on
portfolios (in this paper) both lead to similar conclusions. Thus, the relationship
between size, book-to-market, beta and stock returns in IT stocks appears to be
different from that previously observed in non-financial stocks.
The risk-return tests using the three-factor model show that expected returns
are strongly related to the size, book-to-market, and market factors. Unlike the
multivariate model where size and book-to-market factors are directly measured
from the market capitalization and book-equity to market-equity ratio of a stock,
the size and book-to-market factors in the three-factor model are represented
by the returns on the portfolios that are designed to mimic stocks’ market
capitalisation and book-equity to market-equity ratio. The results show that the
three-factor model works very well for IT stocks over the July 1990–June 2001
period as well as the sub-periods before (July 1990–March 2000) and after the
technology crash (May 2001–June 2001) in April 2000.
Why do the multivariate model and the three-factor model produce different
results? One possible explanation is that the multivariate model uses annual
data while the three-factor model uses monthly data. We argue that the use of
monthly data instead of annual data in the three-factor model is not a serious
problem that could lead to any spurious conclusions. Examining whether the
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 89
three-factor model produces different results due to the alternative uses of annual
data and monthly data could provide a basis for future research. Another possible
reason for the differing results produced by the two models is that they are
simply not comparable. SMB, HML, and RM–RF might not be good proxies for
direct measurements of size (ln(ME)), book-to-market ratio (ln(BM)), and beta
respectively. In other words, the mimicking returns for the size (SMB), book-to-
market (HML), and market factors (RM–RF), might not be good indicators of
movements in the characteristics (ln(ME), ln(BM), and beta) of stocks.
The sub-period results (the periods before and after the technology crash in
April 2000) show that the nature of the relationship between stock returns, size,
book-to-market, and market factors, or the magnitude of the size, book-to-market,
and market premiums, is on average unchanged for both of the sub-periods. The
coefficients on SMB, HML, and RM–RF for the July 1990–March 2000 (the pre-
crash coefficients) are not statistically different from those for the May 2000–June
2001 period. This is particularly interesting since this finding is inconsistent with
the view that the technology stock crash in April 2000 was a correction of stock
prices. As it has been known, the factors in the three-factor model are the “returns”
on some specially designed portfolios. It is possible that the behavior of a return on a
particular portfolio has been changed after the technology crash (in fact it has as the
statistics for SMB, HML, RM–RF, and the size – book-to-market portfolios’ excess
returns have changed from the pre-crash period to post-crash period, Table 12);
but because all the returns have changed by the same magnitude, the relationship
between these returns is unchanged. This conjecture, if true, would offer great
support for the three-factor model, since it could work over any time frame and be
used for tests of structural change (i.e. relationship change) after an event.
This paper makes three important contributions to theory as well as practice.
First, it provides more evidence on the risk-return relationship in the information
technology sector. The evidence in this paper brings a step towards finding a
satisfactory explanation for the deviation from the CAPM. Second, the paper
documents in detail the behavior of IT stock prices before and after the technology
stock crash in April 2000. This information could be used as a reference for
investment decisions. And finally, the paper opens a new direction for future
research in the asset-pricing field: looking for (the risk-return) evidence in
individual sectors rather than in markets as a whole.
NOTES
1. “Old Economy” stocks are those other than IT stocks.
2. Note that a firm might not have beta for a certain period but might have it at a different
time. We account for this by including the firm when its beta is available and excluding it
90 QUANG-NGOC NGUYEN ET AL.
when its beta is unavailable. This is done subject to whether the firm fits into our portfolio
selection procedure (to be described in the methodology section).
3. A testing year is from July of year t to June of year t + 1.
4. It makes more sense to establish a relationship between SMB, HML, and systematic
risk than to find a relationship between size, book-to-market equity, and systematic risk.
This is due to the fact that size and book-to-market equity are company-specific factors.
5. Data on the 1-month Treasury bill rate in DataStream is not available after May 1996.
6. Note that the SUR and OLS techniques produce identical estimators but different
t-statistics for these estimators. As this section is only concerned with the differences in the
pre-crash and after-crash coefficients, the use of OLS is as significant as the use of SUR.
REFERENCES
Aldous, D. (1989). Probability approximations via the Poisson clumping heuristic. New York: Springer.
Alford, A. W., Jones, J. J., & Zmijewski, M. E. (1994). Extension and violation of the statutory SEC
form 10-K filing requirements. Journal of Accounting and Economics, 17, 229–254.
Banz, R. W. (1981). The relationship between return and market value of common stocks. Journal of
Financial Economics, 9, 3–18.
Barber, B. M., & Lyon, J. D. (1997a). Detecting long-run abnormal stock returns: The empirical power
and specification of test statistics. Journal of Financial Economics, 43, 341–372.
Barber, B. M., & Lyon, J. D. (1997b). Firm size, book-to-market ratio, and security returns: A holdout
sample of financial firms. Journal of Finance, 52, 875–883.
Black, F. (1993). Beta and return. Journal of Portfolio Management, 20, 8–18.
Black, F., Jensen, M. C., & Scholes, M. (1972). The capital asset pricing model: Some empirical tests.
In: M. Jensen (Ed.), Studies in the Theory of Capital Market. New York: Praeger.
Capaul, C., Rowley, I., & Sharpe, W. F. (1993). International value and growth stock returns. Financial
Analysts Journal, 49, 27–36.
Chan, K. C., & Chen, N. F. (1988). An unconditional asset-pricing test and the role of firm size as an
instrumental variable for risk. Journal of Finance, 43, 309–325.
Chan, L. K. C., Hamao, Y., & Lakonishok, J. (1991). Fundamentals and stock returns in Japan. Journal
of Finance, 46, 1739–1789.
Chui, A. C. W., & Wei, K. C. J. (1998). Book-to-market, firm size, and the turn-of-the-year effect:
Evidence from Pacific-Basin emerging markets. Pacific-Basin Finance Journal, 6, 275–293.
Daniel, K., & Titman, S. (1997). Evidence on the characteristics of cross sectional variation in stock
returns. Journal of Finance, 52, 1–33.
Davis, J. (1994). The cross-section of realized stock returns: The pre-COMPUSTAT evidence. Journal
of Finance, 49, 1579–1593.
Davis, J. L., Fama, E. F., & French, K. R. (2000). Characteristics, covariances, and average returns:
1929 to 1997. Journal of Finance, 55, 389–406.
DeBondt, W. F. M., & Thaler, R. H. (1987). Further evidence on investor overreaction and stock market
seasonality. Journal of Finance, 42, 557–581.
Dimson, E. (1979). Risk measurement when shares are subject to infrequent trading. Journal of
Financial Economics, 7, 197–216.
Fama, E. F., & French, K. R. (1992). The cross-section of expected stock returns. Journal of Finance,
47, 427–465.
Size and Book-to-Market Effects in the Returns on Information Technology Stocks 91
Fama, E. F., & French, K. R. (1993). Common risk factors in the returns on stocks and bonds. Journal
of Financial Economics, 33, 3–56.
Fama, E. F., & French, K. R. (1995). Size and book-to-market factors in earnings and returns. Journal
of Finance, 50, 131–155.
Fama, E. F., & French, K. R. (1996). Multifactor explanations of asset pricing anomalies. Journal of
Finance, 51, 55–84.
Fama, E. F., & French, K. R. (1997). Industry costs of equity. Journal of Financial Economics, 43,
153–193.
Fama, E. F., & MacBeth, J. D. (1973). Risk, return, and equilibrium: Empirical tests. The Journal of
Political Economy, 81, 607–636.
Franklin, C. (2000, September). Incubating real options: Can option pricing theory help investors make
sense of the business models and valuations of internet incubators? Working Paper, University
of Stirling.
Haugen, R. (1995). The new finance: The case against efficient markets. Englewood Cliffs, NJ: Prentice-
Hall.
Jegadeesh, N. (1992). Does market risk really explain the size effect. Journal of Financial and
Quantitative Analysis, 27, 337–351.
Keim, D. B. (1983). Size-related anomalies and stock return seasonality: Further empirical evidence.
Journal of Financial Economics, 12, 13–32.
Kothari, S. P., Shanken, J., & Sloan, R. (1995). Another look at the cross-section of expected stock
returns. Journal of Finance, 50, 185–224.
Lakonishok, J., Shleifer, A., & Vishny, R. W. (1994). Contrarian investment, extrapolation, and risk.
Journal of Finance, 49, 1541–1578.
Lintner, J. (1965). The valuation of risk assets and the selection of risky investments in stock portfolios
and capital budgets. Review of Economics and Statistics, 47, 13–37.
Lo, A. W., & MacKinlay, A. C. (1990a). When are contrarian profits due to stock market overreaction?
Review of Financial Studies, 3, 175–205.
Lo, A. W., & MacKinlay, A. C. (1990b). Data-snooping biases in tests of financial asset pricing models.
Review of Financial Studies, 3, 431–467.
MacKinlay, A. C. (1995). Multifactor models do not explain deviations from the CAPM. Journal of
Financial Economics, 38, 3–28.
Mech, T. (1993). Portfolio return autocorrelation. Journal of Financial Economics, 34, 307–344.
Newey, W., & West, K. (1987). A simple, heteroskedastic and autocorrelation consistent covariance
matrix. Econometrica, 55, 703–708.
Rajgopal, S., Kotha, S., & Venkatachalam, M. (2000, January). The relevance of web traffic for internet
stock prices. Research Paper, Stanford University.
Rosenberg, B., Reid, K., & Lanstein, R. (1985). Persuasive evidence of market inefficiency. Journal of
Portfolio Management, 11, 9–17.
Sadorsky, P. (2003). The macroeconomic determinants of technology stock price volatility. Review of
Financial Economics, 12, 191–205.
Scholes, M., & Williams, J. (1977). Estimating betas from nonsynchronous data. Journal of Financial
Economics, 5, 309–327.
Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk.
Journal of Finance, 19, 425–442.
Trueman, B., Wong, F., & Zhang, X. (2001). Back to basics: Forecasting the revenues of internet firms.
Review of Accounting Studies, 6, 305–329.
IMPLIED VOLATILITIES
AND AUDITOR REPUTATION:
THE ANDERSEN CASE
ABSTRACT
Audits are a means of reducing the information asymmetry between managers
and investors. If the quality of the audit is in question, outside investors may
face a larger informational disadvantage. We test the hypothesis that this
informational disadvantage is manifested in the implied volatilities associated
with the equity options of the audited firms. We find that volatilities increased
for Andersen audited firms relative to firms audited by other Big Five
accounting firms. This finding is consistent with the view that auditors help
lessen the information asymmetry problem and that some of this reduction is
accomplished by auditor reputation.
1. INTRODUCTION
The reputation of a firm’s auditor is widely seen as an important determinant
in signaling audit quality to investors. In literature, the importance of auditors
and their reputation is shown by Watts and Zimmerman (1983). This importance
has subsequently been examined and discussed in both finance and accounting
literature. In the financial literature, much of the research into auditor reputation has
focused on IPO underpricing and the prestige of the firms’ auditor. The accounting
Research in Finance
Research in Finance, Volume 21, 93–111
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21004-8
93
94 JONATHAN M. GODBEY AND JAMES W. MAHAR
literature shows a consistent view, but from a different perspective, focusing on the
relationship between audit pricing and reputation as well as event-studies showing
market reaction as a function of auditor reputation. A common explanation for
much of this literature is that auditor reputation serves as a means of reducing
information asymmetries between investors and the firm’s management.
The Enron-related troubles that befell Arthur Andersen (Andersen) provide a
unique opportunity to test the importance of auditor reputation. Over the course
of approximately six months, Andersen went from being a highly respected firm
to being a symbol of corporate malfeasance. The importance of the decline of
Andersen on client stock prices has been examined previously by Chaney and
Philipich (C&P, 2002) who report that firms audited by Andersen experienced
statistically significant price declines “surrounding dates on which Andersen’s
audit procedures and independence were under severe scrutiny.” (p. 1221) There
are two possible explanations for the market-adjusted price decline: (1) a decrease
in expected cash flow of Andersen audited clients, that is, a change in the mean of
the cash flow distribution; and (2) an increase in risk stemming from either higher
information asymmetries due to poor audit quality that would allow mangers to
hide important information from investors, or an increased likelihood of incurring
the costs of changing auditors.
One way to gauge whether investors believe that the risk of Andersen audited
firms had increased is to examine the volatility implied by the pricing of equity
options. Merton (1973), Latane and Rendlemand (1976), and Hull and White
(1987) document that the implied volatility of a particular option represents the
anticipated risk over the life of the option. Information asymmetries have also
been shown to play an important role in the option pricing. For example Patell and
Wolfson (1981) show that the implied volatility on equity options drops following
firm’s earnings releases. As pointed out by Merton (1987) and later discussed by
Bellalah and Aboura (2002), as the amount of, or quality of, available information
decreases, the implied volatility, or risk, increases, which in turn increases the
required rate of return and decreases the stock price, raising the cost of capital.
In this paper, we extend the findings of C&P (2002) by examining how option
market participants use auditor reputation in their pricing of equity options. If
auditor reputation affects the informational asymmetry between investors and the
firm, then the events that undermined Andersen’s reputation should be associated
with an increase in the implied volatility of the options of Andersen audited clients.
If auditor reputation is unimportant, then there should be no change in the implied
volatilities relative to that of firms audited by other Big Five accounting firms.
For both the short and long time periods, our results demonstrate that the implied
volatility at S&P 500 firms audited by Andersen increased significantly relative
to firms audited by other Big Five accounting firms during the time period in
Implied Volatilities and Auditor Reputation 95
which Enron and Andersen actions became publicly known. Specific events where
the average implied volatility increased include: (1) the restatement of Enron’s
financial statements and the announcement that the SEC was investigating the case
(November 8, 2001); (2) the Congressional testimony of Andersen CEO Joseph
Berardino (December 12, 2001); (3) Andersen’s admission to the shredding of
documents while under SEC investigation (January 10, 2002); (4) the dual events of
the release of the Powers Report and Andersen’s hiring of former Federal Reserve
Chairman Paul Volcker (February 2 and 3, 2002); and (5) the admission by David
Duncan that he had ordered Enron documents to be destroyed. These findings are
consistent with a view that the quality of auditor reputation is an important factor
in reducing the information asymmetry problem.
The remainder of the paper is organized as follows: Section 2 provides a review
of existing research and an introduction to implied volatilities, Section 3 provides
a description of our data and methodology, Section 4 examines the results of the
tests, and Section 5 gives a summary and conclusion.
2. LITERATURE REVIEW
Auditor reputation is an important means of signaling the quality of a client’s
financial reporting and thereby reducing the information asymmetry problem.
This area of research was established by Watts and Zimmerman (1983), but
has been examined extensively in both financial and accounting literature. The
common thread through this auditor reputation literature is that investors use
auditor reputation as a signal of quality that reduces informational asymmetry.
In financial literature, much of the auditor reputation work has been done within
an IPO framework. Titman and Trueman (1986) find that audit quality can help
explain IPO underpricing. Similarly, Balvers, McDonald and Miller (1988), and
Beatty (1989) find less IPO underpricing when a Big 8 auditor is used. More
recent work by Clarkson and Simunic (1994) find that there is less underpricing
of Canadian IPOs when a high quality auditor is used.
The importance of auditor reputation has also been shown by looking at the
pricing of audits. A partial list of those who have examined the relationship between
auditor quality and their pricing includes Francis and Simon (1987), Beatty (1993),
Dye (1993), and Craswell, Francis and Taylor (1995) who all find that audit firms
with higher perceived reputation, price their audits higher than those with lower
perceived quality. This pricing may be partially due to the fact that larger (and more
highly perceived) auditors have more assets (deeper pockets) and can therefore
better withstand potential liabilities in the event of a lawsuit. However, Datar and
Alles (1999) show that audit firms understand the importance of their reputation
96 JONATHAN M. GODBEY AND JAMES W. MAHAR
and strive to maintain their reputational capital even when legal liability is reduced
by their financial assets.
Auditor reputation has also been investigated with event-study type tests that
measure the market reaction for clients audited by the same auditor. Teoh and
Wong (1993) find that the perceived audit quality influences the market’s response
to earnings announcements. Franz and Crawford (1998) document that audit clients
that are not involved with auditor related litigation lose market value as a result
of litigation brought against their auditor which is seen as damaging the auditor’s
reputation. Moreover, C&P (2002) report that firms audited by Arthur Andersen
experienced market-adjusted stock price declines as the reputation of Andersen
fell.
Our paper extends the work of C&P (2002) by investigating the cause of the
stock price decline that they find for Andersen audited firms following the Enron-
related events. There are two possible explanations for any abnormal decline in
stock price: (1) an unexpected change in expected future cash flows available to
shareholders; or (2) an increase in risk. Since the stock price represents the present
value of the future cash flows, both the expected future cash flows (the mean) as
well as the risk of the cash flows are important. Moreover, by focusing on stock
price alone, it is impossible to determine which of these factors is present. The
implied volatility on the other hand, is not directly influenced by the mean of
the expected cash flows, but rather is the option market’s consensus forecast of
expected volatility of the underlying stock over the life of the option.
The significance of this distinction can be demonstrated by examining the
distribution of expected cash flows. Even if operational cash flows are held
constant, a decline in audit quality may reduce the expected cash flows available
to shareholders due to an increased probability of over-reported earnings or
under-reported liabilities. Both possibilities would lower the expected cash flows
available to shareholders. In addition, sufficient damage to the auditor’s reputation
would necessitate a costly change in auditors. This change reduces the cash flow
available to shareholders. Therefore, it is possible that shareholders face a reduction
in expected cash flows as well as an increase in the risk of cash flows. This
differentiation, which can only be theorized by using stock price data, can be
documented with implied volatility data. Thus, two important distinctions exist
between our analysis and that of C&P (2002). First, we examine the implied
volatility associated with equity option contracts rather than stock returns. This
allows us to determine if option traders use auditor reputation in their pricing
decisions. Second, and more importantly, we are able to partially disentangle
risk from the expected cash flow consequences of Andersen’s loss of prestige,
which gives us a better representation of the impact of the auditor’s reputation on
informational asymmetries.
Implied Volatilities and Auditor Reputation 97
The expected volatility of a given asset cannot be directly observed in the market.
However, given the price of an option and the price of the underlying asset, one may
use an iterative process to calculate the expected volatility implied by a particular
option-pricing model. This implied volatility is thus the market’s expectation of
the future volatility of the underlying asset, in this case the stock price. In this
paper, the implied volatility for each stock is calculated from at-the-money options
with at least 30 days to expiration using the Black-Scholes Option Pricing Model
(BSOPM) (Black & Scholes, 1973) adjusted for dividends (Merton, 1973). This
model is as follows.
C = (e−DT )S 0 N(d 1 ) − Xe−rT N(d 2 ) (1)
where: d1 = [ln((e−DT )S 0 /X) + (r +2 /2)T]/( sqrt(T)); d2 = d1 − ( sqrt(T));
S0 = stock price at time zero; X = strike price; T = time to expiration; r = risk-
free rate; = volatility of the underlying asset; D = Annualized dividend yield;
N(x) = the cumulative probability distribution function for a variable that is
normally distributed with a mean of zero and a standard deviation of 1.0.
Merton (1973), Latane and Rendlemand (1976), and Hull and White (1987) show
that the implied volatility of a particular option represents the mean anticipated
daily volatility, that is the anticipated risk, over the life of the option. Canina and
Figlewski (1993) summarize much of the research on the informational content
of implied volatilities and document the “volatility-smile” whereby the implied
volatility increases as the time to maturity gets small. Dumas, Fleming and Whaley
(1998) confirm this finding and show that volatilities are not constant across time.
The impact of events, both scheduled and unscheduled, on implied volatilities
has been the topic of much research. Patell and Wolfson (1981) as well as Isakov
and Perignon (2001) examine changes in implied volatility around scheduled
earnings announcements and find that implied volatility drops once uncertainty
is resolved. French and Dubofsky (1986) study the effect of unexpected stock
splits on the implied volatility. Finally, Levy and Yoder (1993) examine the impact
of mergers and acquisitions on implied volatilities. Each of these finds that the
option market quickly incorporates the new information and the implied volatilities
change accordingly.
options with at least 30 days to expiration. The 30-day constraint limits the effect
of a “volatility-smile” influencing the findings. At-the-money options are used
because they are the most liquid and therefore less likely to be influenced by stale
data problems brought about by infrequent trading. This infrequent trading could
introduce a bias if the stock, which is more frequently traded than the option,
falls in value after the last recorded trade of the option contract. This would be
a problem since the implied volatility would be computed with an option trade
that occurred before the stock fell in value. As C&P (2002) provide evidence that
Andersen audited clients fell in value, this systematic price drop could lead to
higher implied volatilities. However, by using liquid at-the-money options, the
stale trading problem is reduced. Additionally, the implied volatility is calculated
with using the average of the bid and ask quotes which further reduces any stale
pricing problem.
Each firm’s auditor information was drawn from the firms’ websites as of
October 16, 2002 when the news of Enron’s problems became known. The
breakdown of firms by auditors and sectors is provided in Table 1. For consistency
purposes, only firms audited by one of the Big Five accounting firms (Andersen,
Ernst and Young, Deloitte and Touche, KPMG, and PriceWaterhouseCoopers)
were studied. This resulted in a final sample size of 495 firms audited by a Big
Five accounting firm.
If auditor reputation is important in reducing the information asymmetry
problem, then the Enron-Andersen-related events of late 2001 and early 2002
Consumer Discretionary 12 14 24 9 26 2 87
Consumer Staples 5 7 5 7 10 0 34
Energy 5 2 5 4 8 0 24
Financials 9 13 24 10 18 1 75
Health Care 5 5 13 5 16 1 45
Industrials 13 13 17 8 16 1 68
Information Technology 6 11 24 13 26 0 80
Materials 10 4 5 3 15 0 37
Telecommunications 3 1 3 2 4 0 13
Utilities 14 11 1 0 11 0 37
Total 82 81 121 61 150 5 500
Note: AA = Arthur Andersen; DT = Deloitte & Touche; EY = Ernst & Young; PWC = Price
Waterhouse Coopers.
Sectors are from Standard and Poors website. Auditor information is from firms’ websites.
Implied Volatilities and Auditor Reputation 99
should result in higher implied volatilities for Andersen audited firms. If auditor
reputation is not important in reducing information asymmetries, then the implied
volatility ratios should be the same in the pre and post period. This leads to the
null hypothesis:
H0 . The implied volatility of Andersen audited firms increases as events likely
damaging to Andersen’s reputation occur.
Unlike traditional event studies, the information release effecting Enron and
Andersen was not completed at one time. From August 14, 2001 (when CEO
Jeffrey Skilling resigned) to June 15, 2002 (when Andersen was found guilty of
obstruction of justice and ordered to exit the audit business), news that could be
seen as damaging was released at irregular intervals. The news that was most
directly tied to Andersen was concentrated after November 8, 2001, when Enron
announced it would have to restate earnings.
Within the aforementioned time period, other specific dates that could be judged
as being detrimental to Andersen’s reputation include the dates investigated by
C&P as well as other dates after C&P’s February 4, 2002 cutoff. These include
November 8, 2001 when the SEC began the investigation into Enron, November
29, 2001 when the SEC expanded its investigation to include Andersen, December
12, 2001 when Andersen CEO Joe Berardino testified in front of Congress,
January 10, 2002 when Andersen admitted to shredding documents, January
24, 2002 when Andersen partner David Duncan refused to testify in front of
Congressional hearings, and the early February release of the Powers Report as
well as Andersen’s hiring of former Federal Reserve Chairman Paul Volcker in an
attempt to improve the auditor’s tarnished reputation. Additionally we investigated
several dates that were not examined by C&P 2002. These events include March
14, 2002 when the U.S. Justice Department charged Andersen with obstruction of
justice, March 27, 2002 when Andersen’s CEO Joseph Berardino resigned, and
April 9 when David Duncan admitted that he ordered the shredding of documents
(Table 2).
As time progressed and the seriousness of the problems became known, many
Andersen clients dropped Andersen as their auditor. The magnitude of these
departures introduces further problems in the investigation of event dates. The
extreme case of this is shown at the end of the event period (the August 30th
court mandated end of Andersen’s audit practices) when there were no firms still
being audited by Andersen. However prior to this definitive end point, departures
from Andersen were so pronounced that any analysis is prone to capture firm
specific risk factors. For instance, on June 15, 2002 a Houston jury found Andersen
guilty of obstruction of justice. This event was not examined because only 2 of
the original sample of 82 firms were still with Andersen. To avoid these sample
100 JONATHAN M. GODBEY AND JAMES W. MAHAR
August 14, 2001 Enron CEO Jeffrey Skilling resigns citing personal reasons
October 12, 2001 Arthur Andersen’s in-house lawyer emails Houston office regarding
document destruction policy
October 16, 2001 Enron reports $638 million loss and a $1.2 billion reduction in shareholder
equity
October 17, 2001 SEC requests information from Enron regarding losses
October 22, 2001 SEC begins official inquiry into Enron
October 31, 2001 Enron announces the SEC inquiry has been upgraded to a formal
investigation
November 8, 2001a Enron revises its last five years of financial statements, admitting net losses
of $586 million
Andersen receives a federal subpoena for Enron documents
November 9, 2001 Dynegy announces takeover of Enron for $8 billion
November 19, 2001 Enron again restates third quarter earnings and warns on debt repayments
November 28, 2001 Dynegy pulls out of takeover deal and Enron’s credit rating downgraded to
junk status
November 29, 2001 SEC investigation expanded to include Andersen
December 2, 2001 Enron files for Chapter 11 bankruptcy protection
December 12, 2001a Andersen CEO, Joseph Berardino, testifies before Congress and admits the
auditing firm may have violated securities laws
January 9, 2002 Justice department begins criminal investigation of Enron
January 10, 2002a Andersen acknowledges that it had destroyed documents
January 15, 2002 Andersen Partner, David Duncan, fired
January 22, 2002 Enron employee claims shredding
January 23, 2002 Kenneth Lay resigns as Enron’s CEO
FBI begins investigation into shredding
January 24, 2002 David Duncan refuses to testify in front of Congressional hearings
regarding shredding of “Enron-related documents”
January 25, 2002 Clifford Baxter, who had reportedly “clashed” with Jeff Skilling over
“accounting practices,” is found dead of an apparent suicide
February 2, 2002a Powers report released
February 3, 2002a Andersen hires former Fed Chairman, Paul Volcker
March 14, 2002 U.S. Justice Department charges Andersen with obstruction of justice for
allegedly shredding vital Enron documents
March 20, 2002 Andersen pleads not guilty to obstruction of justice charges
March 27, 2002 CEO Joseph Berardino announces his resignation
April 8, 2002 Andersen reports it will lay off 7,000 employees
April 9, 2002 Former Andersen auditor David Duncan enters a plea bargain with
prosecutors, admitting he ordered the shredding of incriminating Enron
documents
May 6, 2002 Andersen trial begins in a federal courtroom in Houston
June 15, 2002 Jurors find Andersen guilty of obstruction of justice charges and the firm
agrees to cease auditing public companies by August 31, 2002 as ordered
by the SEC
Implied Volatilities and Auditor Reputation 101
Table 2. (Continued )
Date Event Description
June 27, 2002 SEC issues restraining order to prevent Andersen from destroying any
documents relating to its telecommunication client WorldCom
August 16, 2002 Andersen has its Texas accounting license revoked
August 27, 2002 Andersen Worldwide agrees to pay a $60 million settlement to the plaintiffs
in the Enron class-action suit and to creditors of the bankrupt Energy firm
August 30, 2002 Andersen “closes its books,” as its public auditing business is terminated
October 16, 2002 Andersen receives maximum sentence of five years probation and a fine of
$500,000 for obstruction of justice relating to Enron scandal
size problems we report individual events only up to April 9, 2002 when former
Andersen partner David Duncan admitted that he had ordered the destruction of
Enron related documents. The rationale for halting the analysis at this point was
this was the last major event where at least 50% of the original sample was still
with Andersen (Table 3).
ratio to determine whether the events surrounding the Enron-Andersen case had
differential effects on the implied volatilities of Andersen’s clients versus clients of
other auditors. This methodology also accounts for differences in the audit firms’
initial reputation and changes in market-wide volatility.
Copeland, Poon and Stapleton (2000) model implied volatility as a function of
firm specific factors. To control for these firm specific factors, which include the
“duration of the firm’s profits” and the firm’s leverage, we examine the changes
in the IV ratio for the portfolio of Andersen audited S&P 500 firms. Additionally
we end our individual event analysis on April 9, 2002 when the size of the sample
drops to 39 firms (approximately 50% of the original sample).
The August 14, 2001 resignation of Jeffery Skilling has frequently been cited
as the beginning of the end for Enron. It would therefore be theoretically possible
that the markets could have inferred the troubles with Andersen from this initial
announcement. As a result, we create a baseline IV ratio to compare Andersen
audited S&P 500 firms to other Big Five audited S&P 500 firms. This ratio is
computed on a daily basis by dividing the average implied volatility for Andersen
audited firms by the average implied volatility for firms audited by other Big Five
firms. The resulting ratio is found for the 20 trading days preceding August 14. This
baseline is then compared to the daily IV ratio for the 20 trading days following
the release of the Powers report on February 2, 2002.
To insure that any changes in implied volatility were not directly caused by the
events of September 11, 2001, the procedure is repeated using the ten trading day
period immediately prior to the October 16, 2001 announcement by Enron where
they first publicly acknowledged large amounts of off balance sheet financing and
a major loss. These dates were chosen because by October 1, 2001, market-wide
volatility, as measured by the VIX index, had returned to September 10th
levels.
Once the daily implied volatility ratios were computed, the next step in the
investigation was to check the IV ratio data for normalcy. This was accomplished
with the Jarque-Bera test, which showed that the data was likely normally
distributed (p-value of 0.41). The actual investigation is performed in two ways.
First t-tests are used to examine the average implied volatility ratio before and after
the event windows. Following the t-tests, a regression analysis was performed
with a dummy independent variable to capture the marginal impact of the
events.
Black (1976) shows that implied volatilities are inversely related to stock price.
As C&P (2002) report that Andersen audit clients experienced negative abnormal
returns following the reputation damaging events associated with Andersen and
their role in the Enron collapse. Consequently, a stock price variable is used
to control for any potentially confounding results stemming from the negative
104 JONATHAN M. GODBEY AND JAMES W. MAHAR
relationship between the underlying stock price and implied volatility. We regress
the change in the implied volatility ratio from before each event to after each event.
That is we regress the model:
IV ratio = c + b1(Stock Price Ratio) + b2(Time Dummy) (2)
where: IV Ratio = Average Implied Volatility for Andersen audited firms divided
by Average Implied Volatility for firms audited by other Big Five audit firms; Price
Ratio = Average stock price for Andersen audited firms divided by average stock
price for firms audited by other Big Five audit firms; Time Dummy = 1 in the post
event time period.
Once the overall (both the 6 and 8 month event windows) impact of the events
has been determined, the same methodology is used to examine the impact of
the individual events that would be likely to have been interpreted as damaging
to Andersen’s reputation. Post event changes in the relative implied volatility are
examined for the days 0, +1 and days 0, +4 event windows. T-tests are performed
by comparing the daily implied volatility ratio to the average ratio of the 20
days preceding the event. As with the longer event-window tests, we inspect the
IV ratio
Call data 0.9007 0.9439 0.0432 0.0000
Put data 0.9050 0.9473 0.0423 0.0000
Panel B: 6-Month Window
IV ratio
Call data 0.8955 0.9335 0.0380 0.0000
Put data 0.9000 0.9361 0.0361 0.0000
Note: Changes in daily implied volatility ratio: The daily implied volatility ratio is computed by
dividing the average implied volatility of Andersen audited firms in the S&P 500 by the average
implied volatility of S&P 500 firms audited by other “Big Five” firms. The “Before” trading
dates in Panel A are based on the 30-day event window prior to Jeffrey Skilling’s August 14th,
2001 resignation. The “Before” dates in Panel B were selected to minimize the effects of the
September 11, 2001 attacks. The average implied volatility on S&P 500 stocks had returned to
its pre-attack levels by October 1, 2001. Therefore, the “Before” period in Panel B has only 11
trading days prior to the October 16, 2001 disclosure by Enron of a significant equity write-off
and net loss. The “After” trading dates in each Panel are composed of the 30-days following the
February release of the Powers report, which was critical of Andersen’s practices.
Implied Volatilities and Auditor Reputation 105
8-Month 6-Month
Note: The IV Ratio is the average implied volatility of Andersen audited firms divided the average
volatility of firms audited by other Big Five firms. The average price ratio is the average price
of Andersen audited firms divided the price of firms audited by other Big Five firms. The results
shown here are for all firms audited by Andersen as of October 16, 2001. The 8-month window
regression uses the twenty trading days before the August 14, 2001 resignation of Jeffrey Skilling
and the twenty trading days following the April 9, 2002 admission by David Duncan that he
had ordered documents to be destroyed. The 6-month window regression uses the ten trading
days before the October 16, 2001 Enron earnings announcement, and ten trading days following
the April 9, 2002 admission by David Duncan that he had ordered documents to be destroyed.
(t-Statistics are in parenthesis.)
106 JONATHAN M. GODBEY AND JAMES W. MAHAR
August 14, 2001: Enron CEO Call data 0.8949 0.8959 0.8887
Jeffrey Skilling resigns 0.0010 −0.0062
(0.4612) (0.1700)
Put data 0.8908 0.8964 0.8878
0.0056 −0.0030
0.3011 (0.3904)
October 16, 2001: Enron reports Call data 0.8662 0.8589 0.8628
loss −0.0073 −0.0034
(0.2483) (0.3102)
Put data 0.8676 0.8656 0.8644
−0.0020 −0.0032
0.3776 0.3784
November 8, 2001: Andersen Call data 0.8615 0.8754 0.8710
receives federal subpoena and 0.0139 0.0097
Enron restates earnings (0.0260) (0.0221)
Put data 0.8620 0.8762 0.8697
0.0142 0.0077
(0.0204) (0.0450)
November 29, 2001: SEC Call data 0.8687 0.8561 0.8616
investigation is expanded to −0.0126 −0.0071
include Arthur Andersen (0.0425) (0.0961)
Put data 0.8683 0.8531 0.8574
−0.0152 −0.0109
0.0429 0.0226
December 12, 2001: Joseph Call data 0.8691 0.8926 0.9040
Berardino appears in front of 0.0235 0.0349
Congress (0.0049) (0.0000)
Put data 0.8683 0.8927 0.9027
0.0244 0.0344
0.0053 0.0000
January 10, 2002: Enron admits Call data 0.8864 0.8889 0.8862
to shredding 0.0035 −0.0002
(0.4224) (0.4879)
Put data 0.8861 0.8889 0.8866
0.0028 0.0005
0.4072 0.4757
January 22, 2002: Enron Call data 0.8833 0.8854 0.8929
employee claims shredding 0.0021 0.0065
(January 23) FBI begins (0.4269) (0000)
investigation into shredding Put data 0.8839 0.8873 0.8928
0.0034 0.0089
0.3663 0.0817
Implied Volatilities and Auditor Reputation 107
Table 6. (Continued )
Date/Event Before (Days 0, 1) Days (0, 4)
(Days Difference Difference
(−20, −1) (p-Value) (p-Value)
January 24, 2002: David Duncan Call data 0.8831 0.8932 0.9159
refuses to testify 0.0101 0.0328
(0.1020) (0.0002)
Put data 0.8851 0.8932 0.9169
0.0081 0.0318
(0.1718) (0.0003)
February 4, 2002: Release of Call data 0.9031 0.9532 0.9470
Powers Report and Paul 0.0501 0.0439
Volcker hired (0.0068) (0.0022)
Put data 0.9000 0.9527 0.9443
0.0527 0.0443
0.0018 0.0001
March 14, 2002: Andersen Call data 0.9236 0.9105 0.9025
charged with obstruction of −0.0131 −0.0211
justice (0.1134) (0.0027)
Put data 0.9264 0.9160 0.9096
−0.0104 −0.0064
(0.1752) (0.0124)
March 27 2002: Andersen CEO Call data 0.9113 0.8862 0.8840
Joseph Berardino resigns −0.0251 −0.0273
(0.0452) (0.0027)
Put data 0.9156 0.8930 0.8907
−0.0226 −0.0249
(0.0539) (0.0037)
April 9, 2002: Former Andersen Call data 0.8940 0.9033 0.9099
auditor Duncan enters plea 0.0093 0.0159
bargain (0.1473) (0.0048)
Put data 0.8996 0.9152 0.9162
0.0156 0.0166
(0.0399) (0.0024)
While the long-event window results provide evidence that there was an increase
in implied volatilities at Andersen audited clients, they do not say specifically when
this increase occurred. As a result, we used the same methodology to examine
shorter event windows around the events that were hypothesized as being damaging
to Andersen’s reputation. The findings in this area, with both the t-tests as well
as the regression analysis, reaffirm that Andersen’s audit clients did experience
108
Table 7. Regressions Around Key Dates.
Panel A: Call Data
5. CONCLUSION
This paper was motivated by the opportunity to use the Enron and Andersen related
troubles to study the impact of auditor reputation on the information asymmetry
problem. Consistent with the findings of C&P (2002), who found that stock prices
of Andersen audited firms decreased over this time period, we studied how auditor
reputation impacts implied volatility at firms audited by Andersen.
Both long-term and short-term event-studies were used to examine the effects
on implied volatility, of events that were deemed as damaging to Andersen’s
reputation. The results of all of the tests yield strong evidence that the implied
volatilities at Andersen audited firms increased relative to the firms audited by
other Big Five accounting firms over the time period surrounding the events that
led to collapse of Andersen. Thus, the results of both tests are consistent with the
hypothesis that auditor reputation plays an important role in reducing information
asymmetries between investors and the audited firm.
By looking at the implied volatility of Andersen audited firms, rather than the
stock price, we are able to show that the stock price declines documented by C&P
(2002), are due, at least in part, to an increase in the risk of the cash flows. This is a
significant contribution because by looking at stock price alone, the decline could
be caused by any combination of a reduction of expected cash flows or an increase
in risk. While the implied volatility does not allow us to say anything about the
110 JONATHAN M. GODBEY AND JAMES W. MAHAR
level of cash flows, we can say definitively that the market expected the risk to
increase.
NOTE
1. In September 2003, the VIX was changed to track the implied volatility of S&P 500
firms.
ACKNOWLEDGMENTS
We would like to thank David Becher, Joseph Coate, Carol Fischer, Stephan Horan,
Lance Nail, Rodney Paul, Jeffrey Peterson, Kenneth Small, and participants at
the Southern Finance Association meetings in Charleston SC. Additionally we
are indebted to the able research assistance of Patricia Dean, Jill Simme, and
Christopher Zimmer. All errors are our own.
REFERENCES
Balvers, R., McDonald, B., & Miller, R. (1988). Underpricing of new issues and the choice of auditor
as a signal of investment banker reputation. The Accounting Review, 63, 605–622.
BBC News (2002, June 17). Andersen’s fall from grace. URL: http://news.bbc.co.uk/1/hi/business/
2049237.stm.
BBC News (2002, February 4). Timeline: Enron’s rise and fall. URL: http://news.bbc.co.uk/
2/hi/business/1759599.stm.
Beatty, R. (1989). Auditor reputation and the pricing of initial public offerings. The Accounting Review,
64, 693–709.
Beatty, R. (1993). The economic determinants of auditor compensation in the initial public offerings
market. Journal of Accounting Research, 31, 294–302.
Bellalah, M., & Aboura, S. (2002). Pricing options on leveraged equity under asymmetric information
(Working Paper).
Binder, J. J. (1985). Measuring the effects of regulation with stock price data. Rand Journal of
Economics, 16, 167–183.
Black, F. (1976). Studies of stock price volatility changes. Proceedings of the American Statistical
Association, 177–181.
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of Political
Economy, 81, 637–654.
Canina, L., & Figlewski, S. (1993). The informational content of implied volatility. Review of Financial
Studies, 6, 659–681.
Chaney, P., & Philipich, K. (2002). Shredded reputation: The cost of audit failure. Journal of Accounting
Research, 40, 1221–1244.
Implied Volatilities and Auditor Reputation 111
Clarkson, P., & Simunic, D. (1994). The association between audit quality, retained ownership, and firm
risk in U.S. vs. Canadian IPO markets. Journal of Accounting and Economics, 17, 207–228.
Copeland, L., Poon, S., & Stapleton, R. (2000). The implied volatility of options: A test using UK data.
Journal of Business Finance and Accounting, 27, 859–885.
Craswell, A. T., Francis, J. R., & Taylor, S. L. (1995). Auditor brand name reputation and industry
specialization. Journal of Accounting and Economics, 20, 297–322.
Datar, S., & Alles, M. (1999). The formation and role of reputation and litigation in the auditor-manager
relationship. Journal of Accounting, Auditing and Finance, 14, 401–428.
Dumas, B., Fleming, J., & Whaley, R. E. (1998). Implied volatility functions: Empirical test. Journal
of Finance, 53, 2059–2106.
Dye, R. (1993). Auditor status, legal liability, and auditor wealth. Journal of Political Economy,
887–914.
Francis, J., & Simon, D. (1987). A test of audit pricing in the small-client segment of the U.S. audit
market. The Accounting Review, 62, 145–157.
Franz, D. R., & Crawford, D. (1998). The impact of litigation against an audit firm on the market value
of nonlitigating clients. Journal of Accounting, Auditing and Finance, 13, 117–134.
French, D., & Dubofsky, D. (1986). Stock splits and implied stock price volatility. Journal of Portfolio
Management, 12, 55–59.
Houston Chronicle (2002, October 22). Timeline of Andersen events. URL: http://www.chron.com/cs/
CDA/story.hts/special/enron/1452911.
Hull, J., & White, A. (1987). Pricing of option assets with stochastic volatilities. Journal of Finance,
42, 281–300.
Isakov, D., & Perignon, C. (2001). Evolution of market uncertainty around earnings announcements.
Journal of Banking and Finance, 25, 1769–1789.
Latane, H. A., & Rendlemand, R. J., Jr. (1976). Standard deviations of stock price ratios implied in
option prices. The Journal of Finance, 31, 369–381.
Levy, H., & Yoder, J. (1993). The behavior of option implied standard deviations around merger and
acquisition announcements. Financial Review, 28, 261–272.
Merton, R. (1973). Theory of rational option pricing. Bell Journal of Economics and Management
Science, 4, 141–183.
Merton, R. (1987). A simple model of capital market equilibrium with incomplete information. Journal
of Finance, 42, 483–510.
Patell, J. M., & Wolfson, M. A. (1981). The ex ante and ex post price effects of quarterly earnings
announcements reflected in option and stock prices. Journal of Accounting Research, 19,
434–458.
Schwert, G. W. (1981). Using financial data to measure the effects of regulation. Journal of Law and
Economics, 121–158.
Teoh, S., & Wong, T. (1993). Perceived auditor quality and the earnings response coefficient. The
Accounting Review, 68, 346–366.
Titman, S., & Trueman, B. (1986). Information quality and the valuation of new issues. Journal of
Accounting and Economics, 8, 159–172.
Washington Post (2002, November 14). Timeline of Enron’s collapse. URL: http://www.
washintonpost.com/ac2/wp-dyn/A25624–2002Jan10?language=printer.
Watts, R., & Zimmerman, J. (1983). Agency problems, auditing, and the theory of the firm: Some
evidence. Journal of Law and Economics, 26, 613–633.
SECONDARY EQUITY OFFERINGS:
THE CASE OF INSTALLMENTS
RECEIPTS
ABSTRACT
This paper examines secondary equity offerings that were done in the
Canadian markets through “installment receipts” (IRs). Previous studies on
seasoned equity offerings tend to focus on the price reaction around the
announcement date. We extend the analysis to cover a longer period so that
the issues of liquidity effect and information asymmetry can be adequately
addressed. We also offer evidence to indicate that the use of IRs in secondary
offerings can reduce the liquidity impact in markets where market depth is
not as substantial as in the U.S.
1. INTRODUCTION
Past studies on block trading and secondary equity offerings have documented the
discounts that the sellers have to concede in order to execute the transactions.1
(See, for example, Holthausen et al. (1990) in the context of block trading, and
Hudson et al. (1993) in the context of secondary offerings.) While the two types of
transactions appear to belong to two distinct literatures (i.e. block trading belongs
to the microstructure literature whereas secondary issues are in the domain of
seasoned security offerings), there are, however, two common dominant themes –
information asymmetry and liquidity effect. Information asymmetry arises because
Research in Finance
Research in Finance, Volume 21, 113–133
© 2004 Published by Elsevier Ltd.
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21005-X
113
114 NARAT CHARUPAT AND C. SHERMAN CHEUNG
block sellers (in the case of block trades) and major shareholders (in the case of
secondary offerings) are traders with possibly superior information about the shares
that they sell. Liquidity effect arises if the demand for a security is not perfectly
elastic. Both factors can explain a decline in price as a result of the two types
of transactions. The major difference between the two factors is that the price
impact of the liquidity effect is temporary, whereas the price effect of information
asymmetry is permanent (Scholes, 1972).
In this paper, we study secondary equity offerings that were done through
“installment receipts” (IRs) in the Canadian markets. An IR is a security that
evidences the purchase of an underlying security on an installment basis. Typically,
buyers are allowed to pay for the underlying securities in two or three installment
payments. The payments are spread over a period of one to two years (more on
this in Section 2). Large secondary equity offerings in Canada are commonly done
through this method, instead of on a fully-paid basis.
We are interested in several issues in this paper. First, we would like to know
the role of an IR arrangement in secondary equity offerings. While IRs generally
cannot be issued on a public-market basis in the U.S. due to Securities & Exchange
Commission’s margin rules, they are quite common in Commonwealth countries
such as England, Australia, New Zealand and Canada.2 Yet there has been very
little theoretical and empirical research on this instrument. Second, to explain the
role of IRs, we trace the price behavior of stocks and, whenever possible, IRs
from the period before the announcements of the secondary offering through the
announcement dates and the issue dates, to the period after the issue dates. Previous
studies on seasoned equity offerings tend to focus on the pricing behavior around
the announcement dates and ignore the short-term pricing behavior around the
issue dates.3 As a result, the liquidity effect around the issue dates has not been
adequately addressed in the literature. We therefore offer a more comprehensive
empirical analysis of secondary offerings. Third, we will try to explain the cross-
sectional behavior around three critical dates – the announcement date, the issue
date and the post-issue period. In particular, we are interested in the pricing of the
issues and the short-term excess return from buying them.
The paper is organized as follows. In the next section, we discuss the institutional
features of IRs and the rationale for using them. Section 3 provides details of
our sample. Section 4 documents the price pattern of the stocks from the period
prior to the announcements of the offerings to post-issue period. It also reports
the magnitude of the underpricing of IRs and the short-term excess return from
buying them. Cross-section results are reported in Section 5. Section 6 examines
the post-issue performance again to address the liquidity effect. Section 7 provides
supporting evidence for the benefits of using IRs in secondary offerings. Finally,
Section 8 concludes the paper.
Secondary Equity Offerings 115
those who want to participate in the offers but cannot do so due to borrowing
constraints. These potential buyers include mutual funds and pension funds, which
commonly are prohibited from borrowing by their charters, and small investors,
who generally find it costly or impossible to obtain leverage. Under this rationale,
IRs are not redundant securities as their use will increase the demand for the
offers.6 The higher the embedded leverage, the more are the potential buyers IRs
will attract.
The increase in demand can help reduce the underpricing of seasoned/secondary
offerings caused by liquidity effects when securities are not perfectly substitutes
for one another. If securities’ demand is not perfectly elastic, investors will require
extra compensation in the form of underpricing for having to hold more of a
particular security than initially desired.7 The increase in demand helps to reduce
the underpricing by raising the markets’ aggregate risk-bearing ability (which is
an increasing function of the number of traders in the markets).8
A closely parallel literature on liquidity-induced underpricing is that of block
trades. A block trader performs similar functions to those of an underwriter in a
public offering; i.e. to facilitate the trade by locating counterparties. It has been
observed that seller-initiated block trades are generally transacted at prices below
both pre- and post-block prices. This suggests that the sellers have to make price
concessions to the buyers. More importantly, the fact that post-issue prices are
higher than transacted prices suggests that these concessions arise due to liquidity
reasons (e.g. to compensate for search cost).9
Our sample contains all secondary offerings of common shares that were done
through installment receipts on the Toronto Stock Exchange (TSE) or the Montreal
Exchange (ME) between 1990 and 1999. In total, there are 30 issues involving
shares of 29 companies.10
Table 1 provides summary statistics for our sample. Total issue values range
from C$ 125 million to C$ 1.8 billion, with a mean of C$ 667 million and a
median of C$ 606 million.11 In terms of the percentage of the outstanding shares,
these offerings account for between 6.67 and 83.26%, with a mean of 38.91% and
a median of 37.32%. These numbers suggest that the issuers in our sample are
major shareholders of their companies. The ratios between the number of shares
offered and the average daily trading volume are also reported. They range from
22 to 9,939, with a mean of 1,031 and a median of 610. Therefore, the offerings in
our sample represent a substantial supply of equity, both in terms of dollar values
and trading impact that the markets have to absorb.
Secondary Equity Offerings 117
the amounts of the 1st installments. The figures are based on dollars at the time of issues.
c The issue values and the number of offered shares do not include shares subsequently issued (if any)
period (from day −150 to day −101) prior to the announcement dates.
e Leverage offered is calculated by dividing the present value of future installment payments by the
sum of: (i) the initial payment; and (ii) the present value of future installment payments.
Table 1 also reports the issue values based only on their first installment
payments, which measures the immediate impact of the offerings in dollar value’s
terms. By using IRs, the issuers are able to reduce the immediate dollar impact by
an average of over 50%. In addition, the last row of Table 1 reports the degree of
leverage offered by the issuers. The degree of leverage for each issue is the ratio
between: (i) the present value of the issue’s future installment payments; and (ii)
the sum of that present value and the issue’s initial payment. In other words, it is
the percentage of the “full” price that investors do not have to pay up front.
In terms of issuers, six of the issues in our sample (20%) were sold by individual
shareholders, twenty one (70%) by corporate shareholders, and four (10%) by
either the provincial or federal governments in the privatization process. The
breakdown of the issues according to the years in which they were done is given
in Table 2.
Details of each secondary issue such as its size, its issuer(s), the number of
shares offered, the number of shares outstanding and the terms of the installment
payments, were obtained from the TSE Monthly Review and, when available, the
118 NARAT CHARUPAT AND C. SHERMAN CHEUNG
1990 0
1991 1
1992 6
1993 6
1994 2
1995 4
1996 2
1997 7
1998 2
1999 0
Total 30
issues’ prospectuses. Closing prices of IRs and their underlying shares, together
with market returns, as proxied by the TSE 300 value-weighted index, were
collected from the TSE Summary Information Database, Canada Stockwatch and
TCE Research Services. We used the Canadian Business and Current Affairs
Database to identify the announcement dates (AD) of the offerings. Since some
announcements may have occurred before the opening or after the end of the
trading on the announcement day, our announcement period consists of AD and
AD + 1. The listing dates (LD), on which IRs started aftermarket trading, were
readily available from the TSE Monthly Review.
The first step in our analysis is to document the price behavior of IRs and
their underlying stocks. Our analysis covers the period from 150 days before the
announcements to 10 days after issue dates. Previous studies tend to focus on
the pricing behavior around the announcement and ignore the behavior around the
issue date. If the liquidity effect is present, its effect will be temporary and price
recovery will occur after issue date. It is therefore important to examine the price
pattern around the issue date.
To examine the price pattern of the underlying shares, we calculate the shares’
raw and excess holding-period returns using closing prices for the following seven
periods. The first three periods are AD − 150 to AD − 101, AD − 100 to AD −
51; and AD − 50 to AD − 1. These three periods show the stock price movements
prior to the announcements. The fourth period is the announcement period (AD and
AD + 1). The fifth period is the “interim” period, which is from AD + 2 to LD − 1.
The sixth period is the listing dates of the receipts, LD, while the last period is the
“post-issuance” period, consisting of 10 trading days after LD. Returns for these
periods are then used to determine whether offering announcements are likely to be
Secondary Equity Offerings 119
made after share prices have substantially risen, whether the announcements create
excess returns on the stocks, and whether the price movements reverse themselves
afterwards.
Excess holding-period returns on security i over any interval j, XRij , are
computed using the “market-adjusted” procedure (see Brown & Warner, 1985),
whereby excess returns are the difference between raw returns and market returns.
That is,
XRij = R ij − M j ,
where Rij is security i’s raw returns over interval j and Mj is market returns over
the same interval.
To measure the magnitude of underpricing of the offerings, we compare IRs’
offer prices to their closing prices on the first trading days. This “offer-to-close”
measure represents the return that one would get if one buys an IR at its offer price
and sells it at the closing price of the first day on which it is listed. This measure
has been used in studies of seasoned/secondary offerings such as Smith (1977) and
Loderer et al. (1991), and particularly in studies of initial public offerings (IPO).
Our reason for using this measure is that each issue of IRs is an offering of a new
security since the issue has not been traded before. Hence, there is no prior closing
price to which to compare.
Note that because IRs are leveraged instruments, our measure of underpricing
will also reflect the embedded leverage effect. Therefore, this measure is
appropriate if one does not plan to hold the IRs until the next installment payment
date. An alternative measure is one that adjusts for the leverage effect. This measure
– which we term “adjusted offer-to-close return” – represents the return that one
would get if, in addition to acquiring an IR, one also makes a risk-free investment
in an amount equal to the present value of its future installment payments. In other
words, adjusted returns approximate returns in case the offerings were done on a
fully-paid basis.
In addition to offer-to-close returns, we also calculate unadjusted and adjusted
raw and excess holding-period returns from holding IRs during the period from
LD + 1 to LD + 10.
Table 3 presents raw and excess returns on the underlying shares and their
associated p-values. In the three 50-day periods prior to the announcements,
120 NARAT CHARUPAT AND C. SHERMAN CHEUNG
Table 3. Raw and Excess Returns (in Percentages) on the Underlying Shares
around the Announcements.
Interval of Trading Days Raw Returns Excess Returns
the average raw and excess returns are all positive. As we get closer to the
announcement dates, both the magnitude of average returns and the proportion of
sample with positive returns increase. Note that the all pre-announcement positive
raw returns are significant at least at the 5% level and there is a clear pattern that
offering announcements are made after sustained substantial run-ups in prices.
The announcements have a significant negative impact on prices. The average
raw and excess returns during the announcement period (AD and AD + 1) are
−2.49 and −2.67% respectively, both of which are statistically significant at the
1% level. More than three quarters of the sample have negative excess returns.
In the interim period (AD + 2 to LD − 1), the average excess return is small
and positive, but not significantly different from zero. Note that this result can be
affected by a selection bias since issuers that experience a substantial decline in
their stock prices during this period may delay or cancel their offerings. (See, for
example, Mikkelson & Partch, 1986.)
On the listing dates of the receipts, the underlying stocks have insignificant
negative excess returns. On the other hand, during the 10-day period after the
issuance, we find strong evidence of price recovery, the magnitude of which closely
matches that of the average price drop on the announcement dates. Altogether, from
the announcement period up to the post-issuance periods (i.e. AD to LD + 10),
the average excess return (not shown) is −0.75% and not significantly different
from zero.
The observed pattern of positive returns prior to the announcement dates
followed by negative returns at the announcements is consistent with the findings
of previous studies of secondary offerings. For example, Korajczyk et al.
report a 10.63% cumulative excess return over a period of 100 days prior to
Secondary Equity Offerings 121
the announcements, and a negative excess return of −2.49% for the two-day
announcement period. Other studies with similar findings include Asquith and
Mullins (1986) and Hudson et al. (1993).12
Secondary share sales after significant price increases are commonly explained
by information asymmetry. As argued by Myers and Majluf (1984), existing
shareholders know more about their firms’ true values than new investors do. If
they recognize that the shares are overvalued, they will attempt to take advantage
of it. Therefore, their decisions to sell their holdings are construed as a signal that
the stocks are overvalued, which implies that the observed price drops is caused
by a change in the market’s perception of the shares’ value.
Nevertheless, the observed price decline at announcements does not necessarily
imply a change in the market’ valuation of the shares. It is possible that at
announcements, investors anticipate that the offered shares will have to be
underpriced when issued to accommodate the liquidity shock. Therefore, these
investors will sell their holdings of the shares at the announcements in the hope
that they can buy them back at a lower price from the offering. This kind of liquidity
effect is well documented in Scholes’ (1972) study.
As a result, we cannot say for certain what causes the observed price behavior
around the announcement period. In actuality, it is likely that both the information
asymmetry and the anticipation of a liquidity shock contribute to the results. The
price behavior after the announcements provides a clue as to the relative importance
of the two factors. If the price drops at announcement are due to information,
then the drops should be permanent and not reverse themselves since the market’s
valuation of the shares has changed (Scholes, 1972). On the other hand, if the drops
are due to anticipation of a liquidity shock, then they will be temporary. Our findings
of the post-issue date recovery suggest that the majority of the announcement effect
in our sample is caused by the liquidity effect.
Table 4 reports the raw offer-to-close returns on IRs, both unadjusted and adjusted
for the leverage effect. The average unadjusted offer-to-close return is 6.08% and
is significant at the 1% level. Although the returns range from −16.25 to 22.69%,
only slightly over a quarter of our sample are negative and 13% of the sample are
zero. These figures suggest that, on average, investors will make a sizable profit if
they buy IRs at the offering and sell them at the end of the IRs’ first trading day.
However, that high return is achieved partly with the added risk from leverage.
Adjusted for the leverage effect, the average return reduces to 2.79%, which is
also statistically significant.
To put our results in perspective, we compare them to the findings of previous
studies. We are aware of only two studies that document offer-to-close returns
from seasoned and/or secondary offerings, both of which are conducted on fully-
paid issues in the U.S. markets. First, Smith (1977) reports a small but significant
average offer-to-close return of 0.82%, based on a sample of seasoned offerings
of firms listed on the New York Stock Exchange (NYSE) and the American
Stock Exchange (Amex) during the period from 1971 to 1975. Secondly, Loderer
et al. (1991) find significant average returns of 0.80% for Amex stocks and 1.94%
for NASDAQ stocks for the period from 1980 to 1984. However, they do not find
significant returns for NYSE stocks. In comparison to these two studies, our sample
exhibits higher offer-to-close returns, both adjusted and unadjusted.13 It appears,
therefore, that the offerings in our sample are more deeply underpriced than in
these previous studies. If, as our results in the previous subsection suggests, the
observed underpricing is due to liquidity shocks, then the deeper discounts may
be the result of the lack of depth in the Canadian markets compared to the U.S.
markets.
5. CROSS-SECTIONAL RESULTS
5.1. Announcement Effects
We investigate the possible factors that can help explain the observed price decline
at announcements. We do so by regressing the (excess) announcement returns of all
issues in our sample on: (i) issue size; (ii) the risk of the offered shares; (iii) stock
(excess) returns prior to the announcements, and a liquidity measure. Specifically,
the regression equation is
Ranni is the excess return on stock i during the announcement period. Proxies
for ISSUE SIZEi include FRACTIONi , which is the number of shares offered
Secondary Equity Offerings 123
divided by the number of shares outstanding for stock i; and VALUEi , which is
the dollar value of issue i, as calculated in Table 1. However, as will be discussed
later in Section 6, the two proxies are correlated. Therefore, they will not appear
in the same regression. We use VOLi , which is the standard deviation of stock
i’s returns during the period from AD − 100 to AD − 51, as our proxy for the
shares’ risk. PRERETi is the excess return on stock i over a period of 100 trading
days prior to the announcement.14 MULTIPLEi , is the ratio between the number
of shares offered and the average daily trading volume over the period from AD
− 150 to AD − 101. Strictly speaking, MULTIPLEi is a proxy for issue size.
However, since it is not correlated with the other two proxies, we treat it as an
independent liquidity measure.
Table 5 presents the results of our regressions. Each column represents
the resulting coefficients of the above regression equation. Each of the above
independent variables was individually run first and then all independent variables
were run together. None of the variables has significant explaining power. The
lack of a significant relationship between PRERETi and announcement period’s
returns is consistent with several previous studies of secondary offerings.15 It
also suggests that the offerings in our sample were, on average, not information-
related. To see this, note that according to information reasons, large price run-
ups before the offerings suggest that the issuers are attempting to sell overvalued
stocks. Hence, the higher the price run-ups, the more negative the impact of the
offering announcements will be. On the other hand, if the sales are not information-
related, no significant relationship should be observed, which is what our results
show. Note also that what we find here is consistent with the argument that we
make in Section 4.1 that the observed post-issue price recovery suggests that the
announcement effect is mainly due to liquidity reasons.
The coefficients for both proxies for issue size – FRACTIONi and VALUEi−
are insignificant, which is consistent with previous studies of seasoned/secondary
offerings.16 The coefficient of VOLi is negative, as predicted by both the
information and the liquidity reasons.17 The lack of significance is similar to the
results in Masulis and Korwar (1986), who calculate variances based on the 60-
day period prior to the announcements, but is in contrast to the results in Hudson
et al. (1993), who calculate variances based on the 250-day period (from day −500
to day −251) prior to the announcements. To make sure that our results are not
influenced by the choice of periods over which return deviation is measured, we
also ran the same regressions with deviation over AD − 150 to AD − 101. All of
the results are not significant. Hence, the lack of significance does not depend on
the choice of periods.
In sum, our regressions of the announcement effect on various independent
variables provide some support for the liquidity explanation, but none for the
124
Table 5. Regression Estimates of Announcement Period’s Excess Returns on Various Independent Variables.
Variablea (1) (2) (3) (4) (5) (6) (7)
Intercept −0.043 (0.024) −0.049 (0.017) −0.029 (0.008) −0.019 (0.227) −0.027 (0.006) −0.027 (0.444) −0.041 (0.095)
Note: p-Values are in parentheses. They are calculated based on standard errors that are computed using the heteroskedasticty consistent estimator of
White (1980).
a VOL is the standard deviation of stock returns during the period from AD − 100 to AD − 51. PRERET is the shares’ excess returns during the
i i
100-day period prior to the announcements. FRACTIONi is the number of shares offered divided by the number of shares outstanding. MULTIPLEi
is the ratio between the number of shares offered and the average daily trading volume over the period from AD − 150 to AD − 101. VALUEi is the
dollar value of issue i.
Secondary Equity Offerings 125
information reason. We also run the above regressions using raw Ranni and
PRERETi . The results lead to the same conclusion and therefore are not produced
here.
Next, we use a regression analysis to ascertain the liquidity effect by testing for
the relationship between the offer-to-close returns on IRs and the proxies for
issue size, liquidity and the risk of the offered shares (i.e. FRACTIONi , VALUEi ,
MULTIPLEi , and VOLi ). As mentioned in Section 2, underpricing is a form of
compensation to investors who have to hold more of a certain security than initially
desired. The larger the size of the offer and/or the riskier the offer, the greater is
the required compensation. As a result, the magnitude of underpricing should be
positively related to the proxies for issue size and risk.
Table 6 reports the results of the regressions. In Panel A, the dependent
variable is unadjusted returns. Again, each of the above independent variables was
individually run first and then all independent variables were run together. First,
we examine the regression results with each independent variable individually
run. Of the two proxies for issue size, VALUEi is significant at the 10% level.
The proxy for liquidity, MULTIPLEi , is significant at the 1% level. However, the
two coefficients have opposite signs to each other. The proxy for risk, VOLi , is
significant at the 10% level, but its negative sign contradicts the predictions of the
liquidity hypothesis (see Note 17). The general regression involving all variables
indicates that none of the variables is statistically significant.
To make sure that the above results are not caused by the leverage effect, we also
report in Panel B the results of the same regressions but with adjusted returns as
the dependent variable. While MULTIPLEi and VOLi continue to be significant
at the 1 and 10% levels respectively, VALUE no longer has explanation power
when regressions were run with only one independent variable. Again, the general
regression involving all variables indicates that none of the variables is statistically
significant.
Panel A
Intercept 0.046 (0.202) 0.013 (0.710) 0.063 (0.001) 0.113 (0.003) 0.085 (0.190) 0.072 (0.254)
FRACTIONi 0.023 (0.773) 0.071 (0.519)
VALUEi 0.000 (0.098) 0.000 (0.231)
Note: p-Values are in parentheses. They are calculated based on standard errors that are computed using the heteroskedasticty consistent estimator of
White (1980).
a VOL is the standard deviation of stock returns during the period from AD − 100 to AD − 51. FRACTION is the number of shares offered divided
i i
by the number of shares outstanding. MULTIPLEi is the ratio between the number of shares offered and the average daily trading volume over the
period from AD − 150 to AD − 101. VALUEi is the dollar value of issue i.
Secondary Equity Offerings 127
Table 7. Raw and Excess Returns from IRs and Underlying Stocks for the
10-Day Holding Period.
Types of Return Mean Median Lowest Highest %
(%) (%) (%) (%) Negative
Raw returns
Unadjusted IR return 5.25 (0.008) 2.38 −12.50 28.57 33
Adjusted IR return 2.51 (0.006) 0.85 −4.17 14.16 33
Underlying stock return 2.57 (0.003) 0.28 −3.42 13.08 37
Excess returns
Unadjusted IR return 4.86 (0.010) 3.07 −8.51 28.81 40
Adjusted IR return 2.12 (0.023) 2.00 −5.57 15.07 37
Underlying stock return 2.18 (0.013) 1.88 −4.07 13.98 33
Table 7 report raw and excess returns on IRs, on both unadjusted and adjusted
bases, for the 10-day holding period (not annualized). The raw unadjusted returns
have a mean of 5.25%, which is significant at the 1% level. This number is
slightly lower than the average offer-to-close return reported in Table 4, while the
percentage of the sample that have negative returns is slightly higher. Combining
the two periods, investors can, on average, earn over 10% (unadjusted) if they buy
an IR from the offering and sell it 10 days after it is listed.
Adjusted for the leverage effect, the average excess adjusted return on IRs over
the 10-day holding period is 2.12%, which is significant at the 5% level. This
figure is approximately the same as the average excess return on the underlying
shares over the same period, which, as reported in Table 3, is 2.18%. This should
come as no surprise since the adjusted return on an IR is derived from removing
the leverage effect and this should result in obtaining the underlying stock. Also,
arbitrage between the two instruments should ensure that the adjusted return on
an IR and the return on the underlying stock are the same. Hence, both of these
two positive excess returns can be regarded as an evidence of price recovery. This
evidence provides strong support for the liquidity effect.
To provide further support for the liquidity effect, we also test for the relationship
between the 10-day excess returns on the two instruments and proxies for issue
size and liquidity as follows:
(IR)RLD10i denotes the 10-day excess (IR) stock return on (IR) stock i. Adjusted
IR returns are not used here because they are equivalent to stock returns as explained
earlier. The results are reported in Tables 8 and 9. The coefficient of MULTIPLEi
128 NARAT CHARUPAT AND C. SHERMAN CHEUNG
Intercept −0.026 (0.477) 0.055 (0.121) −0.300 (0.129) 0.033 (0.381) −0.015 (0.683)
FRACTIONI 0.191 (0.055) 0.135 (0.213)
VALUEI 0.000 (0.816) 0.000 (0.912)
MULTIPLEI 0.000 (0.033) 0.000 (0.037) 0.000 (0.381)
Note: p-Values are in parentheses. They are calculated based on standard errors that are computed
using the heteroskedasticty consistent estimator of White (1980).
a FRACTION is the number of shares offered divided by the number of shares outstanding.
i
MULTIPLEi is the ratio between the number of shares offered and the average daily trading volume
over the period from AD − 150 to AD − 101. VALUEi is the dollar value of issue i.
Intercept −0.125 (0.539) 0.251 (0.114) 0.012 (0.207) 0.013 (0.441) −0.005 (0.810)
FRACTIONI 0.088 (0.077) 0.050 (0.336)
VALUEI 0.000 (0.777) 0.000 (0.906)
MULTIPLEI 0.000 (0.004) 0.000 (0.004) 0.000 (0.123)
Note: p-Values are in parentheses. They are calculated based on standard errors that are computed
using the heteroskedasticty consistent estimator of White (1980).
a FRACTION is the number of shares offered divided by the number of shares outstanding.
i
MULTIPLEi is the ratio between the number of shares offered and the average daily trading volume
over the period from AD − 150 to AD − 101. VALUEi is the dollar value of issue i.
is positive and significant. It appears that large issues produce stronger recovery.
In other words, not only do we have price recovery, but those that potentially have
largest liquidity impact also produce the greatest recovery.
Table 10. Regression Estimates of Leverage Offered on Proxies for Issue Size
and Liquidity.
Variablea (1) (2) (3)
Notes: p-Values are in parentheses. They are calculated based on standard errors that are computed
using the heteroskedasticty consistent estimator of White (1980).
a FRACTION is the number of shares offered divided by the number of shares outstanding.
i
MULTIPLEi is the ratio between the number of shares offered and the average daily trading volume
over the period from AD − 150 to AD − 101. VALUEi is the dollar value of issue i.
relationship between the degree of offered leverage and proxies for issue size and
liquidity.
The regression estimates of leverage offerred on proxies for issue size and
liquidity are reported in Table 10. Both proxies for issue size – VALUEi and
FRACTIONi – are positively related to the levels of leverage offered to the levels
of leverage offered and are significant at the 1% or the 5% levels respectively.19
In other words, firms with large issues tend to offer more leverage by having a
small first installment payment. This supports our conjecture that issuers use IRs
to reduce the price concessions that they have to make.
8. CONCLUSIONS
the issues hit the market. The IR arrangement can therefore be a very useful device
to minimize the liquidity impact for markets that lack the same depth as the U.S.
markets.
NOTES
1. In this paper, the general term “seasoned equity offerings” refers to the sale of either
new or existing shares to the public, while the term “secondary equity offerings” refers to the
sale of existing shares to the public by major shareholders. Hence, in our context, “secondary
equity offerings” share the same characteristics as block trades since they involve trades in
large amounts of shares without changing the number of shares outstanding.
2. For example, in the 1980s, the British government heavily used IRs to facilitate their
privatization of state-owned enterprises. More recently, in 1998, Ameritech Corp. sold its
(substantial) holding in Telecom New Zealand through an IR plan in the New Zealand
market. In Canada, while IRs have been around for decades, it was not until the 1990s that
they became popular among Canadian investors.
3. The exception is the paper by Hudson et al. (1993).
4. Usually, the investors can choose to pay future installments early and receive the
underlying shares right away. However, that would be irrational as long as interest rate is
greater than zero and IRs can be traded in the market. An exception to this is when the buyers
buy IRs as a part of an arbitrage transaction (i.e. together with a short position in the underling
shares), in which case early payments may be made if the cost of maintaining the short-sale
position is greater than interest that can be earned on future installment payments.
5. To the best of our knowledge, there have been two cases of default in the Canadian
market. To avoid default, some issuers chose instead to reduce the amount of future
installment payments.
6. While IRs are riskier than the underlying shares due to the leverage effect, investors
can, as mentioned earlier, in most cases choose to pay the remaining installment payments
immediately and obtain the underlying shares. Alternatively, investors can easily unlever
their IRs by holding long risk-free assets. Therefore, the additional risk from leverage should
not discourage buyers. Note also that the attractiveness of IRs will depend on the implicit
leverage cost, which, in turn, depends on the terms of the installment payments and the
(unobservable) offer price if the issue were sold on a fully-paid basis.
7. Examples of this type of explanation include the price-pressure hypothesis discussed
in Scholes (1972), and the “price-of-immediacy” model proposed by Grossman and Miller
(1988).
8. Underpricing of secondary offerings can also result from information asymmetry
among the issuers, underwriters and different types of traders provided that the prices
observed in the pre-issue markets are not perfectly revealing. However, as we report in
Section 4, our results do not support this hypothesis.
9. See, for example, Holthausen et al. (1990) and Keim and Madhavan (1996).
10. One company in our sample, Hudson Bay Co., had two offerings in the sample
period.
11. These figures were calculated by multiplying the number of underlying shares offered
by the sum of the required installment amounts, without taking into account the payments’
Secondary Equity Offerings 131
present values. This method is the same as that used in the prospectuses. Note that the
figures are based on dollars at the time of issues. That is, no adjustment for inflation was
made to them.
12. Asquith and Mullins (1986) find a 9.60% excess return during the period from 160
days to 10 days prior to the announcement dates, followed by a –2.0% excess return during
the announcement period. Hudson et al. (1993) report a 4.28% cumulative abnormal return
during the 60-day period before the announcements, and a −2.65% excess return during
the announcement period.
13. As there has not been a study of offer-to-close returns from seasoned/secondary
offerings in the Canadian markets, we compare our results to the findings from the Canadian
initial public offerings (IPO) markets. Jog and Srivastava (1995) report an average offer-
to-close return of 5.40% between 1971 and 1992. Using a shorter data period (from
1984 to 1987), Clarkson and Merkley (1994) find an average return of 6.44%. These two
results are similar in magnitude to our unadjusted returns, but are higher than our adjusted
returns.
14. The alternative period of 150 trading days before the announcements was also used,
and yielded similar results.
15. Hudson et al. (1993) find no significant relationship between price drops and pre-
announcement returns, while Asquith and Mullins (1986) and Korajczyk et al. (1990) find
a very weak relationship. The choices of periods over which the pre-announcement returns
are measured appear to influence their results.
16. Scholes (1972), Asquith and Mullins (1986) and Korajczyk et al. (1990) find no
significant relationship between issue size and announcement period’s return. Hess and
Bhagat (1986), Masulis and Korwar (1986) and Barclay and Litzenberger (1988) report
similar findings for a sample of industrial issues.
17. If the negative returns on the announcement days are due to anticipation of a liquidity
shock, a mean-variance analysis such as that in Grossman and Miller (1988) can be used to
show that the magnitude of underpricing is positively related to the variance of the stock.
A similar positive relation exists if underpricing is due to information asymmetry. This is
because return volatility can reflect the degree of information asymmetry between existing
and potential shareholders.
18. A somewhat analogous example is the amount of required margins on futures
contracts. It has been shown that the level of margins is inversely related to the demand
for futures contracts (See, for example, Hartzmark (1986), Fishe and Goldberg (1986) and
Addrangi and Chatrath (1999)). One explanation for it is that margins are committed funds
which reduce traders’ flexibility in case profitable investment opportunities subsequently
exist. Unless the traders unwind their futures position and get back the margin deposits, they
will have to borrow to take advantage of the profitable opportunities. Hence, the flexibility
cost depends on the leverage cost and constraint that the traders face. To apply this logic to
our context, IRs can be thought of as futures contracts where the first installment payment
is the initial margin, while fully-paid offers are contracts where the full price is the initial
margin. Therefore, investing in IRs involves lower flexibility cost.
19. Different versions of leverage such as the ratio of future payments (without
adjustment for the present value factor) to the sum of all payments and the ratio of future
payments over the initial payment. The conclusions are the same.
20. See, for example, Shleifer (1986).
132 NARAT CHARUPAT AND C. SHERMAN CHEUNG
ACKNOWLEDGMENTS
Both authors gratefully acknowledge the financial support from the Social Sciences
and Humanities Research Council of Canada. Please address all correspondence
to C. Sherman Cheung at scheung@mcmaster.ca.
REFERENCES
Addrangi, B., & Chatrath, A. (1999). Margin requirements and futures activity: Evidence from the
soybean and corn markets. Journal of Futures Markets, 19, 433–455.
Asquith, P., & Mullins, D. (1986). Equity issues and offering dilution. Journal of Financial Economics,
15, 61–89.
Barclay, M., & Litzenberger, R. (1988). Announcement effects of new equity issues and the use of
intraday price data. Journal of Financial Economics, 21, 71–99.
Clarkson, P. M., & Merkley, J. (1994). Ex ante uncertainty and the underpricing of initial public
offerings: Further Canadian evidence. Canadian Journal of Administrative Sciences, 11,
54–67.
Fishe, R. P. H., & Goldberg, L. G. (1986). The effects of margins on trading in futures markets. Journal
of Futures Markets, 6, 261–271.
Grossman, S., & Miller, M. (1988). Liquidity and market structure. Journal of Finance, 43, 617–633.
Hartzmark, M. L. (1986). The effects of changing margin levels on futures market activity, the
composition of traders in the market, and price performance. Journal of Business, 59,
S147–S181.
Hess, A. C., & Bhagat, S. (1986). Size effects of seasoned stock issues: Empirical evidence. Journal
of Business, 59, 567–584.
Holthausen, R., Leftwich, R., & Mayers, D. (1990). Large-block transactions, the speed of response,
and temporary and permanent stock-price effects. Journal of Financial Economics, 26, 71–95.
Hudson, C. D., Jensen, M. R. H., & Pugh, W. N. (1993). Information versus price-pressure effects:
Evidence from secondary offerings. Journal of Financial Research, 16, 193–207.
Jog, V., & Srivastava, A. (1995). Underpricing in Canadian IPOs 1971–1992 – an update. Fineco, 4,
81–89.
Keim, D. B., & Madhavan, A. (1996). The upstairs market for large-block transactions: Analysis and
measurement of price effects. Review of Financial Studies, 9, 1–36.
Korajczyk, R. A., Lucas, D., & McDonald, R. L. (1990). Understanding stock price behavior around
the time of equity issues. In: R. G. Hubbard (Ed.), Asymmetric Information, Corporate Finance,
and Investment (pp. 257–277). Chicago: University of Chicago Press.
Loderer, C., Sheehan, D. P., & Kadlec, G. P. (1991). The pricing of equity offerings. Journal of Financial
Economics, 29, 35–57.
Masulis, R., & Korwar, A. (1986). Seasoned equity offerings: An empirical investigation. Journal of
Financial Economics, 15, 91–118.
Mikkelson, W., & Partch, M. (1986). Valuation effects of security offerings and the issuance process.
Journal of Financial Economics, 15, 31–60.
Myers, S., & Majluf, N. (1984). Corporate financing and investment decisions when firms have
information that investors do not have. Journal of Financial Economics, 13, 187–221.
Secondary Equity Offerings 133
Scholes, M. (1972). The market for securities: Substitution versus price pressure and the effects of
information on share prices. Journal of Business, 45, 179–211.
Shleifer, A. (1986). Do demand curves for stocks slope down. Journal of Finance, 41, 579–590.
Smith, C. (1977). Alternative methods for raising capital: Rights versus underwritten offerings. Journal
of Financial Economics, 5, 273–307.
White, H. (1980). A heteroskedasticity-consistent covariance matrix estimator and a direct test for
heteroskedasticity. Econometrica, 48, 817–838.
A NEW APPROACH TO TESTING PPP:
EVIDENCE FROM THE YEN
ABSTRACT
Conventional methods to test for long-term PPP based on the theory of
cointegration are typically undertaken in the framework of vector error
correction models (VECM). The standard approach in the use of VECMs
is to employ a model of full-order, which assumes nonzero entries in all the
coefficient matrices. But, the use of full-order VECM models may lead to
incorrect inferences if zero entries are required in the coefficient matrices.
Specifically, if we wish to test for indirect causality, instantaneous causality, or
Granger non-causality, and employ “overparameterised” full-order VECM
models that ignore entries assigned a priori to be zero, then the power of
statistical inference is weakened and the resultant specifications can produce
different conclusions concerning the cointegrating relationships among the
variables. In this paper, an approach is presented that incorporates zero
entries in the VECM analysis. This approach is a more straightforward and
effective means of testing for causality and cointegrating relations. The paper
extends prior work on PPP through an investigation of causality between the
U.S. Dollar and the Japanese Yen. The results demonstrate the inconsistencies
that can arise in the area and show that bi-directional feedback exists between
prices, interest rates and the exchange rate such that adjustment mechanisms
are complete within the context of PPP.
Research in Finance
Research in Finance, Volume 21, 135–154
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21006-1
135
136 T. J. BRAILSFORD ET AL.
1. INTRODUCTION
The purchasing power parity (PPP) hypothesis implies that, over the long-run,
changes in the exchange rate between the currencies of two countries reflect
changes in the ratio of those countries’ price levels. PPP has important implications
for an explanation of exchange rate movements and for exchange rate forecasting
in the long-term. In general, most studies tend to reject PPP in the short-run while
in the long-run there is both evidence supporting and rejecting PPP.1
One factor contributing to these mixed results relates to the procedure used in
testing. In recent years the theory of cointegration has been widely applied to
tests of PPP. If the individual variables involved in the PPP relationship, namely
the nominal exchange rate and the ratio of domestic to foreign prices, are non-
stationary, but a specific linear combination of them is stationary, then PPP is
claimed to hold. A major criticism of the classical residual-based tests for non-
stationarity, such as the augmented Dickey and Fuller test, is that they lack power to
distinguish between unit root and near unit root stationary processes, and therefore
they have a tendency to accept the null hypothesis of non-stationarity (DeJong
et al., 1989; Hakkio, 1984). This has prompted the use of tests that employ the
null hypothesis of stationarity (Fisher & Park, 1991; Kwiatkowski et al., 1992).
However tests for PPP are sensitive to the null hypothesis employed. That is, PPP
could be rejected under a null hypothesis embodying the presence of a unit root,
but accepted in a test with a null hypothesis of stationarity.
In the context of a vector time-series framework, which indicates the dynamic
relationships among the relevant financial variables, conventional methods to test
for PPP are undertaken in full-order vector error correction modelling (VECM).
However standard full-order VECM models are based on nonzero elements in
all their coefficient matrices. As the number of elements to be estimated in
these possibly over-parameterised models grows with the square of the number
of variables, the degrees of freedom will be heavily reduced. In addition, in
applications of VECM models to financial market data, a priori assumptions of
zero entries may be required. Therefore the use of full-order VECM models may
lead to incorrect inferences. Specifically, if we wish to test for indirect causality, or
Granger non-causality which crucially depends on the positions of zero entries in
the coefficient matrices, and we ignore the entries assigned a priori to be zero and
employ “overparameterised” full-order VECM models, the power of our statistical
inferences is weakened. Further, if the underlying true VECM and the associated
cointegrating and loading vectors contain zero entries, the resultant specifications
can produce conclusions concerning the cointegrating relationships among the
variables which would be different to the conclusions arising where the a priori
zeros are ignored, and give rise to potentially different conclusions concerning PPP.
A New Approach to Testing PPP: Evidence from the Yen 137
E{(t) (t − )} = V, = 0,
E{(t) (t − )} = 0, > 0.
138 T. J. BRAILSFORD ET AL.
where L denotes the lag operator, and Ly(t) = y(t − 1). It is assumed that the roots
of |B q (L)| = 0 lie outside or on the unit circle to ensure that y(t) can contain I(1)
variables.
Of note, y(t) is integrated of order d, I(d), if it contains at least one element which
must be differenced d times before it becomes I(0). Further, y(t) is cointegrated
with the cointegrating vector, , of order g, if  y(t) is integrated of order (d − g),
where y(t) has to contain at least two I(d) variables.2
Under this I(1) assumption:
B q (L) = B q (1)L + (I − L)B q−1 (L)
Following Engle and Yoo (1991), the equivalent VECM for (1) can be expressed
as:
B q (1)y(t − 1) + B q−1 (L) y(t) = (t), (2)
where y(t) contains variables of two types, namely I(0) and I(1) and = (I − L).
Eq. (2) can be rewritten as:
B ∗ y(t − 1) + B q−1 (L) y(t) = (t), (3)
where B ∗ = B q (1) and B ∗ y(t − 1) is stationary and the first term in (3) is the error
correction term. The term B q−1 (L) y(t) is the vector autoregressive part of the
VECM.
Because y(t) is cointegrated of order 1, the long-term impact matrix, B∗ , must
be singular. As a result, B ∗ = ␣ and  y(t − 1) is stationary, where the rank of
B∗ is r, and ␣ and  are matrices of dimensions s × r and r × 2s respectively.
The columns of  are the cointegrating vectors and the rows of ␣ are the loading
vectors.
Model development is more convenient using VECMs, rather than the equivalent
VARs, if the systems under study include integrated time series. Engle and Granger
(1987) note that, for I(1) systems, the VARs in first difference will be mis-specified
and the VARs in levels will ignore important constraints on the coefficient matrices.
Although these constraints may be satisfied asymptotically, efficiency gains and
improvements in forecasts are likely to result by imposing them. The analogous
conclusion applies to I(1) systems, such as those typically encountered in tests
of PPP. Comparisons of forecasting performance of the VECMs versus VARs for
A New Approach to Testing PPP: Evidence from the Yen 139
cointegrated systems have been reported in studies such as Engle and Yoo (1987)
and LeSage (1990). The results of these studies indicate that, while in the short-run
there may be gains in using unrestricted VAR models, the VECMs produce long-
run forecasts with smaller errors when the variables used in the models satisfy the
test for cointegration.
Further to these developments, we consider a hypothesis where every (i, j)-
th element, for specified i and j, is zero in all coefficient matrices in a VAR.
If this hypothesis is framed in the VAR expressed by (1), these zero entries
will also hold in the error-correction terms and in the vector autoregressive part
of the equivalent VECM, say (2). A discussion of this property is provided in
Appendix 1. Analogously we can achieve a result that if all (i, j)-th coefficient
elements in the error-correction terms and all (i, j)-th coefficient elements in the
vector autoregressive part of the VECM are zeros, then every (i, j)-th entry is zero
for all coefficient matrices in a VAR.
The implications of the above outcome are straightforward. If yj does not
Granger-cause yi , then any (i, j)-th entry must be zero for all coefficient matrices in
the VAR. Also all (i, j)-th coefficient elements in the equivalent VECM are zeros.
In a similar way, we can demonstrate that if yj does Granger-cause yi, then
the (i, j)-th element of Bq (L) in (1) is nonzero. Also, at least a single (i, j)-the
coefficient element is nonzero in Bq (1) or Bq-1 (L) in the equivalent VECM. Of
note, an indirect causality from yj to yi through ym indicates yj causing yi but
only through ym . Hence, yj Granger-causes ym , ym Granger-causes yi , and yj does
not Granger-cause yi directly. We can easily demonstrate that the VAR in (1) has
nonzero (m, j)-th and (i, m)-th elements and a zero (i, j)-th element in Bq (L). This
indirect causality can also be shown in the equivalent VECM, which has at least a
single nonzero (m, j)-th element and a single nonzero (i, m)-th elements in Bq (1)
and Bq-1 (L). Also all the (i, j)-the elements in the equivalent VECM are zeros.
The above discussion indicates that Granger causality, Granger non-causality
and indirect causality detected from both the ZNZ patterned VECM and its
equivalent ZNZ patterned VAR are identical. Since the use of the VECM is more
convenient, it is obvious the ZNZ patterned VECM is a more straightforward and
effective means of testing for the Granger causal relations. The same benefits will
be present if the ZNZ patterned VECM is used to analyse cointegrating relations.
between the USA and Japan using annual data over the period 1951–1994 and
he finds evidence supporting PPP in the long run.
Cheng (1999) uses a VECM to examine the relationship between the U.S. Dollar
and the Japanese Yen and the following one-sided Granger causality test:
M
N
(1 − L)y t = ␣0 + ␣m (1 − L)y t−m + n (1 − L)x t−n
m=1 n=1
Q
+ k (1 − L)z t−k + t (4)
k=1
Note that neither 0 nor 0 is included in the specification (4). Since a VECM
is equivalent to a VAR model with unit roots, and (4) is valid only when no
instantaneous causality exists among the variables (Geweke, 1982; Hatanaka,
1982), instantaneous causality is implicitly neglected in (4). Moreover, when using
low frequency data, such as annual time-series as in Cheng (1999), the impact of
time aggregation suggests that instantaneous relations should not be ignored.
The use of VECM models in financial time-series is versatile. The models can
be employed as a means of detecting Granger causality, Granger non-causality
and indirect causality. In this section, the focus is on instantaneous causality.
Specifically, the section demonstrates that the one-sided Granger causality test
proposed in Cheng (1999) is valid only when no instantaneous causality exists
among the variables.
To begin with, consider a bivariate VECM of (2) and define
q−1 11
y1 (t) Bi Bi12 i
y(t) = , B q−1
(L) = L,
y2 (t) Bi21 Bi22
i=1
1 (t) ␣1
(t) = ,␣= ,
2 (t) ␣2
where
v11
c
, = 0,
E{(t) (t − )} = c v22
0, > 0.
v11 0 and v22 0.
Both y 1 (t) and y 2 (t) are mutually instantaneously caused if and only if c
= 0.
Also both y 1 (t) and y 2 (t) are not instantaneously caused if and only if c = 0
(Penm & Terrell, 1984).5
1 −cv−1 22
Premultiply (5) by the matrix , then we have
0 1
q−1
−cv−1 y1 (t) B11 − cv−1 B21 Bi12 − cv−1
22 Bi
22
1 i 22 i
22 +
0 1 y2 (t) B 21
i Bi22
i=1
y1 (t − i) ␦ a (t)
+ e(t − 1) = ,
y2 (t − i) ␣2 2 (t)
where ␦ = ␣1 − cv−1 −1
22 ␣2 and a = 1 − cv22 2 .
Now the first equation of this new system becomes
q−1
q−1
y 1 (t) + ␥i y 1 (t − i) + j y 2 (t − j) + ␦e(t − 1) = a (t), (6)
i=1 j=0
where 0 = −cv−1 −1 21
22 . Thus 0 = 0, if and only if c = 0. Also ␥i = B i − cv22 B i
11
−1 22
and j = B 12j − cv22 B j .
If 0
= 0, (6) is different from the one-sided Granger causality test. Only if
0 = 0, is (6) equivalent to the one-sided Granger causality test proposed in Cheng
(1999). Thus this one-sided test is valid only when no instantaneous causality exists
between y 1 (t) and y 2 (t). The same result is obtained in a trivariate system. In
Appendix B we summarise the method of instantaneous causality detection with
all cases for a trivariate system.
5. EMPIRICAL RESULTS
In this section, the PPP hypothesis is re-examined using the ZNZ patterned VECM
modelling proposed earlier. The data are drawn from the U.S.-Japan relation
A New Approach to Testing PPP: Evidence from the Yen 143
and replicate Cheng’s (1999) test. Following Cheng (1999), the following three
variables are studied contemporaneously in a stochastic vector system.
(i) Japanese Yen to U.S. Dollar: exchange rate (E) per U.S. Dollar.
(ii) Japanese CPI to U.S. CPI: ratio of price levels (P).
(iii) (1 + U.S. discount rate)/(1 + Japanese discount rate): interest rate ratio (IR).6
While our interest is in assessing the presence or absence of the instantaneous
causality in the proposed E-P-IR system, using the ZNZ patterned VECM
modelling, we also use the cointegrating vectors detected to assist with the answer
to other questions, such as whether the PPP hypothesis in the long-term is supported
by the system selected.
For comparison purposes the following three data sets relating to the above three
variables are obtained (from DataStream™ ).
(i) Annual data over the period 1951–1994.
(ii) Semi-annual data over the period 1974–2000.
(iii) Quarterly data over the period 1974–2000.
The first data set is identical to that of Cheng (1999). For both the second and the
third data sets, the samples begin in 1974. This point is often chosen as a cut-off
as it coincides with the end of the Bretton Woods system and the first world oil
shock.7 For the post-Bretton Woods era, the number of annual observations for the
E-P-IR system is insufficient to conduct the analysis. Moreover, given the earlier
evidence that shows that tests of PPP are sensitive to the sampling interval, semi-
annual and quarterly data are employed. These samples allow an assessment of
long-term PPP and short-term responses.
The variables are log transformed such that y 1 (t) = log(E), y 2 (t) = log(P), and
y 3 (t) = log(IR). Unit root tests indicate that all transformed series are I(1). We
then conduct the search procedures proposed in Penm et al. (1997) to obtain the
optimal ZNZ patterned VAR models.
In the course of selecting the optimal lag order (p) for the autoregressive part
of the VECM system, the principle used by Chen and Wu (2000) to enhance the
procedure is adopted. That is, we examine whiteness for the residual vectors from
the VECM chosen by the Akaike Information Criterion (AIC). If the residual vector
process proves to be non-white, we sequentially increase p to p + 1, and check the
resultant residual vector process until the process is a vector white noise process.
The optimal ZNZ patterned VECM and the optimal ␣ and  are then selected by
using the Schwarz criterion (SC). The search results for each data set are presented
in Tables 1–3 respectively.
The estimated residual variance-covariance matrices and the selected patterns
of the cointegrating vector produce some interesting results. For the annual
144 T. J. BRAILSFORD ET AL.
Table 1. (Continued )
Note: Variables: y 1 (t) = log(E), y 2 (t) = log(P), y 3 (t) = log(IR). Sample Period: 1951–1994
q−1
VECM: B ∗ y(t − 1) + y(t) + =1 B() y(t − ) = (t).
a The residual analysis confirms the residuals have white noise characteristics. For simplicity, the values
of SC for q > 5 are not presented, but can be supplied to readers upon request.
b x Granger-causes y only and not instantaneously: (Notation: x y). feedback, not
instantaneously: (Notation: x y). instantaneous causality only: (Notation: x y).
data set, the non-diagonal V̂ shown in Table 1 indicates the existence of the
instantaneous causality in the system. This outcome could result from the effect of
time aggregation on instantaneous causality in low frequency data. The presence
of causality in the system is prima facie consistent with PPP and the results of
Cheng (1999). Further, the positive relation between log(E) and log(P) is consistent
with PPP in that an increase in relative price levels in Japan is associated with a
depreciation of the Yen. However, a closer inspection of the cointegrating vector
reveals an opposite sign between log(E) and log(IR) indicating that, ceteris paribus,
an increase in IR leads to an appreciation in the Yen. This latter result is inconsistent
with theory which asserts an opposite relation.8
This result could be due to a number of factors. First, since the number of the
observation vectors is only 44, the poor parameter estimates in the cointegrating
vector may be caused by the small sample size. Second, over the time period
studied, the Japanese economy has undergone major change. In the early part
of the sample, the Japanese economy was subject to considerable controls. For
instance, heavy regulation and macroeconomic controls were prevalent during the
1950s and 1960s which included a fixed foreign exchange rate, restrictions on
foreign investment, subsidized central lending and a targeted industrial policy. It
was not until 1964, when Japan joined the OECD, that there was a relaxation of
foreign investment controls. In contrast, the latter part of the sample represents
a period when the Japanese economy developed into a major and open global
economy.9 The results may simply reflect the outcome of combining data from
different economic frameworks. Further disaggregation is not possible due to the
small number of observations. Hence, while there is some evidence from the annual
data that PPP holds, such a conclusion must be interpreted with caution given the
findings in relation to the interest rate effect.
For the semi-annual data set, the non-existence of instantaneous causality is
detected from the diagonal V̂ . This outcome indicates that the possible effect on
instantaneous causality through time aggregation does not arise in the semi-annual
data. In relation to PPP, there is evidence of causality in the system. Further, the
positive relation between log(E) and log(P), and the positive relation between
log(E) and log(IR) shown in Table 2 indicates that an increase in IR or an increase
146 T. J. BRAILSFORD ET AL.
Table 2. (Continued )
14 3,2 −0.3252
(0.062)
B∗ 2,1 −2.0284 2,2 0.2774 2,3 1.8547
(0.263) (0.039) (0.175)
6.3903E-03 0 0
The type of V̂ selected: 0 2.7613E-05 0
0 0 4.5149E-05
in P leads to a depreciation of the Yen. This result is now consistent with theory.
For instance, when the price level in Japan is increasing, the Yen depreciates in
order to retain PPP. Further, when the relative interest rate in Japan increases, there
is an associated appreciation of the Yen.
In reference to the Granger causal relations among the variables, feedback
relations exist between the pair of log(E) and log(P), the pair of log(P) and
log(IR), and the pair of log(E) and log(IR). Hence, the feedback within the
system is complete and shocks to any one of the variables will be processed
through the system. In addition, Fig. 1 shows that all roots detected lie outside
the unit circle (> 1). This latter finding shows that the ZNZ patterned VECM
can increase the modelling power. To check the adequacy of the model fit, the
results in Table 2 support the hypothesis that the residual vector is a white noise
process.
We now turn to the quarterly data set. The non-existence of instantaneous
causality is also detected from the diagonal V̂ in Table 3. Thus the effect of the time
aggregation on instantaneous causality can be ignored in high(er) frequency data.
The presence of causality and the positive relations between log(E) and log(P),
and log(E) and log(IR) shown in Table 3 are again consistent with theory. Hence,
we conclude that both the semi-annual and quarterly data analyses support the PPP
148 T. J. BRAILSFORD ET AL.
Table 3. (Continued )
Long-term Cointegrating Relationship Identified: log(E) = 0.1551 log(P) + 0.8983 log(IR)
Fig. 1. Histogram of Roots for Semi-Annual Sample. Note: Minimum: 1.002 Median:
1.078 Maximum: 1.895.
Fig. 2. Histogram of Roots for Quarterly Sample. Note: Minimum: 1.015 Median: 1.148
Maximum:1.555.
150 T. J. BRAILSFORD ET AL.
hypothesis in the long run. In examining the Granger causal relations, although
there is no direct Granger causation from log(P) to log(E), there is however indirect
causation from log(P) to log(E) via log(IR). In addition direct Granger causation
exists from log(E) to log(P), and feedback relations exist between the pair of
log(P) and log(IR), and the pair of log(E) and log(IR). We therefore conclude that
Granger causal relation (directly or indirectly) exists between log(E) and log(P)
in both semi-annual and quarterly data samples. Since all detected roots shown
in Fig. 2 lie outside the unit circle, the results give no support to the hypothesis
of instability. The possibility of a structural shift, which is reflected in unstable
roots, is not a problem. Again, to check the adequacy of the model fit, the results
in Table 3 support the hypothesis that the residual vector is a white noise process.
6. CONCLUSION
In this paper we have demonstrated a new approach that involves ZNZ patterned
VECM modelling to examine PPP, and associated causality and cointegration.
Three contributions have been made. First, the paper demonstrates an approach
that allows for zero entries in the VECM and shows how it can be applied to a three
variable system that includes prices, interest rates and the exchange rate. Second,
the paper shows that the one-sided Granger causality test proposed in Cheng (1999)
is valid only when no instantaneous causality exists among the variables. Since
time aggregation can contribute to instantaneous causality in low frequency data
(such as annual data), instantaneous causality cannot be ignored. Indeed, in a test
using annual data between Japan and the USA, evidence is found of instantaneous
causality. Hence, caution is required when interpreting prior evidence. Third, prior
studies have reported inconsistent results in relation to tests of PPP. In this paper,
the sampling interval and the sample period is varied. Support for PPP is strongest
when semi-annual data are employed. The results indicate that bi-directional
feedback exists between prices, interest rates and the exchange rate and hence
sheds light on the adjustment mechanisms through which PPP is achieved.
NOTES
1. A good survey of the literature can be found in MacDonald (1995).
2. In this paper, we consider only the case d = 1, although the procedure can be generally
applied to models where d > 1.
3. For this case, McFarland et al. (1994) propose that the necessary condition for PPP
exists in the long-term. The necessary and sufficient condition means that these two variables
are cointegrated and the cointegrating vector is  = (1, −1).
A New Approach to Testing PPP: Evidence from the Yen 151
4. Cheng (1999) found no evidence of causality between productivity and the terms of
trade against the exchange rate, and thus excluded these variables from the model.
5. Since the VECM of (2) is equivalent to the VAR of (1), thus instantaneous causality
exists between y 1 (t) and y 2 (t) if and only if c
= 0. Also both y 1 (t) and y 2 (t) are not
instantaneously caused if and only if c = 0.
6. Consistent with the Fisher equation, the interest rate ratio is expressed in this form
and is numerically closer than the simple ratio of percentage rates.
7. In time series analysis economic and financial systems are evolving, and although
episodic structural shifts may occur in the test period, model evolution in vector systems
can be handled through time update and order update methods to investigate the structure
changes. These methods utilise each incoming observation to update the model structure
and the model parameters.
8. Cheng (1999) does not report the signs on the coefficients.
9. For a history of the Japanese post-war economy, see Komiya et al. (1988).
REFERENCES
Abauf, N., & Jorion, P. (1990). Purchasing power parity in the long run. Journal of Finance, 45,
157–174.
Chen, S. L., & Wu, J. L. (2000). A re-examination of purchasing power parity in Japan and Taiwan.
Journal of Macroeconomics, 22(2), 271–284.
Cheng, B. S. (1999). Beyond the purchasing power parity: Testing for cointegration and causality
between exchange rates, prices, and interest rates. Journal of International Money and Finance,
18, 911–924.
Corbae, D., & Ouliaris, S. (1988). Cointegration and tests of purchasing power parity. The Review of
Economics and Statistics, 70, 508–511.
Cumby, R., & Obstfeld, M. (1984). International interest rate and price level linkages under flexible
exchange rates, a review of recent evidence. In: Bilson & R. Marston (Eds), Exchange Rate
Theory and Practice. Chicago: University of Chicago Press.
DeJong, D. N., Nankervis, J. C., Savin, N. E., & Whiteman, C. H. (1989). Integration versus trend
stationarity in macroeconomic time-series. Working paper 89–99, University of Iowa.
Edison, H. J. (1987). Purchasing power parity in the long-term: A test of the dollar/pound exchange
rate (1890–1978). Journal of Money, Credit and Banking, 19, 376–387.
Engle, R. F., & Granger, C. W. J. (1987). Cointegration and error correction representation, estimation
and testing. Econometrica, 55, 251–276.
Engle, R. F., & Yoo, B. S. (1987). Forecasting and testing in co-integrated system. Journal of
Econometrics, 35, 143–159.
Fisher, E. O. N., & Park, J. Y. (1991). Testing purchasing power parity under the null hypothesis of
cointegration. The Economic Journal, 101, 1476–1484.
Frenkel, J. A. (1981). The collapse of purchasing power parities during the 1970s. European Economic
Review, 16, 145–166.
Froot, K. A., & Rogoff, K. (1994). Perspectives on PPP and long-run real exchange rates. NBER
working paper 4952.
Geweke, J. (1982). Measurement of linear dependence and feedback between multiple time-series.
Journal of the American Statistical Association, 77, 304–313.
152 T. J. BRAILSFORD ET AL.
APPENDIX A
ZERO ENTRIES IN A VAR AND ITS EQUIVALENT
VECM FOR AN I(1) SYSTEM.
In an I(1) system, the VECM of (2) can be expressed as:
If the (i, j)-th entries of B k (L), B k (1),and B k−1 (L)are b ij (L), b ij (1), and c ij (L)
respectively, we have
If b ij (L) = 0, then b ij (1) will also be zero. From (A.2), we have that c ij (L)(1 − L) =
0. (A.3) produces c 0 = 0, (c 1 − c 0 ) = 0, . . ., c k−1 − c k−2 = 0, c k−1 = 0 which
leads to c i = 0, i = 0, 1, . . . , k − 1, and therefore c ij (L) = 0.
At this point, if the (i, j)-th entry of B q (L)is zero, then the (i, j)-th elements of
both B q (1) and B q−1 (L) are zeros.
Therefore it can be concluded that if yj does not Granger-cause yi , then any
(i, j)-th element must be zero for all coefficient matrices in the VAR. Also all (i, j)-th
coefficient elements in the error-correction terms and in the vector autoregressive
part of the equivalent VECM, will also be zeros.
Further, from (A.1) if the (i, j)-th element of B q (L) is nonzero, then at least the
(i, j)-th element is nonzero in B q (1) or B q−1 (L).
Thus, we have just demonstrated that if yj does Granger-cause yi , then the
(i, j)-th element of B q (L) in the VAR is nonzero. In addition at least a single (i, j)-th
coefficient element is nonzero in B q (1) or B q−1 (L) of the equivalent VECM.
154 T. J. BRAILSFORD ET AL.
APPENDIX B
DETECTING INSTANTANEOUS CAUSALITY IN A
TRIVARIATE SYSTEM
v11 e f
In a trivariate system of (2), with V = e v22 h , we have the following
f h v33
eight different patterns of instantaneous causality.
(1) y 1 (t), y 2 (t)and y 3 (t) are mutually instantaneous directly caused if and
only if e
= 0, f
= 0 and h
= 0.
(2) y 1 (t) and y 2 (t) are instantaneously indirectly caused via y 3 (t) if and only
if e = 0, f
= 0 and h
= 0.
(3) y 1 (t) and y 3 (t) are instantaneously indirectly caused via y 2 (t) if and only
if f = 0, h
= 0 and e
= 0.
(4) y 2 (t) and y 3 (t) are instantaneously indirectly caused via y 1 (t) if and only
if h = 0, e
= 0 and f
= 0.
(5) y 1 (t) does not cause instantaneously y 2 (t) and y 3 (t) if and only if
e = f = 0, and h
= 0.
(6) y 2 (t) does not cause instantaneously y 1 (t) and y 3 (t) if and only if
h = e = 0, and f
= 0.
(7) y 3 (t) does not cause instantaneously y 1 (t) and y 2 (t) if and only if
f = h = 0, and e
= 0.
(8) No instantaneous causality exists among y 1 (t), y 2 (t) and y 3 (t) if and
only if e = f = h = 0.
The pattern of instantaneous causal relations can be detected by selecting the case
that minimises the model selection criterion.
CORRELATION AMONG STOCK
MARKETS UNDER DIFFERENT
EXCHANGE RATE SYSTEMS
Paul Sarmas
ABSTRACT
This study investigates the linkage between the Hong Kong stock market
and Singapore stock market and the U.S. stock market during the pre-
and post-East Asia Financial Crisis in 1997 and 1998. It uses multivariate
regression models to study the impact of Hong Kong’s fixed exchange rate
system and Singapore’s free-floating exchange rate system on their respective
stock markets. The results indicate that the exchange rate is not a significant
determinant of linkage between the U.S. and the two Asian stock markets, but
the evidence suggests that stronger post-crisis relationships between the U.S.
and the two Asian stock markets. The evidence also supports a stronger short-
run relationship between the U.S. and Hong Kong stock markets relative to
that between the U.S. and Singapore stock markets.
1. INTRODUCTION
Since the collapse of the Bretton Woods fixed exchange rate system in the early
1970s, most countries have switched from a fixed exchange rate system to a floating
exchange rate system. However, some countries insist on keeping their fixed
exchange rate systems in order to achieve economic stability and development.
Research in Finance
Research in Finance, Volume 21, 155–173
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21007-3
155
156 PAUL SARMAS
Hong Kong and Singapore are dubbed “twin cities” and share many similar
economic features. For example, both emerged under British governance, and
both are small and open metropolitan economies with well-functioning domestic
private sectors and highly rated public services. The two governments had enjoyed
healthy budget surpluses and accumulated large official foreign reserves before
the Asian economic crisis in 1997–1998. Both economies successfully pursued
an export-oriented strategy and upgraded their industrial structures toward higher
value-added activities. Each of them has become a regional financial center and the
operational headquarters of many multinational corporations in the Pacific Asia
region.
Remarkable contrasts, however, exist between the two economies’ monetary
systems and exchange rate regimes.1 Hong Kong has had a currency board regime
since 1983, with the HK dollar pegged to the U.S. dollar at a fixed rate. For
more than a half century up to the 1970s, Singapore had a classical sterling-based
currency board system. Since then, it has evolved a managed floating system that
maintains the value of the Singapore dollar against a trade-weighted basket of
currencies within an undisclosed band.
This study investigates the relationships between daily returns of two Asian
countries’ stock markets and the U.S. stock market under fixed and flexible
exchange rate systems. The motivation for this study has come about as a result of
developing a strategic plan for a diversified portfolio investment in the prosperous
Asian financial markets. The study proposes to analyze the effect of daily changes
of exchange rates and the U.S. stock market on the daily returns of Hong Kong and
Singapore stock markets, and it also tries to assess the movement of the Asian stock
markets under the two different exchange rate systems. Although other traditional
economic factors are also important, it is the perception of many financial analysts
that exchange rate systems would serve as vital links between the U.S. stock market
and the two Asian countries.
shocks finally made it impossible to have all the currencies pegged to each other
(Award, 1998).
Some countries, including Hong Kong, adopted the a rather rigid currency board
principle, maintaining a fixed exchange regime, while still maintaining the fixed
exchange rate policy either pegged to the U.S. dollar or to other major currencies
(Tornell, 2000). Under this system, the local monetary authority guaranteed the
complete and free convertibility of the local currency against the pound sterling at a
fixed exchange rate. The local note issue had 100% backing by sterling deposits in
London. Continuous speculative attacks could smash market confidence, resulting
in a breakdown of the peg. However, even under such adverse circumstances, the
Hong Kong Monetary Authority (HKMA) has been committed to maintaining
the fixed exchange rate system adopted in October 1983, and this has proven to
be effective in maintaining the confidence during the transition of Hong Kong to
Chinese control in 1997 (Yam, 1999).
Amid the international monetary uncertainty in 1973, as with many other
nations, the Singapore government adopted a managed floating system.
Meanwhile, the authorities continued to relax their foreign exchange controls
gradually, and finally liberalized controls by 1978.3 Under the free-floating
exchange rate system, Singapore successfully withstood the test of time, even
during the Asian Crisis in 1987, and achieved stable money policy and prosperous
business development (Wu, 2000).
markets are fairly independent of the U.S. market. The weak linkage between the
financial markets across nations would suggest that there are potential benefits in
international diversification.
Al Awad and Goodwin (1998) examined short-run and long-run dynamic
linkages among weekly real interest rates for G-10 countries using a variety of
time-series tests. The authors focused on the time-series properties of nominal
interest rates, ex-ante expected rates of inflation, and real interest rates. In-sample
and out-of-sample Granger causality tests were also conducted to evaluate lead/lag
relationships among real interest rates. The results provided strong support for well-
integrated markets, particularly in the long run. Their study indicated that the U.S.
financial markets might have a certain impact on Hong Kong’s and Singapore’s
money markets.
Numerous empirical studies have focused on the relationship between foreign
exchange rates and stock markets. However, the validity of their findings seems
to be sensitive to underlying exchange rate regime and economic conditions. For
instance, Bahmani-Oskooee and Sohrabian (1992) found that there was a two-way
relationship between the changes in levels of exchange rate and the fluctuation of
a stock market. Their empirical results show that there is a dual causal relationship
between stock price and the effective exchange rate of the dollar in the short
run. However, the long-run relationship between the two variables failed to show
statistical significance. Yet another study by Mok (1993) used an ARIMA approach
and the Granger causality test to explore the causality of daily interest rates,
exchange rates and stock prices in Hong Kong for period 1986 to 1991. The result
of that study concluded that the interest rate and exchange rate information was
efficiently incorporated in the stock market prices, both at the daily market close
and at the opening. The author also pointed out that there is a weak bi-directional
causality between stock prices and the exchange rate.
Following the same premise, Ying Wu (2000) used an error correction model to
explore the asymmetric effects of four different exchange rates on Singapore stock
prices, and the effects’ sensitivity to economic instability. Both the Singapore
currency appreciation against the U.S. dollar and Malaysian ringgit and the
depreciation against the Japanese yen and Indonesian rupiah led to a long-run
increase in stock prices for most selected periods of the 1990s. However, the
effect associated with the U.S. dollar exchange rate has a sign reversal between
the 1997–1998 crisis periods and the 1999–2000 recovery periods. Ying Wu’s
study indicated that the influence of exchange rates on stock prices increased in
chronological order in the 1990s. So the positive relationship between the stock
market and the exchange rate tends to receive more attention in recent years.
In a more recent study, Phylaktis and Ravazzolo (2001) studied the long run
and short-run dynamics between stock price and exchange rate, and channels
Correlation Among Stock Markets 159
through which great shocks impacted on these markets. They used cointegration
methodology and multivariate Granger causality tests, and applied the model to a
group of Pacific Basin countries over the period from 1980 to 1998. The evidence
suggested that stock and foreign exchange markets are positively related, and the
U.S. stock market acts as a conduit for these links, where these links seemed not
to be affected by foreign exchange restrictions. Thus, it showed once again that
an inconsistent relationship exists between stock markets and exchange rates, no
matter whether the exchange rate system is fixed or flexible.
Despite the recommendation by Hung and Cheung (1995) to use weekly instead
of daily data to avoid the interference of synchronous trading, this study uses daily
data to capture the effects of market shocks. Also, the results don’t show any
significant improvement when weekly data rather than daily data is used in the test
procedure. This research studies and compares the effect of the daily returns on the
U.S. stock market on Hong Kong and Singapore stock markets under the different
exchange rates system before and after the Asian Financial Crisis in 1997–1998.
It was the intent of this study to provide a more recent and specific picture by
contrasting the interrelationship between the Hong Kong stock market and the
U.S. stock market and between the Singapore stock market and the U.S. stock
market from the perspective of different exchange rate systems.
This empirical study was structured to address a number of specific issues.
First, the stationarity of all the time series variables is tested, because only the
stationary data can be employed in a regression test. The second task performed
was to examine collinearity of the explanatory variables: Hong Kong, Singapore,
and the U.S.4 A collinearity test will show the correlation among the independent
variables of three countries’ data in the regression model. Only the variables with no
collinearity were selected for regression analysis. Third, multiple regression tests
were used to determine the relationship between the dependent variable (daily
return of Hong Kong or Singapore stock market) and independent variables (U.S.
stock market return and the exchange rate changes before 1997 and after 1998).
Finally, application of the Granger causality tests was expected to determine the
pair-wise short-run relationship between those variables.
The structural design of this article can be characterized as an inferential study. It
uses multivariate analysis to assess the statistical significance of various predictors
about a single dependent variable (Zikmund, 2000). The independent variables
are: (1) the daily return of the U.S. stock market; and (2) the daily changes of
the exchange rate for the Hong Kong dollar against the U.S. dollar or the daily
160 PAUL SARMAS
changes of the exchange rate for the Singapore dollar against the U.S. dollar. The
dependent variables are: the daily return of the Hong Kong stock market or the
daily return of the Singapore stock market.
Many time series analyses have been based on the assumption that the underlying
data series are stationary. The test of stationarity is often referred to as a unit root
test. The Dickey-Fuller test is known as one of the early approaches to test unit
roots and the degree of integration. If time series data has no unit root, then it
is said to be “integrated of order zero” or I(0). If data has one or more than one
unit roots, that means it is not stationary. It has to be differenced5 once or more to
become stationary, expressed by I(1) and I(t) (Watsham & Parramore, 1997). The
unit root tests will be applied to HSI (the daily change of the Hang Seng Index of
Hong Kong), HGX (the daily change of the exchange rate for Hong Kong dollar
against the U.S. dollar), SP500 (the daily change of the Standard and Poors 500
Index of the U.S.), STI (the daily change of the Straits Times Index of Singapore),
SGX (the daily change of the exchange rate for the Singapore dollar against the
U.S. dollar).
For a time series, it is appropriate to formulate the following regression equation
for the Augmented Dickey-Fuller (ADF) test. The test will be conducted separately
for each time series data and for pre- and post-Asia Financial Crisis periods.
n
(1 − L)Y t = ␣0 + Y t−1 + Y i (1 − L)Y t−i + e (1)
i=1
where: Y is the series of daily variables being tested, L is the lag operator, ␣0 is an
estimated constant,  is regression coefficients, n is the number of lags differences
for E in approximately white noise,6 and e is the errors or residuals.
The null hypothesis to be tested for unit roots is that the series are non-stationary
in the data levels. For the Dickety-Fuller (ADF) test, if the absolute value of
calculated t-statistic is larger than the critical value provided by the ADF test,
the null hypothesis of unit roots (not stationary) is rejected, and the time series
variables are stationary.
Next, the natural log of the variables will be used to express relations that deal
with proportional changes in the variables for the Multiple Regression Analysis.
In this format, the coefficient of the independent variables will be the percentage
changes in the dependent variables per 1% change in each independent variable.
In addition, the variance inflationary factor (VIF) test will be conducted to test
the collinearity of the independent variables in order to set up better regression
models. If a set of explanatory variables is uncorrelated, then optimal VIF will be
equal to 1.7
Correlation Among Stock Markets 161
The following multiple regression model will be tested separately for each
country’s historical data and for the pre- and post-Asia Financial Crisis period.
Y = ␣0 + 0 X + 1 SP500 + e i (2)
where: Y is the percentage change in the daily stock index in Hong Kong or
Singapore, ␣0 = an estimated constant, 0 and 1 are regression coefficients, X is
the percentage change of daily exchange rate, SP500 is the percentage change in
the daily S&P 500 index, and ei is the errors or residuals.
Based on the observed U.S., Hong Kong and Singapore daily stock indices and
foreign exchange rates, the regression models are as follows:
where: lnHSI is the logarithm of daily Hang Seng Index measured by the percentage
changes in the daily closed Hong Kong Stock Market Index, lnSTI is the logarithm
of daily Straits Times Index measured by the percentage changes in the daily closed
Singapore Stock Market Index, ␣1 , ␣2 are intercept, an estimated constant, 1 , 4
are the coefficients of the percentage change in the Hong Kong dollar or Singapore
dollar’s exchange rates and the daily return of the S&P 500, HKX is the percentage
change in HK$ daily exchange rate against US$, SGX is the percentage change in
SG$ daily exchange rate against US$, lnSP500 is the logarithm of S&P 500 index
proxied for percentage change in the daily closed U.S. stock market, and e is the
error or residual term.
In addition to examining the correlation of variables for the exchange rate and the
stock market, this study investigates the short-run dynamics by Granger causality
tests. Because correlation does not necessarily imply causation in any meaningful
sense of that world, Granger causality tests have been used frequently to investigate
short run relationships among two or more variables. A high degree of causality
from one variable to another indicates that the two variables are integrated and
that changes in one variable will tend to cause changes in the other. Thus, the
Granger approach to the question of whether X causes Y is to see how much of the
current Y can be explained by the past value of Y and then to see whether adding
lagged values of X can improve the explanation. Y is said to be Granger-caused by
X if X helps to predict Y, or equivalently if the coefficients on the lagged Xs are
statistically significant.
The causality issue between stock markets and the exchange rate is determined
with the following equation,
HSI A10 A11 (L)
A12 (L) A13 (L) HSIt−1
HKX = 20 + A21 (L)
A A22 (L) A23 (L) × HKXt−1
SP A30 A31 (L) A32 (L) A33 (L) SPt−1
eHSI
+ eHKX (5)
eSP
STI A10 A11 (L)
A12 (L) A13 (L) STIt−1
SGX = A20 + A21 (L) A22 (L) A23 (L) × SGXt−1
SP A30 A31 (L) A32 (L) A33 (L) SPt−1
eSTI
+ eSGX (6)
eSP
where Aio are the parameters representing intercept terms and Aij are the
polynomials in the lag operator L. The lag structure will be arbitrarily tested from
2 to12 lags.
The following six simple hypotheses will be tested to find the pair-wise short-run
relationship (Granger Causality) between the variables for Hong Kong,
H0 : HKX does not cause HSI, H1 : HKX does cause HSI. (7)
H0 : HSI does not cause HKX, H1 : HSI does cause HKX. (8)
H0 : SP does not cause HSI, H1 : SP does cause HSI. (9)
H0 : HSI does not cause SP, H1 : HSI does cause SP. (10)
H0 : SP does not cause HKX, H1 : SP does cause HKX. (11)
H0 : HKX does not cause SP, H1 : HKX does cause SP. (12)
Correlation Among Stock Markets 163
The following six simple hypotheses will be tested to find the pair-wise short-run
relationship (Granger Causality) between the variables for Singapore,
H0 : SGX does not cause STI, H1 : SGX does cause STI. (13)
H0 : STI does not cause SGX, H1 : STI does cause SGX. (14)
H0 : SP does not cause STI, H1 : SP does cause STI. (15)
H0 : STI does not cause SP, H1 : STI does cause SP. (16)
H0 : SP does not cause SGX, H1 : SP does cause SGX. (17)
H0 : SGX does not cause SP, H1 : SGX does cause SP. (18)
The null hypotheses will be rejected if the calculated F-statistics are significant.
It is normal to say that there exists a causation that one variable Granger causes
the other.
6. EMPIRICAL RESULTS
The unit root test was performed to check the stationarity of all the variables. The
unit root test results, as shown in Table 1, indicate that all the absolute values of the
calculated ADF test statistics are larger than the absolute value of the MacKinnon
critical values at the 1, 5 and 10% significance level, and thus, the null hypothesis
of a unit root will be rejected. This suggests that all the time series variables for
the regression model are stationary.
The variance inflationary factor (VIF) was used to measure collinearity of the
variable data. The VIF statistics and variance analysis as shown in Tables 2–5
are in the acceptable range from 1.00 to 1.80 for the pre- and post-crisis period.
Therefore, there is no reason to suspect any colliearity for the set of explanatory
variables, and the appropriate regression analysis can be performed.
The multiple regression estimates, as show in Tables 2 and 3, provide the
following relationship between Hang Seng Index (HSI) and daily changes in
exchange rate (HKX) and the U.S. stock market (SP500 ). The multiple regression
analysis of the data, shown in Tables 4 and 5, indicate the strength of the linear
relationship between Singapore’s STI Index and the foreign exchange rate (SGX)
and the U.S. stock market index (S&P500 ). The results can be summarized as shown
below:
Source DF SS MS F P
Source DF SS MS F P
Source DF SS MS F P
Source DF SS MS F P
In the case of Hong Kong, the estimated coefficients of the foreign exchange
rate (−16.62) for the pre-crisis period, and in the case of Singapore, the estimate
coefficient of the foreign exchange rate (−3.19) for the post crisis period may
appear somewhat controversial. The inverse relationship between changes in HSI
and changes in exchange rate in the pre-crisis period, and between changes in
STI and changes in exchange rate during the post-crisis period, can be explained
by the inflation factor. A higher rate of inflation in the home country forces the
domestic currency to lose its value, and hence, investors will demand a higher risk
premium and a higher rate of return, which causes stock prices to fall. Therefore,
depreciating the Hong Kong dollar (or the Singapore dollar) represents higher
inflationary pressure in Hong Kong (or Singapore), which could explain each
country’s sluggish economy and depressed stock market.
However, in the pre-crisis period of the Singapore case and the post-crisis
period of the Hong Kong case, there exists a positive correlation, and this agrees
with the previous findings of positive relationship. In the macro economic view,
depreciating the Hong Kong (or Singapore) dollar will increase exports of the
corresponding country, which usually stimulates their export-oriented economy,
and results in a booming Singapore (or Hong Kong) stock market. In this scenario,
lack of statistical significance is consistent with findings in similar studies involving
different currencies and stock markets.
The correlation coefficients for the U.S. S&P stock index are quite different from
the coefficients of the exchange rate. These two variables resulted in a significant
positive correlation between the Asian Stock markets and the U.S. stock market.
In addition, the movements of the U.S., Hong Kong, and Singapore stock markets
grew stronger during post-crisis period.
The standard errors of the regression, which describes the dispersion of data
points above and below the regression line, indicate that there would little variance
between the predicted values and the actual values. The residual analysis reveals
that the errors are normally distributed, as they are linearly related to HSI and STI,
and the linear assumption holds for this model.
The R2 and adjusted R2 are high enough to conclude that a large portion of
changes in Hang Seng Index can be explained by the Hong Kong dollar’s exchange
rate movements and the S&P 500 daily return during both sub-periods. However,
both statistics indicate a somewhat weaker relationship between the explanatory
variable and the Straits Times Index (Singapore’s stock market index).
The F-statistic is used to determine the significance or existence of the regression
line. At a 5% level of significance, the output F-statistics for both of the Hang Seng
Index (HSI) and the Straits Times Index (STI) exceed the critical value on the F
distribution (with 2 and 453 degree of freedom) which is 3.07. In addition, the
P-value of the F-test for both regression estimates during the pre- and post-crisis
168
Table 6. Granger Causality Test for Hong Kong.
Null Hypothesis Lags: 2 Lags: 4 Lags: 6 Lags: 8 Lags: 10 Lags: 12
F-Stat. Prob. F-Stat. Prob. F-Stat. Prob. F-Stat. Prob. F-Stat. Prob. F-Stat. Prob.
PAUL SARMAS
Correlation Among Stock Markets
Table 7. Granger Causality Test for Singapore.
Null Hypothesis Lags: 2 Lags: 4 Lags: 6 Lags: 8 Lags: 10 Lags: 12
F-Stat. Prob. F-Stat. Prob. F-Stat. Prob. F-Stat. Prob. F-Stat. Prob. F-Stat. Prob.
169
170 PAUL SARMAS
periods are 0.000 and less than critical value at 5% level. This indicates that at least
one of the explanatory variables is related to daily changes in HSI or STI, and the
regression lines are significant for the regression model to predict the movements
in these indices.
The results of the Granger causality test for the Hang Seng Index are illustrated
in Table 6. During the pre-crisis period, given the arbitrary 2, 4, 6, 8, 10 and 12 lags,
the calculated larger F-statistics and zero probabilities reject the null hypothesis
that the daily return of the S&P 500 index does not Granger-cause the daily return of
the Hang Seng Index. This depicts a unidirectional causality from the daily return
of the U.S. equity market to the daily return of the Hong Kong equity market.
In the post-crisis period, the unidirectional causality starts to appear in the lag
interval of 6, 8, 10 and 12, which reveal the postponed but stronger influence of
the daily change of the U.S. stock market on that of the Hong Kong stock market
after the crisis. Lack of unidirectional or bi-directional causality is evident between
the exchange rate changes and the two equities markets. However, there exists a
unidirectional causality from the S&P 500 daily return to the exchange rate daily
return given 2 and 4 lags period.
The results of the Granger causality test for the Straight Times Index are
summarized in Table 7, which shows no unidirectional or bi-directional causality
between the daily changes of the exchange rate and the two equity markets, either
pre-crisis or post-crisis given the arbitrary 2, 4, 6, 8, 10 and 12 lags. As Hong Kong,
the calculated larger F-statistics and zero probabilities reject the null hypothesis
that the daily return of the S&P 500 index does not Granger cause the daily
return of the Straits Times Index. This indicates a unidirectional causality from
the daily return of the U.S. equity market to the daily return of the Singapore
equity market, either before or after the crisis. In the post-crisis period, there is
a stronger unidirectional causality from the U.S. equity market to the Singapore
equity market, given the same lag interval of 2, 4, 6, 8, 10 and 12. This supports
the multiple regression test result.
under the different foreign exchange rate systems pre- and post-Asian Financial
Crisis in the mid 1997 and 1998.
The initial step involved examining this relationship by applying a Multiple
Regression analysis, which tests for the correlation between these countries’ stock
markets and their exchange rates. Then, the Granger Causality test was performed
to study the further interactions between the various markets under the different
exchange rate systems.
The empirical results show that different exchange rates systems have no signifi-
cant role in determining the linkage between the two Asian countries stock markets
and the U.S. stock market. The multiple regression analysis indicates that the
relationship between the stock markets and the exchange rate were inconsistent and
conflicting both before and after the crisis. The exchange rate had a positive effect
on the Hong Kong stock market in the pre-crisis period, but a negative effect in the
post-crisis period. The exchange rate had a small inverse effect on the Singapore
stock market in the pre-crisis period, but positive effect in the post-crisis period.
The further investigation by the Granger Causality test indicates no short-run re-
lationship for the Hong Kong and Singapore exchange rates and their equity market
or the U.S. stock market before the crisis. Nevertheless, after the crisis, there existed
a short-run linkage for the daily changes of the Hong Kong exchange rate and the
daily return of S&P 500. This also supports the inconsistent relationship between
the exchange rates and the equity markets. Both positive and negative effects may
be explained with the inflationary disturbance and the macro economic view.
The complex relationship between stock prices and foreign exchanges can be
supplemented by the Fisher effect and financial theories. However, the impact
of change in the fixed exchange rate on the stock market was too great when
interpreted from the regression output. It is impractical to see the dramatic effect on
the stock market after the change of the exchange rate because consequent dynamic
market adjustments eliminate the results as predicted by the efficient markets theory
and the equilibrium theory. Additionally, Hong Kong’s fixed-linked exchange rate
system limits such change. Therefore, the inference of the exchange rate effect on
the equity market is invalid because the system is shown to be incomplete without
considering other important factors.
The empirical investigation also implies that both of the Hong Kong and
Singapore stock markets were stable and positively correlated with the U.S. stock
market in the periods of pre- and post-crisis. This reflects the strong effect of the
U.S. stock market on the movements of the Asian stock markets. The two tests
also reveal that the movements of the Hong Kong and Singapore stock markets
were closer to those of the U.S. before and after East Asia Financial Crisis.
The causality test reveals only the short-run unidirectional causality from the
U.S. equity market to the two Asian equity markets. In Singapore, there were
172 PAUL SARMAS
consistent effects of the daily U.S. stock market fluctuations on the daily changes
of the Singapore market, either pre-crisis or post-crisis. In contrast, in Hong Kong,
the daily S&P 500 return had an immediate effect on the daily Hang Seng return
before the crisis. However, after the crisis, this effect was delayed until the influence
of the daily exchange rate changes given more lag periods. All these effects support
the previous findings that the linkage was increasing in a chronological order
between the Asian equity markets and the U.S. equity market in the 1990s. The
study further reveals that there were different degrees of the linkage between the
two Asian countries and the U.S. stock markets, and the East Asian Crisis appears
to differ in the degree of that linkage.
NOTES
1. For a detailed discussion refer to Zheng (2001).
2. For further study, please refer to Eiteman (1997) and Shapiro (2003).
3. Please refer to Monetary Authority of Singapore web site (www.mas.gov.sg).
4. There are evidence that both the Hong Kong and the Singapore financial markets
are influenced by the stock market movements in the U.S., the UK and Japan. For further
examination refer to the article by Chan, Cup and Pan (1992). However, this study focuses
on the correlation between the U.S. stock market and those of Hong Kong and Singapore,
in order to account for the effects of the two exchange rate systems: Hong Kong Dollar’s
fixed rate system versus Singapore Dollar’s floating rate system.
5. Differencing is the process of calculating the change in the value of a variable in
successive time periods. Integration refers to the degree of differencing that a datum requires
for it to be transformed into a stationary series (Solnik, Bourcelle & Le Fur, 1996).
6. The equation assumes that a time series is “white noise”, which means the variables
have zero mean, a constant variance, and zero correlation between successive observations
(Leady & Ormrod, 2003).
7. A more conservative criterion would employ alternatives to least-square regression if
the maximum VIF exceeds 5 (Levine, Berenson & Stephan, 1999).
8. Vestor Autoregressive (VAR) is a system in which every equation has the same right
hand variables, and those variables include lagged values of all the endogenous variables.
(Levine, Berenson & Stephan, 1999).
REFERENCES
Arshanapalli, B., & Doukas, J. (1995). Pre and post-October 1987 stock market linkages between U.S.
& Asian markets. Pacific-Basin Finance Journal, 3, 1–14.
Awad, M. A., & Goodwin, B. K. (1998). Dynamic linkages among real interest rates in international
capital markets. Journal of International Money and Finance, 453–519.
Bahmani-Oskooee, M., & Sohrabian, A. (1992). Stock prices and the effective exchange rate of the
Dollar. Applied Economics, 459–464.
Correlation Among Stock Markets 173
Chan, K. C., Cup, B. E., & Pan, M. S. (1992). An empirical analysis of stock prices in major Asian
markets and the United States. The Financial Review, 27, 289–308.
Eiteman, D. K. (1997). Multinational business finance (8th ed.). Reading, MA: Addison-Wesley.
Federal Reserve Bank of St. Louis (FRBS). Internet address at URL: www.stls.frb.org/fred/data/
exchange.html.
Hong Kong Monetary Authority Monthly Bulletin. Internet address at URL: www.info.gov.hk/
hkma/eng/statistics.
Hung, B., & Cheung, Y. L. (1995). Interdependence of Asian emerging equity markets. Journal of
Economic Dynamics and Control, 4, 243–259.
Levine, D. M., Berenson, M. L., & Stephan, D. (1999). Statistics for managers, using Microsoft Excel,
Phstat Excel ‘98 Add-In (2nd ed.). New York: Prentice Hall.
Leady, P. D., & Ormrod, J. E. (2003). Practical research, planning and design (8th ed.). Upper Saddle
River, NJ: Prentice Hall PTR.
Mok, H. M. K. (1993). Causality of interest rate, exchange rate and stock prices at stock market open
and close in Hong Kong. Asia Pacific Journal of Management, 123–143.
Monetary Authority of Singapore. Internet address at URL: www.mas.gov.sg.
Phylaktis, K., & Ravazzolo, F. (2001). Stock price and exchange rate dynamics. Department of Banking
and Finance Working Papers. City University Business School, 321–244.
Shapiro, A. C. (2003). Foundations of multinational financial management (4th ed.). New York: John
Wiley and Sons, Inc.
Solnik, B., Bourcelle, C., & Le Fur, Y. (1996). International market correlation and volatility. Financial
Analysts Journal, 1734–1745.
Wong, K. A. (1995). Is there an intra-month effect on stock returns in developing stock markets?
Applied Financial Economics, 5, 285–289.
Wu, Y. (2000). Stock prices and exchange rates in a VEC model – the case of Singapore in the 1990s.
Journal of Economics and Finance, 36–48.
Zheng, W. Y. S. (2001). Hong Kong and Singapore equity market’s U.S. correlation under the different
exchange rate system. MBA Thesis. Pomona, CA: California State Polytechnic University.
Zikmund, W. G. (2000). Business research methods (6th ed.). New York: Dryden Press, Harcourt
College.
MULTIPLE BANKING AS A
COMMITMENT NOT TO RESCUE
Paul Povel
ABSTRACT
We show why investors may prefer not to be a firm’s unique lender, even if
they are in a strong bargaining position. Some firms need additional funds
after a first investment: providing additional funds is rational after the first
investment is sunk, but together the two investments are unprofitable. A unique
lender will always provide additional funds and make losses. Two creditors
can commit not always to provide funds: inefficient negotiations over debt
forgiveness may end with a project’s liquidation, which is harmful ex post,
but helpful ex ante, if it keeps entrepreneurs with nonpromising projects from
initially requesting funds.
1. INTRODUCTION
This paper analyzes a bank’s incentives to forgive debt and refinance a distressed
firm. We compare the decision of a unique lender with that of two banks, which
have jointly provided a loan to the firm. We show that banks may prefer such co-
financing, even if they enjoy a strong bargaining position relative to the firm. The
main difference between single and multiple banking lies in the negotiations that
are necessary, if the firm cannot repay its debt but it could profitably be refinanced.
Suppose that refinancing is profitable, once an initial investment is sunk, but
that ex ante it is not. Some firms will need refinancing, others not, and the
Research in Finance
Research in Finance, Volume 21, 175–199
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21008-5
175
176 PAUL POVEL
creditors would like to finance the latter, only. The entrepreneurs of the respective
firms, however, who are informed about their prospective financial needs, are only
interested in receiving a loan, irrespective of whether it will be performing well or
badly. If the creditors could commit not to refinance a firm, the entrepreneurs with
ex ante unprofitable firms would prefer to be inactive, instead of being forced to
liquidate their firm prematurely. A single lender cannot credibly commit to being
tough, as it is always sequentially rational to refinance a distressed firm, once the
initial loan is sunk. We argue that introducing multiplicity on the side of the lenders
can make such a commitment possible. Even if they agree on the need to rescue
the firm, two lenders will have to bargain about the distribution of the overall
loss. Asymmetric information between the banks is the cause of inefficiencies in
the rescue decision: with positive probability the firm is not refinanced, and it is
liquidated, instead.
There is a large literature now, which analyzes the effects of single or multiple
lending on the decisions of a firm. One strand of the literature analyzes the effects
that the structure of the creditors’ claims has on the possibilities to reorganize
a distressed firm. Gertner and Scharfstein (1991) and Detragiache (1994) for
instance assume that bonds are held by atomistic investors and therefore cannot
be renegotiated. They analyze the effects of different bankruptcy regimes on the
possibilities to reorganize a distressed firm.
These effects can be used strategically by a firm, i.e. different financial structures
can be used to achieve different goals. Several papers have asked the question why
a firm may prefer to have one or many creditors. The difference between the
market-based financial system in the U.S. and the bank-based system in Germany
and Japan are striking, and an analysis of the relative advantages of the two systems
is an important research program.
A frequently stated advantage of the “main bank” financial system in Germany
and Japan is that distressed firms are rescued more frequently (see e.g. Hoshi et al.,
1990, for the case of Japan, and Edwards & Fischer, 1994, for the case of Germany).
Some theoretical papers have analyzed the conditions under which “main bank”
finance is more efficient than a system with multiple lenders (see e.g. Dewatripont
& Maskin, 1995; Fischer, 1990; von Thadden, 1995). As Edwards and Fischer
(1990) conclude, however, these models are not compatible with the empirical
evidence for the German case. While in the models at most one “main bank” can
emerge, in reality a German firm has more than one “Hausbank.” The question to
analyze is thus why we may observe more than one nonatomistic lender. Several
answers are possible.
First, one could argue that banks are risk averse and want to spread out their
risk exposure by sharing risks with their competitors. This is certainly true, but not
a very satisfying explanation from a theoretical point of view. Banks are usually
Multiple Banking as a Commitment Not to Rescue 177
thought of as “large,” compared with the size of the average firm. They should
therefore be able to diversify away most of their risks, as was modeled in Diamond
(1984). This makes them de facto risk neutral, and they should not suffer from risk
exposure. After all, it is the banks’ business to deal with risks and to allocate them
optimally, and not to avoid risks. Additionally, it would be interesting to know
whether there is more behind multiple banking than mere risk-sharing.
Second, a bank may lack the funds to finance a project. Dewatripont and
Maskin (1995) suggested that such smallness could be a solution to the soft
budget constraint problem in centralized economies. Inability to finance a project
exclusively may be a real problem when firms are very large. However, even in cases
when the firms are very small, compared with their banks, we find multiplicity. As
before, there is a need for additional explanations.
Third, firms may want to have many banks because this protects them from
being exploited by too strong a partner, as was suggested in von Thadden (1992).
This third rationale for multiple banking implies that neither the banks nor the
firms enjoy exceptionally strong bargaining positions in their relationship. This
contrasts with the general perception that in bank-dominated financial systems,
banks are in a stronger position. Many situations can occur in which a firm has
to rely on its bank or banks and in which the bank can cheaply “punish” earlier
unfriendly behaviour.
Finally, some authors analyze the use of multiple claimants, holding different
types of securities, in solving agency problems: the investors may have poor
incentives either to really monitor their debtor, or to make proper use of their
information (e.g. to liquidate a firm). See e.g. Diamond (1993), Berglöf and von
Thadden (1995), Dewatripont and Tirole (1994), Rajan and Winton (1995), and
Repullo and Suarez (1995).
The present paper offers a rationale for multiplicity, which complements the
explanations above. We argue that multiplicity is requested by the banks, who use
it as a commitment device for eventual renegotiations of the lending contracts.
The inefficiencies that arise in rescue negotiations (the banks have to determine
their respective degrees of debt forgiveness) are a threat for entrepreneurs with bad
projects. If the inefficiencies are sufficiently strong, this allows the banks to deter
nonprofitable projects, and to finance high quality ones, only.
The idea that multiplicity can serve as a commitment device was first stated in
Hellwig (1991). Dewatripont and Maskin (1995) analyze the role of “multiple
lending” in hardening the “soft budget constraint” of a firm. In their model,
however, multiplicity is a credible commitment not to rescue only because of
the assumption that lenders are “small,” and cannot provide both an initial and a
refinancing loan. Bolton and Scharfstein (1996) analyze a renegotiation problem
that is similar in spirit to ours. In their model, too, multiplicity is used as a
178 PAUL POVEL
The models in the three papers, however, are somewhat technical, too, and do not
generate more elegant results than our model. We believe, therefore, that the war
of attrition is a good compromise between the requirements for the analysis and
the tractability of the results.
The rest of the paper is structured as follows: In Section 2, the projects and the
entrepreneurs are introduced, and the difficulties of a single bank are discussed.
The model is extended in Section 3, where two banks finance a firm, and renegotiate
if it must be refinanced. These renegotiations are modeled as a war of attrition.
Section 4 solves this model to find the equilibrium of the renegotiation stage, as
well as that of the whole game. Section 5 presents some empirical evidence, and
discusses implications and extensions of the model. Section 6 concludes. Proofs
are in the Appendix A.
There is a large number of entrepreneurs who can start one project each. Each
entrepreneur privately knows the type of project that he can start, either “good”
or “bad.” The proportion of entrepreneurs with “good” projects, ␥, is common
knowledge.
The timing of a project is the following. In the first period, an investment I must
be sunk. In period 2 the project types become publicly observable. Payoffs are
earned in the third (the last) period. A “good” project earns R > I, while a “bad”
project earns zero. Both project types can be liquidated, which earns r, where
0 ≤ r < R. A “bad” project can be “rescued” in period 2: if an additional amount
J is invested, a payoff R̄ is earned, instead of zero.
Assumption 1. It is profitable to rescue a “bad” project in period 2, as
R̄ − J > r. However, it is not profitable to finance a “bad” project ex ante:
R̄ − J − I < 0. Neither should a random sample of projects be financed:
␥(R − I) + (1 − ␥)(R̄ − J − I) < 0.
The entrepreneurs’ payoffs depend on whether their projects were started and
completed. If a project was not started, the entrepreneur earns zero utility. If the
project was started, and either completed successfully (if “good”) or rescued (if
“bad”), his utility is M > 0. If a project was started and then liquidated, this causes
harm to the entrepreneur, and his payoff is −m (where m > 0).
The entrepreneurs have no wealth of their own, and need outside finance to start
their projects. We assume that a project cannot be separated from its entrepreneur.
“Good” projects cannot be continued without him, and “bad” projects cannot
be rescued – both types would have to be liquidated. The entrepreneurs are
180 PAUL POVEL
The model with two banks incorporates some observations about private
workouts and bankruptcy negotiations that are reported in the business press, in
empirical and descriptive papers (e.g. Edwards & Fischer, 1994; Fischer, 1990),
in studies on the banking system and insolvency procedures in Germany, and
in the large literature on the reform of the bankruptcy laws in Germany. These
observations, or “stylized facts,” are:
(1) Banks seem to have a strong bargaining position.
(2) The parties involved try to keep the negotiations secret.
(3) The banks want to terminate the negotiations quickly.
(4) It is likely that customers and suppliers are lost if they hear that there are rescue
negotiations.
(5) Whether to rescue or not is rarely subject to dispute.
(6) The parties rather bargain about who is to sacrifice how much.
We have used these observations to construct a model of debt renegotiations,
such that it captures important elements of an existing financial system, and it
generates results which can again be compared with reality. To do so, we must
expand the model with a single bank, by adding some assumptions. Two comments
will be helpful before this is done. First, all additional assumptions could have been
added to the model with a single bank, without changing any of the results. This has
not been done, as it would have complicated the exposition unnecessarily. Second,
we will make assumptions that are much more restrictive than is necessary to
generate the results. Again, this is done to simplify the notation. Where assumptions
are “extreme,” we mention this fact, and discuss weaker alternatives.
We model the renegotiation process between the two banks of a firm as a war of
attrition. Each of the two banks tries to convince its opponent to carry the burden
of refinancing. An outside observer of the negotiations will find that no progress
is being made for a while: the banks fail to come to an agreement on how to split
the overall loss R̄ − I − J, if there should be a rescue. The negotiations can end in
two different ways. Either one of the banks gives in, i.e. it accepts the rescue plan
of its opponent. Or fate turns against the firm: a rescue becomes impossible for
exogenous reasons, and it must be liquidated. In the latter case, each bank incurs a
loss (additional to the financial loss). The size of this loss is privately known by the
respective bank. In equilibrium, the higher it is, the more a bank fears liquidation,
and the less it is willing to reject its opponent’s rescue plan.
We now introduce the extensions of the single banking model, incorporating the
observations listed above. The equilibrium of the war of attrition will be analyzed
in Section 4.
The first observation above states that banks are the main players in rescue
negotiations. This is captured by assuming that they are the only bargaining parties,
182 PAUL POVEL
and by assuming that the courts strictly enforce Absolute Priority Rules. These rules
specify that no party may receive any of the returns of the firm, if the banks have
neither been repaid in full, nor have agreed to such a payment.
Observation 2 describes how the banks want to keep the negotiations secret. It
is helpful in achieving this goal to conclude an agreement as quickly as possible
(see Observation 3). The reason for this wish for secrecy lies in the bankruptcy
laws, which in most countries favour the banks (France is a notable exception). The
assets of the firm usually are used as collateral for the loans from the banks, and
absolute priority rules enforce the need to repay these claims first. The customers
and suppliers are the parties who typically do badly in bankruptcy. Similar to a
bank run, they have every incentive to request what they are owed, as soon as they
discover the firm’s problems, and not to engage in any new trades (except possibly
on a cash-only basis). We model this sensitivity of a rescue to the cooperation of
these parties as a heavily reduced form of Observation 4.
Assumption 2. At any time during the rescue negotiations, the public can
discover that there are such negotiations going on. This happens by the time t
with probability F(t). If the negotiations have been discovered, a rescue becomes
immediately impossible, and the project must be liquidated.
conclude that restructuring is the more likely, the more debt is owed to banks. This
may be caused by the banks’ superior skills and capabilities in attempting to rescue
a firm, but it may also signal that banks are more willing to rescue a firm than other
creditors. In the model this willingness to rescue is captured by the valuation i .
Assumption 4 and the next assumption jointly capture Observation 6, that the
banks bargain about who has to bear how much of the loss. The set of outcomes
that the banks can achieve is restricted to simplify the analysis, that is how the net
surplus s (the returns R̄ minus the cost J and the opportunity cost r) from rescuing
can be split (it is positive because of Assumption 1).
Assumption 5. The banks fight for the whole surplus s := (R̄ − r − J). No
offer to share the surplus is made or accepted. If one bank gives in it receives its
share r i in the liquidation value r of the firm from the other bank, where r 1 and
r 2 are specified in the initial contract. The winning bank is committed to rescue
the firm immediately, but may keep the returns for itself.
As before (in Assumption 2), the formulation of Assumption 5 is much stronger
than necessary. A sharing rule saying that the gross surplus R̄ − J can only be
shared in proportions ␣ and (1 − ␣), where ␣ = (1/2), would be sufficient. This
would lead to significant complications of the analysis, however, which are not
rewarded by the additional insight that one gets.
This completes the introduction of the model with two banks. As one can easily
see, the assumptions that have been added in this section could also have been
introduced in the single bank model, without changing anything. The loss of a
valuation i if the firm is liquidated would make a single lender even more willing
to refinance a “bad” project. This rescue happens already without the valuation,
however.
In the model with one bank, a strategy for the bank consisted of a financing and
a refinancing decision. In the case with two banks it is slightly more complicated.
We first consider the part of the strategy which is used in the rescue negotiations.
If a firm needs refinancing, the sequence of events is the following. First, the banks
decide whether they want to give in immediately. If none of the banks has given in,
the negotiations are discovered with probability , and the firm must be liquidated.
With probability (1 − ) the continuous time war of attrition starts. We assume
that if both banks give in simultaneously, each “wins” with probability 1/2.
A strategy is a function T i : + → [2, ], which determines for each moment of
time whether a bank B i with valuation i should give in or not. It will be shown
in Section 4, that if the equilibrium strategy tells this bank to stop at time T i (i ),
it will stop at every later time, as well. Thus, we will define T i as determining the
first time at which a bank plans to stop. This includes the static lottery which is
played because of the mass point in F at t = 2.
Multiple Banking as a Commitment Not to Rescue 185
One may wonder why the banks cannot renegotiate the lending contract, after
it has been signed. Both are fully aware of the inefficiency that will arise, if the
contract is renegotiated using the war of attrition. Why cannot one bank (or a third
bank) take over all debt for a flat price? Suppose B 1 would make such an offer to
B 2 . B 2 would claim to have a valuation 2 = 0 and not to fear the war of attrition,
in order to increase the takeover price. B 1 would claim to have the same valuation,
to decrease the price. None of the two has any incentive to admit having a positive
valuation, until a rescue is really needed. In this case, however, the war of attrition
will start. The time that passes by is the only credible information about one’s
valuation, as talk is “cheap,” and neither before nor during the war of attrition the
parties can renegotiate more efficiently. Even a bank with valuation i = ∞ would
wait until a rescue is necessary, as it might be that the opponent gives in. Nothing
is lost by waiting until t = 2, at which time both banks can prevent a liquidation
with probability one by giving in.
4. EQUILIBRIUM STRATEGIES
The first step in solving the renegotiation game is to determine which types
would want to start the war of attrition, and which types would prefer to give
in immediately, in order to secure the rescue of the firm. If no bank gives in
immediately, the negotiations are discovered with probability (the mass point in
F), and the firm is liquidated. With probability (1 − ) the continuous time war of
attrition starts.
A bank with a very high valuation at stake will not want to gamble for the
surplus s, and stop immediately. We must determine which is the lowest valuation,
for which this is still true. Denote this cut-off value of bank B i with i . If its
valuation is i > i , it should strictly prefer to give in immediately, while if
it is i < i , it should want to start the war of attrition, and plan to stop later
than t = 2.
We define i as the valuation with which a bank B i is indifferent between giving
in immediately, and starting the war of attrition, if it is sure that the opponent will
either give in immediately (with probability 1 − G(2 )), or will start the war of
attrition without giving in (with probability G(2 )).
Consider the bank with valuation i = i − , where > 0. Given the definition
of i , there must be a ␦ > 0, such that it will strictly prefer to start the negotiations,
if the probability that the opponent gives in immediately, as soon as the negotiations
have started, is ␦. Thus, a bank with a valuation below i has an incentive to hold
out for a strictly positive amount of time. A bank with a valuation higher than i ,
however, strictly prefers to give in immediately.
186 PAUL POVEL
not “bygones” in this game: every second that passes by signals information about
a bank’s valuation, and is relevant for the present and future decisions of the
opponent.
As was mentioned above, the finiteness of is not a necessary condition for the
tractability of the model. If the function f had an infinite support, then Lemma 2.(iv)
would state that banks with zero liquidation loss never stop, and banks with strictly
positive loss levels plan to stop at some finite time.
L i , the inverse of the strategy function T i , is the lowest cost level that would make
bank B i want to stop at time t. It will be helpful for characterising the equilibrium
strategies in the following. These are determined by finding for each moment t 1
a valuation L 1 (t 1 ), such that B 1 is exactly indifferent between stopping at t 1 , and
waiting for a small amount of time , and giving in then (the derivation is similar
to that of the cut-off values i ).
If the bank gives in at time t 1 , its payoff is r 1 for sure. We require this payoff to
be equal to the expected payoff, if it decides to wait until t 1 + :
G(L 2 (t 1 + )) (F(t 1 + ) − F(t 1 )) (1 − F(t 1 + ))
r1 = (r 1 − L 1 (t 1 )) + r1
G(L 2 (t 1 )) 1 − F(t 1 ) 1 − F(t 1 )
G(L 2 (t 1 )) − G(L 2 (t 1 + )) F(t 1 + ) − F(t 1 )
+ (r 1 − L 1 (t 1 ))
G(L 2 (t 1 )) 1 − F(t 1 )
1 − F(t 1 + )
+ (R̄ − r 2 − J) . (3)
1 − F(t 1 )
The second expected payoff (on the right-hand side of Eq. (3)) has four components.
The opponent may have a low valuation, and plan to give in later than t 1 + .
By this time, the negotiations may have been discovered, and the firm must be
liquidated. The bank receives its share r 1 of the liquidation value r, but loses
L 1 (t 1 ). If the negotiations are not discovered, it will give in at time t 1 + , which
earns r 1 . On the other hand, the opponent may plan to give in between t 1 and
t 1 + . As before, the negotiations may be discovered, or they may not. In the
latter case, the firm is rescued. The bank pockets the surplus R̄ − J, and pays r 2
to the opponent. We abstract from the possibility that both may give in at t 1 +
simultaneously, as the probability that this happens is negligible.
Equation (3) can be simplified by rearranging, subtracting r 1 on both sides, and
by substituting s for (R̄ − r 1 − r 2 − J). A division of both sides by leads to
G(L 2 (t 1 )) − G(L 2 (t 1 + )) 1 − F(t 1 + )
s
G(L 2 (t 1 )) 1 − F(t 1 )
F(t 1 + ) − F(t 1 )
=− L 1 (t 1 ). (4)
(1 − F(t 1 ))
188 PAUL POVEL
Since the strategies are differentiable everywhere it is possible to take the limit as
goes to zero. The same procedure can be repeated for the second bank, and we
get a system of two differential equations:
G(L 2 (t 1 )) f(t 1 ) L 1 (t 1 )
L 2 (t 1 ) = − , (5)
g(L 2 (t 1 )) 1 − F(t 1 ) s
G(L 1 (t 2 )) f(t 2 ) L 2 (t 2 )
L 1 (t 1 ) = − . (6)
g(L 1 (t 2 )) 1 − F(t 2 ) s
Given the strategy of the opponent, Eq. (5) determines the optimal response of
bank B 1 , if it has loss level L 1 (t 1 ) = 1 (the two are equivalent, if the equilibrium
strategy tells bank B i with cost level i to stop at time t i ) and bank B 2 plays strategy
L 2 (·). If Eq. (5) were an inequality, B 1 would either want to wait longer than t 1
(if <), or it would want to have stopped earlier (if >).
Since by Assumption 2 the probability density function g is strictly positive on
−1 : [0, 1] →
+ , G has an inverse function G + . Equations (5) and (6) can be
integrated, and this leads to the following reaction function for bank B i :
ti
f(t) L i (t)
L j (t i ) = G −1 G()exp − dt . (7)
2 1 − F(t) s
Equation (7) implicitly describes the strategy of bank B j that makes bank B i
exactly indifferent between stopping at t i and stopping at t i + (where is a
small amount of time), given its cost level L i (t i ). The analogous can be done
to derive the strategy of the other bank. The solution to these two equations
will give us the equilibrium strategies for the banks. We will continue with the
differential equations (5) and (6), and show that there is a unique equilibrium. The
reaction functions will be helpful in Section 5, where we present some comparative
statics.
With the help of the differential equations and the boundary conditions it is
now possible to describe the equilibrium strategies of the players for the whole
renegotiation game.
Proposition 1. The renegotiation game has a unique symmetric Bayesian
equilibrium, which is implicitly described by the system of differential equations
(5) and (6), and the boundary conditions T 1 () = T 2 () = 0. The equilibrium
strategy for bank B i is to stop at time t if and only if i ≥ L i (t), where L i (t) is
determined in Eq. (7).
We can now find the equilibrium strategies for the whole game with two banks,
including the financing decision. Whether an entrepreneur with a bad project
applies for a loan in the first period depends on the probability with which his
Multiple Banking as a Commitment Not to Rescue 189
project is rescued in the second period. In Eq. (1) we determined an upper bound
q̄ to this probability, such that “bad” projects are not financed.
Proposition 2. If the probability of non-rescue due to bargaining delays is high
enough,
M
2 F(T 1 (1 ))G(1 )g(1 ) d1 ≥ , (8)
0 M+m
the entrepreneurs will apply for the initial loan if and only if the project is of
the “good” type.
Proposition 2 is the main result of the paper. There are cases in which a financial
system with multiple banking performs strictly better than one with single bank
lending. If the condition in Eq. (8) is met, the banks prefer to require co-financing
by a second bank to being a single lender.
5. EMPIRICAL IMPLICATIONS
The main result of the paper is that banks might want to syndicate a loan to a firm,
if they fear to find themselves in a harmfully weak bargaining position if the firm
has to be refinanced. The loan is shared for strategic reasons, and the banks propose
to share even if they have all bargaining power. There can be other reasons for why
loans are syndicated, however, like (see Section 1) risk aversion, the sheer size
of the loan, or because the strong competition on the lenders’ side. These reasons
complement each other, and it is not clear which one was the most important if a
loan has been shared.
There is some empirical work on this question for the U.S. and for Germany. For
the U.S., Gilson et al. (1990) have analyzed the performance of private workouts.
One of their results is that debt restructurings are more likely if the number of
lenders is small, which could support the result above. For the case of Germany,
Fischer (1990) and Edwards and Fischer (1994) report that all but the very small
firms have several “main banks,” which could be interpreted as supporting the
conclusions in this paper.
Interesting evidence is reported in Armendariz (1999). She analyzes the
performance of several development banks, i.e. the default rates of their loans.
Some of these banks require that projects are co-financed by commercial banks,
while others usually are the unique providers of capital. The former enjoy
considerably less arrears in the repayment of their loans. Her interpretation of these
facts is that the requirement of co-financing hardens the Soft Budget Constraint of
development projects, exactly what the results above suggest.
190 PAUL POVEL
A similar observation can be made if firms grow: suppose that for a small firm
R̄s − J s > I s , while for a larger firm R̄l − J l < I l . Then a “main bank” could
require that a growing firm finds a second main lender, for instance by committing
to finance only a fraction of a major investment. Similarly, a bank could require
co-financing if fixed costs of rescuing a firm are higher than the net surplus s for
small firms, but lower for larger firms.
We now analyze other implications of the model. The equilibrium strategies of
all parties are unique, and therefore we can analyze the effects of varying some of
the parameters of the model.
Proposition 3. A higher expected value of the firm R̄, a lower liquidation value
r and a lower additional loan J lead to later concessions. This in turn implies
that the liquidation of a “bad” firm becomes more likely.
The intuition behind Proposition 3 is clear: if the prize is increased, and the
expected costs of fighting remain unchanged, the banks have an incentive to fight
longer. The implications for rescue negotiations are surprising, however. Of two
otherwise identical candidates for a rescue, the one with a higher post-rescue
return R̄, i.e. the more profitable, is more likely to be liquidated. Similarly, the one
with a lower liquidation value is more likely to be liquidated. This seems to be
counterintuitive, as usually we would expect a valuable rescue to be undertaken.
The result follows from two modeling assumptions. First, the negotiations are
inefficient, as the “cake” that is to be split can disappear at any time. Second,
the banks’ valuations for the surplus from a rescue and for the rescue itself are
independent. Suppose that s depends on the number of employees of the firm, and
that the banks’ public relations suffer if they cause unemployment by not assisting
a distressed debtor (they lose i ). In this case we would expect a bank to be more
willing to rescue if the firm is larger.
A “valuable” firm could therefore be rescued for different reasons, either because
a rescue is profitable (large s), or because failing to rescue would cause indirect
costs (large i ). The second reason is an incentive problem that is similar to the one
underlying our assumption: once a project has been financed, its investors have
too strong incentives to refinance it (see Mitchell, 1993, or Aghion et al., 1999, on
the problems that this can cause for banking regulation).
The result should hold, however, in situations in which the valuations i
are small, compared with the surplus from a rescue, s. One could analyze the
refinancing decisions of foreign banks, that care less about their public image
outside their home country. Similarly, one could analyze these decisions in sectors,
regions, or during time periods, in which unemployment and bankruptcies are not
considered as being major problems.
Multiple Banking as a Commitment Not to Rescue 191
that have to be incurred if the negotiations continue are relevant for the decision
to stop.
Finally, the entrepreneurs’ utility functions are relevant. As m, the utility loss
that an entrepreneur incurs if his project is liquidated, increases, funds become
available for more parameter settings. Thus, there is a use in this model for the
stigma that is attached to a business failure. While we do not want to suggest
that this is a good way of solving incentive problems, we can conclude from the
model that the financing patterns of two regions or industries should be different
if bankruptcy is “not a big deal” in one of them, while it has strong negative
connotations in the other.
6. CONCLUSIONS
This paper studies the difference between single and multiple banking. It
concentrates on renegotiation problems, which are shown to be solved better in the
case of multiple banking. We assume that entrepreneurs ask banks for loans, such
that they can start projects. These may be of a “good” or “bad” type, where the type
of a project can be observed by the respective entrepreneur, only. “Bad” projects
need refinancing at an intermediate stage, which makes them nonprofitable from
an ex ante perspective. However, once the initial loan is lost, refinancing is better
than the only alternative, liquidation.
A single bank cannot commit not to refinance a bad project, which would
keep entrepreneurs with “bad” projects from applying for a loan. Two banks,
however, can commit not to refinance with some probability. The reason for
this are inefficiencies in the negotiations between the banks, when they have
to agree on their respective degree of debt forgiveness. If the probability of
liquidation is sufficiently high, entrepreneurs with “bad” projects do not ask for a
loan at all.
We model the negotiations as a war of attrition. Each of the two banks incurs
a privately known loss, if the firm is liquidated, and therefore would like to have
it refinanced. Additionally, refinancing is profitable, once the initial loan is sunk.
The banks have to agree on how to split the costs and revenues, if they refinance
the firm. These negotiations take time, and the longer they last, the more likely it
becomes that a rescue becomes impossible (for exogenous reasons). In order to
prevent this, the banks plan to “give in” after a while, i.e. to let the opponent pocket
the gain from rescuing, only to make sure that the firm is refinanced. There is a
unique equilibrium in this game: the higher the potential loss, the earlier a bank
decides to give in. The negotiations can last for a while, if both banks’ potential
losses are low, and therefore the firm is liquidated with positive probability.
194 PAUL POVEL
The model is designed to isolate the advantage of multiplicity for the lenders.
We thus abstract from many aspects which are relevant for the choice between
bilateral and multilateral finance, as well as for reorganisation procedures. One of
these is the tradeoff between single and multiple banking. Bolton and Scharfstein
(1996) analyze a case where either single or multiple lending may be optimal,
and also derive results for voting rules, as well as for the optimal use of assets as
collateral. Similarly, the effects of different bankruptcy laws need further analysis.
In the model the two banks decide to share the highest priority rank. It would be
interesting to analyze a model in which their claims have different ranks. A further
topic for future analysis is whether and how a distressed firm is rescued, if the
banks do not enjoy the highest priority rank.
NOTES
1. A simple alternative to the mass point assumption will be discussed below, see
Lemma 1.
2. A sufficient condition for uniqueness can be found by inverting 2 (1 ), and requiring
that the slope of this inverse is never equal to the slope of 1 (2 ). It is, however, difficult to
interpret:
2
g() g(G −1 [s/(2 + s)]) 2
=
∀ ∈ + .
[G()]2 [G(G −1 [s/(2 + s)])]2 s
REFERENCES
Abreu, D., & Gul, F. (2000). Bargaining and reputation. Econometrica, 68, 85–117.
Admati, A., & Perry, M. (1991). Joint projects without commitment. Review of Economic Studies, 58,
259–276.
Aghion, P., Bolton, P., & Fries, S. (1999). Optimal design of bank bailouts: The case of transition
economies. Journal of Institutional and Theoretical Economics, 155, 51–70.
Armendariz, B. (1999). Development banking. Journal of Development Economics, 58, 83–100.
Berglöf, E., & von Thadden, E.-L. (1995). Short-term versus long-term interests: Capital structure with
multiple investors. Quarterly Journal of Economics, 109, 1055–1084.
Bhattacharya, S., & Chiesa, G. (1995). Proprietary information, financial intermediation, and research
incentives. Journal of Financial Intermediation, 4, 328–357.
Birckhoff, G., & Rota, G.-C. (1978). Ordinary differential equations. New York: Wiley.
Bolton, P., & Scharfstein, D. (1996). Optimal debt structure with multiple creditors. Journal of Political
Economy, 104, 1–25.
Dewatripont, M., & Maskin, E. (1995). Credit and efficiency in centralized and decentralized
economies. Review of Economic Studies, 62, 541–555.
Dewatripont, M., & Tirole, J. (1994). A theory of debt and equity: Diversity of securities and
manager–shareholder congruence. Quarterly Journal of Economics, 109, 1028–1054.
Multiple Banking as a Commitment Not to Rescue 195
Diamond, D. (1984). Financial intermediation and delegated monitoring. Review of Economic Studies,
51, 393–414.
Diamond, D. (1993). Seniority and maturity of debt contracts. Journal of Financial Economics, 33,
341–368.
Edwards, J., & Fischer, K. (1994). Banks, finance and investments in Germany. Cambridge: Cambridge
University Press.
Fernandez, R., & Glazer, J. (1991). Striking for a bargain between two completely informed agents.
American Economic Review, 81, 240–252.
Fischer, K. (1990). Hausbankbeziehungen als Instrument der Bindung zwischen Banken und
Unternehmen. Unpublished doctoral dissertation, University of Bonn.
Gertner, R., & Scharfstein, D. (1991). A theory of workouts and the effects of reorganization law.
Journal of Finance, 46, 1189–1222.
Gilson, S., John, K., & Lang, L. (1990). Troubled debt restructurings. An empirical study of private
reorganization of firms in default. Journal of Financial Economics, 27, 315–353.
Hellwig, M. (1991). Banking, financial intermediation and corporate finance. In: A. Giovannini &
C. Mayer (Eds), European Financial Integration (pp. 35–63). Cambridge: Cambridge
University Press.
Hoshi, T., Kashyap, A., & Scharfstein, D. (1990). The role of banks in reducing the costs of financial
distress in Japan. Journal of Financial Economics, 27, 67–88.
Mitchell, J. (1993). Creditor passivity and bankruptcy: Implications for economic reform. In: C. Mayer
& X. Vives (Eds), Capital Markets and Financial Intermediation (pp. 197–224). Cambridge:
Cambridge University Press.
Rajan, R., & Winton, A. (1995). Covenants and collateral as incentives to monitor. Journal of Finance,
50, 1113–1146.
Repullo, R., & Suarez, J. (1995). Monitoring, liquidation, and security design. Review of Financial
Studies, 11, 163–187.
Schneider-Lenné, E. (1992). Corporate control in Germany. Oxford Review of Economic Policy, 8,
11–23.
von Thadden, E.-L. (1992). The commitment of finance, duplicated monitoring, and the investment
horizon. Working Paper, University of Basel.
von Thadden, E.-L. (1995). Long-term contracts, short-term investment and monitoring. Review of
Economic Studies, 62, 557–575.
Yosha, O. (1995). Information disclosure costs and the choice of financing source. Journal of Financial
Intermediation, 4, 3–20.
196 PAUL POVEL
APPENDIX A: PROOFS
Proof of Lemma 1
(A.1) compares the respective payoffs for bank B 1 with valuation 1 , given 2 :
R̄ − J − r
G(2 )r 1 + (1 − G(2 )) + r1
2
= (1 − G(2 ))(R̄ − r 2 − J) + G(2 ) (1 − )(R̄ − r 2 − J) + (r 1 − 1 ) .
(A.1)
The left-hand side of Eq. (A.1) is the expected payoff if bank B 1 gives in
immediately. With probability G(2 ) the opponent has a low valuation and does not
give in. The firm is rescued, and the bank receives r 1 . With probability 1 − G(2 )
the opponent gives in, as well, and the net surplus is shared (in expected terms).
The right-hand side of (A.1) is the payoff if the bank gives in as soon as the war of
attrition has started. With probability 1 − G(2 ) the opponent has a high valuation
and will give in immediately. The bank rescues, pockets the surplus R̄ − J, and
pays r 2 to the opponent. With probability G(2 ) the war of attrition starts. It
is discovered with probability , and the firm is liquidated. With probability
(1 − ), the game could continue, but by definition the bank plans to stop,
which earns r 1 .
Some simplifications of (A.1) and of an analogous equation for bank B 2 lead to
the two equations in Lemma 1. There is always an interior solution for the cut-off
levels: If i goes to zero, then j (i ) goes to infinity, while if i goes to infinity it
goes to zero.
Proof of Lemma 2
(i) We first show that T i is nonincreasing, and then that it is strictly decreasing. By
utility-maximisation it must be the case that
and
where Vi (t i , T j (·), i ) is the expected payoff of bank B i with cost level i , if it stops
at t i , and bank B j plays strategy T j (·):
The payoff of a bank depends on the chosen stopping time t i , the opponent’s
strategy T j and the (privately known) loss i of losing the firm. With probability
Pr{T j (j ) ≥ t i } the opponent plans to stop later than t i . If the public discovered
the negotiations (this happens with probability F(t i )), the payoff is (r i − i ). If the
secret was kept well, the bank receives r i from bank B j who rescues the firm. The
second term of Eq. (A.4)is the equivalent if the opponent plans to stop earlier. Here
the bank receives (R̄ − J) if the firm can be rescued and pays r j to the opponent.
We can rewrite the two inequalities Eqs (A.2) and (A.3) using Eq. (A.4).
Subtracting the RHS of Eq. (A.3) from the LHS of Eq. (A.2), and the LHS of
Eq. (A.3) from the RHS of Eq. (A.2), we get
Pr{T 2 (2 ) ≥ t 1 }F(t 1 )( 1 − 1 ) + F(T 2 (2 ))( 1 − 1 )g(2 ) d2
{2 |T 2 (2 )<t 1 }
≥ Pr{T 2 (2 ) ≥ t 1 }F(t 1 )( 1 − 1 ) + F(T 2 (2 ))( 1 − 1 )g(2 ) d2
{2 |T 2 (2 )<t 1 }
or, rearranging,
× ( 1 − 1 ).
Proof of Proposition 1
t ∈ [2, ], the (expected) payoffs from stopping or non-stopping can be compared,
as was done in deriving Eq. (5). Since the strategies are strictly decreasing, at
t < T 1 (1 ), i.e. if L 1 (t) > 1 , the payoff to bank B i with loss level 1 will be
higher if it waits. The opposite holds for L 1 (t) < 1 . For all t ≥ t 1 , type L 1 (t 1 ) can
only decrease his payoff by waiting, and will stop whenever possible.
The players constantly update their beliefs using Bayes’ Rule. If a player stops
at the wrong time (this is the only deviation that is possible) the opponent will
have no difficulties in updating his beliefs: If a player stops too early, the game is
over and beliefs are not relevant anymore. If a player waits too long, the strategy
tells him to stop immediately: Type i stops at any time t if t > T i (i ). Again, the
opponent can update his beliefs without problems.
Proof of Proposition 2
Proof of Proposition 3
The reaction curves L i (see Eq. (7)) are shifted outward, if s is increased. The
indirect effect via the cut-off value goes in the same direction: i (j ) is shifted
outward, as well (see Lemma 1).
As Proposition 3: analyze the equilibrium conditions Eqs (5) and (6), and the
indirect effect via the cutoff value in Lemma 1.
Proof of Proposition 7
ABSTRACT
This paper develops a model that explains how the creation of a futures
clearinghouse allows traders to reduce default and economize on margin.
We contrast the collateral necessary between bilateral partners with that
required when multilateral netting occurs. Optimal margin levels balance the
deadweight costs of default against the opportunity costs of holding additional
margin. Once created, it may be optimal for the clearinghouse to monitor
the financial condition of its members. If undertaken, monitoring will reduce
the amount of margin required but need not affect the probability of default.
Once created, it becomes optimal for the clearinghouse membership to expel
defaulting members. This reduces the probability of default. Our empirical
tests suggest that the opportunity cost of margin plays an important role in
clearinghouse behavior particularly their determination of margin amounts.
The relationship between volatility and margins suggests that participants
face an upward-sloping opportunity cost of margin. This appears to dominate
the effects that monitoring and expulsion might have on margin setting.
Research in Finance
Research in Finance, Volume 21, 201–227
© 2004 Published by Elsevier Ltd.
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21009-7
201
202 HERBERT L. BAER ET AL.
The field of competition . . . consists of all the individuals who are willing and able to recontract
about the articles under consideration.
F. Y. Edgeworth
Contract terms that specify posting of collateral effect a limit on the exposure to loss
from counterparty nonperformance. This outcome is obtained in two ways. First,
seizure of deposited collateral inherently reduces losses resulting from contract
nonperformance. Second, posting of collateral encourages contract performance.1
This paper studies decision making for instances where collateral requirements are
costly. We show that cost-avoidance within a competitive market reliably predicts
certain outcomes.
1.1. Introduction
low (and too risky).2 Time series evidence concludes that coverage ratios increase
(decrease) when coverage ratios are lower (higher) than their unconditional means.
Examining a cross section of margin-coverage ratios, we find that the opportunity
cost of margin deposits significantly influences the level of magin required from
exchange members. Our regressions suggest a negative relationship between
economy-wide shifts in the opportunity cost of margin deposits and levels of margin
coverage. In addition, we find a negative relationship between margin coverage
and participant-specific shifts in participants’ borrowing needs as proxied by levels
of implied standard deviation. These suggest that opportunity costs are important
and that market participants face upward-sloping schedules of opportunity costs
for their margin deposits.
They show that the values of exchange seats secure under-margined positions held
during the day. Their model complements ours by explaining the dynamics of
collateralizing against default during the trading day.
We develop the monitoring activities of exchanges as part of their intermediating
role. In this respect, the model resembles the delegated monitoring described by
Diamond (1984), and the modeling of risk management and financial guarantees
by Merton and Bodie (1992) and Hsieh (1993). We extend the earlier work by
explicitly incorporating certain institutional features of clearinghouses. These
features include expulsion from the clearinghouse, clearinghouse monitoring of
members’ financial condition and the possibility that members face increasing
costs for external funds. These have very different effects on optimal margin
setting and the probability of default. Our model also parallels Gorton’s (1985)
modeling of bank clearinghouses, particularly concerning expulsion and risk
mutualization.
two, the contract is settled. If x is positive and less than M(h, j), x is transferred
from the account of the short to the account of the long. Thus the short now has
M(h, j) – x; the long now has M(j, h) + x. If x is negative and |x| is less than
M(j, h) then x is transferred from the long to the short.
After contracts are settled, traders immediately restore their collateral-account
balances to M(j, h) and M(h, j), by either depositing cash when they are on the
losing side or withdrawing excess balances on gains. These account adjustments
restore equal values for the default options and counterparty risk of both
counterparties.
Figlewski (1984) shows that contract counterparties implicitly give each other
options to default. In the simplest case, contract default occurs whenever losses
exceed margin-account balances. Thus, if x is positive and greater than M(h, j),
the short rationally defaults on the contract and the long takes possession of the
margin assets M(h, j). Similarly, if x is negative and |x| is greater than M(j, h), the
long rationally defaults and the short takes possession of the margin assets M(j, h).
Remaining losses include recontracting costs, higher borrowing costs arising from
liquidity problems, and costs of financial distress. These are deadweight losses.
Agent j’s expectation of these losses is:
∞
D(j, h) = ␣N (x − M(h, j)f(x, ) dx) (1)
M(h,j)
where N is the net number of contracts j has open with h; i.e. the absolute value
after summing across N(j, h).
Agents jointly minimize the cost of contracting as would be obtained in a
perfectly competitive market. This is realistic provided agents freely choose among
a large number of counterparties each willing to minimize joint contracting costs.
Post-trade bargaining problems are included in the cost-recovery factor subsumed
in ␣.
Contracting entails three costs: the total opportunity cost of margin deposits
I(j); counterparty risk, that is, the expected difference between the promised
and the actual payment when h defaults on j, L(j, h); and expected deadweight
losses incurred when h defaults on j, D(j, h). Offsetting these costs, each party
also receives an option to default O(j, h). The two parties jointly minimize the
206 HERBERT L. BAER ET AL.
following costs:
I(j) + I(h) Opportunity Costs
D(j, h) + D(h, j) Deadweight Losses
(2)
L(j, h) + L(h, j) Counterparty Risk
O(j, h) + O(h, j) Default Options
Default involves both a loss to one party and a corresponding gain to the other.
Alternately stated, one party’s default option is another party’s counterparty
risk, that is, that L(j, h) = O(h, j). In our representative-agent framework, the
expression for joint contracting costs reduces to:
I(j) + I(h) + D(j, h) + D(h, j) (3)
The first order conditions for minimization of (3) with respect to M(j, h) and
M(h, j) are:
i i
[1 − F(M ∗ (j, h), )] = and F(M ∗ (j, h), ) = (4)
␣ ␣
where i is the opportunity cost for an additional dollar of margin. Equation (4)
implies that counterparties optimally collateralize when their default probabilities
equal their ratios of opportunity cost of additional margin to their deadweight
loss rate. The higher this ratio, the lower is the optimal collateral level. Nonzero
collateral requirements are optimal when i/␣ < 1. Should i/␣ exceed unity,
counterparties set margin at zero, the losing trader always defaults, and contracts
are unenforceable.
Since the objective function is linear in the number of contracts, the opportunity
cost of additional margin is constant. This implies that the collateral per unit of
exposure is independent of the aggregate level of exposure, and that collateral
amounts can be set on a per-contract basis. Further, if the distribution of price
changes is symmetric, equal collateral amounts are required for both long and
short positions.
Proposition II. when the distribution of price changes is uniquely invertible
and the opportunity cost of collateral deposits is constant, optimal coverage
ratios do not vary with volatility.
When a unique inverse exists for the price-change distribution depicted in
Eq. (4), the coverage ratio giving the level of collateral to exposure to price
changes is:
M∗ −1 i
=F ,1 (5)
␣
Inspection of (5) confirms that margin increases proportionately with .
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 207
This implies that optimal coverage ratios do not vary with volatility. For instance,
Proposition II holds if the distribution of price changes is normal.
Proposition III. If members have the same i and ␣, the clearinghouse sets
margin at the same level as though contracts were cleared and settled bilaterally.
Let party j’s open interest nh=1 N(j, h) be denoted by N(j). If we assume that
f, the distribution of price changes, is symmetric then the clearinghouse chooses
208 HERBERT L. BAER ET AL.
In our representative framework, i and ␣ are the same for all members and the
solution to this problem is given by Eq. (4). Thus, per contract, margin will be
the same whether contracts are cleared and settled bilaterally by counterparty
pairs or multilaterally through a clearinghouse. We argue this representative-agent
framework serves as a useful starting point in the following sense. Exchanges
screen members for their financial ability to fulfill contracts. That effort tends to
reduce member heterogeneity in their credit risk dimensions. Later sections of the
paper consider the monitoring efforts undertaken by exchanges as they attempt to
retain low levels of heterogeneity.
Because a clearinghouse will set the same margin rate that these agents willingly
negotiate between themselves, analyzing the benefits derived from forming a
clearinghouse is straightforward. The essential benefit of the clearinghouse is that
it permits its members to economize on margin while also reducing expected
deadweight losses. Clearinghouses economize on margins and deadweight loss
because, for the same set of contracts, participants’ net positions are less risky.
Consequently, total margin deposits required by the clearinghouse are smaller
than totals required for a comparable set of bilateral transactions. In addition, pro
rata expected deadweight losses are also smaller.
Proposition IV. Total margin deposits posted by each member will be the same
or lower under a clearinghouse system than under a system of bilateral collateral
deposits.
Under a clearinghouse system, j posts margin against the net of his position
with the rest of the market, that is M |N(j)|. In effect, a multilateral clearinghouse
secures the losing positions of a potential defaulter with its winning positions. That
is, members are prevented from “cherry picking” among their contracts realizing
gains while defaulting on those contracts where losses have been realized. The
margin posted by each member will be the same or lower under a clearinghouse
system.
Proposition V. Under an appropriate loss-sharing arrangement, total expected
deadweight losses are lower under a clearinghouse system than under a system
of bilateral collateral deposits.
Similarly, no counterparty’s expected deadweight loss is greater under a
clearinghouse system and for some expected deadweight losses will be smaller.10
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 209
The cost of funds function may be increasing as the amount of required margin
increases. Thus, an increase in margins drives up the marginal cost of funds.
Proposition VI. Where the marginal cost of funds is increasing, the optimal
coverage ratios decrease with volatility when prices are normally distributed
and the opportunity cost of margin assets is increasing.
Margin data are from the clearing organizations for eighteen contracts trading
on the following futures exchanges: the Chicago Board of Trade, the Chicago
Mercantile Exchange, the Coffee, Sugar and Cocoa Exchange, the Commodity
Exchange, and the New York Mercantile Exchange. These eighteen contracts are
the most heavily traded contracts having options on the underlying futures contract.
During the sample period, with the exception of contracts listed by the New
York Mercantile Exchange, exchange affiliation was the basis for determining
margin requirements. The speculative positions of non-clearing members are
assessed the highest levels of margin.12 The initial margin requirement for clearing
members is usually the same as the initial margin amount for the hedge positions of
non-clearing members. Finally, the maintenance margin requirements of clearing
members are the same as their initial requirements. Thus, our assumption that
periodic settlement restores the account to the level M gives a lower bound for a
clearing member’s margin account. Members must always have at least the amount
of the current initial margin, and may choose to allow excess balances to remain
in the account.
Table 1 summarizes our sample. Listed under each exchange are the contracts
trading on that exchange. The start date is the first date used in the sample; generally,
this is the beginning of options trading on the respective futures contracts. In each
case, the sample extends through June 1991. Sample dates are the last Thursday of
every contract month. The number of available observations ranges from 29 for the
Treasury bond and Deutschemark contracts to 15 for the Heating Oil contract. We
report mean margin levels for positions classed as initial nonmember speculative
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 211
Note: Table reports summary statistics for a sample of futures contracts. Start date is the first sample
date. Mean margin is the average of initial speculative or initial member margin required on the
sample dates. Mean ISD is the average implied standard deviation for options trading on the
sample dates. Margin coverage is respective mean level of margin divided by the dollar volatility
of the contract.
and for clearing members (or nonmember hedgers) on the above-indicated sample
dates.
For each sample date, we impute volatilities for the respective contracts. The
needed data are from various issues of the Wall Street Journal. These data are: prices
for call options expiring in the next delivery month at each strike price traded on that
date, futures settlement prices for corresponding delivery months, and Treasury
bill rates with maturities most closely matching the time until expiration of the
option contracts. The Barone-Adesi and Whaley (1987) model was used to impute
212 HERBERT L. BAER ET AL.
volatilities for each of the option contracts. For each contract, we calculate a time
series of representative implied standard deviations (ISDs) on each sample date
using a Taylor-series approximation based on iterated regressions as described by
Whaley (1982). The table reports mean ISDs. These range from a low of 0.01 for
the Eurodollar contract to 0.53 for the sugar contract.13
Margin coverage ratios divide the respective margin amounts by dollar-price
volatility. Dollar-price volatility is the product of the ISD and the notional value
of the contract – futures prices times number of deliverable units – after adjusting
the annualized volatility for the length of the holding period. This gives a market-
based forecast of holding period volatility. Dividing initial speculative and member
margin requirements by their respective dollar volatilities gives coverage ratios.
Columns 6 and 8 list mean coverage ratios for the member and nonmember
categories. The peak of the frequency distribution for mean nonmember speculative
margin coverage ratios is about five. This implies that margin levels most often
cover five times the expected single-day price deviation. Comparison of the means
of nonmember speculative and member margin requirements reveals that clearing
members’ margin is about 80% of that for speculative positions. The exception is
the New York Mercantile Exchange where they are equal.
Notably, the coverage ratio for the S&P 500 contract well exceeds the typical
level obtained for nonmember speculative positions, averaging 10.17 during the
sample period. In contrast to the coverage obtained by nonmember margin levels,
the S&P 500 members’ margin, generally around four, is within the range obtained
for other contracts. The difference between coverage ratios for the S&P and other
nonmember speculative margins probably reflects considerations unique to the
sample period. Market breaks in 1987 and 1989 increased debate over the need for
regulatory intervention in the determination of stock-index margin requirements
may have resulted in higher margins than the clearinghouse would have set for
purely prudential reasons.
The contrast between margin for the S&P contract and the others is more
noticeable on recognizing that during part of this period, the S&P 500 contract
settled twice per day. Other contracts settled only once per day throughout the
period. Since we calculate coverage ratios with daily standard deviations, the
coverage ratio for the S&P 500 should be smaller, not larger. Other things
equal Fenn and Kupiec’s (1993) analysis suggests coverage ratios should be
approximately half as large.
Assuming price changes are normally distributed, the coverage ratios for
clearing members imply that the probability of a price change exceeding required
margin from one settlement period to the next is much less than 1%. The “excess”
of coverage suggests that actual distributions are kurtotic, a result that is consistent
with the findings of Kofman (1993).
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 213
We use daily data for four of the eighteen contracts to gain further insight into the
margin-setting process. These contracts are: Deutschemark, S&P 500, Soybean
and Treasury Bond. Daily implied volatilities, computed as previously described,
were matched with required margin levels on these dates and margin coverage
ratios computed. We examine the time series of these quantities.
The first test considers whether the coverage ratio for a contract tends to revert
to its long-run, unconditional mean. Denoting coverage ratios CRt , our model
implies that shocks to these ratios initiate intervention that restores them to desired
levels. Specifically, prudential concerns dictate that coverage ratios regarded as too
small should lead to increasing margin coverage and opportunity cost concerns
dictate that excessively large ratios lead to reducing margin coverage. Our model
equilibrates these pressures, and predicts a mean-reverting time series of coverage
ratios.
We employ the augmented Dickey-Fuller (ADF) procedure to consider this
hypothesis. Changes in coverage ratios are regressed on the first lag of their levels
and lags of changes in the coverage ratio. The specification is:
K
CRi,t = ␣i,0 + ␣i,1 CRi,t−1 + ␣i,1+j CRi,t−j + u i,t (9)
j=1
Deutschemark−0.004579 −3.52
S&P 500 −0.004704 −2.88
Soybean −0.012160 −4.04
−0.017178 −6.84
Treasury Bond
CRt = ␣0 + 4l=1 ␣l1 Q l CRt –1 + K
i=1 ␣4+i CRt –i + u t
Note: Table reports results for the two time series specifications listed below. CRt is the time-t ratio
of initial member margin to the option-implied volatility stated in dollars. Ql is the coverage
quartile for margin coverage during the sample period. K, the number of lagged changes in
coverage ratio included in the specification, is determined by AIC. Critical values are from
Fuller (1976): −1.95 at the 5% level and −2.58 at the 1% level. Lower values of t are indicative
of reversion to the mean; i.e. rejects the null of no mean reversion.
coefficient estimates from the volatility specification and for the coverage-ratio
specification. In no case do these ranges overlap. Hence, we reject the alternative
hypothesis that adjustments to coverage ratios stem from volatility reverting to its
long-run mean rather than from exchange action.15 We can be confident at the 5%
level that exchanges actively adjust margin levels in response to coverage-ratio
shocks.
We extend these tests to detect if reversion to the mean is more rapid when
coverage ratios are above or below their long-run averages. The prudential
hypothesis of previous authors such as Gay, Hunter and Kolb (1986) predicts that
clearinghouses respond to low coverage ratios by raising margin requirements.
Previous models of prudentiality do not predict clearinghouse response to shocks
resulting in excess margin coverage. In contrast, the model of this paper predicts
that a high cost of margin coverage induces clearinghouses to lower margin
coverage with the provision that they meet prudentiality objectives. The ADF test
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 215
is modified to test for differential slopes on the lagged level of the coverage ratio.
Quartiles are determined for the sample of coverage ratios and indicator variables,
denoted Ql , used to classify observations into quartiles.
Lagged coverage ratios are interacted with quartile-indicator variables to
measure differential responses by the clearinghouses. This specification is:
4 K
CRt = ␣0 + ␣l1 Q l CRt−1 + ␣i CRt –i + u t (10)
l=1 j=1
We report these results in the lower panel of Table 2. Most coefficients differ
reliably from zero. The exception is the speculative margin requirement of the
soybean contract where response to low coverage ratios has the correct sign
but is not statistically significant. However, in every case, coefficients on the
highest quartile classification differ reliably from zero. This is consistent with
clearinghouse policies that lower margin requirements when margin coverage
ratios exceed their long-run averages and implies an internalization of the costs of
high margins born by the exchange membership. Fenn and Kupiec (1993) predict
a similar cost internalization.
Comparing the coefficients on the low and high coverage quartiles adds evidence
for the presence of tradeoffs between prudentiality and margin costs. Coefficients
that are larger (in absolute value) imply quicker responses to shocks to the coverage
ratio. In every case, the coefficients on the low-coverage quartiles are larger in
absolute value than those on the high-coverage quartiles. This implies that these
clearinghouses respond more quickly to surety lost when coverage ratios decline
than to the increase in costs borne by clearinghouse members when coverage ratios
rise.16
The opportunity cost of margin is the difference between the cost of financing an
additional dollar of margin assets and the return on those assets. If margin deposits
were non-interest-bearing cash, movements in firms’ short-term borrowing costs
would suitably measure changes in the opportunity cost of margin. However,
members most often fulfill margin requirements with securities or standby letters
of credit. For securities, the appropriate measure of opportunity cost is the risk-
adjusted difference between the yield on the margin assets and an additional dollar
of credit. During our sample period, the sampled exchanges accepted government
and agency-debt securities as margin, Treasury bills being the most widely posted
form of margin.17
216 HERBERT L. BAER ET AL.
Ideally, we need a time series on the spread between the risk-adjusted borrowing
costs of market participants and rates on Treasury bills. Lacking this, we adopt
a proxy for the cost of borrowing. The proxy should capture economy-wide
shifts in the borrowing costs. In addition, borrowing-cost changes should be
correlated across borrowers even though members may be at different points on
their respective credit supply curves.
Commercial banks are a significant source of credit to futures market
participants. Consequently, the prime rate measures credit costs. When the
prime rate rises, firms with prime-based loan agreements experience a change in
borrowing costs. To isolate credit premia we use differences between the prime
rate and the Treasury bill rate to represent changes in the opportunity cost of
margin.
Our model implies a negative correlation between coverage ratios and volatility
levels when members face inelastic supplies of external finance. Holding constant
the coverage ratio, open interest, and the other assets of clearing members, a
volatility increase implies higher margin deposits and greater reliance on external
finance. With an upward-sloping supply of external funds, a higher margin
requirement raises the opportunity cost for margin deposits. A clearinghouse which
is optimizing as our model suggests will respond to this higher opportunity cost by
reducing its coverage ratios. Thus, a constant cost for borrowing implies a positive
correlation between volatility and individual member borrowing costs. Hence, our
model implies a negative correlation between volatility and coverage ratio.
This discussion suggests the following test specification:
CRit = ␣0i + ␣1 R t + ␣i2 ISDit + it (11)
where i denotes the ith contract, Rt is a proxy variable capturing variation in the
opportunity cost of borrowing from economy-wide changes in bank financing,
and ISDit is the implied standard deviation for the particular contract. Including
these implied standard deviations incorporates both intertemporal and cross-
sectional differences in market participants’ opportunity cost that might result from
differences in the demand for credit to finance margin positions. The increasing
opportunity cost model implies the following coefficient restrictions: ␣1 ≤ 0 and
␣i2 ≤ 0.
Table 3 presents estimation results for Eq. (11). The first three columns of data
in Table 3 present the results from OLS estimation of Eq. (11) for the eighteen
contracts included in our sample. With the exception of the British pound, gold
and silver, the coefficients on ISD are negative and differ reliably from zero. The
coefficient on the interest-rate-spread variable differs significantly from zero only
for the contracts on the British pound and wheat. From these results we conclude
that after controlling for the opportunity costs imposed by margin deposits, the
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 217
Table 3. (Continued )
Unrestricted Intercept Opportunity ISD Within-Exchange ␣ij1 = ␣ij1 ∃j
Coefficients Cost Coefficient Restrictions:
␣ij2 = ␣ij2 ∃j
Intercept Opportunity ISD
Cost
Note: Standard errors in parentheses. Table reports results for two methods of estimating the following
time-series and cross-sectional specification: CRijt = ␣ij0 + ␣ij1 Rt + ␣ij2 ISDit + ijt . Variable
definitions are: CRijt is the coverage ratio for contract i traded at exchange j on date t, Rt
is the excess of the prime rate over the 3-month treasury bill rate, and ISDit is the implied
standard deviation for options trading on contract i at date t. Unrestricted coefficients are from
separate OLS regressions for each contract. The within-exchange restricted coefficients are
from estimating those equations as a system of seemingly unrelated regressions. Regressions
include equality restrictions on the two right-hand side variables. Within the five exchanges, the
coefficients on R are restricted to equality and the coefficients on ISD are restricted to equality.
See Table 1 for contracts traded on the five exchanges.
N(d, j)
|x − M| < C(d, d) (12)
j=1
(1 + ␣)
N(d, j)
C(d, CH)
<
N(d, j)
j=1
Proposition VII. When the clearinghouse can make a credible expulsion threat,
the optimal margin coverage ratio increases as volatility increases.
The first order condition for minimization of (15) with respect to M(j, h) and
M(h, j) is:
C C C i
1 − F M ∗∗ + , + f M ∗∗ + , =
|N(j)| |N(j)| |N(j)| ␣
Whenever clearinghouse membership is valuable, then C > 0 and the final term
on the hand side of (16) is strictly positive. This means that a policy of expelling a
defaulting member reduces the probability of default F(·) to less than i/␣, the level
prevailing absent expulsion. Since a credible expulsion threat acts as a substitute
for costly margin, the clearinghouse will choose a higher level of protection than
it otherwise would.
Differentiating Eq. (16) with respect to M** and C/|N(j)|, implies that when
margin is taken dM** /d(C/|N(j)|) < 0. Thus, increasing the value placed on
membership decreases the amount of margin required to obtain a given level of
safety. Intuitively, as membership and the volume cleared expands over time, the
threat of expulsion becomes more serious because clearing privileges are more
valuable, and the clearinghouse can place greater reliance on the expulsion threat
as a deterrent to default, thus decreasing the optimal level of margin.
222 HERBERT L. BAER ET AL.
If the threat of expulsion is an important part of the guarantee system, how would
coverage rations change when volatility changes? On the one hand, if we hold the
value of clearing C constant, then optimal margin coverage ratio M** / increases
as volatility increases. Thus, margin must increase more rapidly than volatility to
supply the same level of protection. This result contrasts sharply with our analytic
results for constant or increasing costs of margin deposits.
On the other hand, it may be that the value of clearing C changes systematically
with changes in volatility. The most likely conjecture is that the value of clearing
increases as volatility increases, because increased volatility is likely to be
associated with increased volume of trading. This could reverse the result cited in
the previous paragraph: if clearing becomes more valuable in volatile markets, it
is possible that optimal margin could drop.
The above suggests the benefits of clearinghouse creation can go beyond margin
economies and avoiding deadweight default costs. Creation of a clearinghouse
assures contract performance at levels beyond that obtained by margin deposits.
Further, it can be the case that reliance on membership value is more cost effective
than relying simply on margin deposits.
Relaxing the assumption that only collateral can be attached in case of default, we
allow counterparties to grant senior claims on unencumbered assets k(j).20 Each
party knows its own k(j), however we assume counterparties incur an examination
cost to learn k(j). This cost is denoted e. Traders choose to be monitored when the
savings from these senior claims against k(j) exceed examination costs.
Examination, at most, saves the firm the opportunity cost of holding k(j), that
is, the maximum savings is ik(j). If the quantity ik(j) is less than e, then inspection
does not pay. However, failure of this condition is not sufficient for inspection
to occur. If the optimal margin M*|N(j)| without inspection is less than k(j), the
opportunity cost savings from granting a senior claim against k(j) is iM*|N(j)|.21
Inspecting a member firm reveals one of two conditions. Members post no margin
when unencumbered assets k(j) exceed M* |N(j)|. Alternately, if k(j) is less than
M* |N(j)|, then the clearinghouse’s problem is of the same form as Eq. (6) with
M + k(j)/|N(j)| substituted for M.
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 223
∗∗∗ k(j) i
M + = F −1 ,s (16)
|N(j )| ␣
Because k(j) is less than M* |N(j)|, parties still post margin. Thus, for a constant
opportunity cost of margin, the optimal default probability is identical to the no-
examination case because firms substitute claims against unencumbered assets for
more costly forms of margin. Unlike the case where expulsion acts as a deterrent
to default, if default occurs, the other assets can be seized, making them a perfect
substitute for margin deposits.
However, when opportunity costs are increasing in total required margin,
examination decreases the optimal default rate. When other assets are substituted
for margin, the cost of additional margin decreases, and the optimal level of
protection can thus be increased.
Equation (16) implies that with monitoring, coverage ratios increase as volatility
increases, as unencumbered assets k(j) decrease, and as the number of open
contracts increases. The result that the optimal coverage ratio declines as firms
become less able to substitute unencumbered assets for collateral deposits differs
from the predictions of the constant opportunity cost model (Eq. (4)) and the
increasing opportunity cost model (Eq. (8)).
Since firm asset holdings are dissimilar, reliance on unencumbered assets
entails scrutiny of member positions, with margin setting on a member-by-
member basis. Instead, margin requirements are uniform across the memberships
of organized clearinghouses. This uniformity arises for several reasons. First,
payment delays may be the principal cause of deadweight losses for members of
the clearinghouse. The presence of unencumbered-but-illiquid assets may not be
useful for time-critical settlement requirements. Second, timely verification of the
existence of k may significantly raise clearinghouse costs, making is uneconomical
to monitor at all. Third, netting may reduce each party’s net exposure to such
low levels that intensive monitoring can not be cost-effective. In any event, the
uniformity of margins across members suggests that if clearinghouses do engage
in extensive monitoring, it must be for purposes other than controlling risk between
members.
The prediction of a positive correlation between volatility and the coverage ratio
also contrasts sharply with the independence of the coverage ratio and volatility.
Thus, this result is distinct from the simple netting model of Eq. (5) and the negative
correlation generated by the increasing opportunity cost of funds model of Eq. (10).
224 HERBERT L. BAER ET AL.
6. SUMMARY
We incorporate the cost of external funds and the deadweight losses associated with
counterparty default into collateral decisions. Given their scale of activities, it is not
surprising that clearinghouses internalize these costs into their margin decisions.
Thus, clearinghouse pursuit of prudentiality through margin is constrained by
the costs that members incur by carrying these balances. When margin is set
without regard to additional information about the condition of the clearinghouse
members, the coverage ratio is either uncorrelated or negatively correlated with
volatility.
The time series of coverage ratios supports the conclusion that clearinghouse
determination of margin incorporates prudential concerns. Our empirical results
demonstrate that clearinghouses respond to high levels of margin by adjusting
coverage ratios downward. This behavior is not consistent with prudentiality alone,
but is consistent with exchanges optimizing across the costs of placing margin
deposits and the costs from incurring deadweight losses.
Our pooled-regression results show that futures clearinghouses set margin in a
cost-minimizing fashion, balancing the risk of loss against the greater opportunity
costs associated with higher margins. Those results suggest that at least part of
these opportunity costs arise because market participants have imperfect access to
capital markets for their general financing.
We extend our model to capture other institutional features. We expect positive
correlations between coverage ratios and volatility when clearinghouses actively
monitor their members for risk-management purposes. Our emphasis on the
foundations of the clearinghouse makes clear that membership is valuable. Because
membership is valuable, expelling defaulting members is credible and effective for
the clearinghouse. This means that members will perform on their contracts even
when price moves exceed the value of margin on deposit.
NOTES
1. The term collateral is used for all instances where a deposit is required to reduce
exposure to credit risk. The term “margin” when used here is the special case applying
when collateral deposits are uniform within broad categories of counterparties.
2. Baer, France and Moser (1995b) re-examine some previous tests of prudentiality.
3. For-profit exchanges operating in a competitive market can also be expected to cost
minimize. Hence, our model also applies to for-profit organizations.
4. For example, courts may be unable to force the transfer of collateral quickly enough
to avoid deadweight default costs and will involve significant legal costs.
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 225
15. Another possibility not considered here is that the margin responses of exchanges
induce subsequent changes in volatility. The evidence does not generally support this claim,
however for an alternative view, see Hardouvelis (1988).
16. An F test indicates that the difference between the coefficients on the high and low
quartiles of the S&P and Deutschemark contracts is significant at better than the 95% level.
17. Other clearinghouses, for instance the Options Clearing Corporation, have long
accepted equity as margin. Futures clearinghouses are increasingly adopting this practice.
18. A minimal number of clearing members may be necessary to insure that none are
too valuable to expel.
19. This presumes that the value of an exchange seat reflects trading rather than clearing
privileges, as suggested by Bhasin and Brown (1997). If clearing memberships were a
separately traded asset, their value would reflect the value of C(d, d) for the marginal
member, and that amount could be recouped by the other members in a default. The model
would then closely resemble that in the next section, where counterparties grant senior
claims on assets.
20. Relaxing this assumption implicitly assumes that courts are effective in seizing
collateral and that the speed of payment is not an issue. If payment delay is the principal
reason that default imposes a deadweight loss on the membership, then the existence of
unencumbered assets may be irrelevant.
21. We assume that the inspection process includes assessing the probable value of k
in the default state. Clearinghouses “haircut” non-cash assets by valuing them at less than
current market value.
ACKNOWLEDGMENTS
Much of this paper was completed while France was visiting at the Chicago
Mercantile Exchange and the University of Chicago. Susanne Malek and Jan
Napoli provided valuable research assistance. The authors thank John Conley,
Ramon DeGennaro, Mark Flannery, Gary Koppenhaver, Todd Petzel, Will
Roberds, Jerry Roberts, an anonymous referee, and seminar participants at the
University of Illinois at Urbana-Champaign, the Federal Reserve Bank of Chicago,
the University of Tennessee, and the Chicago Risk Management Workshop.
Opinions expressed are entirely those of the authors and do not reflect concurrence
by the Federal Reserve, the Chicago Mercantile Exchange, or the World Bank.
REFERENCES
Baer, H. L., France, V. G., & Moser, J. T. (1993). Opportunity cost and prudentiality: A representative-
agent model of futures clearinghouse behavior. Faculty Working Paper #93–0142, University
of Illinois.
Baer, H. L., France, V. G., & Moser, J. T. (1995a, Spring). What does a clearinghouse do? Derivatives
Quarterly, 1(3), 39–46.
Opportunity Cost and Prudentiality: An analysis of Collateral Decisions 227
Baer, H. L., France, V. G., & Moser, J. T. (1995b). Determination of collateral deposits by bilateral
parties and clearinghouses. In: Proceedings of the 31st Annual Conference on Bank Structure
and Competition (pp. 525–546). Chicago: Federal Reserve Bank of Chicago.
Baer, H. L., & McElravey, J. (1994, January). Capital adequacy and the growth of U.S. banks. In:
C. Stone & A. Zissu (Eds), Global Risk Based Capital Regulations: Management and Funding
Strategies. Homewood, IL: Dow-Jones-Irwin.
Barone-Adesi, G., & Whaley, R. E. (1987). Efficient analytic approximation of American option values.
Journal of Finance, 42, 301–320.
Bhasin, V., & Brown, D. P. (1997, January). The organization of competitive financial exchanges.
Working Paper, Indiana University. Journal of Financial Markets (Forthcoming).
Bollerslev, T., Chou, R. Y., Jayaraman, N., & Kroner, K. F. (1991). ARCH modeling in finance: A
selective review of the theory and empirical evidence with suggestions for future research.
Journal of Econometrics, 52, 5–59.
Brennan, M. J. (1986). A theory of price limits in futures markets. Journal of Financial Economics,
16, 49–56.
Calomiris, C. W., & Hubbard, R. G. (1995, October). Internal finance and investment: Evidence from
the undistributed profits tax of 1936–1937. Journal of Business, 68(4), 443–482.
Cornes, R., & Sandler, T. (1986). The theory of externalities, public goods, and club goods. Cambridge:
Cambridge University Press.
Craine, R. (1997). Valuing the futures market performance guarantee. Macroeconomic Dynamics, 1(4),
701–719.
Diamond, D. (1984). Financial intermediation and delegated monitoring. Review of Economic Studies,
51, 393–414.
Fazzari, S. M., Hubbard, R. G., & Peterson, B. C. (1988). Financing constraints and corporate
investment. Brookings Papers on Economic Activity, 141–195.
Fenn, G., & Kupiec, P. (1993). Prudential margin policy in a future-style settlement system. Journal
of Futures Markets, 13, 389–408.
Figlewski, S. (1984). Margins and market integrity: Margin setting for stock index futures and options.
Journal of Futures Markets, 4(Fall), 385–416.
Fuller, W. A. (1976). Introduction to statistical time series. New York: Wiley.
Gay, G. D., Hunter, W. C., & Kolb, R. W. (1986). A comparative analysis of futures contract margins.
Journal of Futures Markets, 6(Summer), 307–324.
Gorton, G. (1985). Clearinghouses and the origin of central banking in the U.S. Journal of Economic
History, 45(June), 277–283.
Hardouvelis, G. A. (1988, Summer). Margin requirements and stock market volatility. Federal Reserve
Bank of New York Quarterly Review, 13(2), 80–89.
Hsieh, D. A. (1993). Implication of non-linear dynamics for financial risk management. Journal of
Financial and Quantitative Analysis, 28(March), 41–64.
Hubbard, R. G., & Kashyap, A. K. (1992). Internal net worth and the investment process: An application
to U.S. agriculture. Journal of Political Economy, 100(June), 506–534.
Kofman, P. (1993). Optimizing futures margins with distribution tails. Advances in Futures and Options
Research, 6, 263–278.
Laffont, J.-J. (1988). Fundamentals of public economics. Cambridge, MA: MIT Press.
Merton, R. C., & Bodie, Z. (1992). On the management of financial guarantees. Financial Management,
21(Winter), 87–109.
Whaley, R. E. (1982). Valuation of American call options on dividend-paying stocks: Empirical tests.
Journal of Financial Economics, 10, 29–57.
COLLATERALIZATION AND THE
NUMBER OF LENDERS IN PRIVATE
DEBT CONTRACTS: AN EMPIRICAL
ANALYSIS
ABSTRACT
Using the Dealscan database of large, U.S. corporate loans, we examine the
determinants of the number of bank relationships and the presence or absence
of collateral. Consistent with prior studies, we find that important explanatory
variables are firm quality, desire for financial flexibility, the probability of
financial distress, growth opportunities and firm size. Higher quality firms
as well as firms with a stronger desire for financial flexibility are less likely
to collateralize and borrow from more lenders. Larger firms as well as those
with lower probabilities of financial distress and greater growth opportunities
prefer multiple lenders.
1. INTRODUCTION
Why do some firms choose to borrow from a single bank while others engage in
multiple banking relationships? What factors determine whether a firm secures its
loans and how do the two decisions interact? A number of recent empirical studies
analyze multiple banking relationships in Europe: Detragiache et al. (2000) and
Research in Finance
Research in Finance, Volume 21, 229–252
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21010-3
229
230 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
D’Auria et al. (1999) – Italy, Degryse and Ongena (2001) – Norway, Farinha and
Santos (2002) – Portugal, and Harhoff and Korting (1998) – Germany. Studies on
U.S. banking relationships have focused on small firms (Petersen & Rajan, 1994)
and the role of multiple bank relationships in mitigating hold-up problems and
creating financing flexibility (Houston & James, 1996, 2001).
The present paper examines the determinants of bank relationships and collateral
for large U.S. loans in the Loan Pricing Corporation’s DealScan database for
the period 1990–1999. Consistent with prior studies, we find that important
explanatory variables are firm quality, desire for financial flexibility, the probability
of financial distress, growth opportunities and firm size. Higher quality firms
as well as firms with a stronger desire for financial flexibility are less likely to
collateralize and borrow from more lenders. Larger firms as well as those with
lower probabilities of financial distress and greater growth opportunities prefer
multiple lenders. Although there is a common set of explanatory variables, tests
for interdependence effects suggest that the number of bank relationships and
whether to offer collateral are independent decisions.
Section 2 reviews relevant literature. Section 3 presents the models and variables
used. The data sample is presented in Section 4, which is followed by a discussion
of our simultaneous modeling technique analysis in Section 5. Section 6 presents
our statistical analysis and Section 7 concludes paper.
2. PRIOR LITERATURE
Prior research indicates that firm quality, the probability of financial distress,
growth opportunities, the desire for managerial flexibility, and firm size are relevant
determinants of both the number of lenders and whether loans are collateralized.
We discuss each decision in turn. We can summarize the testable hypotheses from
these studies with respect to the collateralization and the number of creditors in
private debt contracts as shown in Table 1.
We expect firms with greater growth opportunities to borrow from multiple lenders
for moral hazard reasons (Repullo & Suarez, 1998) as well as to limit the single
lender’s bargaining power and to mitigate any information monopoly (Houston &
James, 1996; Rajan, 1992). A similar argument applies to larger firms. Detragiache,
Garella and Guiso (2000) present a model in which borrowing from multiple banks
can ensure a more stable supply of credit. Avoiding rationing by a single lender
Collateralization and the Number of Lenders in Private Debt Contracts 231
Growth − Gilson and Warner (2000) + Rajan (1992), Houston and James
(1996), Repullo and Suarez (1998)
Probability of + Gorton and Kahn (1993), Rajan − Berlin and Mester (1992),
distress and Winton (1995) Chemmanur and Fulghieri (1994)
+ Degryse and Ongena (2001)
Firm quality + Bolton and Scharfstein (1996) + Berlin and Loeys (1988), Houston
and James (1996), Rajan (1992),
Bolton and Scharfstein (1996)
− Rajan and Winton (1995) − Detragiache et al. (2000), Degryse
and Ongena (2001)
Firm size + Repullo and Suarez (1998),
Detragiache et al. (2000)
Loan purpose + Gilson and Warner (2000) + Gilson and Warner (2000)
Note: The table summarizes the hypotheses and findings of previous studies on collateralization
and number of creditors in private debt contracts with respect to the five firm characteristics
mentioned. A “+” in front of the study indicates a positive hypothesized relationship between
the debt contract and the firm characteristic, and conversely a “–” in front of the study indicates
a negative relationship. Collateralization and Number of Creditors are binary variables, with “1”
indicating presence of collateral, or multiple creditors, and “0” indicating absence of collateral,
or single creditor, respectively.
reduces the risk that a profitable project will have to be prematurely liquidated.
If the adverse selection problem is very severe because outside banks have less
information than the borrower’s inside bank and suspect the firm to be a bad risk,
it may pay for the firm to establish multiple relationships from the beginning.
Hence, they theorize that larger and less profitable firms are less likely to have
single banking relationships.
Degryse and Ongena (2001) examine how the number of bank relationships
affects firm performance for a sample of Norwegian firms. Firms with valuable
proprietary information may use fewer creditors in order to prevent information
leakage. At the same time, financially distressed firms facing credit rationing at
their main bank may be forced to seek additional, more costly financing elsewhere.
Farhina and Santos (2002) obtain a similar result for Portuguese firms. A positive
association between the probability of financial distress and the number of bank
relationships is also present in Germany according to Harhoff and Korting (1998).
Firms that are looking to increase managerial discretion and flexibility, as
indicated by the loan purpose, will also borrow from multiple creditors (Gilson &
Warner, 2000). In contrast, firms with greater probability of financial distress, or
higher credit risk firms find the option to renegotiate more valuable and hence
232 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
may prefer to borrow from fewer lenders (Berlin & Mester, 1992). Further,
Chemmanur and Fulghieri (1994) find that banks are able to use reputation as a
commitment device to promise firms credibly that they will devote more resources
toward evaluating them and thereby make better renegotiation-versus-liquidation
decisions in the event of financial distress.
2.2. Collateralization
Firm quality We proxy firm quality by Standard and Poor’s bond rating;
the ratings of AAA, AA through C are translated to an
ordinal scale ranging from 2 to 21, as recorded by
COMPUSTAT. Blume et al. (1998) point out that rating
agencies, such as Standard and Poor’s, employ both
publicly available information, such as accounting
statements, and non-public information, such as confidential
interviews with management, to assign quality ratings as a
measure of the “creditworthiness” of a corporation with
respect to a particular debt security. Their survey of prior
studies concludes that quality ratings do have some
informational content and that these ratings contain
information beyond what is publicly available. Their own
empirical study also concludes that bond rating standards
have become more stringent over time.
Probability of This is proxied by the Z score value, calculated using
distress Altman’s (1968) model. The Z score model utilizes multiple
discriminant analysis to obtain a linear function of variables
that is able to discriminate membership between two
populations, bankrupt and non-bankrupt. The Z score
indicates how close the firm is to being classified as being in
one of the groups, i.e., the score is a measure of the
probability of being in financial distress.
Growth The standard market to book ratio is used to proxy for
opportunities growth opportunities.
Loan purpose Following Carey et al. (1998) and Dennis et al. (2000), we
place into groups the 16 stated uses of the debt by the
borrowers, as recorded by the Loan Pricing Corporation, to
indicate the purpose of the loan. We utilize Carey et al.’s
(1998) four groups and categorize loans as being made for
the following purposes: (1) general corporate purposes; (2)
234 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
Aside from the above firm characteristics, we also control for the effects of
leverage in our regressions. The importance of controlling for leverage is
demonstrated by Lang, Ofek and Stulz (1996), who find that future growth and
investments are negatively related to leverage. Johnson (1998) also found that
leverage was statistically and economically significantly higher for firms with
bank debt due to the fact that bank debt attenuates negative effects on leverage of
potential asset substitution problems. Hence, omitting this variable would cause
our model to be misspecified. Leverage is measured by the book value of debt
divided by the market value of equity plus the book value of debt. We control for
the type of private debt issued since this may directly affect the need for collateral
as well as the number of creditors. The facility type is also categorized into one
of four groups as follows: (1) short-term revolver; (2) long-term revolver; (3)
term loan; and (4) other (including leases, bridge loans, notes, standby letters of
credit, etc.).
For the collateral regression, we also control for the amount of fixed assets
available to put up as collateral, as well as the type of lead lender, whether bank
or finance company. Carey et al. (1998) and Houston and James (2001) describe
finance companies as “asset-based” lenders and banks as “cash flow” lenders. In
making a loan, an asset-based lender emphasizes collateral as a source of ultimate
repayment whereas a cash flow lender relies more heavily on projected cash flow
from operations. The amount of fixed assets is measured by the ratio of fixed
assets to total assets, while the lender type is a binary variable, with 1 for a finance
company as lead lender and 0 for a bank as lead lender. A detailed formulation of
all the regression variables is in Appendix A.
We utilize a simultaneous model capable of assessing the interdependencies
between the choice of the number of creditors as well as the amount of collateral
offered hypothesized by Bolton and Scharfstein (1996). Building on the framework
of Dennis et al. (2000), we model our regressions as follows:
where the ␣’s are coefficients of the interdependence effects. This permits a
direct test for the significance of interdependence effects. Both Collateral and
Creditors are binary variables, with 1 indicating presence of collateral or multiple
Collateralization and the Number of Lenders in Private Debt Contracts 235
4. DATA
We started construction of our sample set from the DealScan 5.6 database compiled
by the Loan Pricing Corporation (LPC). In this area of the literature, Shockley and
Thakor (1997) were the first to utilize the LPC database to empirically analyze
the fee structure of bank loan commitment contracts.1 The LPC database provides
detailed market information on commercial loans and private placements made
to publicly held U.S. companies that are required to file such information with
the Securities and Exchange Commission (SEC). The database also includes deal
information obtained directly from banks, which is later confirmed after the deal is
recorded with the SEC. The data include details such as the name of the borrower,
the names of all lenders party to the loan contract at origination, the type, purpose,
maturity, amount, material restrictions and contract date of the loan as well as other
details. An illustration of relevant sample data from LPC’s DealScan database is
presented in Table 2.
Note: The database contains over 200 searchable fields. Only some of the relevant fields are illustrated
in this sample. Each deal (or package) may consist of one or more facilities, with each facility
separately listed.
236 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
Note: All dollar figures are in millions of dollars, except the share price. The complete definition of
the variables is found in Appendix A.
∗∗∗ All mean values were found to be statistically significantly different from 0 at the 1% level.
Collateralization and the Number of Lenders in Private Debt Contracts 237
5. MODEL ESTIMATION
Following Dennis et al. (2000), we apply Mallar’s (1977) two-stage estimation
procedure for simultaneous equation models with limited dependent variables to
our model.3 We note that Eqs (1) and (2) both involve discrete binary choices and
hence use a simultaneous probit model. In the first stage, a reduced form model
for each of the two endogenous variables is estimated:
Collateral = 1 X + 1 (3)
Creditors = 2 X + 2 (4)
N
1
W2 = A 2 XX
N
1
N
1 2
W3 = A 1 ␣1 ZX
N 1
1
N
1
W4 = a 1 a 2 E[(y 1 − 1 )(y 2 − 2 )]XZ
N
1
Then the covariance of N 1/2 (␥ˆ 1 − ␥01 )(where ␥01 is the true value of ␥1 and ␥ˆ 1
is the two-stage estimator) is
W −1 −1 −1 −1 −1
1 [W 1 − W 3 W 2 W 4 − W 4 W 2 W 3 + W 3 W 2 W 3 ]W 1 (7)
The covariance of ␥ˆ 2 is similarly derived as follows.
1 2
a1 = , a2 =
1 (1 − 1 ) 2 (1 − 2 )
∗
1 X
A 1 = 1 a 1 , A 2 = 2 a 2 , Z =
X
N
1
W5 = A 2 ZZ
N
1
N
1
W6 = A 1 XX
N
1
N
1 1
W7 = A 2 ␣2 ZX
N 2
1
N
1
W8 = a 1 a 2 E[(y 1 − 1 )(y 2 − 2 )]XZ
N
1
Then the covariance of N 1/2 (␥ˆ 2 − ␥02 ) (where ␥ˆ 02 is the true value of ␥2 and ␥ˆ 2
is the two-stage estimator) is
W −1 −1 −1 −1 −1
5 [W 5 − W 7 W 6 W 8 − W 8 W 6 W 7 + W 7 W 6 W 7 ]W 5 (8)
Following Carey et al. (1998), the fit of the probit model is estimated by
McKelvey and Zavoina’s (1975) pseudo-R2 measure, calculated as in Greene
(1999).
Collateralization and the Number of Lenders in Private Debt Contracts 239
6. ANALYSIS
6.1. Univariate Analysis
Note: All dollar figures are in millions of dollars, except the share price. The complete definition of
the variables is found in Appendix A.
∗∗ All mean values were found to be statistically significantly different from 0 at the 5% level, except
for RELEPS and CUMPROF, which were not significantly different from 0.
240 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
sufficient incentive to monitor and force liquidation when efficient, higher quality
firms decrease collateral. Our finding that lower quality firms are more likely to
offer collateral supports Pozzolo (2002) for Italian firms while contradicting both
Bolton and Scharfstein (1996) and John, Lynch and Puri (2002). Furthermore, those
offering collateral are not as profitable and have higher probabilities of financial
distress in the near future (OITA, ZSCORE). This is also in accordance with
the hypotheses of Rajan and Winton (1995) and Gorton and Kahn (1993), that
firms which expect to face higher probabilities of financial distress provide more
collateral to give lenders greater control as well as the ability to threaten liquidation
of inefficient projects. Finally, we find that firms borrowing from multiple lenders
have greater growth potential (MKTBOOK) than those borrowing from single
lenders. This supports the theory of Gilson and Warner (2000) that high growth
firms put up less collateral to increase managerial discretion and flexibility.
In looking at Table 5, which compares firms using multiple creditors with those
using single creditors, we find that firms using a single lender have statistically
significantly higher leverage (DEBTASST, MKTLEV) in accordance with the
findings of Detraigaiche et al. (2000), Degryse and Ongena (2001), and Harhoff and
Note: All dollar figures are in millions of dollars, except the share price. The complete definition of
the variables is found in Appendix A.
∗∗∗ All mean values were found to be statistically significantly different from 0 at the 1% level.
Collateralization and the Number of Lenders in Private Debt Contracts 241
The results of the simultaneous regressions on the collateralization and the number
of creditors in private debt are presented in Table 6.6 With pseudo-R2 of 70% and
44% respectively, both of which are significant at the 1% level, the regressions fit
reasonably well.
We find that the two decisions are not interrelated since the coefficients of the
interdependence effects are insignificant. The quality and loan purpose proxies
are significant for the collateralization regression in column (1). The positive
coefficient on the quality proxy indicates that firms of lower quality put up more
242 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
Note: Results shown are those from the second stage of a 2-stage simultaneous equations model, using
corrected asymptotic variances. The variables are as defined in Appendix A. Column 1 is a probit
model where 1 = Collateral, and 0 = No Collateral. Column 2 is also a probit model where
1 = Multiple Creditors, and 0 = Single Creditor.
∗ Coefficients significant at the 10% level.
∗∗ Coefficients significant at the 5% level.
∗∗∗ Coefficients significant at the 1% level.
collateral and firms of higher quality put up less collateral for private debt.7 This
is in support of Rajan and Winton’s (1995) theory that private lenders only take
collateral in the bad state as incentives to monitor and force efficient liquidation.
In contrast, our result here does not support the argument by Bolton and
Scharfstein (1996) and John, Lynch and Puri (2002) that higher quality borrowers
minimize strategic defaults by increasing collateralization. The significant loan
purpose proxy is in support of Gilson and Warner’s (1999) finding that firms that
are looking for managerial freedom and “flexibility” put up less collateral. The
loan purposes labeled “1” and “2” are for general purposes and those labeled “3”
and “4” are those for special purposes like takeovers and acquisitions. Hence,
managers are looking for flexibility, freedom and discretion in everyday affairs,
while offering collateral only in special, risky projects like takeovers. We also find
that collateralization depends on the type of debt issued. Collateral is offered for
term loans, but not for revolvers. Neither the growth nor the probability of distress
proxies are significant in the collateralization regression.
In column (2) of the probit regression analyzing multiple creditors, the leverage
and size proxies are significant. The positive sign on the size proxy is in accordance
Collateralization and the Number of Lenders in Private Debt Contracts 243
with the hypothesis of Repullo and Suarez (1998) that larger firms would prefer
multiple lenders since single lenders would be unable to enforce a credible threat
of liquidation. It is also in support of Detragiache et al.’s (2001) hypothesis that
larger firms maintain multiple creditors to avoid liquidity risk due to the lender and
to ensure a stable supply of credit. The negative sign on the leverage proxy implies
that more highly leveraged firms prefer single creditors to multiple creditors, in
support of the findings of Degryse and Ongena (2001). They state that “highly
leveraged firms choose a single relationship since firms judged fit to carry a lot of
debt do not face credit rationing by their single bank.” Hence this finding is quite
robust to regime changes, and validated in both the Norwegian and U.S. samples.
This is also in support of Johnson (1998) who found that leverage was
statistically and economically significantly higher for firms with bank debt due
to the fact that bank debt attenuates negative effects on leverage of potential asset
substitution problems. However it is inconsistent with Houston and James’ (1996)
finding of a negative relationship between leverage and the proportion of bank
financing. They explain their results as consistent with either economies of scale
in issuing public claims or the notion that bank monitoring is a public good and
firms with more leverage may use the same level of bank debt.
Degryse and Ongena (2001) and Detragiache et al. (2002) find that profitability
impacts the number of bank relationships. Since our analysis focuses on
incremental decisions that may be based on relationship or transaction lending
(Boot & Thakor, 2000), specifically adjusting for profitability should not have any
impact on the robustness of our results. To test this, we run a set of simultaneous
regressions on collateralization and the number of creditors, but adding a fitted
profitability variable (ratio of operating income to total assets) in the second stage
of the model to analyze this effect.8 The results are presented in Table 7.
As expected, our base results are materially unchanged. We re-ran the
regressions using alternative measures for profitability, namely the EBIT to total
assets ratio, and Altman’s (1993) cumulative profitability measure, calculated as
retained earnings divided by total assets. Again, we did not find any difference in
results.
We also evaluated alternative specifications of the base model and did not
find any major difference in results. Since the correlation matrix (Table B.1 in
Appendix B) indicates that some variables are highly and significantly correlated
with others (for example market leverage), we drop each one of the variables
from the model in turn to test for robustness of the results. Two sample results
244 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
Note: Results shown are those from the second stage of a 2-stage simultaneous equations model, using
corrected asymptotic variances, with a fitted profitability variable inserted in the second stage.
The variables are as defined in Appendix A. Column 1 is a probit model where 1 = Collateral,
and 0 = No Collateral. Column 2 is also a probit model where 1 = Multiple Creditors, and 0 =
Single Creditor.
∗∗ Coefficients significant at the 5% level.
∗∗∗ Coefficients significant at the 1% level.
are included in the appendix (Tables C.1 and C.2, with market leverage and the
market to book ratio dropped, respectively) that show that even when these highly
correlated variables are dropped, the results do not change materially.
7. CONCLUSIONS
This paper empirically assesses theoretical predictions on two aspects of debt
contracts, the number of creditors and use of collateral, employing a current
database of private U.S. corporate debt transactions. We simultaneously model
the two aspects of debt contracts, since there may exist trade-off effects not
captured in non-simultaneous models. Further, we focus on these two aspects
of debt contracts from the borrower’s strategic perspective, and not from a lender’s
pricing perspective.
We do not find support for Bolton and Scharfstein’s (1996) theory that the two
decisions are made jointly to minimize strategic defaults and the loss of value
Collateralization and the Number of Lenders in Private Debt Contracts 245
in liquidity defaults. Rather, we find that the two decisions are not interrelated.
Firms of lower quality put up more collateral and firms of higher quality put up
less collateral possibly because private lenders only take collateral in the bad state
as incentives to monitor and force efficient liquidation (Rajan & Winton, 1995).
Furthermore, those offering collateral also give lenders greater control as well as
the ability to threaten liquidation of inefficient projects (Gorton & Kahn, 1993;
Rajan & Winton, 1995). We also find that firms that are looking for managerial
freedom and “flexibility” put up less collateral (Gilson & Warner, 2000).
Large U.S. firms in our sample with greater probability of financial distress,
or higher credit risk firms find the option to renegotiate more valuable and hence
borrow from fewer lenders (Berlin & Mester, 1992; Chemmanur & Fulghieri,
1994). Due to moral hazard reasons as well as to limit the single lender’s bargaining
power and mitigate any information monopoly, firms with greater growth oppor-
tunities borrow from multiple creditors (Rajan, 1992; Repullo & Suarez, 1998).
Larger firms prefer multiple creditors for the same moral hazard reasons (Repullo &
Suarez, 1998), or to avoid liquidity risk due to the lender and to ensure a stable
supply of credit (Detragiache et al., 2001). Finally, we find that more leveraged
firms choose a single lender, perhaps due to the fact that they do not face credit
rationing from their relationship lender (Degryse & Ongena, 2001), or due to
the attenuation of the negative effects on leverage of potential asset substitution
problems (Johnson, 1998). Overall, larger, less leveraged firms choose multiple
lenders and smaller, more leveraged firms choose a single lender.
We also discover differences and similarities in the behavior of U.S. and
European firms, perhaps due to the different financial systems in which they
operate. More leveraged firms in both environments prefer single lenders to
multiple creditors. However, profitable European firms operating in a highly
concentrated financial market maintain single relationships, while large, profitable
U.S. firms working in a large and diffused financial market, maintain multiple
relationships to mitigate lenders’ informational monopolies (Houston & James,
1996; Rajan, 1992).
Viewed broadly, our results show that the variables identified in prior studies,
firm quality, the probability of financial distress, growth opportunities, the desire
for managerial flexibility, and firm size, play major roles in explaining two
important features of loan contracts in the U.S.
NOTES
1. Other users of the LPC database include Carey, Post and Sharpe (1998), who use the
database to empirically establish the basic difference between bank and non-bank lenders
246 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
and borrowers. Angbazo, Mei and Saunders (1998) use the database to analyze credit spreads
on highly leveraged transaction loans. Hubbard et al. (1999) use it in analyzing bank effects
in borrower’s cost of funds. Strahan (1999) uses it to analyze the pricing of borrower risk.
Dennis et al. (2000) use it to analyze the determinants of contract terms in bank revolving
credit agreements.
2. Strahan (1999) compared the DealScan firms with those found in the COMPUSTAT
universe and found that the median DealScan firm was not only larger (by a factor of four)
but also more profitable (almost double) than the median COMPUSTAT firm.
3. See Chapter 8 of Maddala (1983) for more details on the two-stage estimation
procedure.
4. We appreciate the help of William Greene in resolving some econometric issues with
the model.
5. Statistical significance for ratio differences were calculated using the non-parametric
Wilcoxon sum rank test since ratios are not normally distributed and other tests would not
be valid.
6. As suggested by Gujarati (1995) and Greene (1997), since some of the variables are
significant in the regression, and the R2 is not very high (less than 75%), we conclude that
correlation amongst regressors is not a problem. The full correlation matrix is presented in
Appendix B along with some alternate regressions in Appendix C.
7. Recall that COMPUSTAT records S&P bond ratings in an “inverse” fashion: 2 for
AAA and 21 for C.
8. Using the method explained in Section 5, the profitability variable is estimated in the
first stage as an independent variable. The predicted value is then inserted into the second
stage as a fixed value to estimate the rest of the variables.
ACKNOWLEDGMENTS
We thank James Darroch, Vasil Mihov, Kamphol Panyagometh, Pauline Shum,
John Smithin, and participants at the 2000 FMA Annual Meeting for comments
that helped to improve the paper. We would also like to acknowledge the help
of William Greene in resolving econometric issues and Sumon Mazumdar and
Yuxing Yan in accessing the DealScan database.
REFERENCES
Altman, E. I. (1968). Financial ratios. Discriminant analysis and the prediction of corporate bankruptcy.
Journal of Finance.
Altman, E. I. (1993). Corporate financial distress and bankruptcy: A complete guide to predicting and
avoiding distress and profiting from bankruptcy. New York: Wiley.
Angbazo, A., Mei, J., & Saunders, A. (1998). Credit spreads in the market for highly leveraged
transaction loans. Journal of Banking and Finance, 22, 1249–1282.
Berlin, M., & Mester, L. (1992). Debt covenants and renegotiation. Journal of Financial Intermediation,
2, 95–133.
Collateralization and the Number of Lenders in Private Debt Contracts 247
Blume, M. E., Lim, F., & Mackinlay, A. C. (1998). The declining credit quality of U.S. corporate debt:
Myth or reality? Journal of Finance, 53, 1389–1412.
Bolton, M., & Scharfstein, D. (1996). Optimal debt structure and the number of creditors. Journal of
Political Economy, 104(1), 1–25.
Boot, A. W., & Thakor, A. V. (2000). Can relationship banking survive. Journal of Finance, 55,
679–713.
Carey, M., Post, M., & Sharpe, S. A. (1998). Does corporate lending by banks and finance companies
differ? Evidence on specialization in private debt contracting. Journal of Finance, 53, 3.
Chemmanur, T. J., & Fulghieri, P. (1994). Reputation, renegotiation and the choice between bank loans
and publicly traded debt. Review of Financial Studies, 7, 475–506.
D’Auria, C., Foglia, A., & Reedtz, P. M. (1999). Bank interest rates and credit relationships in Italy.
Journal of Banking and Finance, 23, 1067–1093.
Degryse, H., & Ongena, S. (2001). Bank relationships and firm profitability. Financial Management,
30(1), 9–34.
Dennis, S., Nandy, D., & Sharpe, I. G. (2000). The determinants of contract terms in bank revolving
credit agreements. Journal of Financial and Quantitative Analysis, 35, 87–110.
Detragiache, E., Garella, P. G., & Guiso, L. (2000). Multiple versus single banking relationships:
Theory and evidence. Journal of Finance, 55, 1133–1161.
Farinha, L., & Santos, J. A. C. (2002). Switching from single to multiple bank lending relationships:
Determinants and implications. Journal of Financial Intermediation, 11, 124–151.
Gilson, S. C., & Warner, J. B. (2000). Private versus public debt: Evidence from firms that replace bank
loans with junk bonds. Working Paper. Harvard Business School.
Gorton, G., & Kahn, J. (1993). The design of bank loan contracts, collateral, and renegotiation. Working
Paper. National Bureau of Economic Research.
Greene, W. H. (1999). Econometric analysis. Upper Saddle River, NJ: Prentice-Hall.
Gujarati, D. N. (1995). Basic econometrics. McGraw-Hill.
Harhoff, D., & Korting, T. (1998). How many creditors does it take to Tango? Berlin:
Wissenschaftszentrum.
Houston, J., & James, C. (1996). Bank information monopolies and the mix of private and public debt
claims. Journal of Finance, 51, 1863–1890.
Houston, J., & James, C. (2001). Do relationships have limits? Banking relationships, financial
constraints, and investment. Journal of Business, 74, 347–374.
Hubbard, R. G., Kuttner, K. N., & Palia, D. N. (1999). Are there “bank effects” in borrowers’ cost
of funds?: Evidence from a matched sample of borrowers and banks. Working Paper. Federal
Reserve Bank of New York.
John, K., Lynch, A. W., & Puri, M. (2002). Credit ratings, collateral and loan characteristics:
Implications for yield. Journal of Business.
Johnson, S. A. (1998). Effect of bank debt on optimal capital structure. Financial Management, 27(1),
47–56.
Lang, L., Ofek, E., & Stulz, R. (1996). Leverage, investment and firm growth. Journal of Financial
Economics, 40, 3–29.
Maddala, G. S. (1983). Limited dependent and qualitative variables in economics. New York, NY:
Cambridge University Press.
Mallar, C. D. (1977). The estimation of simultaneous probability models. Econometrica, 45(7),
1717–1722.
McKelvey, R., & Zavoina, W. (1975). A statistical model for the analysis of ordinal level dependent
variables. Journal of Mathematical Sociology, 4, 103–120.
248 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
Petersen, M., & Rajan, R. (1994). The benefits of lending relationships: Evidence from small business
data. Journal of Finance, 49, 3–37.
Pozzolo, A. F. (2002). Secured lending and borrowers’ riskiness. Working Paper. Banca d’ Italia.
Rajan, R. (1992). Insiders and outsiders: The choice between informed and arm’s-length debt. Journal
of Finance, 47, 1367–1400.
Rajan, R., & Winton, A. (1995). Covenants and collateral as incentives to monitor. Journal of Finance,
50, 1113–1146.
Repullo, R., & Suarez, J. (1998). Monitoring, liquidation, and security design. Review of Financial
Studies, 11(1), 163–187.
Shockley, R. L., & Thakor, A. V. (1997). Bank loan commitment contracts: Data, theory, and tests.
Journal of Money, Credit, and Banking, 29(4), 517–534.
Strahan, P. E. (1999). Borrower risk and the price and nonprice terms of bank loans. Working Paper.
Federal Reserve Bank of New York.
APPENDIX A
DETAILED FORMULATION OF VARIABLES
Variable Definition (and Data Source)
BONDRAT Standard and Poor’s bond rating; the ratings of AAA, AA
through C are translated to an ordinal scale ranging from 2
to 21 by COMPUSTAT.
CUMPROF Cumulative Profitability = Retained Earnings/Total Assets
(COMPUSTAT)
DEBTASST Total Debt/Total Assets (COMPUSTAT)
EBITSALE EBIT/Total Sales (COMPUSTAT)
EBITTA EBIT/Total Assets (COMPUSTAT)
FACMATUR Loan Maturity in Days (LPC)
FACSIZE Loan Size (LPC)
FACTA Loan Size/Total Assets (COMPUSTAT; LPC)
FACTD Loan Size/Total Debt (COMPUSTAT; LPC)
Loans are categorized into one of four groups depending on
their stated purpose as follows.
PURPGRP (1) General corporate purposes.
(2) Debt repayment/consolidation.
(3) Take over/acquisition.
(4) Other.
The Loan (or facility) type is also categorized into one of
four groups as follows.
FACGRP (1) Short term revolver.
(2) Long term revolver.
Collateralization and the Number of Lenders in Private Debt Contracts 249
Note: The numbers below the correlation coefficients are p-values indicating significance of the correlation.
∗ Coefficients significant at the 10% level.
∗∗ Coefficients significant at the 5% level.
∗∗∗ Coefficients significant at the 1% level.
Collateralization and the Number of Lenders in Private Debt Contracts 251
APPENDIX C
ROBUSTNESS CHECKS FOR REGRESSION RESULTS
Note: Results shown are those from the second stage of a 2-stage simultaneous equations model, using
corrected asymptotic variances, with the market leverage variable dropped. The variables are as
defined in Appendix A. Column 1 is a probit model where 1 = Collateral, and 0 = No Collateral.
Column 2 is also a probit model where 1 = Multiple Creditors, and 0 = Single Creditor.
∗ Coefficients significant at the 10% level.
∗∗ Coefficients significant at the 5% level.
∗∗∗ Coefficients significant at the 1% level.
252 GORDON S. ROBERTS AND NADEEM A. SIDDIQI
Note: Results shown are those from the second stage of a 2-stage simultaneous equations model, using
corrected asymptotic variances, with the growth potential variable dropped. The variables are as
defined in Appendix A. Column 1 is a probit model where 1 = Collateral, and 0 = No Collateral.
Column 2 is also a probit model where 1 = Multiple Creditors, and 0 = Single Creditor.
∗ Coefficients significant at the 10% level.
∗∗ Coefficients significant at the 5% level.
∗∗∗ Coefficients significant at the 1% level.
AN EXAMINATION OF THE
EFFICIENCY OF SINGLE VS.
MULTIPLE COMMON BOND
CREDIT UNIONS
ABSTRACT
As of 1982, federal credit unions were allowed to add select employee groups
and thus create institutions with multiple-group common bonds. We examine
the efficiency of single bond and multiple bond federal-chartered credit
unions by using data envelopment analysis (DEA), a non-parametric, linear
programming methodology. Results indicate that multiple bond credit unions
have better pure technical efficiency than single bond credit unions. However,
single bond credit unions appear to be more scale efficient than the multiple
bond credit unions. Our results also indicate that members of multiple bond
credit unions may derive greater wealth gains than members of single bond
credit unions.
1. INTRODUCTION
In 1981, approximately 500 federal-chartered credit unions closed their doors
due to failure or liquidation. The U.S. economy was suffering through one of
its worst recessions since the early 1930s. The number of business failures rose
Research in Finance
Research in Finance, Volume 21, 253–263
Copyright © 2004 by Elsevier Ltd.
All rights of reproduction in any form reserved
ISSN: 0196-3821/doi:10.1016/S0196-3821(04)21011-5
253
254 JAMES D. TRIPP ET AL.
2. RELEVANT LITERATURE
Over the past twenty years, many researchers have studied the issue of efficiency
in financial institutions. The bulk of this research has focused on the commercial
banking industry. Studies focusing on the banking industry generally conclude
that inefficiencies found are primarily due to the inefficient utilization of resources
(poor management) rather that scale inefficiencies related to inappropriate size (see
Aly et al., 1990; Elysasiani & Mehdian, 1990; Ferrier & Lovell, 1990; Grabowski
et al., 1994; Miller & Noulas, 1996; Rangan et al., 1988).
The efficiency of one type of financial institution, credit unions, has been less
frequently considered. In addition to the Emmons and Schmid (1999) research,
a general study of credit union efficiency was undertaken by Fried, Lovell and
Vanden Eeckaut (1993). They reported an average of 20% productive inefficiency
in the U.S. credit unions they examined. Other credit union studies have focused
on efficiency issues in the Australian market. Worthington (1999, 2000) reported
overall technical inefficiency at 46% in one study and 30% in another. However, it
should be noted that there are substantial differences between U.S. and Australian
256 JAMES D. TRIPP ET AL.
credit unions. For example, Australian credit unions pay a federal income tax while
U.S. credit unions do not. In order to compete effectively with other credit unions
and other financial institutions, it is imperative for credit unions to be as efficient
as possible.
3. DATA
Data were collected for the years 1998–2002 from the National Credit Union
Administration’s on-line database (NCUA, 2003) that contains financial and
statistical reports filed by U.S. credit unions. For our study, we chose occupational
credit unions, both single bond (Type 13) and multiple bond (Type 43), whose
primary sponsors operated in the machine manufacturing industry. As mentioned
earlier, credit unions with sponsors operating in the more cyclical business sectors
face more concentration risk and therefore would be more likely to diversify by
adding SEGs to the existing membership. We restricted our sample to credit unions
operating in the machine-manufacturing sector to limit the impact of economic
variability across industries on our results. The resulting data set included 36
single bond credit unions and 58 multiple bond.
In order to test for efficiency, it must first be determined what specifically do
credit unions produce, or what is their output. Two approaches used in previous
studies regarding financial institutions were considered: the intermediation
approach and the production approach. Under the intermediation approach,
financial institutions are considered producers of services that collect deposits and
purchase funds that are intermediated into loans and other assets. Additionally,
deposits are viewed as inputs, along with labor and capital, while measures of
output are defined as the financial institution’s various dollar volumes of earning
assets. Finally, under the intermediation approach, total cost is defined as total
operating costs and interest expense.
Alternatively, under the production approach, financial institutions are viewed
as producers of services related to individual deposit and loan accounts. Both
capital and labor are used to produce these account services. Output is defined as
the number of deposit and loan accounts and total cost is defined as total operating
costs and does not include interest expense. Also, according to Humphrey
(1985), some control factor for differences in the average sizes of accounts across
different-sized financial institutions must be incorporated under this approach.
According to Clark (1988), reasonable arguments have been made for both
approaches to modeling a financial institution’s behavior. Clark also noted that the
empirical results do not appear to be sensitive to the method used in defining costs
and outputs.
An Examination of the Efficiency of Single vs. Multiple Common Bond 257
For purposes of this paper, the intermediation approach is utilized. Using this
method, input and output variables comparable to those used by Miller and
Noulas (1996) in their study of bank performance were selected. The input variables
(in dollars) are:
(1) Total shares and deposits (TSD)
(2) Total interest expense (TIE = interest on borrowed money + dividends on
shares + interest on deposits)
(3) Total non-interest expense (TNIE = total operating expenses less provision
for loan losses).
The outputs variables (in dollars) are:
(1) Total loans (TL = unsecured credit card loans + all other unsecured loans +
new vehicle loans + used vehicle loans + total all other loans to members +
total first mortgage real estate loans + total other real estate loans)
(2) Total investments (TIN)
(3) Total income (TI = interest on loans – interest refunded + income from
investments + income from trading securities + fee income + other operating
income).
Sample statistics are displayed in Table 1.
4. METHODOLOGY
These input and output variables are utilized in a data envelopment analysis model
(DEA) in order to examine efficiency of multiple common bond and single bond
credit unions. The use of DEA for purposes such as this was first introduced by
Charnes, Cooper and Rhodes (CCR) in 1978. DEA models seek to determine
258 JAMES D. TRIPP ET AL.
which decision making units (DMUs) under consideration are efficient given the
use of m inputs for generating s outputs. The efficient DMUs out of the set of n
DMUj (where j = 1, . . ., n) will lie on the envelopment surface. Inefficient DMUs
do not lie on the surface. In developing the envelopment surface through solving
mathematical programming models, each DMUj is examined relative to the other
DMUs in the sample. Consequently, the measured efficiency is sample specific.
The first model utilized examines an input-oriented data envelopment frontier
with variable returns to scale as originally developed by Banker, Charnes, and
Cooper (BCC model, 1984). The model uses n DMUj (j = 1, 2, . . ., n) observations
with m inputs represented by xij (i = 1, 2, . . ., m) and s outputs represented by yrj
(r = 1, 2, . . ., s). With variable returns to scale the model can be formulated as
follows (Zhu, 2003):
m s
min− s−i + s+
r
i=1 r=1
subject to
n
j x ij + s −
i = x io , i = 1, 2, . . . , m;
j=1
n
j y rj − s +
r = y ro , r = 1, 2, . . . , s;
j=1
j ≥ 0, j = 1, 2, . . . , n
and
n
j = 1
j=1
with s − +
i representing the input slack, s r representing the output slack, ??, is a
non-archimedean constant, and λj being nonnegative scalars. This BCC model
provides a measure of pure technical efficiency.
For the constant returns to scale input-oriented data envelopment model as
originally developed
by Charnes, Cooper and Rhodes (CCR model, 1978), the last
condition above, nj=1 j = 1, is dropped. The CCR model provides an overall
technical efficiency measure.
The desired output from the models is an efficiency score. The efficient target
with variable or constant returns to scale is given by (Zhu, 2003):
∗
x̂io = ∗ x io − s −
i i = 1, 2, . . . , m
An Examination of the Efficiency of Single vs. Multiple Common Bond 259
∗
ŷro = y ro + s +
r r = 1, 2, . . . , s
5. RESULTS
The causes of inefficiency for a DMU can stem from two sources. The DMU’s
inefficiency may be a result of inefficient operations (as measured by pure technical
efficiency) or be attributed to disadvantageous conditions (as measured by scale
efficiency) related to the environment in which it operates (Cooper et al., 2000).
The overall technical efficiency (OTE), pure technical efficiency (PTE), and
scale efficiency (SE) scores for the 36 single bond (Type 13) and 52 multiple bond
(Type 43) credit unions are displayed in Table 2.
The OTE scores for both the single bond and the multiple bond credit unions
showed relatively high levels of global efficiency under the assumption of constant-
returns-to-scale. The OTE scores for single bond credit unions ranged from a low
260 JAMES D. TRIPP ET AL.
of 0.895255 in 1999 to a high of 0.950947 in 1998 and averaged 0.916051 for the
five years studied. The multiple bond credit unions’ OTE scores ranged from a low
of 0.877354 in 1999 to a high of 0.952796 in 2001 and averaged 0.919742. These
results indicate that the inefficiency was approximately 8% for both types and
is smaller than reported in other U.S. credit union studies. However, sample and
methodological differences when utilizing DEA make it very difficult to compare
our results to other credit unions studies.
The PTE scores reported in Table 2 provide efficiency measures using a local
measure of scale under the assumption of variable returns to scale. The findings
show that the multiple bond credit unions were more efficient on a pure technical
basis in all five years examined. These differences were significant at the 0.10 level
in 2002 and the 0.05 level in 2000. This finding is indicative of greater managerial
efficiency in the multiple bond credit unions.
However, the scale efficiency (SE) scores displayed in Table 2 highlight an
efficiency advantage for the single bond credit unions. As reported earlier in
Table 1, the average single bond credit union was substantially smaller than
the average multiple bond credit union over the 1998–2002 time period (total
assets of $2.3 million vs. $28.9 million). According to Cooper et al. (2000), this
An Examination of the Efficiency of Single vs. Multiple Common Bond 261
may indicate that these smaller credit unions are operating under advantageous
conditions relative to larger multi-bond credit unions. SE results in Table 2 reveal
that single bond credit unions had significantly higher efficiency scores in four of
the five years studied.
Table 3 presents the results regarding differences in member wealth gains
between single bond and multiple bond credit unions over the 1998–2002 time-
period. The yield spread is lower in all five years for the multiple bond credit unions
and significantly so in 2001 (0.10 level) and 2002 (0.01 level). Multiple bond credit
unions had significantly lower loan rates in all five years studied while the single
bond credit unions provide higher share rates. However, as interest rates in the
economy trended lower in 2001 and 2002, there was no significant difference
in the share rates of the two types of credit unions. This finding indicates that
multiple bond credit unions may be improving their members’ financial welfare
by providing better interest rates than single bond credit unions.
6. CONCLUSIONS
Prior credit union research has shown that the creation of multiple bond credit
unions has reduced the amount of concentration risk in the industry. The results of
262 JAMES D. TRIPP ET AL.
our study reveal additional benefits of this diversification. As measured by the pure
technical efficiency scores, multiple bond credit unions appear to offer superior
managerial expertise. These larger multiple bond credit unions are quite likely
managed by individuals with higher levels of education and greater experience
resulting in greater operational efficiency. In addition, these multiple bond credit
unions appear to pass along the financial gains of this managerial efficiency to
their members by offering superior rates over their single bond counterparts. The
only negative finding regarding multiple bond credit unions concerns their scale
efficiency. Our results indicate that the larger multi-bond credit unions are less
efficient on a size basis than the smaller single bond credit unions.
From a regulatory perspective, the results of our study should be encouraging
since both single and multiple bond credit unions appear to be operating at relatively
high levels of overall efficiency. Additionally, the finding that multiple bond credit
unions may offer better managerial efficiency and wealth gains to their members
might justify regulators promoting further expansion in this area.
REFERENCES
Aly, H. Y., Grabowski, R., Pasurka, C., & Rangan, N. (1990). Technical, scale and allocative efficiencies
in U.S. banking: An empirical investigation. Review of Economics and Statistics, 72, 211–218.
Banker, R. D., Charnes, A., & Cooper, W. W. (1984). Some models for estimating technical and scale
inefficiencies in data envelopment analysis. Management Science, 30(9), 1078–1092.
Charnes, A., Cooper, W. W., & Rhodes, E. (1978). Measuring efficiency of decisions making units.
European Journal of Operational Research, 2, 429–444.
Clark, J. A. (1988). Economies of scale and scope at depository financial institutions: A review of the
literature. Economic Review, Federal Reserve Bank of Kansas City, September/October, 16–23.
Cooper, W. C., Seiford, L. M., & Tone, K. (2000). Data envelopment analysis: A comprehensive text
with models, applications, references and DEA-solver software. Boston: Kluwer.
Elysasiani, E., & Mehdian, S. M. (1990). A nonparametric approach to measurement of efficiency and
technological change: The case of large U.S. commercial banks. Journal of Financial Services
Research, 4, 157–168.
Emmons, W. R., & Schmid, F. A. (1999). Credit unions and the common bond. The Federal Reserve
Bank of Saint Louis Review, 81(September/October), 41–64.
Ferrier, G. D., & Lovell, C. A. K. (1990). Measuring cost efficiency in banking: Econometric and linear
programming evidence. Journal of Econometrics, 46(October/November), 229–245.
Frame, W. S., Karels, G. V., & McClatchey, C. (2002). The effect of the common bond and membership
expansion on credit union risk. Financial Review, 37(November), 613–636.
Fried, H. O., Lovell, C. A. K., & Vanden Eeckaut, P. (1993). Evaluating the performance of U.S. credit
unions. Journal of Banking & Finance, 17, 251–265.
Grabowski, R., Rangan, N., & Rezvanian, R. (1994). The effect of deregulation on the efficiency of
U.S. banking firms. Journal of Economics and Business, 46, 39–54.
Hinson, Y. L., & Juras, P. E. (2002). Performance evaluation of credit unions: Reaping the benefit of
tax status. Journal of Managerial Issues, 14, 145–161.
An Examination of the Efficiency of Single vs. Multiple Common Bond 263
Humphrey, D. B. (1985). Cost and scale economies in bank intermediation. In: R. Aspinwall & R.
Eisenbeis (Eds), Handbook for Bank Strategy. New York: Wiley.
Johnson, E. (1999). Multiple-group memberships, http://www.cuna.org/data/newsnow.
Kohers, T. (1986). Parent organization’s activities as a cause for differential credit union operating
characteristics. Southwest Journal of Business and Economics, 4(Fall), 11–21.
Miller, S. M., & Noulas, A. G. (1996). The technical efficiency of large bank production. Journal of
Banking & Finance, 20, 495–509.
Mosheim, R. (2004). The efficiency of credit unions in Hawaii and Puerto Rico.
http://www.ceterisparibus.uprm.edu/articulos/vol3/articulo6.htm.
NCUA (2003). Call report data files under FOIA. http://www.ncua.gov/data/FOIA/foia.html,
Alexandria: NCUA.
Rangan, N., Grabowski, R., Aly, H. Y., & Pasurka, C. (1988). The technical efficiency of U.S. banks.
Economic Letters, 28, 169–175.
Worthington, A. C. (1999). Measuring technical efficiency in Australian credit unions. The Manchester
School, 67(March), 231–248.
Worthington, A. C. (2000). Cost efficiency in Australian non-bank financial institutions: A non-
parametric approach. Accounting and Finance, 40, 75–97.
Zhu, J. (2003). Quantitative models for performance evaluation and benchmarking: Data envelopment
analysis with spreadsheets and DEA Excel solver. Boston: Kluwer.