2015 Book HandbookOfFinancialEconometric PDF
2015 Book HandbookOfFinancialEconometric PDF
John C. Lee
Editors
Handbook of
Financial
Econometrics
and Statistics
1 3Reference
Handbook of Financial Econometrics
and Statistics
Cheng-Few Lee • John C. Lee
Editors
Handbook of
Financial Econometrics
and Statistics
John C. Lee
Center for PBBEF Research
North Brunswick, NJ, USA
Financial econometrics and statistics have become very important tools for empirical
research in both finance and accounting. Econometric methods are important tools for
doing asset pricing, corporate finance, options and futures, and conducting financial
accounting research. Important econometric methods used in this research include:
single equation multiple regression, simultaneous regression, panel data analysis,
time series analysis, spectral analysis, non-parametric analysis, semi-parametric
analysis, GMM analysis, and other methods.
Portfolio theory and management research have used different statistical distri-
butions, such as normal distribution, stable distribution, and log normal distribu-
tion. Options and futures research have used binomial distribution, log normal
distribution, non-central chi square distribution, Poission distribution, and others.
Auditing research has used sampling survey techniques to determine the sampling
error and non-sampling error for auditing.
Based upon our years of experience working in the industry, teaching classes,
conducting research, writing textbooks, and editing journals on the subject of
financial econometrics and statistics, this handbook will review, discuss, and
integrate theoretical, methodological, and practical issues of financial econometrics
and statistics. There are 99 chapters in this handbook. Chapter 1 presents an
introduction of financial econometrics and statistics and shows how readers can
use this handbook. The following chapters, which have been contributed by
accredited authors, can be classified by the following 14 topics.
i. Financial Accounting (Chapters 2, 9, 10, 61, 97)
ii. Mutual Funds (Chapters 3, 24, 25, 68, 88)
iii. Microstructure (Chapters 4, 44, 96, 99)
iv. Corporate Finance (Chapters 5, 21, 30, 38, 42, 46, 60, 63, 75, 79, 95)
v. Asset Pricing (Chapters 6, 15, 22, 28, 34, 36, 39, 45, 47, 50, 81, 85, 87, 93)
vi. Options (Chapters 7, 32, 37, 55, 65, 84, 86, 90, 98)
vii. Portfolio Analysis (Chapters 8, 26, 35, 53, 67, 73, 80, 83)
viii. Risk Management (Chapters 11, 13, 16, 17, 23, 27, 41, 51, 54, 72, 91, 92)
ix. International Finance (Chapters 12, 40, 43, 59, 69)
x. Event Study (Chapters 14)
xi. Methodology (Chapters 18, 19, 20, 29, 31, 33, 49, 52, 56, 57, 58, 62, 74, 76,
77, 78, 82, 89)
v
vi Preface
vii
About the Editors
ix
x About the Editors
author of the book on how to use MINITAB and Microsoft Excel to do statistical
analysis which is a companion text to Statistics of Business and Financial
Economics, 2nd and 3rd, of which he is one of the co-authors. In addition, he has
also coauthored the textbooks Financial Analysis, Planning and Forecasting,
2ed (with Cheng F. Lee and Alice C. Lee), and Security Analysis, Portfolio
Management, and Financial Derivatives (with Cheng F. Lee, Joseph Finnerty,
Alice C. Lee, and Donald Wort). John has been a Senior Technology Officer at
the Chase Manhattan Bank and Assistant Vice President at Merrill Lynch.
Currently, he is the Director of the Center for PBBEF Research.
Contents
Volume 1
xi
xii Contents
Volume 2
Volume 3
59 Pre-IT Policy, Post-IT Policy, and the Real Sphere in Turkey . . . 1647
Ahmed Hachicha and Cheng-Few Lee
Volume 4
xix
xx Contributors
Contents
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2 Financial Econometrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.1 Single Equation Regression Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2.2 Simultaneous Equation Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2.3 Panel Data Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.2.4 Alternative Methods to Deal with Measurement Error . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.2.5 Time Series Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.6 Spectral Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.3 Financial Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.1 Important Statistical Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.2 Principle Components and Factor Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.3 Nonparametric and Semi-parametric Analyses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.3.4 Cluster Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
1.4 Applications of Financial Econometrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.5 Applications of Financial Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.6 Overall Discussion of Papers in this Handbook . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
1.7 Summary and Conclusion Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
Appendix 1: Brief Abstracts and Keywords for Chapters 2 to 99 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
Abstract
The main purposes of this introduction chapter are (i) to discuss important
financial econometrics and statistics which have been used in finance and
accounting research and (ii) to present an overview of 98 chapters which have
been included in this handbook. Sections 1.2 and 1.3 briefly review and
discuss financial econometrics and statistics. Sections 1.4 and 1.5 discuss
application of financial econometrics and statistics. Section 1.6 first classifies
98 chapters into 14 groups in accordance with subjects and topics. Then this
section has classified the keywords from each chapter into two groups:
finance and accounting topics and methodology topics. Overall, this chapter
gives readers of this handbook guideline of how to apply this handbook to
their research.
1.1 Introduction
Financial econometrics and statistics have become very important tools for empir-
ical research in both finance and accounting. Econometric methods are important
tools for asset-pricing, corporate finance, options, and futures, and conducting
financial accounting research. Important econometric methods used in this research
include: single equation multiple regression, simultaneous regression, panel data
analysis, time-series analysis, spectral analysis, nonparametric analysis, semi-
parametric analysis, GMM analysis, and other methods.
Portfolio theory and management research have used different statistics
distributions, such as normal distribution, stable distribution, and log-normal dis-
tribution. Options and futures research have used binomial distribution, log-normal
distribution, non-central Chi-square distribution, Poisson distribution, and others.
Auditing research has used sampling survey techniques to determine the sampling
error and non-sampling error for auditing. Risk management research has used
Copula distribution and other distributions.
Section 1.1 is the introduction. Section 1.2 discusses financial econometrics.
In this section, we have six subsections. These subsections include single equation
regression methods, simultaneous equation models, panel data analysis, as well
as alternative methods to deal with measurement error, time-series analysis,
and spectral analysis. In the next section, Sect. 1.3, we discuss financial statistics.
Within financial statistics, we discuss six subtopics, including statistical
distributions; principle components and factor analysis; nonparametric, semi-
parametric, and GMM analyses; and cluster analysis. After exploring these topics,
we discuss the applications of financial econometrics and financial statistics in
Sects. 1.4 and 1.5. In Sect. 1.6, we discuss the overview of all papers included
in this handbook in accordance with the subject and methodologies used in the
papers. Finally in Sect. 1.7, we summarize all the chapters in this handbook and add
our concluding remarks.
1 Introduction to Financial Econometrics and Statistics 3
There are important issues related to single equation regression estimation method.
They are (a) Heteroskedasticity, (b) Specification error, (c) Measurement error,
(d) Skewness and kurtosis effect, (e) Nonlinear regression and Box-Cox transfor-
mation, (f) Structural change, (g) Chow test and moving Chow test, (h) Threshold
regression, (i) Generalized fluctuation, (j) Probit and Logit regression for credit risk
analysis, (k) Poisson regression, and (l) Fuzzy regression. These issues are briefly
discussed as follows:
(a) Heteroskedasticity
– White (1980) and Newvey and West (1987) are two important papers
discussing how the heteroskedasticity test can be performed. The latter
paper discusses heteroskedasticity when there are serial correlations.
(b) Specification error
– Specification error occurs when there is missing variable in a regression analysis.
To test the existence of specification error, we can refer to the papers by Thursby
(1985), Fok et al. (1996), Cheng and Lee (1986), and Maddala et al. (1996).
(c) Measurement error
– Management error problem is when there exists imprecise independent
variable in a regression analysis. Papers by Lee and Jen (1978), Kim
(1995, 1997, 2010), Miller and Modigliani (1966), and Lee and Chen
(2012) have explored how measurement error methods can be applied to
finance research. Lee and Chen have discussed alternative errors in variable
estimation methods and their application in finance research.
(d) Skewness and kurtosis effect
– Both skewness and kurtosis are two important measurement variables to pre-
pare stock variation analysis. Papers by Lee (1976a), Sears and Wei (1988), and
Lee and Wu (1985) discuss the skewness and kurtosis issue in asset pricing.
(e) Nonlinear regression and Box-Cox transformation
– Nonlinear regression and Box-Cox transformation are important tools for
finance, accounting, and urban economic researches. Papers by Lee (1976,
1977), Lee et al. (1990), Frecka and Lee (1983), and Liu (2006) have
discussed how nonlinear regression and Box-Cox transformation techniques
can be used to improve the specification of finance and accounting research.
Kau and Lee (1976), and Kau et al. (1986) have explored how Box-Cox
transformation can be used to conduct the empirical study of urban structure.
(f) Structural change
– Papers by Yang (1989), Lee et al. (2011b, 2013) have discussed how the
structural change model can be used to improve the empirical study of
dividend policy and the issuance of new equity.
(g) Chow test and Moving Chow test
– Chow (1960) has proposed a dummy variable approach to examine the
existence of structure change for regression analysis. Zeileis et al. (2002)
1 Introduction to Financial Econometrics and Statistics 5
have developed software programs to perform the Chow test and other
structural change models which has been frequently used in finance and
economic research.
(h) Threshold regression
– Hansen (1996, 1997, 1999, 2000a, and 2000b) have explored the issue of
threshold regressions and their applications in detecting structure change for
regression.
(i) Generalize fluctuation test
– Kuan and Hornik (1995) have discussed how the generalized fluctuation test
can be used to perform structural change to regression.
(j) Probit and Logit regression for credit risk analysis
– Probit and Logit regressions are frequently used in credit risk analysis.
Ohlson (1980) used the accounting ratio and macroeconomic data to do
credit risk analysis. Shumway (2001) has used accounting ratios and
stock rate returns for credit risk analysis in terms of Probit and
Logit regression techniques. Most recently, Hwang et al. (2008, 2009)
and Cheng et al. (2010) have discussed Probit and Logit regression for credit
risk analysis by introducing nonparametric and semi-parametric techniques
into this kind of regression analysis.
(k) Poisson regression
– Lee and Lee (2012) have discussed how the Poisson Regression can be
performed, regardless of the relationship between multiple directorships,
corporate ownership, and firm performance.
(l) Fuzzy regression
– Shapiro (2005), Angrist and Lavy (1999), and Van Der Klaauw (2002) have
discussed how Fuzzy Regression can be performed. This method has the
potential to be used in finance accounting and research.
In this section, we will discuss important issues related to panel data analysis. They are
(a) fixed effect model, (b) random effect model, and (c) clustering effect model.
Three well-known textbooks by Wooldridge (2010), Baltagi (2008) and Hsiao
(2003) have discussed the applications of panel data in finance, economics, and
accounting research. Now, we will discuss the fixed effect, random effect, and
clustering effect in panel data analysis.
(a) Fixed effect model
– Chang and Lee (1977) and Lee et al. (2011a) have discussed the role of
the fixed effect model in panel data analysis of dividend research.
(b) Random effect model
– Arellano and Bover (1995) have explored the random effect model and its role
in panel data analysis. Chang and Lee (1977) have applied both fix effect and
random effect model to investigating the relationshipbetween price per share,
dividend per share, and retained earnings per share.
(c) Clustering effect model
– Papers by Thompson (2011), Cameron et al. (2006), and Petersen (2009)
review the clustering effect model and its impact on panel data analysis.
In this section, we will discuss important models in time-series analysis. They are
(a) ARIMA, (b) ARCH, (c) GARCH, (d) fractional GARCH, and (e) combined
forecasting.
– Two well-known textbooks by Anderson (1994) and Hamilton (1994) have
discussed the issues related to time-series analysis. We will discuss some
important topics in time-series analysis in the following subsections.
– Myers (1991) discloses ARIMA’s role in time-series analysis: Lien and Shrestha
(2007) discuss ARCH and its impact on time-series analysis: Lien (2010)
discusses GARCH and its role in time-series analysis: Leon and Vaello-Sebastia
(2009) further research into GARCH and its role in time series in a model called
Fractional GARCH.
– Granger and Newbold (1973), Granger and Newbold (1974), Granger and
Ramanathan (1984) have theoretically developed combined forecasting
methods. Lee et al. (1986) have applied combined forecasting methods to
forecast market beta and accounting beta. Lee and Cummins (1998) have
shown how to use the combined forecasting methods to perform cost of capital
estimates.
Anderson (1994), Chacko and Viceira (2003), and Heston (1993) have discussed
how spectral analysis can be performed. Heston (1993) and Bakshi et al. (1997)
have applied spectral analysis in the evaluation of option pricing.
8 C.-F. Lee and J.C. Lee
Ait-Sahalia and Lo (2000), and Hutchison et al. (1994) have discussed how
nonparametric can be used in risk management and derivative securities evaluation.
Hwang et al. (2010), and Hwang et al. (2007) have used semi-parametric to conduct
credit risk analysis.
The detailed procedures to discuss how cluster analysis can be used to find
groups in data can be found in the textbook by Kaufman and Rousseeuw (1990).
Brown and Goetzmann (1997) have applied cluster analysis in mutual fund
research.
1 Introduction to Financial Econometrics and Statistics 9
In this section, we classify 98 papers (chapters 2–99) which have been presented in
Appendix 1 in accordance with (A) Chapter titles and (B) Keywords.
(A) Chapter title classification in terms of Chapter Titles
1 Introduction to Financial Econometrics and Statistics 11
This chapter has discussed important financial econometrics and statistics which
have been used in finance and accounting research. In addition, this chapter has
presented an overview of 98 chapters which have been included in this handbook.
In Sect. 1.2 “Financial Econometrics,” we have six subsections which are: a single
equation regression methods, Simultaneous equation models, Panel data analysis,
Alternative methods to deal with measurement error, Time-series analysis, and
18 C.-F. Lee and J.C. Lee
Spectral Analysis. Section 1.3 “Financial Statistics” has four subsections: Important
Statistical Distributions, Principle components and factor analysis, Nonparametric
and Semi-parametric analyses, Cluster analysis review and discuss financial
econometrics and statistics. In Sect. 1.4 “Applications of financial econometrics,”
we briefly discuss how different methodologies of financial econometrics will
be applied to the topics of finance and accounting. These methods include: Asset-
pricing Research, Corporate Finance Research, Financial Institution Research,
Investment and Portfolio Research, Option Pricing Research, Future and Hedging
Research, Mutual Fund Research, and Credit Risk Modeling. Section 1.5, “Appli-
cations of Financial Statistics,” states that financial statistics is an important tool
to conduct research in the areas of (1) Asset-pricing Research, (2) Investment
and Portfolio Research, (3) Credit Risk Management Research, (4) Market
Risk Research, (5) Operational Risk Research, (6) Option Pricing Research,
(7) Mutual Fund Research, (8) Hedge Fund Research, (9) Value-at-risk Research,
and (10) Auditing. Section 1.6 is an “Overall Discussion of Papers in this
Handbook.” It classifies 98 chapters into 14 groups in accordance to Chapter title
and keywords.
This chapter uses a Bayesian model of updating forecasts in which the bias in
forecast endogenously determines how the forecaster’s own estimates weigh into
the posterior beliefs. The model used in this chapter predicts a concave relationship
between accuracy in forecast and posterior weight that is put on the forecaster’s self-
assessment. This chapter then uses a panel regression to test the analytical findings
and find that an analyst’s experience is indeed concavely related to the forecast error.
Keywords: Financial analysts, Forecast accuracy, Information asymmetry,
Forecast bias, Bayesian updating, Panel regressions, Rational bias, Optional bias,
Analyst estimation, Analyst experience
This chapter uses simulation to gain insights into trading and market structure topic
by two statistical methods. The statistical methods we use include experimental
design, and careful controls over experimental parameters such as the instructions
given to participants. The first is discrete event simulation and the model of
computer-generated trade order flow that we describe in Sect. 3. To create a realis-
tic, but not ad hoc, market background, we use draws from a log-normal returns
distribution to simulate changes in a stock’s fundamental value, or P*. The model
uses price-dependent Poisson distributions to generate a realistic flow of computer-
generated buy and sell orders whose intensity and supply-demand balance vary over
time. The order flow fluctuations depend on the difference between the current
market price and the P* value. In Sect. 4, we illustrate the second method, which is
experimental control to create groupings of participants in our simulations that have
the same trading “assignment.” The result is the ability to make valid comparisons
of trader’s performance in the simulations.
Keywords: Trading simulations, Market microstructure, Order flow models,
Random walk models, Experimental economics, Experimental control
This paper is the first to examine the motivations for issuing putable bonds in which
the embedded put option is not contingent upon a company-related event. We find
that the market favorably views the issue announcement of these bonds that we
refer to as bonds with European put options or European putable bonds. This
response is in contrast to the response documented by the literature to other bond
issues (straight, convertible, and most studies examining poison puts), and to the
response documented in the current paper to the issue announcements of poison-put
bonds. Our results suggest that the market views issuing European putable bonds as
helping mitigate security mispricing. Our study is an application of important
statistical methods in corporate finance, namely, Event Studies and the use of
General Method of Moments for cross-sectional regressions.
20 C.-F. Lee and J.C. Lee
Interest rate sensitivity of bank stock returns has been studied using an augmented
CAPM: a multiple regression model with market returns and interest rate as
independent variables. In this chapter, we test an asset-pricing model in which
the CAPM is augmented by three orthogonal factors which are proxies for the
innovations in inflation, maturity risk, and default risk. The methodologies used in
this chapter are multiple regression and factor analysis.
Keywords: CAPM, APT, Bank stock return, Interest rate risk, Orthogonal
factors, Multiple regression
This chapter derives a new nonparametric lower bound and provides an alternative
interpretation of Ritchken’s (1985) upper bound to the price of the European option.
In a series of numerical examples, our new lower bound is substantially tighter than
previous lower bounds. This is prevalent especially for out-of-the-money (OTM)
options where the previous lower bounds perform badly. Moreover, we present that
our bounds can be derived from histograms which are completely nonparametric in
an empirical study. We first construct histograms from realizations of S&P
500 index returns following Chen et al. (2006), calculate the dollar beta of the
option and expected payoffs of the index and the option, and eventually obtain our
bounds. We discover violations in our lower bound and show that those violations
present arbitrage profits. In particular, our empirical results show that out-of-the-
money calls are substantially overpriced (violate the lower bound). The methodol-
ogies used in this chapter are nonparametric, option pricing model, and histograms
methods.
Keywords: Option bounds, Nonparametric, Black-Scholes model, European
option, S&P 500 index, Arbitrage, Distribution of underlying asset, Lower bound,
Out-of-the-money, Kernel pricing
This chapter examines the accuracy of the earnings forecasts by the following test
methodologies. Multiple Regression Models are used to examine the effect of six
factors: firm size, market volatility, trading volume turnover, corporate earnings
variances, type of industry, and experience. If the two-sample groups are related,
Wilcoxon Two-Sample Test will be used to determine the relative earnings forecast
accuracy. Readers are well advised and referred to the chapter appendix for
methodological issues such as sample selection, variable definition, regression
model, and Wilcoxon tow-sample test.
Keywords: Multiple regression, Wilcoxon two-sample test, Corporate earnings,
Forecast accuracy, Management earnings, Firm size, Corporation regulation,
Volatility, Trade turnover, Industry
This study uses standard models such as earnings level and earnings changes, among
others. Models that fit better to the data drawn from companies listed on the Athens
Stock Exchange have been selected employing autoregressive integrated moving
average with exogenous variables (ARIMAX) models. Models I (price on earnings
model) “II (returns on change in earnings divided by beginning-of-period price and
prior period)” V (returns on change in earnings over opening market value), VII
(returns deflated by lag of 2 years on earnings over opening market value), and IX
(differenced-price model) have statistically significant coefficients of explanatory
variables. These models take place with backward looking information instead of
forward looking information that recent literature is assessed. The methodologies
used in this chapter are price on earnings model, return models, autoregressive
moving average with exogenous variables (ARIMAX), minimum value of squared
residuals (MSE loss function), and Dickey-Fuller test.
22 C.-F. Lee and J.C. Lee
This study uses a mixture of the dynamic factor model of Duffee (1999) and a
contagious effect in the specification of a Hawkes process, a class of counting
processes which allows intensities to depend on the timing of previous events
(Hawkes 1971). Using the mixture factor- contagious-effect model, Monte Carlo
simulation is performed to generate default times of two hypothesized firms. The
goodness-of-fit of the joint distributions based on the most often used copula
functions in literature, including the Normal, t-, Clayton, Frank, and Gumbel
copula, respectively, is assessed against the simulated default times. It is demon-
strated that as the contagious effect increases, the goodness-of-fit of the joint
distribution functions based on copula functions decreases, which highlights the
deficiency of the copula function approach.
Keywords: Static factor model, Dynamic factor model, Correlated defaults,
Contagious effect, Hawkes process, Monte Carlo simulation, Normal copula,
t-copula, Clayton copula, Frank copula, Gumbel copula
This chapter uses simultaneous equation modeling and uses Hausman test to
determine whether to report fixed or random-effects estimates. We first report
random-effects estimates based on the estimation procedure of Baltagi (Baltagi
1981; Baltagi and Li 1995; Baltagi and Li 1994). We consider that the error
component two-stage least squares (EC2SLS) estimator of Baltagi and Li (1995) is
more efficient than the generalized two-stage least squares (G2SLS) estimator
of Balestra and Varadharajan-Krishnakumar (1987). For our second estimation
procedure, for comparative purposes, we use the dynamic panel modeling estimates
recommended by Blundell and Bond (1998). We employ the model of Blundell and
Bond (1998), as these authors argue that their estimator is more appropriate than the
Arellano and Bond (1991) model for smaller time periods relative to the size
of the panels. We also use this two-step procedure, use as an independent
variable the first lag of the dependent variable, reporting robust standard errors of
Windmeijer (2005). Thus, our two different panel estimation techniques place
differing emphasis on cross-sectional and time-series effects, with the Baltagi-Li
1 Introduction to Financial Econometrics and Statistics 23
This chapter employs stochastic frontier analysis to create a new type of instru-
mental variable. The unrestricted frontier model determines the highest possible
profitability based solely on the book value of assets employed. We develop
a second frontier based on the level of bank holding company capital as well as
the amount of assets. The implication of using the unrestricted model is that we are
measuring the unconditional inefficiency of the banking organization. This chapter
applies generalized method of moments (GMM) regression to avoid the problem
caused by departure from normality. To control for the impact of size on a bank’s
risk-taking behavior, the book value of assets is considered in the model.
The relationship between the variables specifying bank behavior and the use of
equity is analyzed by GMM regression.
Keywords: Bank capital, Generalized method of moments, Stochastic frontier
analysis, Bank risks, Bank holding companies, Endogeneity of variables
This chapter examines the performance of more than 20 different testing procedures
that fall into two categories. First, the buy-and-hold benchmark approach uses
a benchmark to measure the abnormal buy-and-hold return for every event firm,
and tests the null hypothesis that the average abnormal return is zero. Second, the
calendar-time portfolio approach forms a portfolio in each calendar month
consisting of firms that have had an event within a certain time period prior to the
month, and tests the null hypothesis that the intercept is zero in the regression of
monthly portfolio returns against the factors in an asset-pricing model. This chapter
also evaluates the performance of bootstrapped Johnson’s skewness-adjusted t-test.
This computation-intensive procedure is considered because the distribution of
long-horizon abnormal returns tends to be highly skewed to the right. The
bootstrapping method uses repeated random sampling to measure the significance
of relevant test statistics. Due to the nature of random sampling, the resultant
measurement of significance varies each time such a procedure is used. We also
evaluate simple nonparametric tests, such as the Wilcoxon signed-rank test or the
Fisher’s sign test, which are free from random sampling variation.
24 C.-F. Lee and J.C. Lee
This chapter employs the ANST-GARCH model that is capable of capturing the
asymmetric volatility effect of a positive and negative return shock. The key feature
of the model is the regime-shift mechanism that allows a smooth, flexible
transition of the conditional volatility between different states of volatility
persistence. The regime-switching mechanism is governed by a logistic transition
function that changes values depending on the level of the previous return shock.
With a negative (positive) return shock, the conditional variance process is
described as a high (low)-persistence-in-volatility regime. The ANST-GARCH
model describes the heteroskedastic return dynamics more accurately and generates
better volatility forecasts.
Keywords: Intertemporal risk-return relation, Unexpected volatility shocks,
Time-varying rational expectation hypothesis, Stock market overreaction,
Expected market risk premium, Volatility feedback effect, Asymmetric mean
reversion, Asymmetric volatility response, Time-varying volatility, Volatility
regime switching, ANST-GARCH model
This chapter uses a novel method, the Logical Analysis of Data (LAD), to reverse-
engineer and construct credit risk ratings which represent the creditworthiness of
financial institutions and countries. LAD is a data-mining method based on com-
binatorics, optimization, and Boolean logic that utilizes combinatorial search tech-
niques to discover various combinations of attribute values that are characteristic of
the positive or negative character of observations. The proposed methodology is
applicable in the general case of inferring an objective rating system from archival
data, given that the rated objects are characterized by vectors of attributes taking
numerical or ordinal values. The proposed approaches are shown to generate
transparent, consistent, self-contained, and predictive credit risk rating models,
closely approximating the risk ratings provided by some of the major rating
agencies. The scope of applicability of the proposed method extends beyond the
rating problems discussed in this study, and can be used in many other contexts
where ratings are relevant. This study also uses multiple linear regression to derive
the logical rating scores.
1 Introduction to Financial Econometrics and Statistics 25
This chapter uses denoising analysis which imposes new challenges for financial data
mining due to the irregularities and roughness observed in financial data, particularly,
for instantaneously collected massive amounts of tick-by-tick data from financial
markets for information analysis and knowledge extraction. Inefficient decomposition
of the systematic pattern (the trend) and noises of financial data will lead to erroneous
conclusions since irregularities and roughness of the financial data make the applica-
tion of traditional methods difficult. The methodologies used in this chapter are linear
filters, nonlinear filters, time-series analysis, trend extraction, and wavelet.
Keywords: Jump detection, Linear filters, Nonlinear filters, Time-series analy-
sis, Trend extraction, Wavelet
This chapter presents a set of tools, which allow gathering information about
the frequency components of a time-series. In the first step, we discuss spectral
26 C.-F. Lee and J.C. Lee
analysis and filtering methods. Spectral analysis can be used to identify and to
quantify the different frequency components of a data series. Filters permit to
capture specific components (e.g. trends, cycles, seasonalities) of the original
time-series. In the second step, we introduce wavelets, which are relatively new
tools in economics and finance. They take their roots from filtering methods and
Fourier analysis, but overcome most of the limitations of these two methods.
Their principal advantages derive from: (i) combined information from both
time-domain and frequency-domain and (ii) their flexibility as they do not
make strong assumptions concerning the data generating process for the series
under investigation.
Keywords: Filtering methods, Spectral analysis, Fourier transform, Wavelet
filter, Continuous wavelet transform, Discrete wavelet transform, Multi-resolution
analysis, Scale-by-scale decomposition, Analysis of variance, Case-Shiller home
price indices
This chapter derives the exact formulae for several goodness-of-fit statistics that
should be applied to loss models with left-truncated data where the fit of a distribution
in the right tail of the distribution is of central importance. We apply the proposed
tests to real financial losses, using a variety of distributions fitted to operational loss
and the natural catastrophe insurance claims data. The methodologies discussed in
this chapter are goodness-of-fit tests, loss distribution, ruin probability, value-at-risk,
Anderson-Darling statistic, Kolmogorov-Smirnov statistic.
Keywords: Goodness-of-fit tests, Left-truncated data, Minimum recording thresh-
old, Loss distribution, Heavy-tailed data, Operational risk, Insurance, Ruin probabil-
ity, Value-at-risk, Anderson-Darling statistic, Kolmogorov-Smirnov statistic
This chapter identifies and defines variables for merger synergy analysis followed
by principal component factor analysis, variability percentage adjustment, and
performance score calculation. Finally, Wilcoxon sign rank test is used for hypoth-
esis testing. We extract principle component factors from a set of financial ratios.
Percentage of variability explained and factor loadings are adjusted to get
a modified average weight for each financial ratio. This weight is multiplied by
the standardized Z value of the variable, and summed a set of variables get a firm’s
performance score. Performance scores are used to rank the firm. Statistical signif-
icance of difference in pre- and post-merger rank is tested using the Wilcoxon
sign rank.
1 Introduction to Financial Econometrics and Statistics 27
This chapter uses a factor copula approach to evaluate basket credit derivatives with
issuer default risk and demonstrate its application in a basket credit linked note
(BCLN). We generate the correlated Gaussian random numbers by using the
Cholesky decomposition, and then, the correlated default times can be decided by
these random numbers and the reduced-form model. Finally, the fair BCLN coupon
rate is obtained by the Monte Carlo simulation. We also discuss the effect of
issuer default risk on BCLN. We show that the effect of issuer default risk cannot
be accounted for thoroughly by considering the issuer as a new reference entity in
the widely used one factor copula model, in which constant default correlation is often
assumed. A different default correlation between the issuer and the reference entities
affects the coupon rate greatly and must be taken into account in the pricing model.
Keywords: Factor copula, Issuer default, Default correlation, Reduced-form
model, Basket credit derivatives, Cholesky decomposition, Monte Carlo simulation
This chapter finds that for closed-end country funds, the international CAPM can be
rejected for the underlying securities (NAVs) but not for the share prices. This
finding indicates that country fund share prices are determined globally, whereas
the NAVs reflect both global and local prices of risk. Cross-sectional variations in the
discounts or premiums for country funds are explained by the differences in the risk
exposures of the share prices and the NAVs. Finally, this chapter shows that the share
price and NAV returns exhibit predictable variation, and country fund premiums
vary over time due to time-varying risk premiums. The chapter employs Generalized
Method of Moments (GMM) to estimate stochastic discount factors and examines if
the price of risk of closed-end country fund shares and NAVs is identical.
Keywords: Capital markets, Country funds, CAPM, Closed-end funds, Market
segmentation, GMM, Net asset value, Stochastic discount factors, Time-varying
risk, International asset pricing
This study uses three different risk measurements: the Mean-variance model, the
Mean Absolute Deviation model, and the Downside Risk model. Meanwhile short
selling is also taken into account since it is an important strategy that can bring
a portfolio much closer to the efficient frontier by improving a portfolio’s risk-
return trade-off. Therefore, six portfolio rebalancing models, including the MV
model, MAD model and the Downside Risk model, with/without short selling, are
compared to determine which is the most efficient. All models simultaneously
consider the criteria of return and risk measurement. Meanwhile, when short
selling is allowed, models also consider minimizing the proportion of short selling.
Therefore, multiple objective programming is employed to transform multiple
objectives into a single objective in order to obtain a compromising solution. An
example is used to perform simulation, and the results indicate that the MAD
model, incorporated with a short selling model, has the highest market value and
lowest risk.
1 Introduction to Financial Econometrics and Statistics 29
This chapter first utilizes the ordered logit and the ordered probit models. Then, we use
ordered logit combining method to weight different techniques’ probability measures,
as described in Kamstra and Kennedy (1998) to form the combining model.
The samples consist of firms in the TSE and the OTC market, and are divided into
three industries for analysis. We consider financial variables, market variables as well as
macroeconomic variables and estimate their parameters for out-of-sample tests. By
means of Cumulative Accuracy Profile, the Receiver Operating Characteristics, and
McFadden, we measure the goodness-of-fit and the accuracy of each prediction model.
The performance evaluations are conducted to compare the forecasting results, and we
find that combing technique does improve the predictive power.
Keywords: Bankruptcy prediction, Combining forecast, Credit rating, Credit
risk, Credit risk index, Forecasting models, Logit regression, Ordered logit,
Ordered probit, Probability density function
The aim of this chapter is to check whether certain playing rules, based on the
undervaluation concept arising from the CAPM, could be useful as investment
strategies, and can therefore be used to beat the Market. If such strategies work, we
will be provided with a useful tool for investors, and, otherwise, we will obtain
a test whose results will be connected with the efficient Market hypothesis (EMH)
and with the CAPM. The methodology used is both intuitive and rigorous: analyz-
ing how many times we beat the Market with different strategies, in order to check
whether when we beat the Market, this happens by chance.
Keywords: ANOVA, Approximately normal distribution, Binomial distribu-
tion, CAPM, Contingency tables, Market efficiency, Nonparametric tests, Perfor-
mance measures
This chapter adopts an Analytic Hierarchy Process (AHP) approach to solve various
accounting or finance problems such as developing a business performance evalu-
ation system and developing a banking performance evaluation system. AHP uses
30 C.-F. Lee and J.C. Lee
This chapter focuses on various structural change and monitoring tests for a class of
widely used time-series models in economics and finance, including I(0), I(1), I(d)
processes and the co-integration relationship. In general, structural change tests can be
categorized into two types: One is the classical approach to testing for structural
change, which employs retrospective tests using a historical data set of a given length;
the other one is the fluctuation-type test in a monitoring scheme, which means for
given a history period for which a regression relationship is known to be stable,
1 Introduction to Financial Econometrics and Statistics 31
we then test whether incoming data are consistent with the previously established
relationship. Several structural changes such as CUSUM squared tests, the QLR test,
the prediction test, the multiple break test, bubble tests, co-integration breakdown
tests, and the monitoring fluctuation test are discussed in this chapter, and we further
illustrate all details and usefulness of these tests.
Keywords: Co-integration breakdown test, Structural break, Long memory
process, Monitoring fluctuation test, Boundary function, CUSUM squared test,
Prediction test, Bubble test, Unit root time series, Persistent change
This chapter develops econometric models for seasonal patterns in both price returns
and proportional changes in demand for Australian electricity. Australian Electricity
spot prices differ considerably from equity spot prices in that they contain an extremely
rapid mean reversion process. The electricity spot price could increase to a market cap
price of AU$12,500 per Megawatt Hour (MWh) and revert back to a mean level
(AUD$30) within a half hour interval. This has implications for derivative pricing and
risk management. We also model extreme spikes in the data. Our study identifies both
seasonality effects and dramatic price reversals in the Australian electricity market.
The pricing seasonality effects include time-of-day, day-of-week, monthly, and yearly
effects. There is also evidence of seasonality in demand for electricity.
Keywords: Electricity, Spot price, Seasonality, Outlier, Demand, Econometric
modeling
32 C.-F. Lee and J.C. Lee
This chapter uses both parametric and nonparametric approaches to evaluate private-
and public-equity timberland investments in the United States. Private-equity timber-
land returns are proxied by the NCREIF Timberland Index, whereas public-equity
timberland returns are proxied by the value-weighted returns on a dynamic portfolio
of the US publicly traded forestry firms that had or have been managing timberlands.
Static estimations of the capital asset-pricing model and Fama-French three-factor
model are obtained by ordinary least squares, whereas dynamic estimations are
obtained by state-space specifications with the Kalman filter. In estimating the
stochastic discount factors, linear programming is used.
Keywords: Alternative asset class, Asset pricing, Evaluation, Fama-French
three-factor model, Nonparametric analysis, State-space model, Stochastic discount
factor, Timberland investments, Time series, Time-varying parameter
This chapter derives and characterizes optimal orthogonal portfolios in the presence
of conditioning information in the form of a set of lagged instruments. In this
setting, studied by Hansen and Richard (1987), the conditioning information is used
to optimize with respect to the unconditional moments. We present an empirical
illustration of the properties of the optimal orthogonal portfolios. The methodology in
this chapter includes regression and maximum likelihood parameter estimation, as
well as method of moments estimation. We form maximum likelihood estimates of
nonlinear functions as the functions evaluated at the maximum likelihood parameter
estimates.
Keywords: Asset-pricing tests, Conditioning information, Minimum variance
efficiency, Optimal portfolios, Predicting returns, Portfolio management, Stochas-
tic discount factors, Generalized, Method of moments, Maximum likelihood, Para-
metric bootstrap, Sharpe ratios
This chapter uses a multifactor, multi-indicator approach to test the capital asset-
pricing model (CAPM) and the arbitrage pricing theory (APT). This approach is
able to solve the measuring problem in the market portfolio in testing CAPM, and it
is also able to directly test APT by linking the common factors to the macroeco-
nomic indicators. We propose a MIMIC approach to test CAPM and APT. The beta
estimated from the MIMIC model by allowing measurement error on the market
1 Introduction to Financial Econometrics and Statistics 33
portfolio does not significantly improve the OLS beta, while the MLE estimator
does a better job than the OLS and GLS estimators in the cross-sectional regressions
because the MLE estimator takes care of the measurement error in beta. Therefore,
the measurement error problem on beta is more serious than that on the market
portfolio.
Keywords: Capital asset-pricing model, CAPM, Arbitrage pricing theory, Mul-
tifactor multi-indicator approach, MIMIC, Measurement error, LISREL approach,
Ordinary least square, OLS, General least square, GLS, Maximum likelihood
estimation, MLE
This chapter will first demonstrate how Microsoft Excel can be use to create the
Decision Trees for the Binomial Option Pricing Model. At the same time, this
chapter will discuss the Binomial Option Pricing Model in a less mathematical
fashion. All the mathematical calculations will be taken care by the Microsoft Excel
program that is presented in this chapter. Finally, this chapter also uses the Decision
Tree approach to demonstrate the relationship between the Binomial Option Pricing
Model and the Black-Scholes Option Pricing Model.
Keywords: Binomial option pricing model, Decision trees, Black-Sholes option
pricing model, Call option, Put option, Microsoft Excel, Visual Basic for applica-
tions, VBA, Put-call parity, Sigma, Volatility, Recursive programming
This chapter uses the econometric methodology to deal with the dynamic inter-
relationships between dividend payments and share repurchases and investigate
endogeneity of certain explanatory variables. Identification of the model parameters
is achieved in such models by exploiting the cross-equations restrictions on the
coefficients in different time periods. Moreover, the estimation entails using
nonlinear optimization methods to compute the maximum likelihood estimates of
the dynamic random-effects models and for testing statistical hypotheses using
likelihood ratio tests. This study also highlights the importance of developing
comprehensive econometric models for these interrelationships. It is common in
finance research to spell out “specific hypotheses” and conduct empirical research
to investigate validity of the hypotheses.
Keywords: Compustat database, Corporate policies, Dividends, Dynamic
random-effects models, Econometric methodology, Endogeneity, Maximum like-
lihood, Intangible assets, Model formulation, Nonlinear optimization, Panel data,
Share repurchases
34 C.-F. Lee and J.C. Lee
In this chapter, we illustrate some recent developments in the yield curve modeling
by introducing a latent factor model called the dynamic Nelson-Siegel model. This
model not only provides good in-sample fit, but also produces superior out-of-sample
performance. Beyond Treasury yield curve, the model can also be useful for other
assets such as corporate bond and volatility. Moreover, the model also suggests
generalized duration components corresponding to the level, slope, and curvature
risk factors. The dynamic Nelson-Siegel model can be estimated via a one-step
procedure, like the Kalman filter, which can also easily accommodate other variables
of interests. Alternatively, we could estimate the model through a two-step process by
fixing one parameter and estimating with ordinary least squares. The model is flexible
and capable of replicating a variety of yield curve shapes: upward sloping, downward
sloping, humped, and inverted humped. Forecasting the yield curve is achieved
through forecasting the factors and we can impose either a univariate autoregressive
structure or a vector autoregressive structure on the factors.
Keywords: Term structure, Yield curve, Factor model, Nelson-Siegel curve,
State-space model
The literature has so far focused on the risk-return trade-off in equity markets and
ignored alternative risky assets. This chapter examines the presence and signifi-
cance of an intertemporal relation between expected return and risk in the foreign
exchange market. This chapter tests the existence and significance of a daily risk-
return trade-off in the FX market based on the GARCH, realized, and range
volatility estimators. Our empirical analysis relies on the maximum likelihood
estimation of the GARCH-in-mean models, as described in Appendix A. We also
use the seemingly unrelated (SUR) regressions and panel data estimation to inves-
tigate the significance of a time-series relation between expected return and risk on
currency.
Keywords: GARCH, GARCH-in-mean, Seemingly unrelated regressions
(SUR), Panel data estimation, Foreign exchange market, ICAPM, High-frequency
data, Time-varying risk aversion, High-frequency data, Daily realized volatility
ARCH models, GARCH models, and the CaViaR models is systematically intro-
duced. Comparing the proposed QR method with traditional methods based on
distributional assumptions, the QR method has the important property that it is
robust to non-Gaussian distributions. Quantile estimation is only influenced by the
local behavior of the conditional distribution of the response near the specified
quantile. As a result, the estimates are not sensitive to outlier observations. Such
a property is especially attractive in financial applications since many financial data
like, say, portfolio returns (or log returns), are usually not normally distributed. To
highlight the importance of the QR method in estimating VaR, we apply the QR
techniques to estimate VaRs in International Equity Markets. Numerical evidence
indicates that QR is a robust estimation method for VaR.
Keywords: ARCH, Copula, GARCH, Non-normality, QAR, Quantile regres-
sion, Risk management, Robust estimation, Time series, Value-at-risk
Using a sample of listed firms in Southeast Asia countries, this chapter examines the
association among board structure and corporate ownership structure in affecting
earnings quality. The econometric method employed is regressions of panel data.
In a panel data setting, I address both cross-sectional and time-series dependence.
Following Gow et al. (2010), I employ the two-way clustering method where the
standard errors are clustered by both firm and year in my regressions of panel data.
Keywords: Earnings quality, Board structure, Corporate ownership structure,
Panel data regressions, Cross-sectional and time-series dependence, Two-way
clustering method of standard errors
This chapter examines the rationality and diversity of industry-level forecasts of the
yen-dollar exchange rate collected by the Japan Center for International Finance.
We compare three specifications for testing rationality: the “conventional” bivariate
regression, the univariate regression of a forecast error on a constant and other
information set variables, and an error correction model (ECM). We extend the
analysis of industry-level forecasts to a SUR-type structure using an innovative
GMM technique (Bonham and Cohen 2001) that allows for forecaster cross-
correlation due to the existence of common shocks and/or herd effects. Our
GMM tests of micro-homogeneity uniformly reject the hypothesis that forecasters
exhibit similar rationality characteristics.
Keywords: Rational expectations, Unbiasedness, Weak efficiency, Micro-
homogeneity, Heterogeneity, Exchange rate, Survey forecasts, Aggregation bias,
GMM, SUR
36 C.-F. Lee and J.C. Lee
This chapter examines the riskiness of the Taiwan stock market by determining the
VaR from the expected return distribution generated by historical simulation.
Value-at-risk (VaR) measures the worst expected loss over a given time horizon
under normal market conditions at a specific level of confidence. VaR is determined
by the left tail of the cumulative probability distribution of expected returns. Our
result indicates the cumulative probability distribution has a fatter left tail, com-
pared with the left tail of a normal distribution. This implies a riskier market. We
also examined a two-sector asset allocation model subject to a target VaR con-
straint. The VaR-efficient frontier of the TAIEX traded stocks recommended,
mostly, a corner portfolio.
Keywords: Value-at-risk, Asset allocation, Cumulative probability distribution,
Normal distribution, VaR-efficient frontier, Historical simulation, Expected return dis-
tribution, Two-sector asset allocation model, Delta, gamma, Corner portfolio, TAIEX
The most common valuation model is the dividend growth model. The growth rate is
found by taking the product of the retention rate and the return on equity. What is less well
understood are the basic assumptions of this model. In this paper, we demonstrate that the
model makes strong assumptions regarding the financing mix of the firm. In addition,
1 Introduction to Financial Econometrics and Statistics 37
we discuss several methods suggested in the literature on estimating growth rates and
analyze whether these approaches are consistent with the use of using a constant discount
rate to evaluate the firm’s assets and equity. The literature has also suggested estimating
growth rate by using the average percentage change method, compound-sum method,
and/or regression methods. We demonstrate that the average percentage change is very
sensitive to extreme observations. Moreover, on average, the regression method yields
similar but somewhat smaller estimates of the growth rate compared to the compound-
sum method. We also discussed the inferred method suggested by Gordon and Gordon
(1997) to estimate the growth rate. Advantages, disadvantages, and the interrelationship
among these estimation methods are also discussed in detail.
Keywords: Compound sum method, Discount cash flow model, Growth rate,
Internal growth rate, Sustainable growth rate
A security is liquid to the extent that an investor can trade significant quantities of the
security quickly, at or near the current market price, and bearing low transaction costs.
As such, liquidity is a multidimensional concept. In this chapter, I review several
widely used econometrics or statistics-based measures that researchers have devel-
oped to capture one or more dimensions of a security’s liquidity (i.e., limited depen-
dent variable model (Lesmond et al. 1999) and autocovariance of price changes (Roll
1984)). These alternative proxies have been designed to be estimated using either
low-frequency or high-frequency data, so I discuss four liquidity proxies that are
estimated using low-frequency data and two proxies that require high-frequency data.
Low-frequency measures permit the study of liquidity over relatively long time
horizons; however, they do not reflect actual trading processes. To overcome this
limitation, high-frequency liquidity proxies are often used as benchmarks to determine
the best low-frequency proxy. In this chapter, I find that estimates from the effective
tick measure perform best among the four low-frequency measures tested.
Keywords: Liquidity, Transaction costs, Bid-ask spread, Price impact, Percent
effective spread, Market model, Limited dependent variable model, Tobin’s model,
Log-likelihood function, Autocovariance, Correlation analysis
models based on the normal innovations appear not to be as well suited as infinitely
divisible models for predicting financial crashes. In our empirical analysis, we
forecast 1 % value-at-risk in both spot and futures markets using normal and
tempered stable GARCH models following a quasi-maximum likelihood estimation
strategy. In order to determine the accuracy of forecasting for each specific model,
backtesting using Kupiec’s proportion of failures test is applied.
Keywords: Infinitely divisible models, Tempered stable distribution, GARCH
models, Value-at-risk, Kupiec’s proportion of failures test, Quasi-maximum likeli-
hood estimation strategy
This chapter examines the core computing competence for financial services.
Securities trading is one of the few business activities where a few seconds of
processing delay can cost a company big fortune. Grid and Cloud computing
will be briefly described. How the underlying algorithm for financial analysis
can take advantage of Grid environment is chosen and presented. One of the
most popular practiced algorithms Monte Carlo Simulation is used in our cases
study for option pricing and risk management. The various distributed com-
putational platforms are carefully chosen to demonstrate the performance issue
for financial services.
Keywords: Financial service, Grid and cloud computing, Monte Carlo simula-
tion, Option pricing, Risk management, Cyberinfrastructure, Random number
generation, High end computing, Financial simulation, Information technology
This chapter introduces the long-run stock return methodologies and their statistical
inference. The long-run stock return is usually computed by using a holding strategy
more than 1 year but up to 5 years. Two categories of long-run return methods are
illustrated in this chapter: the event-time approach and calendar-time approach. The
event-time approach includes cumulative abnormal return, buy-and-hold abnormal
return, and abnormal returns around earnings announcements. In former two methods,
it is recommended to apply the empirical distribution (from the bootstrapping method)
to examine the statistical inference, whereas the last one uses classical t-test.
In addition, the benchmark selections in the long-run return literature are introduced.
Moreover, the calendar-time approach contains mean monthly abnormal return,
factor models, and Ibbotson’s RATS, which could be tested by time-series volatility.
Keywords: Long-run stock return, Buy-and-hold return, Factor model, Event-
time, Calendar-time, Cumulative abnormal return, Ibottson’s RATS, Conditional
market model, Bootstrap, Zero-investment portfolio
1 Introduction to Financial Econometrics and Statistics 39
This study utilizes the parametric approach (GARCH-based models) and the semi-
parametric approach of Hull and White (1998) (HW-based models) to estimate the
Value-at-Risk (VaR) through the accuracy evaluation of accuracy for the eight stock
indices in Europe and Asia stock markets. The measure of accuracy includes the
unconditional coverage test by Kupiec (1995) as well as two loss functions, quadratic
loss function and unexpected loss. As to the parametric approach, the parameters of
generalized autoregressive conditional heteroskedasticity (GARCH) model are esti-
mated by the method of maximum likelihood and the quantiles of asymmetric distri-
bution like skewed generalized student’s t (SGT) can be solved by composite trapezoid
rule. Sequentially, the VaR is evaluated by the framework proposed by Jorion (2000).
Turning to the semi-parametric approach of Hull and White (1998), before performing
the traditional historical simulation, the raw return series is scaled by a volatility ratio
where the volatility is estimated by the same procedure of parametric approach.
Keywords: Value-at-risk, Semi-parametric approach, Parametric approach,
Generalized autoregressive conditional heteroskedasticity, Skewed generalized
student’s t, Composite trapezoid rule, Method of maximum likelihood, Uncondi-
tional coverage test, Loss function
This chapter illustrates the superiority of the mean-variance ratio (MVR) test
over the traditional SR test by applying both tests to analyze the performance
of the S&P 500 index and the NASDAQ 100 index after the bursting of the
Internet bubble in 2000s. This shows the superiority of the MVR test statistic
in revealing short-term performance and, in turn, enables investors to make
better decisions in their investments. The methodologies used in this chapter
are mean-variance ratio, Sharpe ratio, hypothesis testing, and uniformly most
powerful unbiased test.
Keywords: Mean-variance ratio, Sharpe ratio, Hypothesis testing, Uniformly
most powerful unbiased test, Internet bubble, Fund management
This chapter uses the CoVaR (Conditional VaR) framework to obtain accurate
information on the interdependency of risk factors. The basic technical ele-
ments of CoVaR estimation are two levels of quantile regression: one on
market risk factors; another on individual risk factor. Tests on the functional
form of the two-level quantile regression reject the linearity. A flexible semi-
parametric modeling framework for CoVaR is proposed. A partial linear model
(PLM) is analyzed. In applying the technology to stock data covering the crisis
period, the PLM outperforms in the crisis time, with the justification of the
backtesting procedures. Moreover, using the data on global stock markets
indices, the analysis on marginal contribution of risk (MCR) defined as the
local first order derivative of the quantile curve sheds some light on the source
of the global market risk.
Keywords: CoVAR, Value-at-risk, Quantile regression, Locally linear quantile
regression, Partial linear model, Semi-parametric model
This chapter uses Monte Carlo simulation to investigate the impacts of managerial
overconfidence on the optimal strike prices of executive incentive options.
Although it has been shown that optimally managerial incentive options should
be awarded in-the-money, in practice, most firms award them at-the-money. We
show that the optimal strike prices of options granted to overconfident executive are
directly related to their overconfidence level, and that this bias brings the optimal
strike prices closer to the institutionally prevalent at-the-money prices. The Monte
Carlo simulation procedure uses a Mathematica program to find the optimal effort
by managers and the optimal (for stockholders) contract parameters. An expanded
discussion of the simulations, including the choice of the functional forms and the
calibration of the parameters, is provided.
1 Introduction to Financial Econometrics and Statistics 41
This chapter uses densities and conditional distributions analysis to carry out
consistent specification testing and model selection among multiple diffusion
processes. In this chapter, we discuss advances to this literature introduced by
Corradi and Swanson (2005), who compare the cumulative distribution (marginal
or joint) implied by a hypothesized null model with corresponding empirical
distributions of observed data. In particular, parametric specification tests in the
spirit of the conditional Kolmogorov test of Andrews (1997) that rely on block
bootstrap resampling methods in order to construct test critical values are discussed.
The methodologies used in this chapter are continuous time simulation methods,
single process specification testing, multiple process model selection, and multi-
factor diffusion process, block bootstrap, and jump process.
Keywords: Multifactor diffusion process, Specification test, Out-of-sample
forecasts, Conditional distribution, Model selection, Block bootstrap, Jump process
This chapter describes different procedures to deal with measurement error in linear
models, and assess their performance in finite samples using Monte Carlo simulations,
and data on corporate investment. We consider the standard instrumental variables
approach proposed by Griliches and Hausman (1986) as extended by Biorn
(2000) [OLS-IV], the Arellano and Bond (1991) instrumental variable estimator, and
the higher-order moment estimator proposed by Erickson and Whited (2000, 2002).
Our analysis focuses on characterizing the conditions under which each of these
estimators produces unbiased and efficient estimates in a standard “errors in variables”
setting. In the presence of fixed effects, under heteroscedasticity, or in the absence of
a very high degree of skewness in the data, the EW estimator is inefficient and
returns biased estimates for mismeasured and perfectly measured regressors. In con-
trast to the EW estimator, IV-type estimators (OLS-IV and AB-GMM) easily handle
individual effects, heteroskedastic errors, and different degrees of data skewness.
The IV approach, however, requires assumptions about the autocorrelation structure
of the mismeasured regressor and the measurement error. We illustrate the application
of the different estimators using empirical investment models.
Keywords: Investment equations, Measurement error, Monte Carlo simulations,
Instrumental variables, GMM, Bias, Fixed effects, Heteroscedasticity, Skewness,
High-order moments
42 C.-F. Lee and J.C. Lee
Chapter 59: Pre-IT Policy, Post-IT Policy, and the Real Sphere
in Turkey
We estimate Two SVECM (Structural Vector Error Correction) Models for the
Turkish economy based on imposing short run and Long-run restrictions that
accounts for examining the behavior of the real sphere in the Pre-IT policy
(before Inflation-Targeting adoption) and Post-IT policy (after Inflation-
Targeting Adoption). Responses reveals that an expansionary interest policy
shock leads to a decrease in price level, a fall in output, an appreciation in the
exchange rate, an improvement in the share prices in the very short run for the
most of Pre-IT period.
Keywords: SVECM models, Turkish economy, Short run, Long run, Restric-
tions, Inflation targeting, Pre-IT policy, Post-IT policy, Share prices, Exchange rate,
Monetary policy shock, Output, Price level, Real sphere
This chapter uses Jones Model (1991) which projects the expected level of discre-
tionary accruals and demonstrates specific test methodology for detection of earnings
management in the oil and gas industry. This study utilized several parametric and
nonparametric statistical methods to test for such earnings management. By compar-
ing actuals versus projected accruals, we are able to compute the total unexpected
accruals. We also correlate unexpected total accruals with several difficult to manip-
ulate indicators that reflect company’s level of activities.
Keywords: Earning management, Jones (1991) model, Discretionary accruals,
Income from operations, Nonrecurring items, Special items, Research and devel-
opment expense, Write-downs, Political cost, Impression management, Oil and gas
industry
This chapter examines two asymmetric stochastic volatility models used to describe
the volatility dependencies found in most financial returns. The first is the
autoregressive stochastic volatility model with Student’s t-distribution (ARSV-t),
and the second is the basic Svol of JPR (1994). In order to estimate these models,
our analysis is based on the Markov Chain Monte Carlo (MCMC) method. Therefore,
44 C.-F. Lee and J.C. Lee
the technique used is a Metropolishastings (Hastings 1970), and the Gibbs sampler
(Casella and George 1992; Gelfand and Smith 1990; Gilks et al. 1993). The empirical
results concerned on the Standard and Poor’s 500 composite Index (S&P), CAC40,
Nasdaq, Nikkei, and Dow-Jones stock price indexes reveal that the ARSV-t model
provides a better performance than the Svol model on the Mean Squared Error (MSE)
and the Maximum Likelihood function.
Keywords: Autoregression, Asymmetric stochastic volatility, MCMC, Metro-
polishastings, Gibbs sampler, Volatility dependencies, Student’s t-distribution,
SVOL, MSE, Financial returns, Stock price indexes
This chapter uses the Ordinary Least-Squares (OLS) methodology in the main tests
to examine the impact of internal control material weaknesses (ICMW hereafter) on
sell side analysts. We match our ICMW firms with non-ICMWs based on industry,
sales, and assets. We re-estimate the models using rank regression technique to
assess the sensitivity of the results to the underlying functional form assumption
made by OLS. We use Cook’s distance to test the outliers.
Keywords: Internal control material weakness, Analyst forecast accuracy, Ana-
lyst forecast bias, Brokerage reputation, Sarbanes-Oxley act, Ordinary least squares
regressions, Rank regressions, Fixed effects, Matching procedure, Cook’s distance
This chapter applies Logit and OLS econometric techniques to analyze the Federal
Reserve Y-9C report data. We show that short-term financing is a response to the adverse
economic shocks rather than a cause of the recent crisis. The likelihood of financial crisis
actually stems from the illiquidity and low creditworthiness of the investment. Our
results are robust to endogeneity concerns when we use a difference-in-differences
(DiD) approach with the Lehman bankruptcy in 2008 proxying for an exogenous shock.
Keywords: Financial crisis, Short-term financing, Debt maturity, Liquidity risk,
Deterioration of bank asset quality
formulae for barrier options are derived by repeatedly applying the reflection
principle. Besides, our formulae can be applied to extend the applicability of the
first passage model – a branch of structural credit risk model. The stock price falls
due to the dividend payout in the option pricing problem is analog to selling the
firm’s asset to finance the loan repayment or dividend payout in the first passage
model. Thus, our formulae can evaluate vulnerable bonds or the equity values given
that the firm’s future loan/dividend payments are known.
Keywords: Barrier option, Option pricing, Stock option, Dividend, Reflection
principle, Lognormal, Credit risk
This chapter uses the criteria measurements to evaluate investment style and
investigate multiple criteria decision-making (MCDM) problem. To achieve this
objective, first, we employ factor analysis to extract independent common factors
from those criteria. Second, we construct the evaluation frame using hierarchical
system composed of the above common factors with evaluation criteria, and then
derive the relative weights with respect to the considered criteria. Third, the
synthetic utility value corresponding to each investment style is aggregated by
the weights with performance values. Finally, we compare with empirical data and
find that the model of MCDM predicts the rate of return.
Keywords: Investment strategies, Multiple Criteria Decision Making (MCDM),
Hierarchical system, Investment style, Factor analysis, Synthetic utility value,
Performance values
This chapter investigates carry trade strategy in the currency markets whereby
investors fund positions in high interest rate currencies by selling low interest rate
currencies to earn the interest rate differential. In this chapter, we first provide an
overview of the risk and return profile of currency carry trade; second, we introduce
two popular models, the regime-switch model and the logistic smooth transition
regression model, to analyze carry trade returns because the carry trade returns are
highly regime dependent. Finally, an empirical example is illustrated.
Keywords: Carry trade, Uncovered interest parity, Markov chain Monte Carlo,
Regime-switch model, Logistic smooth transition regression model
This study adopts the survival analysis framework (Allison 1984) to examine
issuer-heterogeneity and time-heterogeneity in the rating migrations of fallen
angels (FAs) and their speculative grade-rated peers (FA peers). Cox’s hazard
model is considered the pre-eminent method to estimate the probability that
an issuer survives in its current rating grade at any point in time t over the
time horizon T. In this study, estimation is based on two Cox’s hazard
models, including a proportional hazard model (Cox 1972) and a dynamic
hazard model. The first model employs a static estimation approach and time-
independent covariates, whereas the second uses a dynamic estimation
approach and time-dependent covariates. To allow for any dependence
among rating states of the same issuer, the marginal event-specific method
(Wei et al. 1989) was used to obtain robust variance estimates. For validation
purpose, the Brier score (Brier 1950) and its covariance decomposition
(Yates 1982) were applied to assess the forecast performance of estimated
models in forming time-varying survival probability estimates for issuers
out-of-sample.
Keywords: Survival analysis, Hazard model, Time-varying covariate, Recurrent
event, Brier score, Covariance decomposition, Rating migration, Fallen angel,
Markov property, Issuer-heterogeneity, Time-heterogeneity
This chapter provides a brief overview of the basic Vector autoregression (VAR)
approach by focusing on model estimation and statistical inferences. VAR models
have been used extensive in finance and economic analysis. Applications of VAR
models in some finance areas are discussed, including asset pricing, international
finance, and market microstructure. It is shown that such approach provides
a powerful tool to study financial market efficiency, stock return predictability,
exchange rate dynamics, and information content of stock trades and market
quality.
1 Introduction to Financial Econometrics and Statistics 49
The chapter examines three specific, different hedonic specifications: the linear,
semi-log, and Box-Cox transformed hedonic models and applies them to real estate
data. It also discusses recent innovations related to hedonic models and how these
models are being used in contemporary studies. This provides a basic overview of
the nature and variety of hedonic empirical pricing models that are employed in
the economics literature. It explores the history of hedonic modeling and summa-
rizes the field’s utility-theory-based, microeconomic foundations. It also provides
a discussion of and potential solutions for common problems associated with
hedonic modeling.
Keywords: Hedonic models, Regression, Real estate, Box-Cox, Pricing, Price
indexes, Semi-log, Least squares, Housing, Property
introduce growth rate, systematic risk, and total risk variables into the theoretical
model. We use a panel data collected in the USA from 1969 to 2009 to empirically
investigate the impact of growth rate, systematic risk, and total risk on the optimal
payout ratio in terms of the fixed-effects model. Furthermore, we implement the
moving estimates process to find the empirical breakpoint of the structural change
for the relationship between the payout ratio and risks and confirm that the
empirical breakpoint is not different from our theoretical breakpoint. Our theo-
retical model and empirical results can therefore be used to identify whether
flexibility or the free cash flow hypothesis should be used to determine the
dividend policy.
Keywords: Dividends, Payout policy, Optimal payout ratio, Flexibility hypoth-
esis, Free cash flow hypothesis, Signaling hypothesis, Fixed effect, Clustering
effect, Structural change model, Moving estimates processes, Systematic risk,
Total risk, Market perfection
This chapter uses an exponential generalized beta distribution of the second kind
(EGB2) to model the returns on 30 Dow-Jones industrial stocks. The model
accounts for stock return characteristics, including fat tails, peakedness
(leptokurtosis), skewness, clustered conditional variance, and leverage effect.
The goodness-of-fit statistic provides supporting evidence in favor of EGB2
distribution in modeling stock returns. The EGB2 distribution used in this
chapter is a four parameter distribution. It has a closed-form density function,
and its higher-order moments are finite and explicitly expressed by its parameters.
The EGB2 distribution nests many widely used distributions such as normal
distribution, log-normal distribution, Weibull distribution, and standard logistic
distribution.
Keywords: Expected stock return, Higher moments, EGB2 distribution, Risk
management, Volatility, Conditional skewness, Risk premium
This chapter performs dominance test to show that revenue surprises, earnings
surprises, and prior returns, each lead to significant momentum returns that cannot
be fully explained by the others, suggesting that each convey some exclusive
and unpriced information content. Also, the joint implications of revenue surprises,
earnings surprises, and prior returns are underestimated by investors, particularly
when information variables point in the same direction. Momentum
1 Introduction to Financial Econometrics and Statistics 51
This chapter compares two types of GARCH models, namely, the VG-NGARCH
and the GARCH-jump model with autoregressive conditional jump intensity, i.e.,
the GARJI model, to make inferences on the log of stock returns when there are
irregular substantial price fluctuations. The VG-NGARCH model imposes
a nonlinear asymmetric structure on the conditional shape parameters in a variance-
gamma process, which describes the arrival rates for news with different degrees of
influence on price movements, and provides an ex ante probability for the occur-
rence of large price movements. On the other hand, the GARJI model, a mixed
GARCH-jump model proposed by Chan and Maheu (2002), adopts two indepen-
dent autoregressive processes to model the variances corresponding to moderate
and large price movements, respectively.
Keywords: VG-NGARCH model, GARCH-jump model, Autoregressive
conditional jump intensity, GARJI model, Substantial price fluctuations, Shape
parameter, Variance-gamma process, Ex ante probability, Daily stock price,
Goodness-of-fit
This chapter compares three major approaches to solve the numerical instability
problem inherent in the fundamental solution of the Heston model. In this chapter,
we used the fundamental transform method proposed by Lewis to reduce the
number of variables from two to one and separate the payoff function from the
calculation of the Green function for option pricing. We show that the simple
adjusted-formula method is much simpler than the rotation-corrected angle method
of Kahl and Jäckel and also greatly superior to the direct integration method of
Shaw if taking computing time into consideration.
Keywords: Heston, Stochastic volatility, Fourier inversion, Fundamental trans-
form, Complex logarithm, Rotation-corrected angle, Simple adjusted formula,
Green function
This chapter considers two time-scales and uses the “short” time-scale to define
GARCH dynamics and the “long” time-scale to incorporate parameter jumps. This
leads to a Bayesian change-point ARX-GARCH model, whose unknown parameters
may undergo occasional changes at unspecified times and can be estimated by explicit
recursive formulas when the hyperparameters of the Bayesian model are specified.
Efficient estimators of the hyperparameters of the Bayesian model can be developed.
The empirical Bayes approach can be applied to the frequentist problem of partitioning
the time series into segments under sparsity assumptions on the change-points.
Keywords: ARX-GARCH, Bounded complexity, Contemporaneous jumps,
Change-point models, Empirical Bayes, Frequentist segmentation, Hidden Markov
models, Hyperparameter estimation, Markov chain Monte Carlo, Recursive filters,
Regression models, Stochastic volatility
This chapter first reviews the recent literature on the Unspanned Stochastic Volatil-
ities (USV) documented in the interest rate derivatives markets. The USV refers to
the volatilities factors implied in the interest rate derivatives prices that have little
correlation with the yield curve factors. We then present the result in Li and Zhao
(2006) that a sophisticated DTSM without USV feature can have serious difficulties
in hedging caps and cap straddles, even though they capture bond yields well.
Furthermore, at-the-money straddle hedging errors are highly correlated with
cap-implied volatilities and can explain a large fraction of hedging errors of all
caps and straddles across moneyness and maturities. We also present a multifactor
1 Introduction to Financial Econometrics and Statistics 53
term structure model with stochastic volatility and jumps that yields a closed-form
formula for cap prices from Jarrow et al. (2007). The three-factor stochastic volatility
model with Poisson jumps can price interest rate caps well across moneyness and
maturity. The econometric methods in this chapter include extended Kalman filter-
ing, maximum likelihood estimation with latent variables, local polynomial method,
and nonparametric density estimation.
Keywords: Term structure modeling, Interest rate volatility, Heath-Jarrow-
Morton model, Nonparametric density estimation, Extended Kalman filtering
This chapter examines alternative equity valuation models and their ability to
forecast future stock prices. We use simultaneous equations estimation technique
to investigate the stock price forecast ability of Ohlson’s model, Feltham and
Ohlson’s Model, and Warren and Shelton’s (1971) model. Moreover, we use the
combined forecasting methods proposed by Granger and Newbold (1973) and
Granger and Ramanathan (1984) to form combined stock price forecasts from
individual models. Finally, we examine whether comprehensive earnings can
provide incremental price-relevant information beyond net income.
Keywords: Ohlson model, Feltham and Ohlson model, Warren and Shelton
model, Equity valuation models, Simultaneous equations estimation, Fundamental
analysis, Financial statement analysis, Financial planning and forecasting, Com-
bined forecasting, Comprehensive earnings, Abnormal earnings, Operating earn-
ings, Accounting earnings
Chapter 88: Time Series Models to Predict the Net Asset Value (NAV)
of an Asset Allocation Mutual Fund VWELX
This research examines the use of various forms of time-series models to predict
the total net asset value (NAV) of an asset allocation mutual fund. The first set of
model structures included simple exponential smoothing, double exponential
smoothing, and the Winter’s method of smoothing. The second set of predictive
models used represented trend models. They were developed using regression
estimation. They included linear trend model, quadratic trend model, and an
exponential model. The third type of method used was a moving average method.
The fourth set of models incorporated the Box-Jenkins method, including an
autoregressive model, a moving average model, and an unbounded autoregressive
and moving average method.
Keywords: NAV of a mutual fund, Asset allocation fund, Combination of
forecasts, Single exponential smoothing, Double exponential smoothing, Winter’s
method, Linear trend model, Quadratic trend model, Exponential trend model,
Moving average method, Autoregressive model, Moving average model,
Unbounded autoregressive moving average model
54 C.-F. Lee and J.C. Lee
This chapter reviews the different theoretical methods used to estimate implied
standard deviation and to show how the implied volatility can be estimated in
empirical work. The OLS method for estimating implied standard deviation is first
introduced and the formulas derived by applying a Taylor series expansion method
to Black-Scholes option pricing model are also described. Three approaches of
estimating implied volatility are derived from one, two, and three options, respec-
tively. Because of these formulas with the remainder terms, the accuracy of these
formulas depends on how an underlying asset is close to the present value of
exercise price in an option. The formula utilizing three options for estimating
implied volatility is more accurate rather than other two approaches. In this chapter,
we use call options on S&P 500 index futures in 2010 and 2011 to illustrate how
MATLAB can be used to deal with the issue of convergence in estimating implied
volatility of future options.
Keywords: Implied volatility, Implied standard deviation (ISD), Option pricing
model, MATLAB, Taylor series expansion, Ordinary least-squares (OLS), Black-
Scholes Model, Options on S&P 500 index futures
This chapter uses a new approach to estimate future credit risk on target portfolio based
on the framework of CreditMetricsTM by J.P. Morgan. However, we adopt the
perspective of factor copula and then bring the principal component analysis concept
into factor structure to construct a more appropriate dependence structure among
credits. In order to examine the proposed method, we use real market data instead of
a virtual one. We also develop a tool for risk analysis which is convenient to use,
1 Introduction to Financial Econometrics and Statistics 55
especially for banking loan businesses. The results show the fact that people assume
dependence structures are normally distributed will indeed lead to risks underestimate.
On the other hand, our proposed method captures better features of risks and shows the
fat-tail effects conspicuously even though assuming the factors are normally distributed.
Keywords: Credit risk, Credit VaR, Default correlation, Copula, Factor copula,
Principal component analysis
This chapter conducts some simulation tests to justify the effectiveness of the
Fourier transform method. Malliavin and Mancino (2009) proposed a nonparamet-
ric Fourier transform method to estimate the instantaneous volatility under the
assumption that the underlying asset price process is a semi-martingale. Two
correction schemes are proposed to improve the accuracy of volatility estimation.
By means of these Fourier transform methods, some documented phenomena such
as volatility daily effect and multiple risk factors of volatility can be observed.
Then, a linear hypothesis between the instantaneous volatility and VIX derived
from Zhang and Zhu (2006) is investigated.
Keywords: Information content, Instantaneous volatility, Fourier transform
method, Bias reduction, Correction method, Local volatility, Stochastic volatility,
VIX, Volatility daily effect, Online estimation
This chapter first theoretically extends Black’s CAPM, and then uses price, dividend
per share, and earnings per share to test the existence of supply effect with US equity
data. A simultaneous equation system is constructed through a standard structural
form of a multi-period equation to represent the dynamic relationship between supply
and demand for capital assets. The equation system is exactly identified under our
specification. Then, two hypotheses related to supply effect are tested regarding
the parameters in the reduced-form system. The equation system is estimated by the
Seemingly Unrelated Regression (SUR) method, since SUR allows one to estimate
the presented system simultaneously while accounting for the correlated errors.
Keywords: CAPM, Asset, Endogenous supply, Simultaneous equations
This chapter proposes the generalized hyperbolic distribution as the joint log-return
distribution of the spot and futures. Using the parameters in this distribution, we
derive several most widely used optimal hedge ratios: minimum variance,
56 C.-F. Lee and J.C. Lee
This chapter reviews the instrumental variables (IV) approach to endogeneity from
the point of view of a finance researcher who is implementing instrumental variable
methods in empirical studies. This chapter is organized into two parts. Part I
discusses the general procedure of the instrumental variable approach, including
Two-Stage Least Square (2SLS) and Generalized Method of Moments (GMM), the
related diagnostic statistics for assessing the validity of instruments, which are
important but not used very often in finance applications, and some recent advances
in econometrics research on weak instruments. Part II surveys corporate finance
applications of instrumental variables. We found that the instrumental variables
used in finance studies are usually chosen arbitrarily, and very few diagnostic
statistics are performed to assess the adequacy of IV estimation. The resulting IV
estimates thus are questionable.
Keywords: Endogeneity, OLS, Instrumental variable (IV) estimation,
Simultaneous equations, 2SLS, GMM, Overidentifying restrictions, Exogeneity
test, Weak instruments, Anderson-Rubin statistic, Empirical corporate finance
This research first discusses the evolution of probability of informed trading in the
finance literature. Motivated by asymmetric effects, e.g., return and trading volume
in up and down markets, this study modifies a mixture of the Poisson distribution
model by different arrival rates of informed buys and sells to measure the
probability of informed trading proposed by Easley et al. (1996). By applying
the expectation–maximization (EM) algorithm to estimate the parameters of the
model, we derive a set of equations for maximum likelihood estimation and these
equations are encoded in a SAS Macro utilizing SAS/IML for implementation of
the methodology.
1 Introduction to Financial Econometrics and Statistics 57
This chapter uses ordinary least squares estimation to investigate the relations
between CEO stock options and analysts’ earnings forecast accuracy and bias.
Our OLS models relate forecast accuracy and forecast bias (the dependent
variables) to CEO stock options (the independent variable) and controls for earn-
ings characteristics, firm characteristics, and forecast characteristics. In addition,
the models include controls for industry and year. We use four measures of options:
new options, existing exercisable options, existing unexercisable options, and total
options (sum of the previous three), all scaled by total number of shares outstand-
ing, and estimate two models for each dependent variable, one including total
options and the other including new options, existing exercisable options, and
existing unexercisable options. We also use both contemporaneous as well as
lagged values of options in our main tests.
Keywords: CEO stock options, Analysts’ forecast accuracy, Analysts’ forecast
bias, CEO compensation, Agency costs, Investment risk taking, Effort allocation,
Opportunistic earnings management, Opportunistic disclosure management,
Forecasting complexity
This chapter fills this gap by first developing an implementable option model in
closed form that admits both stochastic volatility and stochastic interest rates and that
is parsimonious in the number of parameters. Based on the model, both delta-neutral
and single-instrument minimum variance hedging strategies are derived analytically.
Using S&P 500 option prices, we then compare the pricing and hedging performance
of this model with that of three existing ones that, respectively, allow for (i) constant
volatility and constant interest rates (the Black-Scholes), (ii) constant volatility but
stochastic interest rates, and (iii) stochastic volatility but constant interest rates.
Overall, incorporating stochastic volatility and stochastic interest rates produces the
best performance in pricing and hedging, with the remaining pricing and hedging
errors no longer systematically related to contract features. The second performer in
the horse-race is the stochastic volatility model, followed by the stochastic interest
rates model and then by the Black-Scholes.
Keywords: Stock option pricing, Stochastic volatility, Stochastic interest rates,
Hedge ratios, Hedging, Pricing performance, and Hedging performance
58 C.-F. Lee and J.C. Lee
This chapter purports to provide a theoretical underpinning for the problem of the
Investment Company Act. The theory of the Le Chatelier Principle is well known in
thermodynamics: The system tends to adjust itself to a new equilibrium as far as
possible. In capital market equilibrium, added constraints on portfolio investment in
each stock can lead to inefficiency manifested in the right-shifting efficiency
frontier. According to the empirical study, the potential loss can amount to millions
of dollars coupled with a higher risk-free rate and greater transaction and informa-
tion costs.
Keywords: Markowitz model, Efficient frontiers, With constraints, Without
constraints, Le Chatelier principle, Thermodynamics, Capital market equilibrium,
Diversified mutual funds, Quadratic programming, Investment company act
References
Aitchison, J., & Brown, J. A. C. (1973). Lognormal distribution, with special reference to its uses
in economics. London: Cambridge University Press.
Ait-Sahalia, Y., & Lo, A. W. (2000). Nonparametric risk management and implied risk aversion.
Journal of Econometrics, 94, 9–51.
Amemiya, T. (1974). A note on a fair and Jaffee model. Econometrica, 42, 759–762.
Anderson, T. W. (1994). The statistical analysis of time series. New York: Wiley-Interscience.
Anderson, T. W. (2003). An introduction to multivariate statistical analysis. New York: Wiley-
Interscience.
Angrist, J. D., & Lavy, V. (1999). Using Maimonides’ rule to estimate the effect of class size on
scholastic achievement. Quarterly Journal of Economics, 14(2), 533–575.
Arellano, M., & Bover, O. (1995). Another look at the instrumental variable estimation of error-
components models. Journal of Econometrics, 68(1), 29–51.
Bakshi, G., Cao, C., & Chen, Z. (1997). Empirical performance of alternative option pricing
models. Journal of Finance, 52(5), 2003–2049.
Baltagi, B. (2008). Econometric analysis of panel data (4th ed.). Chichester: Wiley.
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of
Political Economy, 81(3), 637–654.
Black, F., Jensen, M. C., & Scholes, M. (1972). The capital asset pricing model: Some empirical tests.
In M. C. Jensen (Ed.), Studies in the theory of capital markets. New York: Praeger.
Blume, M. E., & Friend, I. (1973). A new look at the capital asset pricing model. Journal of
Finance, 28, 19–33.
Brown, S. J., & Goetzmann, W. N. (1997). Mutual fund styles. Journal of Financial Economics,
43(3), 373–399.
Cameron, A. C., Gelbach, J. B., & Miller, D. L. (2006). Robust inference with multiway clustering
(Working Paper).
Chacko, G., & Viceira, L. M. (2003). Spectral GMM estimation of continuous-time processes.
Journal of Econometrics, 116(1), 259–292.
Chang, C. F. (1999). Determinants of capital structure and management compensation: The
partial least squares approach. Ph.D. dissertation, Rutgers University.
Chang, H. S., & Lee, C. F. (1977). Using pooled time-series and cross section data to test the firm
and time effects in financial analysis. Journal of Financial and Quantitative Analysis,
12, 457–471.
1 Introduction to Financial Econometrics and Statistics 59
Chang, C., Lee, A. C., & Lee, C. F. (2009). Determinants of capital structure choice: A structural
equation modeling approach. Quarterly Review of Economic and Finance, 49(2), 197–213.
Chen, H. Y. (2011). Momentum strategies, dividend policy, and asset pricing test.
Ph.D. dissertation, Rutgers University.
Chen, S. N., & Lee, C. F. (1981). The sampling relationship between Sharpe’s performance
measure and its risk proxy: Sample size, investment horizon and market conditions. Manage-
ment Science, 27(6), 607–618.
Chen, K. H., & Shimerda, T. A. (1981). An empirical analysis of useful finance ratios. Financial
Management, 10(1), 51–60.
Chen, W. P., Chung, H., Lee, C. F., & Liao, W. L. (2007). Corporate governance and equity
liquidity: Analysis of S&P transparency and disclosure rankings. Corporate Governance: An
International Review, 15(4), 644–660.
Cheng, D. C., & Lee, C. F. (1986). Power of alternative specification errors tests in identifying
misspecified market models. The Quarterly Review of Economics and Business, 16(3), 6–24.
Cheng, K. F., Chu, C. K., & Hwang, R. C. (2010). Predicting bankruptcy using the discrete-time
semiparametric hazard model. Quantitative Finance, 10, 1055–1066.
Chow, G. C. (1960). Tests of equality between sets of coefficients in two linear regressions.
Econometrica, 28, 591–605.
Core, J. E. (2000). The directors’ and officers’ insurance premium: An outside assessment of the
quality of corporate governance. Journal of Law, Economics, and Organization, 16(2), 449–477.
Cox, J. C., Ross, S. A., & Rubinstein, M. (1979). Option pricing: A simplified approach. Journal of
Financial Economics, 7(3), 229–263.
Fair, R. C., & Jaffee, D. M. (1972). Methods of estimation of market in disequilibrium.
Econometrica, 40, 497–514.
Fama, E. F. (1971). Parameter estimates for symmetric stable distributions. Journal of the
American Statistical Association, 66, 331–338.
Fama, E. F., & MacBeth, J. D. (1973). Risk, return, and equilibrium: Empirical tests. Journal of
Political Economy, 81, 607–636.
Fok, R. C. W., Lee, C. F., & Cheng, D. C. (1996). Alternative specifications and estimation
methods for determining random beta coefficients: Comparison and extensions. Journal of
Financial Studies, 4(2), 61–88.
Frecka, T. J., & Lee, C. F. (1983). Generalized financial ratio adjustment processes and their
implications. Journal of Accounting Research, 21, 308–316.
Gordon, J., & Gordon, M. (1997). The fiinite horizon expected return model. The Financial
Analyst Journal, 53, 52–61.
Granger, C. W. J., & Newbold, P. (1973). Some comments on the evaluation of economic
forecasts. Applied Economics, 5(1), 35–47.
Granger, C. W. J., & Newbold, P. (1974). Spurious regressions in econometrics. Journal of
Econometrics, 2, 111–120.
Granger, C. W. J., & Ramanathan, R. (1984). Improved methods of combining forecasts. Journal
of Forecasting, 3, 197–204.
Hamilton, J. D. (1994). Time series analysis. Princeton: Princeton University Press.
Hansen, L. P. (1982). Large sample properties of generalized method of moments estimators.
Econometrica, 50(4), 1029–1054.
Hansen, B. E. (1996). Inference when a nuisance parameter is not identified under the null
hypothesis. Econometrica, 64(2), 413–430.
Hansen, B. E. (1997). Approximate asymptotic p-values for structural change tests. Journal of
Business and Economic Statistics, 15, 60–67.
Hansen, B. E. (1999). Threshold effects in non-dynamic panels: Estimation, testing, and inference.
Journal of Econometrics, 93, 345–368.
Hansen, B. E. (2000a). Sample splitting and threshold estimation. Econometrica, 68(3), 575–603.
Hansen, B. E. (2000b). Testing for structural change in conditional models. Journal of Economet-
rics, 97, 93–115.
60 C.-F. Lee and J.C. Lee
Heston, S. L. (1993). A closed-form solution for options with stochastic volatility with applica-
tions to bond and currency options. Review of Financial Studies, 6(2), 327–343.
Hsiao, C. (2003). Analysis of panel data (2nd ed.). Cambridge MA: Cambridge University Press.
Hutchinson, J. M., Lo, A. W., & Poggio, T. (1994). A nonparametric approach to pricing and
hedging derivative securities via learning networks. Journal of Finance, 49(3), 851–889.
Hwang, R. C., Cheng, K. F., & Lee, J. C. (2007). A semiparametric method for predicting
bankruptcy. Journal of Forecasting, 26, 317–342.
Hwang, R. C., Wei, H. C., Lee, J. C., & Lee, C. F. (2008). On prediction of financial distress using
the discrete-time survival model. Journal of Financial Studies, 16, 99–129.
Hwang, R. C., Cheng, K. F., & Lee, C. F. (2009). On multiple-class prediction of issuer crediting
ratings. Journal of Applied Stochastic Models in Business and Industry, 25, 535–550.
Hwang, R. C., Chung, H., & Chu, C. K. (2010). Predicting issuer credit ratings using
a semiparametric method. Journal of Empirical Finance, 17(1), 120–137.
Ittner, C. D., Larcker, D. F., & Rajan, M. V. (1997). The choice of performance measures in annual
bonus contracts. Accounting Review, 72(2), 231–255.
Jones, J. (1991). Earnings management during import relief investigations. Journal of Accounting
Research, 29, 193–228.
Kao, L. J., & Lee, C. F. (2012). Alternative method for determining industrial bond ratings: theory
and empirical evidence. International Journal of Information Technology & Decision Making,
11(6), 1215–1236.
Kau, J. B., & Lee, C. F. (1976). Functional form, density gradient and the price elasticity of
demand for housing. Urban Studies, 13(2), 181–192.
Kau, J. B., Lee, C. F., & Sirmans, C. F. (1986). Urban econometrics: Model developments and
empirical results. Greenwich: JAI Press. Research in Urban Economics, Vol. 6.
Kaufman, L., & Rousseeuw, P. J. (1990). Finding groups in data: An introduction to cluster
analysis (9th ed.). New York: Wiley-Interscience.
Kim, D. (1995). The errors in the variables problem in the cross-section of expected stock returns.
Journal of Finance, 50(5), 1605–1634.
Kim, D. (1997). A reexamination of firm size, book-to-market, and earnings price in the cross-
section of expected stock returns. Journal of Financial and Quantitative Analysis, 32(4),
463–489.
Kim, D. (2010). Issues related to the errors-in-variables problems in asset pricing tests. In
Handbook of quantitative finance and risk management (Part V, Chap. 70, pp. 1091–1108).
Kuan, C. M., & Hornik, K. (1995). The generalized fluctuation test: A unifying view. Econometric
Reviews, 14, 135–161.
Lambert, R., & Larcker, D. (1987). An analysis of the use of accounting and market measures of
performance in executive compensation contracts. Journal of Accounting Research, 25, 85–125.
Lee, C. F. (1973). Errors-in-variables estimation procedures with applications to a capital asset
pricing model. Ph.D. dissertation, The State University of New York at Buffalo.
Lee, C. F. (1976a). A note on the interdependent structure of security returns. Journal of Financial
and Quantitative Analysis, 11, 73–86.
Lee, C. F. (1976b). Functional form and the dividend effect of the electric utility industry. Journal
of Finance, 31(5), 1481–1486.
Lee, C. F. (1977a). Functional form, skewness effect and the risk-return relationship. Journal of
Financial and Quantitative Analysis, 12, 55–72.
Lee, C. F. (1977b). Performance measure, systematic risk and errors-in-variable estimation
method. Journal of Economics and Business, 39, 122–127.
Lee, A. (1996). Cost of capital and equity offerings in the insurance industry. Ph.D. dissertation,
The University of Pennsylvania in Partial.
Lee, C. F., & Chen, H. Y. (2013). Alternative errors-in-variable models and their application in
finance research (Working Paper). Rutgers University.
Lee, A. C., & Cummins, J. D. (1998). Alternative models for estimating the cost of capital for
property/casualty insurers. Review of Quantitative Finance and Accounting, 10(3), 235–267.
1 Introduction to Financial Econometrics and Statistics 61
Lee, C. F., & Jen, F. C. (1978). Effects of measurement errors on systematic risk and performance
measure of a portfolio. Journal of Financial and Quantitative Analysis, 13(2), 299–312.
Lee, K. W., & Lee, C. F. (2012). Multiple directorships, corporate ownership, firm performance
(Working Paper). Rutgers University.
Lee, C. F., & Wu, C. C. (1985). The impacts of kurtosis on risk stationarity: Some empirical
evidence. Financial Review, 20(4), 263–269.
Lee, C. F., & Zumwalt, J. K. (1981). Associations between alternative accounting
profitability measures and security returns. Journal of Financial and Quantitative Analysis,
16, 71–93.
Lee, C. F., Newbold, P., Finnerty, J. E., & Chu, C. C. (1986). On accounting-based, market-
based and composite-based beta predictions: Method and implications. Financial Review,
21, 51–68.
Lee, C. F., Wei, K. C. J., & Bubnys, E. L. (1989). The APT versus the multi-factor CAPM:
Empirical evidence. Quarterly Review of Economics and Business, 29, 6–16.
Lee, C. F., Wu, C. C., & Wei, K. C. J. (1990). Heterogeneous investment horizon and capital asset
pricing model: Theory and implications. Journal of Financial and Quantitative Analysis,
25, 361–376.
Lee, C. F., Tsai C. M., & Lee A. C. (2011a). Asset pricing with disequilibrium price adjustment:
Theory and empirical evidence. Quantitative Finance.
Lee, C. F., Gupta, M. C., Chen, H. Y., & Lee, A. C. (2011b). Optimal payout ratio under
uncertainty and the flexibility hypothesis: Theory and empirical evidence. Journal of Corpo-
rate Finance, 17(3), 483–501.
Lee, C. F., Chien, C. C., Lin, H. C., Lin, Y. C., & Lin, F. C. (2013). Tradeoff between reputation
concerns and economic dependence for auditors – Threshold regression approach (Working
Paper). Rutgers University.
Leon, A., & Vaello-Sebastia, A. (2009). American GARCH employee stock option valuation.
Journal of Banking and Finance, 33(6), 1129–1143.
Lien, D. (2010). An note on the relationship between the variability of the hedge ratio and hedging
performance. Journal of Futures Market, 30(11), 1100–1104.
Lien, D., & Shrestha, K. (2007). An empirical analysis of the relationship between
hedge ratio and hedging horizon using wavelet analysis. Journal of Futures Market, 27(2),
127–150.
Liu, B. (2006). Two essays in financial econometrics: I: Functional forms and pricing of country
funds. II: The term structure model of inflation risk premia. Ph.D. dissertation, Rutgers University.
Maddala, G. S., Rao, C. R., & Maddala, G. S. (1996). Handbook of statistics 14: Statistical
methods in finance. Amsterdam: Elsevier Science & Technology.
Martin, C. (1990). Corporate borrowing and credit constrains: Structural disequilibrium estimates
for the U.K. Review of Economics and Statistics, 72(1), 78–86.
Mayer, W. J. (1989). Estimating disequilibrium models with limited a priori price-adjustment
information. Journal of Econometrics, 41(3), 303–320.
Miller, M. H., & Modigliani, F. (1966). Some estimates of the cost of capital to the utility industry,
1954-7. American Economic Review, 56(3), 333–391.
Myers, R. J. (1991). Estimating time-varying optimal hedge ratios on futures markets. Journal of
Futures Markets, 11(1), 39–53.
Newey, W. K., & West, K. D. (1987). A simple, positive semi-definite, heteroskedasticity and
autocorrelation consistent covariance matrix. Econometrica, 55(3), 703–708.
Ohlson, J. S. (1980). Financial ratios and the probabilistic prediction of bankruptcy. Journal of
Accounting Research, 19, 109–131.
Petersen, M. A. (2009). Estimating standard errors in finance panel data sets: Comparing
approaches. Review of Financial Studies, 22, 435–480.
Pinches, G. E., & Mingo, K. A. (1973). A multivariate analysis of industrial bond ratings. Journal
of Finance, 28(1), 1–18.
Quandt, R. E. (1988). The econometrics of disequilibrium. New York: Basil Blackwell Inc.
62 C.-F. Lee and J.C. Lee
Rendleman, R. J., Jr., & Barter, B. J. (1979). Two-state option pricing. Journal of Finance,
24, 1093–1110.
Rubinstein, M. (1994). Implied binomial trees. Journal of Finance, 49, 771–818.
Schroder, M. (1989). Computing the constant elasticity of variance option pricing formula.
Journal of Finance, 44(1), 211–219.
Sealey, C. W., Jr. (1979). Credit rationing in the commercial loan market: Estimates of a structural
model under conditions of disequilibrium. Journal of Finance, 34, 689–702.
Sears, R. S., & John Wei, K. C. (1988). The structure of skewness preferences in asset pricing
model with higher moments: An empirical test. Financial Review, 23(1), 25–38.
Shapiro, A. F. (2005). Fuzzy regression models (Working Paper). Penn State University.
Shumway, T. (2001). Forecasting bankruptcy more accurately: A simple hazard model. The
Journal of Business, 74, 101–124.
Thompson, S. B. (2011). Simple formulas for standard errors that cluster by both firm and time.
Journal of Financial Economics, 99(1), 1–10.
Thursby, J. G. (1985). The relationship among the specification error tests of Hausman, Ramsey
and Chow. Journal of the American Statistical Association, 80(392), 926–928.
Titman, S., & Wessels, R. (1988). The determinants of capital structure choice. Journal of
Finance, 43(1), 1–19.
Tsai, G. M. (2005). Alternative dynamic capital asset pricing models: Theories and empirical
results. Ph.D. dissertation, Rutgers University.
Van Der Klaauw, W. (2002). Estimating the effect of financial aid offers on college enrollment:
A regression-discontinuity approach. International Economic Review, 43(4), 1249–1287.
Wei, K. C. (1984). The arbitrage pricing theory versus the generalized intertemporal capital asset
pricing model: Theory and empirical evidence. Ph.D. dissertation, University of Illinois at
Urbana-Champaign.
White, H. (1980). A heteroskedasticity-consistent covariance matrix estimator and a direct test for
heteroskedasticity. Econometrica, 48(4), 817–838.
Wooldridge, J. M. (2010). Econometric analysis of cross section and panel data (2nd ed.).
Cambridge, MA: The MIT Press.
Yang, C. C. (1989). The impact of new equity financing on firms’ investment, dividend and debt-
financing decisions. Ph.D. dissertation, The University of Illinois at Urbana-Champaign.
Yang, C. C., Lee, C. F., Gu, Y. X., & Lee, Y. W. (2010). Co-determination of capital structure and
stock returns – A LISREL approach: An empirical test of Taiwan stock markets. Quarterly
Review of Economic and Finance, 50(2), 222–233.
Zeileis, A., Leisc, F., Hornik, K., & Kleiber, C. (2002). Strucchange: An R package for testing for
structural change in linear regression models. Journal of Statistical Software, 7, 1–38.
Experience, Information Asymmetry, and
Rational Forecast Bias 2
April Knill, Kristina L. Minnick, and Ali Nejadmalayeri
Contents
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 64
2.2 Theoretical Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 67
2.3 Empirical Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 71
2.4 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 75
2.5 Empirical Findings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 78
2.6 Robustness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
2.7 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
Appendix 1: Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
The Objective Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
Proof of Proposition 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 97
Proof of Proposition 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Appendix 2: Alternate Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
Appendix 3: Alternate Proxies for Information Asymmetry for Table 2.5 . . . . . . . . . . . . . . . . . . . . 98
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
Abstract
We use a Bayesian model of updating forecasts in which the bias in forecast
endogenously determines how the forecaster’s own estimates weigh into the
posterior beliefs. Our model predicts a concave relationship between accuracy in
forecast and posterior weight that is put on the forecaster’s self-assessment.
A. Knill (*)
The Florida State University, Tallahassee, FL, USA
e-mail: aknill@cob.fsu.edu
K.L. Minnick
Bentley University, Waltham, MA, USA
e-mail: kminnick@bentley.edu
A. Nejadmalayeri
Department of Finance, Oklahoma State University, Oklahoma, OK, USA
e-mail: ali.nejadmalayeri@okstate.edu
We then use a panel regression to test our analytical findings and find that an
analyst’s experience is indeed concavely related to the forecast error.
This study examines whether it is ever rational for analysts to post biased
estimates and how information asymmetry and analyst experience factor into the
decision. Using a construct where analysts wish to minimize their forecasting
error, we model forecasted earnings when analysts combine private information
with consensus estimates to determine the optimal forecast bias, i.e., the devi-
ation from the consensus. We show that the analyst’s rational bias increases with
information asymmetry, but is concavely related with experience. Novice ana-
lysts post estimates similar to the consensus but as they become more experi-
enced and develop private information channels, their estimates become biased
as they deviate from the consensus. Highly seasoned analysts, who have superior
analytical skills and valuable relationships, need not post biased forecasts.
Keywords
Financial analysts • Forecast accuracy • Information asymmetry • Forecast bias •
Bayesian updating • Panel regressions • Rational bias • Optional bias • Analyst
estimation • Analyst experience
2.1 Introduction
Extant evidence suggests an intimate link between an analyst’s experience and her
forecasting performance. Analysts who are experienced and highly specialized
often forecast better than others (Clement and Tse 2005; Bernhardt et al. 2006).
One way they do so is by posting an optimistic bias (Mest and Plummer 2003; Gu
and Xue 2007). Novice analysts with limited resources tend to herd with others,
which results in almost no bias (Bernhardt et al. 2006). In theory, superior fore-
casters produce better estimates either by resolving information asymmetry or by
offering a better assessment. Lim (2001) suggests that analysts can improve fore-
cast accuracy by strategically biasing their forecasts upwards, which placates
management, and in essence purchases additional information.1 Analysts with
1
Beyer (2008) argues that, even without incentives to appease management, analysts may still post
forecasts that exceed median earnings because managers can manipulate earnings upward to
prevent falling short of earnings forecasts. Moreover, Conrad et al. (2006) find support for the
idea that analysts’ “. . . recommendation changes are “sticky” in one direction, with analysts
reluctant to downgrade.” Evidence also indicates that analysts rarely post sell recommendations
for a stock, suggesting that losing a firm’s favor can be viewed as a costly proposition. At the
extreme, firms even pursue legal damages for an analyst’s unfavorable recommendations. In
a 2001 congressional hearing, president and chief executive officer of the Association for Invest-
ment Management and Research told the US House of Representatives Committee on Financial
Services, Capital Markets Subcommittee, that “. . .In addition to pressures within their firms,
analysts can also be, and have been, pressured by the executives of corporate issuers to issue
favorable reports and recommendations. Regulation Fair Disclosure notwithstanding, recent
history. . .has shown that companies retaliate against analysts who issue ‘negative’
2 Experience, Information Asymmetry, and Rational Forecast Bias 65
long histories of examining firms in a particular industry can also offer a unique
perspective, and they signal their ability by posting biased estimates to signal
superior ability (Bernhardt et al. 2006).
Of course, analysts do not indefinitely and indiscriminately bias forecasts to
appease the firm or signal their ability. This is mainly because analysts can learn
from the forecasts of other analysts (Chen and Jiang 2006). By incorporating
information from other forecasts, analysts can improve the accuracy of their own
forecasts without posting biased estimates. The important question then is: given
the analyst’s own assessment ability and her efficacy in procuring private informa-
tion versus using consensus information, how should she construct an optimal
forecast? Does that optimal forecast ever include bias, and how does information
asymmetry affect this decision? We address these questions by analytically
and empirically examining how an analyst’s experience and information asymme-
try affect her forecasting. In so doing, we also account for the role of the consensus
estimate in an analyst’s forecast. We begin by modeling the problem of
optimal forecasting. To be specific, we combine key features of current rational
forecasting models by Lim (2001) and Chen and Jiang (2006). As in Chen and Jiang
(2006), analysts in our model form rational (i.e., minimum squared error) forecasts
by weighing both public and private information.2 Following Lim (2001), our
analysts post rational forecasts that deviate from the consensus to purchase
private information from managers. Motivated by Lim (2001) and Bernhardt
et al. (2006), we also allow analysts to post biased forecasts because they have
more expertise than the consensus. The novelty of our approach is that we
directly model how information asymmetry and analyst experience combine to
affect the purchase of private information for use in the forecast deviation.
We are also able to model how analysts with different levels of experience, i.e.,
novice, moderately experienced, and highly seasoned analysts, construct their
forecast.
We analytically derive the optimal deviation from the consensus as one that
minimizes the mean squared error while allowing for rational Bayesian updating
based on public and private knowledge. Our analysis shows that even in a rational
forecast framework, analysts’ forecast deviation depends on the observed consen-
sus deviation of other analysts. When analysts observe others deviating from the
consensus (especially those with more experience), they gain enough insight to
avoid posting a large deviation themselves. Our results confirm the findings of both
Bernhardt et al. (2006) and Chen and Jiang (2006) – that analysts can rationally
herd. In the presence of the informative consensus, analysts choose to herd with
each other, rather than post estimates that are biased.
Our theory suggests that the likelihood of posting biased estimates, conditional
on the consensus, is significantly influenced by the analyst’s ability to process
information. Consistent with Hong et al. 2000), we show that novice analysts
essentially herd. Without either honed analytical abilities or valuable relationships,
these analysts follow the premise that the consensus is more accurate than their own
private information. Second, we show that a moderately experienced analyst relies
more on her sources inside the firm than her analytical skills. This manifests itself as
a biased forecast since she must appease management to tap into those sources. This
link between experience and deviation is not monotonic. Highly seasoned analysts
do not purchase as much additional information (or they purchase the information at
a reduced price), either because they possess superior analytical skills or because
their valuable relationships with firms afford them beneficial information without
the optimistic forecast. This preferential treatment is akin to that which is afforded
to companies in relationship lending with banks (Petersen and Rajan 1994). Indeed,
Carey et al. (1998) argue that the relationship lending of banks can also be ascribed
to some nonbank financial intermediaries. Although the firms that an analyst covers
are not necessarily financial, the same relationship could certainly exist and is based
on information asymmetry.
We further demonstrate that as the analyst-firm information asymmetry
increases, so does the bias. Similar to the results found in Mest and Plummer
(2003), analysts find private channels and analytical ability valuable mitigating
factors when faced with information asymmetry. Moderately experienced analysts,
who begin to tap into reliable private channels without the valuable relationships
that might afford preferential treatment, post larger deviations with the hope of
ascertaining better information. This suggests that both information asymmetry
and experience interactively affect the way analysts balance public and private
information to form forecasts. Our model also shows that the effect of information
asymmetry and analyst experience on rational deviation depends on the
dispersion of and the correlation between public and private signals. The quality
of private information channels, the informativeness of consensus, and the con-
nectedness of public and private signals significantly affect how analysts form
forecasts.
To examine the validity of our analytical findings, we empirically investigate
how analyst experience and information asymmetry affect forecast deviation from
the consensus.3 Our empirical results confirm our theoretical predictions that
rational bias (i.e., deviation from the consensus) is concavely related to analyst
experience and positively associated with information asymmetry. Novice analysts
and highly seasoned analysts post forecasts with smaller bias, while moderately
seasoned analysts post estimates that deviate further from the consensus. Moder-
ately seasoned analysts can benefit from a positive bias if they have the confidence
to separate from the herd and access reliable private sources of information.
3
Clement and Tse (2005) are the closest to our analysis; however while they admit that the
observed link between inexperience and herding can be a complex issue that might have other
roots than just career concerns, they do not provide detailed insight as to what and how this
complexity develops.
2 Experience, Information Asymmetry, and Rational Forecast Bias 67
These results are stronger for earlier forecasts versus later ones. As analysts become
highly seasoned, with external information networks and superior skills in fore-
casting, they find appeasing management to purchase information less useful
(or less costly in the same way that relationship lending affords firms cheaper
capital). As one might expect, when information asymmetry increases, rational
deviation from the consensus also increases. The degree of information
asymmetry faced by seasoned analysts has a significant positive effect on the
forecast deviation (information asymmetry also affects novice analysts but not as
extensively).
This study contributes to a growing literature on rational bias in analyst fore-
casting. Motivated by recent works by Bernhardt et al. (2006), Beyer (2008), Chen
and Jiang (2006), Lim (2001), and Mest and Plummer (2003), we take an integrated
modeling approach to arrive at a rational forecast bias, which is based on Bayesian
updating of public consensus and endogenously acquired private information. Our
approach is unique in that the forecast bias affects how public and private infor-
mation are combined. Specifically, analysts can use bias to improve forecast
accuracy by purchasing private information but can also learn from other analysts
by following the consensus. Our analytical findings, which are empirically
confirmed, show that unlike behavioral models, analysts can rationally herd. Unlike
signaling models, seasoned analysts do not always post biased estimates. It also
shows the value of the relationships that both moderately experienced and
highly seasoned analysts leverage to gain reliable private information. These results
are of practical interest because they provide evidence that analysts can and may
optimally bias their earnings estimates and that this optimal bias differs across
both analyst experience and information asymmetry. This knowledge may be
useful to brokerage houses for training purposes and/or evaluation of analyst
performance, particularly across industries with different levels of information
asymmetry.
Recent studies show that earnings forecasts reflect public information and private
assessments (Boni and Womack 2006; Chen and Jiang 2006; Lim 2001; Ramnath
2002). As Bernhardt et al. (2006) note, analysts’ forecasts of public information
partly relate to the consensus, common information, and unanticipated market-wide
shocks. We thus model analyst earnings forecasts using two components of earn-
ings information: common and idiosyncratic, with some uncertainty about each
component. In doing so, we also assume, as does Lim (2001), that the idiosyncratic
component directly relates to the private information analysts can obtain from the
firm by posting an optimistic view of the firm’s prospects (Nutt et al. 1999). Our
typical analyst also observes a previously posted forecast whereby the analyst can
learn about both the common and idiosyncratic components of earnings by incor-
porating previously disclosed information. Since we assume analysts are Bayesians,
they can learn from previous forecasts as an alternative to posting a positive
68 A. Knill et al.
deviation to purchase private signals from the firm. Since these prior estimates
partially reflect private information, the analysts may heavily weigh them into their
assessment, depending on the perceived information asymmetry and experience of
the previous analysts.4
In this section, we model analyst earnings forecasting in the presence of forecast
uncertainty. Following recent studies (Lim 2001; Ramnath 2002), our analyst has
an unconditional estimate, E ¼ X + e, about earnings, X N(0, s2), with
some uncertainty, e N(0, te2). As in Chen and Jiang (2006), our analyst
also observes a noisy consensus forecast, Ec ¼ X + ec, with a consensus
uncertainty, ec N(0, tc2). As a Bayesian, our analyst forms a conditional
forecast, F ¼ w E + (1 – w) Ec, by combining her unconditional estimate with the
consensus. The optimal weight in her conditional forecast minimizes her squared
forecast error.5
As in Lim (2001), the optimal forecasting, however, is endogenous to private
information acquisition. Our analyst’s forecast precision relates to the private
information analysts can obtain from the firm by posting an optimistic view of
the firm’s prospects. That is to say, while forecast precision, te, consists of a self-
accuracy component, t0, reflecting the analyst expertise, the forecast precision can
be improved by t(b) through placating managers by posting positively biased, b,
forecasts. The marginal precision per bias, ∂t/∂b, reflects the analyst’s information
asymmetry. This is because an analyst faced with greater information asymmetry
should derive a larger marginal benefit from a biased forecast. Since the conditional
forecast partially reflects private information, the analyst’s optimal rational forecast
(and forecast bias) depends on the analyst’s expertise and information asymmetry.
As noted before, the objective of the analyst is to minimize her squared forecast
error, F – X:
h i h i
min E ðF XÞ2 ¼ min ðwE þ ð1 wÞEc Þ2 (2.1)
wjb wjb
where E[•] is the expectation operator. Let’s assume that the analyst’s estimate and
the consensus are correlated, E[e, ec] ¼ r (t tc)1. This correlation reflects the
extent to which the analyst uses common private channels. The analyst objective,
Eq. 2.1, can be rewritten as
4
Here, we focus only on the case of one-period sequential forecasting. However, we believe that
the main implications of our model hold true for a multi-period sequential forecasting setting.
Since we assume that the probabilistic characteristics of different components are known and
analysts can gauge each others’ experience and the amount of information asymmetry perfectly,
there would be no incentive to deviate from posting commensurate optimal, rational forecasts. If
expert analysts intentionally deviate from their optimal forecasts, no other analyst can compensate
for their experience or information asymmetry [for more discussion, see Trueman (1990)].
5
For more details on Bayesian methods of inference and decision making, see Winkler (1972).
2 Experience, Information Asymmetry, and Rational Forecast Bias 69
h i
min w2 ðt0 þ tðbÞÞ2 þ ð1 wÞ2 tc 2 þ 2rwð1 wÞðt0 þ tðbÞÞ1 t1
c : (2.2)
wjb
As information asymmetry increases, i.e., marginal precision per bias rises, a larger
bias is needed to obtain the optimal weight of the analyst’s estimate and the
consensus. Interestingly, as an analyst becomes more experienced, her optimal
weight on her own estimate becomes less sensitive to information asymmetry.
where DFC is the Deviation from the consensus for analyst i in quarter t. We have
several analyst-specific controls: Experience, Experience squared, and the
NumRevisions. Following our analytical model, analysts with more experience
should be better at forecasting earnings and should have better channels of private
information. Following Leone and Wu (2007), we measure Experience in two
ways: using an absolute measure (the natural log of the quarters of experience)
and a relative measure (the natural log of the quarters of experience divided by the
72 A. Knill et al.
average of this measure for analysts following firms in the same industry). We
include Experience2 to test for a nonlinear relationship between Deviation and
Experience. To control for varying degrees of accuracy due to intertemporal
Bayesian updating, we include NumRevisions, which is the number of times the
analyst revises her estimate. Studies such as Givoly and Lakonishok (1979) suggest
that information has been acquired (i.e., purchased). This suggests a positive rela-
tionship between the number of revisions and the resulting deviation; thus we expect
the coefficient, b0, to be positive. Same quarter is an indicator variable for horizon
value, where the variable is equal to one if the estimate occurs in the same quarter as the
actual earnings report is released and zero otherwise. Including Same quarter can help
us understand whether analysts are more likely to be optimistic early in the period than
late in the period (Mikhail et al. 1997; Clement 1999; Clement and Tse 2005).6 We
expect the coefficient on this variable, b1, to be negative. Reg. FD is an indicator
variable equal to one if the quarter date is after Reg. FD was passed (October 23, 2000),
and zero otherwise.7 Although extant literature is divided on whether or not Reg. FD
has decreased the information asymmetry for analysts, if information asymmetry is
decreased and more public information available, analysts will be less likely to
purchase private information, and the coefficient on Reg. FD, b2, would be negative.
Indeed, Zitzewitz (2002) shows that although private information has decreased post
Reg. FD, the amount of public information has improved. Both Brown et al. (2004) and
Irani (2004) show that information asymmetry decreased following Reg. FD.8
We control for firm effects through the inclusion of firm-specific variables such as
Brokerage reputation, Brokerage size, Accruals, Intangible assets, and Return st. dev.
(vector X in Eq. 2.4). Following Barber et al. (2000), we use the number of companies
a brokerage house follows per year in total as a proxy for brokerage size. Brokerage
house reputation may play an integral role in accessing to private information
(Agrawal and Chen 2012). To calculate broker reputation, we start with the Carter
and Manaster (1990), Carter et al. (1998) and the Loughran and Ritter (2004)
rankings. When a firm goes public, the prospectus lists all of the firms that are in
the syndicate, along with their shares. More prestigious underwriters are listed higher
in the underwriting section. Based upon where the underwriting brokerage firm is
listed, they are assigned a value of 0–9, where nine is the highest ranking. As Carter
and Manaster (1990) suggest that prestigious financial institutions provide a lower
level of risk (i.e., lower information asymmetry), we include control variables for
these characteristics of the firm and expect the relationships to be negative.9
6
Horizon value and the NumRevisions are highly correlated at 65 %. We therefore orthogonalize
horizon value in the equation to ensure that multicollinearity is not a problem between these two
variables.
7
http://www.sec.gov/rules/final/33-7881.htm.
8
See Lin and Yang (2010) for a study of how Reg. FD affects analyst forecasts of restructuring
firms.
9
Brokerage reputation and Brokerage size are highly correlated at 67 %. We therefore orthogo-
nalize brokerage reputation in the equation to ensure that multicollinearity is not a problem
between these two variables.
2 Experience, Information Asymmetry, and Rational Forecast Bias 73
where t is the quarter in which we measure deviation and i denotes the ith analyst.
We define Information asymmetry three different ways: (1) the inverse of analyst
coverage, i.e., 1/(number of brokerage houses following the firm), (2) the standard
deviation of the company’s forecasts (102), and (3) the relative firm size, i.e., the
difference between the firm’s assets and the quarterly median assets for
the industry. These definitions are constructed such that the direction of the
expected marginal coefficient is congruent with that of information asymmetry.
10
Following Stangeland and Zheng (2007), we measure Accruals as income before extraordinary
items (Data #237) minus cash flow from operations, where cash flow from operations is defined as
net cash flow from operating activities (Data #308) minus extraordinary items and discontinued
operations (Data #124).
11
Following Hirschey and Richardson (2004), we calculate Intangibles as intangible assets to total
assets (Data 33/Data #6).
12
As an alternate proxy for industry fixed effects, Fama-French 12-industry classifications (Fama
and French 1997) are used. Results using these proxies are available upon request.
74 A. Knill et al.
The first definition, the inverse of analyst coverage, is based on evidence that the
level of financial analyst coverage affects how efficiently the market processes
information (Bhattacharya 2001). Analyst coverage proxies for the amount of
public and private information available for a firm and, therefore, captures
a firm’s information environment (Zhang 2006). The second definition, Forecast
dispersion, is supported in papers such as Krishnaswami and Subramaniam (1998)
and Thomas (2002), among others. Lastly, we use Relative firm size as a proxy for
information asymmetry. This is equally supported by the literature, including but
not limited to Petersen and Rajan (1994) and Sufi (2007). The firm-specific control
variables are the same as in Eq. 2.4, and a fixed effects treatment around analysts is
taken.
As previously explained, our model allows for the interaction of experience and
information asymmetry; the concavity of the relationship between experience and
Deviation may change at different levels of information asymmetry. Thus, we
estimate Eq. 2.4 for analysts with both high and low levels of information asym-
metry (i.e., based on relation to the industry-quarter median forecast dispersion).
We expect the coefficients on Experience and Experience squared to increase when
there is more information asymmetry.
Similarly, we demonstrate that experience affects the link between information
asymmetry and deviation. While the deviation of the novice and highly seasoned
analysts is only marginally affected by information asymmetry, the deviation of the
moderately experienced analyst is highly affected by information asymmetry. Since
Proposition 2 suggests that information asymmetry will affect analysts differently
based on their experience, we segment our sample into three groups: novice,
moderately experienced, and highly seasoned analysts. To form these experience
groups, we create quarterly terciles of analysts based on their experience. The
bottom third comprises the novice analysts; the middle third, moderately seasoned
analysts; and the top third, highly seasoned analysts. We examine the model in
Eq. 2.5 separately for our three experience subsamples. The coefficient on infor-
mation asymmetry should be larger for experienced analysts.
It is possible that a resolution of idiosyncratic uncertainty makes deviation
insensitive to both analyst experience and information asymmetry. Regulations
such as Reg. FD were enacted to reduce the uncertainty around companies
(De Jong and Apilado 2008). If Reg. FD reduced firm-specific uncertainty more
than common uncertainty, then the coefficients on Experience, Experience squared,
and information asymmetry should be smaller after Reg. FD. In other words, if Reg.
FD was effective in “leveling the playing field” for analysts with regard to prefer-
ential treatment of some analysts over others, the implications of this model should
no longer exist.
Extant evidence suggests that by prohibiting exclusive private communication of
pertinent information, Reg. FD would cause overall earnings uncertainty to decline
(Baily et al. 2003). In Eqs. 2.4 and 2.5, we simply control for any effect that Reg.
FD might have. By using interactive terms, however, we can explore this relation-
ship more fully by offering a comparison of the relationship between Experience/
Information asymmetry and Deviation both with and without Reg. FD. If most of
2 Experience, Information Asymmetry, and Rational Forecast Bias 75
2.4 Data
Consistent with Brown and Sivakumar (2003) and Doyle et al. (2003), we define
earnings as the First Call reported actual earnings per share.13 Our sample includes
companies based in the United States with at least two analysts, consists of 266,708
analyst-firm-quarter forecasts from 1995 to 2007, and includes all analysts’ revi-
sions. Variables are winsorized at the 1 % level to ensure that results are not biased
by outliers.
First Call data is well suited for examining analyst revisions because most of the
analysts’ estimates in the data have the date that they were published by the broker.
These revisions are reflected daily, which aids in understanding the changes in
deviation based on changes in the information environment. One limitation with
First Call data is that it identifies only brokerage houses, not individual analysts.
Following Leone and Wu (2007), we make the assumption that for each firm quarter
there is only one analyst in each brokerage house following the firm. It is notewor-
thy that this biases our study against finding a nonlinear impact for analyst’s
experience.
Panel A of Table 2.1 shows the summary statistics of the variables used in our
analysis. The average Deviation from the consensus is 5.16¢. However, there is
wide dispersion, from 54.67¢ to 112.50¢, as compared to an average Forecast
error of 3.79¢, with a range of 131–72¢. The analysts in our sample have on
average 12.71 quarters of Experience (the most experienced analyst has 13 years
experience, and the least experienced analyst has no prior experience following the
firm). Analysts revise their estimates 2.49 times per quarter. The companies they
13
As a robustness test, we use I/B/E/S data. Results may be found in Appendix 2.
76
Reg. FD is in effect and 0 otherwise. NumRevisions is the total number of revisions the analyst made during the quarter for the quarter-end earnings. Accruals
are income before extraordinary items minus cash flow from operations, where cash flow from operations is defined as net cash flow from operating activities
minus extraordinary items and discontinued operations. Intangible assets are intangible assets to total assets. Return std. dev. is the standard deviation of the
12-month returns. Panel A shows the univariate statistics and Panel B shows the pairwise correlations
Asterisks in Panel A represent statistical significance relative to zero
Bolded numbers in Panel B represent 5 % or 1 % significance
77
78 A. Knill et al.
Table 2.2 presents the results of our estimation from Eq. 2.4, which examines how
analyst experience affects analyst deviation from the consensus. The coefficients on
our control variables all exhibit the expected signs. There is a negative relation
between the horizon value (Same quarter) and Deviation, which suggests that
analysts are more likely to be optimistic early in the period than late in the period.
Supporting the contentions of Carter and Manaster (1990), there is a negative
relationship between the brokerage characteristics – Reputation and Size – with
Deviation. There is also a negative relationship between Reg. FD and Deviation,
suggesting that the enactment of Reg. FD and its mandatory indiscriminant infor-
mation revelation have made forecasting easier. NumRevisions is positively related
to Deviation; revisions are made when valuable information content is received
(Givoly and Lakonishok 1979), suggesting that private information is paid for
through incremental deviation of their forecasts over time. We find that higher
Accruals lead to less deviation from the consensus. Firms who actively manage
earnings make forecasting easier; thus analysts do not have to purchase
private information. The positive sign on Intangible assets is expected as more
intangible assets make forecasting more difficult, necessitating the procurement of
private information. Finally, as the standard deviation of Returns increases, future
firm performance is more difficult to predict.
Turning to our variables of interest, we find that there is a highly significant
positive relationship between Experience and Deviation, as evidenced by a 0.829¢
increase in deviation for every additional unit of experience (specification 1) and
a 0.581¢ increase in deviation for every additional unit of relative experience
(specification 2).14 The results in specifications (3) and (4) confirm our contention
that this relationship is not linear. The squared Experience variable
14
Inasmuch as the Experience variable is transformed using the natural logarithm, one unit of
experience is approximately equal to two quarters of experience. For tractability, we refer to this as
a unit in the empirical results.
2
Intangible assets 1.370*** 1.244*** 1.237*** 1.264*** 2.718*** 2.718*** 0.097 0.146 1.415*** 1.411*** 1.974*** 2.054***
[0.232] [0.232] [0.232] [0.231] [0.393] [0.393] [0.241] [0.242] [0.161] [0.161] [0.440] [0.440]
Return std. dev. 19.384*** 17.800*** 19.063*** 17.700*** 26.067*** 26.308*** 7.387*** 7.448*** 7.944*** 7.265*** 17.102*** 15.640***
[0.578] [0.571] [0.578] [0.571] [0.975] [0.976] [0.612] [0.612] [0.448] [0.441] [1.003] [0.991]
Constant 6.239*** 8.269*** 6.501*** 6.913*** 2.193 4.345* 6.268*** 6.633*** 6.286*** 6.423*** 13.895*** 14.185***
[1.439] [1.434] [1.438] [1.437] [2.682] [2.621] [1.467] [1.469] [0.975] [0.974] [3.041] [3.038]
Firm FE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Industry FE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
(continued)
79
80
Table 2.2 presents the results of our estimation from Eq. 2.4. Experience is the natural log of the number of quarters the analyst has followed the firm for which
forecast error is calculated. Relative experience is the analyst experience scaled by the average analyst experience (same industry). Information asymmetry
(IA) is proxied based on the forecast dispersion – the standard deviation of estimates. Same quarter is a dummy variable equal to one if the estimate is in the
same quarter as the actual and zero otherwise. Same quarter is orthogonalized (on NumRevisions) to ensure that multicollinearity is not a problem between
these two variables. NumRevisions is the total number of revisions the analyst made during the quarter for the quarter-end earnings. Reg. FD is an indicator
variable that takes on a value of 1 if Reg. FD is in effect and 0 otherwise. X is a vector of firm-specific variables including Broker reputation, Broker size,
Accruals, Intang. assets, and Return std dev. Brokerage reputation is a ranking of brokerage reputation, where 0 is the worst and 9 is the best. Brokerage size is
the natural log of the number of companies per quarter a Brokerage house follows. Brokerage reputation is orthogonalized (on brokerage size) to ensure that
multicollinearity is not a problem between these two variables. Accruals are the accrued revenue/liabilities utilized for earnings smoothing. Intang. assets are
the covered firm’s intangible assets value relative to its total assets. Return std. dev. is the standard deviation of the covered firm’s return. I is a vector of
one-digit SIC industry dummies. T is a vector of time dummies. A formal fixed effects treatment around analysts is employed. Standard errors are reported in
brackets
* **
, , and *** indicate significance levels of 10 %, 5 %, and 1 %, respectively (two-tailed test). In columns 1–4 we use the full sample. In columns 5–8, we use
the median value of horizon value to divide the firms into early (columns 5–6) and late estimates (7–8) groups to perform the analysis. In columns 9–12, we use
the median value of forecast dispersion to divide the firms into low information asymmetry (columns 9–10) and high information asymmetry (11–12) groups to
perform the analysis
A. Knill et al.
2 Experience, Information Asymmetry, and Rational Forecast Bias 81
15
We are grateful to an anonymous referee for this point.
82 A. Knill et al.
Table 2.3 Time impact of Reg. FD on the relationship between experience/relative experience and deviation
Dependent variable ¼ DFC (deviation from consensus)
1 2 3 4 5 6 7 8
Experience 0.302*** 2.116*** 1.528*** 2.764***
[0.050] [0.157] [0.088] [0.271]
Experience2 0.485*** 0.400***
[0.040] [0.082]
Reg. FD * experience 0.962*** 0.026
[0.100] [0.326]
Reg. FD * experience2 0.168*
[0.092]
Relative experience 0.072 0.511* 0.822*** 7.019***
[0.082] [0.269] [0.149] [0.503]
Relative experience2 0.274** 3.128***
[0.120] [0.243]
Reg. FD * relative experience 0.331* 3.715***
Experience, Information Asymmetry, and Rational Forecast Bias
[0.172] [0.586]
Reg. FD * relative experience2 1.831***
[0.277]
Reg. FD 0.319 2.308*** 1.398*** 0.975***
[0.318] [0.288] [0.362] [0.358]
(continued)
83
84
analytical abilities, which would suggest that perhaps they have to purchase less
information to achieve precision, and valuation relationships, which allow them to
purchase information at reduced “costs” (i.e., deviation from consensus). To ensure
that the marginal effects are significantly different, we perform Wald tests to
confirm that the differences between the coefficients are statistically significant.
As can be seen at the bottom of Panel A, the differences are statistically significant
in every case.
In Panel B, we examine experience-based subsamples to see more intricately
how experience affects the impact of Information asymmetry on Deviation. Results
confirm that Experience affects the link between Information asymmetry and
Deviation. The moderately experienced analyst is the most affected by an increase
in information asymmetry, regardless of the definition of information asymmetry. It
is interesting to note that the results for Analyst coverage and Forecast dispersion
are much stronger than Relative firm size. We use a Hausman test of statistical
significance and find that the experience estimations are all statistically different.
These empirical results nicely complement the theoretical predictions of our
model. Taken collectively, the results suggest that analysts’ rational forecast devi-
ation from the consensus is a result of analysts maximizing their objective function
of minimizing error while taking into consideration the information environment in
which they operate.
To provide evidence that the deviation from the consensus discussed in the rest
of the paper is indeed rationale, we test whether an analyst posting an estimate with
the optimal deviation from consensus, empirically represented by the fitted value of
either Eqs. 2.4 or 2.5, achieves a lower forecast error (defined as estimate minus
actual earnings). Here, a negative association between fitted Deviation from con-
sensus (DFC*) and Forecast error (FE) is indicative of optimal forecasting. Chen
and Jiang (2006) argue that a positive (negative) association between observed
Deviation from consensus (DFC) and Forecast error (FE) is indicative of an analyst
overweighting (underweighting) her own assessment. This is because if analysts
follow a Bayesian updating rule and minimize squared forecast error, there should
be no link between observed deviation from forecast and forecast error. They find
that analysts, on average, overweigh their own belief, and thus the link between
observed deviation from consensus and forecast error is positive. In this paper, we
extend Chen and Jiang’s model to allow for rational biasing of forecasts, similar to
Lim (2001). As such, we find that there should be a concave relationship between
Experience and Deviation from consensus. If our theory is correct and if our
empirical model of deviation from consensus truly captures the essence of rational
forecasting that occurs among analysts, the fitted Deviation from consensus should
be devoid of analysts’ overconfidence about their own information processing. In
fact, if fitted Deviation reflects the tendency toward more rational forecasting, we
should find a negative relationship between fitted Deviation and Forecast error.
Looking to the results, which are found in Table 2.5, Panel A, we see that this is
indeed what we find. In all specifications, the fitted Deviation (DFC) is negatively
related with Forecast error. All specifications lead to a significant decline in the
forecast error of analyst estimation. Specifically, we find that a 1¢ increase in bias
90
Return std. dev. 10.012*** 9.187*** 2.975*** 10.716*** 9.805*** 0.384 18.238*** 17.662*** 5.043*** 19.153*** 17.969*** 1.710***
[0.666] [0.657] [0.664] [2.079] [2.063] [35.868] [0.579] [0.570] [0.568] [0.473] [0.472] [0.436]
Constant 10.593*** 11.004*** 9.734*** 8.532 10.407* 57.535 3.465*** 4.612*** 2.021* 2.682*** 0.828 2.010***
[1.409] [1.405] [1.392] [5.760] [5.770] [165.519] [1.258] [1.267] [1.219] [0.693] [0.694] [0.619]
Firm FE Yes Yes Yes Yes Yes Yes No No No Yes Yes Yes
Industry FE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Time FE Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes Yes
Observations 66,550 66,550 62,014 13,181 13,181 146 266,708 266,708 253,673 295,717 295,717 295,717
# Brokerage houses{ 130 130 130 119 119 22 130 130 130 7708 7708 7708
Model R2 0.03 0.03 0.07 0.06 0.06 0.24 0.06 0.06 0.16 0.07 0.07 0.23
Analyst information comes from First Call (except in Specifications (10)–(12)). Company information comes from Compustat/CRSP. Experience is the
natural log of the number of quarters the analyst has followed the firm for which forecast error is calculated. Relative experience is the analyst experience
scaled by the average analyst experience (same industry). Same quarter is a dummy variable equal to one if the estimate is in the same quarter as the actual and
zero otherwise. Same quarter is orthogonalized (on NumRevisions) to ensure that multicollinearity is not a problem between these two variables. Forecast
dispersion is the standard deviation of estimates. X is a vector of firm-specific variables including Broker reputation, Broker size, Accruals, Intang. assets, and
Return std dev. Brokerage reputation is a ranking of brokerage reputation, where 0 is the worst and nine is the best. Brokerage size is the natural log of the
number of companies per quarter a brokerage house follows. Brokerage reputation is orthogonalized (on brokerage size) to ensure that multicollinearity is not
a problem between these two variables. Accruals are the accrued revenue/liabilities utilized for earnings smoothing. Intang. assets are the covered firm’s
intangible assets value relative to its total assets. Return standard deviation is the standard deviation of the covered firm’s return. I is a vector of one-digit SIC
industry dummies. T is a vector of time dummies. Specifications (1)–(3) include only original estimates (i.e., no revisions). Specifications (4)–(6) include only
firms with one analyst following the firm. Specifications (7)–(9) use random effects. Specifications (10)–(12) use I/B/E/S data. A formal fixed effects treatment
around analysts is employed unless otherwise noted. Standard errors are reported in brackets.
* **
, , and *** indicate significance levels of 10 %, 5 %, and 1 %, respectively
{
In sample derived from I/B/E/S data, this is actually the number of analysts
Experience, Information Asymmetry, and Rational Forecast Bias
93
94
2.6 Robustness
2.7 Conclusions
To understand the exact role experience and information asymmetry play in forming
a rational deviation, we offer an integrated framework in which an optimum forecast
is derived from weighing public and private information, where the resulting
forecast bias (i.e., deviation from the consensus) improves accuracy through the
acquisition of private information. We study how information asymmetry and
analyst experience affect the efficacy of the feedback from bias on private informa-
tion acquisition. Our construct enables a realistic optimization, where rational
forecast bias emerges as an outcome of a trade-off between public and private
information while minimizing forecast squared error. We show both analytically
and empirically that this trade-off and the resulting rational deviation are determined
by experience and information asymmetry. While the optimal bias and information
asymmetry are monotonically positively related, the optimal bias and experience are
concavely linked. Moderately experienced analysts find it optimal to purchase
private information with a positively biased deviation from the consensus. The
superior analytical skills and access to private information (at a lower cost) of highly
seasoned analysts leads them to optimally rely on private information.
We use Reg. FD as an experiment to further illuminate the relationship between
experience and forecast bias and to make a distinction between access to private
information and analytical expertise. We find that, as we would expect, the biases of
experienced analysts, who weigh private information more heavily, are affected
more. The extent to which information asymmetry and analyst experience play
a role in determining the analyst’s bias is directly affected by the dispersion of the
common and idiosyncratic signals of all analysts as well as the extent of their
correlation with each other.
Our results may help to paint a clearer picture of the types of information
produced by analysts. Institutions may be able to use these results to design
continuing education programs and to make hiring/firing decisions. Our results
can also help to explain why and how changes in the information environment
caused by regulatory changes (e.g., Reg. FD), index inclusions, and public disclo-
sures affect the role of analysts as information processors.
2 Experience, Information Asymmetry, and Rational Forecast Bias 97
Appendix 1: Proofs
Given that the analyst’s forecast is a weighted average of the analyst’s uncondi-
tional estimate and the consensus, F ¼ wE + (1 – w)Ec, the objective function can be
expressed as
h i
min w2 ðt0 þ tðbÞÞ2 þ ð1 wÞ2 t2 1 1
c þ 2rwð1 wÞðt0 þ tðbÞÞ tc (2.6)
wjb
Proof of Proposition 1
Clearly, since the denominator of ∂w/∂t0 is positive, then the sign is only
a function of the numerator. This implies that the sign changes when the numerator,
2tc(t0 + t(b)) 2r(t0 + t(b))2 rt2c , is at maximum. To find the maximum,
we solve for t0 that satisfies the first-order conditions of the numerator. The first-
order condition yields tc 2r(t0 + t(b)) 0. Thus, at optimal weight
t0 + t(b) ¼ 0.5r1tc.
98 A. Knill et al.
Proof of Proposition 2
h i
∂t
∂w 2t2c ðt0 þ tðbÞÞ 2rtc ðt0 þ tðbÞÞ2 rt3c ∂b
¼ h i2 (2.9)
∂b
ðt0 þ tðbÞÞ2 þ t2c 2rtc ðt0 þ tðbÞÞ
Clearly, since the denominator of ∂w/∂b is positive, then the sign is only
a function of the numerator. This implies (1) that since ∂t/∂b is positive, then the
optimal weight would be monotonically increasing with ∂t/∂b or information
asymmetry, and (2) that the optimal weight is nonlinearly, concavely related to
private information precision. Since the first term in the numerator is a quadratic
function of analyst’s own precision, the maximum in the function is the point at
which the numerator changes sign. This point, however, is exactly the same point at
which ∂w/∂t0 maximizes. For biases at which t0 + t(b) falls below 0.5r1tc., then
so long as bias increases so does the optimal weight.
References
Agrawal, A., & Chen, M. (2012). Analyst conflicts and research quality. Quarterly Journal of
Finance, 2, 179–218.
Baily, W., Li, H., Mao, C., & Zhong, R. (2003). Regulation fair disclosure and earnings informa-
tion: Market, analysts, and corporate response. Journal of Finance, 58, 2487–2514.
Bannister, J., & Newman, H. (1996). Accrual usage to manage earnings toward financial analysts’
forecasts. Review of Quantitative Finance and Accounting, 7, 259–278.
Barber, B., Lehavy, R., Trueman, B. (2000). Are all brokerage houses created equal? Testing
for systematic differences in the performance of brokerage house stock recommendations.
University of California at Davis and University of California at Berkeley (unpublished),
March.
Bernhardt, D., Campello, M., & Kutsoati, E. (2006). Who herds? Journal of Financial Economics,
80, 657–675.
Beyer, A. (2008). Financial analysts’ forecast revisions and managers’ reporting behavior. Journal
of Accounting and Economics, 46, 334–348.
Bhattacharya, N. (2001). Investors’ trade size and trading responses around earnings announce-
ments: An empirical investigation. Accounting Review, 76, 221–244.
Boni, L., & Womack, K. (2006). Analysts, industries, and price momentum. Journal of Financial
Quantitative Analysis, 41, 85–109.
Brown, L., & Sivakumar, K. (2003). Comparing the quality of two operating income measures.
Review of Accounting Studies, 4, 561–572.
Brown, S., Hillegeist, S., & Lo, K. (2004). Conference calls and information asymmetry. Journal
of Accounting and Economics, 37, 343–366.
Carey, M., Post, M., & Sharp, S. (1998). Does corporate lending by banks and finance companies
differ? Evidence on specialization in private debt contracting. Journal of Finance, 53, 845–878.
Carter, R., & Manaster, S. (1990). Initial public offerings and underwriter reputation. Journal of
Finance, 45, 1045–1068.
Carter, R., Dark, F., & Singh, A. (1998). Underwriter reputation, initial returns, and the long-run
performance of IPO stocks. Journal of Finance, 53, 285–311.
Chen, Q., & Jiang, W. (2006). Analysts’ weighting of private and public information. Review of
Financial Studies, 19, 319–355.
Clement, M. (1999). Analyst forecast accuracy: Do ability, resources, and portfolio complexity
matter? Journal of Accounting and Economics, 27, 285–303.
Clement, M., & Tse, S. (2005). Financial analyst characteristics and herding behavior in forecast-
ing. Journal of Finance, 60, 307–341.
Conrad, J., Cornell, B., Landsman, W., & Rountree, B. (2006). How do analyst recommendations
respond to major news? Journal of Financial and Quantitative Analysis, 41, 25–49.
De Jong, P., & Apilado, V. (2008). The changing relationship between earnings expectations and
earnings for value and growth stocks during Reg FD. Journal of Banking and Finance, 33,
435–442.
Dechow, P., Kothari, S., & Watts, R. (1998). The relation between earnings and cash flows.
Journal of Accounting and Economics, 25, 133–168.
Doyle, J., Lundholm, R., & Soliman, M. (2003). The predictive value of expenses excluded from
pro forma earnings. Review of Accounting Studies, 8, 145–174.
Fama, E., & French, K. (1997). Industry costs of equity. Journal of Financial Economics, 43,
153–193.
Givoly, D., & Lakonishok, J. (1979). The information content of financial analysts’ forecasts of
earnings. Journal of Accounting and Economics, 2, 165–185.
Gu, Z., & Xue, J. (2007). Do analysts overreact to extreme good news in earnings? Review of
Quantitative Finance and Accounting, 29, 415–431.
Han, B., Manry, D., & Shaw, W. (2001). Improving the precision of analysts’ earnings forecasts by
adjusting for predictable bias. Review of Quantitative Finance and Accounting, 17, 81–98.
100 A. Knill et al.
Hirschey, M., & Richardson, V. (2004). Are scientific indicators of patent quality useful to
investors? Journal of Empirical Finance, 11, 91–107.
Hong, H., Kubik, J., & Salomon, A. (2000). Security analysts’ career concerns and herding of
earnings forecasts. RAND Journal of Economics, 31, 121–144.
Hsu, D., & Chiao, C. (2010). Relative accuracy of analysts’ earnings forecasts over time:
A Markov chain analysis. Review of Quantitative Finance and Accounting, 37, 477–507.
Irani, A. (2004). The effect of regulation fair disclosure on the relevance of conference calls to
financial analysts. Review of Quantitative Finance and Accounting, 22, 15–28.
Krishnaswami, S., & Subramaniam, V. (1998). Information asymmetry, valuation, and the corpo-
rate spin-off decision. Journal of Financial Economics, 53, 73–112.
Kwon, S. (2002). Financial analysts’ forecast accuracy and dispersion: High-tech versus low-tech
stocks. Review of Quantitative Finance and Accounting, 19, 65–91.
Leone, A., & Wu, J. (2007). What does it take to become a superstar? Evidence from institutional
investor rankings of financial analysts. Simon School of Business Working Paper No. FR 02-12.
Available at SSRN: http://ssrn.com/abstract=313594 or http://dx.doi.org/10.2139/ssrn.313594.
Lim, T. (2001). Rationality and analysts; forecast deviation. Journal of Finance, 56, 369–385.
Lin, B., & Yang, R. (2010). Does regulation fair disclosure affect analysts’ forecast performance?
The case of restructuring firms. Review of Quantitative Finance and Accounting, 38, 495–517.
Loughran, T., & Ritter, J. (2004). Why has IPO underpricing changed over time? Financial
Management, 33(3), 5–37.
Mest, D., & Plummer, E. (2003). Analysts’ rationality and forecast bias: Evidence from analysts’
sales forecasts. Review of Quantitative Finance and Accounting, 21, 103–122.
Mikhail, M., Walther, D., & Willis, R. (1997). Do security analysts improve their performance
with experience? Journal of Accounting Research, 35, 131–157.
Nutt, S., Easterwood, J., & Easterwood, C. (1999). New evidence on serial correlation in analyst
forecast errors. Financial Management, 28, 106–117.
Petersen, M., & Rajan, R. (1994). The benefits of lending relationships: Evidence from small
business data. Journal of Finance, 49, 3–37.
Ramnath, S. (2002). Investor and analyst reactions to earnings announcements of related firms: An
empirical analysis. Journal of Accounting Research, 40, 1351–1376.
Stangeland, D., & Zheng, S. (2007). IPO underpricing, firm quality, and analyst forecasts.
Financial Management, 36, 1–20.
Sufi, A. (2007). Information asymmetry and financing arrangements: Evidence from syndicated
loans. Journal of Finance, 62, 629–668.
Thomas, S. (2002). Firm diversification and asymmetric information: Evidence from analysts’
forecasts and earnings announcements. Journal of Financial Economics, 64, 373–396.
Trueman, B. (1990). Theories of earnings-announcement timing. Journal of Accounting and
Economics, 13, 285–301.
Waymire, G. (1986). Additional evidence on the accuracy of analyst forecasts before and after
voluntary managed earnings forecasts. Accounting Review, 61, 129–141.
Winkler, R. L. (1972). An introduction to Bayesian inference and decision. New York: Holt,
Rinehart and Winston.
Zhang, F. (2006). Information uncertainty and stock returns. Journal of Finance, 61, 105–137.
Zhou, T., & Lai, R. (2009). Herding and information based trading. Journal of Empirical Finance,
16, 388–393.
Zitzewitz, E. (2002). Regulation fair disclosure and the private information of analysts. Available
at SSRN: http://ssrn.com/abstract=305219 or http://dx.doi.org/10.2139/ssrn.305219.
An Appraisal of Modeling Dimensions for
Performance Appraisal of Global 3
Mutual Funds
Contents
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
3.2 Performance Evaluation Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 102
3.3 A Review on Various Models for Performance Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
3.3.1 Jensen Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
3.3.2 Fama Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 111
3.3.3 Treynor and Mazuy Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
3.3.4 Statman Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
3.3.5 Choi Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 112
3.3.6 Elango Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
3.3.7 Chang, Hung, and Lee Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
3.3.8 MM Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
3.3.9 Meijun Qian’s Stage Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 114
3.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 116
Abstract
A number of studies have been conducted to examine investment performance of
mutual funds of the developed capital markets. Grinblatt and Titman (1989,
1994) found that small mutual funds perform better than large ones and that
performance is negatively correlated to management fees but not to fund size or
expenses. Hendricks, Patel, and Zeckhauser (1993), Goetzmann and Ibbotson
(1994), and Brown and Goetzmann (1995) present evidence of persistence
in mutual fund performance. Grinblatt and Titman (1992) and Elton, Gruber,
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 101
DOI 10.1007/978-1-4614-7750-1_3,
# Springer Science+Business Media New York 2015
102 G.V. Satya Sekhar
and Blake (Journal of Financial Economics 42:397–421, 1996) show that past
performance is a good predictor of future performance. Blake, Elton, and
Grubber (1993), Detzler (1999), and Philpot, Hearth, Rimbey, and Schulman
(1998) find that performance is negatively correlated to fund expense, and
that past performance does not predict future performance. However, Philpot,
Hearth, and Rimbey (2000) provide evidence of short-term performance persis-
tence in high-yield bond mutual funds. In their studies of money market mutual
funds, Domian and Reichenstein (1998) find that the expense ratio is the most
important factor in explaining net return differences. Christoffersen (2001) shows
that fee waivers matter to performance. Smith and Tito (1969) conducted a study
into 38 funds for 1958–1967 and obtained similar results. Treyner (1965)
advocated the use of beta coefficient instead of the total risk.
Keywords
Financial modeling • Mutual funds • Performance appraisal • Global investments •
Evaluation of funds • Portfolio management • Systematic risk • Unsystematic
risk • Risk-adjusted performance • Prediction of price movements
3.1 Introduction
However, they mostly failed to outperform the world index. The sample consisted
of 19 US-based international funds, and the Sharpe measure was used to assess
excess returns.
Barua and Varma (1993b) have examined the relationship between the NAV and
the market price on Mastershares. They conclude that market prices are far more
volatile than what can be justified by volatility of NAVs. The prices also show
a mean reverting behavior, thus perhaps providing an opportunity for discovering
a trading rule to make abnormal profits in the market. Such a rule would basically
imply buying Mastershares whenever the discount from NAV was quite high and
selling Mastershares whenever the discount was low.
Droms and Walker (1994) used a cross-sectional/time-series regression meth-
odology. Four funds were examined over 20 years (1971–1990), and 30 funds were
analyzed for a 6-year period (1985–1990). The funds were compared to the
Standard and Poor’s 500 Index, the Morgan Stanley Europe, Australia, and Far
East Index (EAFE) which proxies non-US stock markets, and the World Index.
Applying the Jensen, Sharpe, and Treynor indices of performance, they found
that international funds have generally underperformed the US market and the
international market. Additionally, their results indicated that portfolio turnover,
expense ratios, asset size, load status, and fund size are unrelated to fund
performance.
Bauman and Miller (1995) studied the persistence of pension and investment
fund performance by type of investment organization and investment style. They
employed a quartile ranking technique, because they noted that “investors pay
particular attention to consultants’ and financial periodicals’ investment perfor-
mance rankings of mutual funds and pension funds.” They found that portfolios
managed by investment advisors showed more consistent performance (measured
by quartile rankings) over market cycles and that funds managed by banks and
insurance companies showed the least consistency. They suggest that this result
may be caused by a higher turnover in the decision-making structure in these less
consistent funds. This study controls for the effects of turnover of key decision
makers by restricting the sample to those funds with the same manager for the entire
period of study.
Volkman and Wohar (1995) extend this analysis to examine factors that impact
performance persistence. Their data consists of 322 funds over the period
1980–1989 and shows performance persistence is negatively related to size and
negatively related to levels of management fees.
Elton et al. (1996) examined the predictability of stock mutual funds perfor-
mance based on risk-adjusted future performance. It also demonstrated application
of modern portfolio techniques on past data to improve selection, which permitted
construction of portfolio funds that significantly outperformed a rule based on the
past rank alone. The portfolio so selected was reported to have small, but statisti-
cally significant, positive risk-adjusted returns during a period when mutual funds
in general had negative risk-adjusted returns.
Jayadeve (1996) paper enlightens performance evaluation based on monthly
returns. His paper focuses on performance of two growth-oriented mutual funds
3 An Appraisal of Modeling Dimensions 105
and beta. The data is obtained from value research. The author finds inconsistencies
between investment style/objective classification and the return obtained by
the fund.
Coates and Hubbard (2007) reviewed the structure, performance, and dynamics
of the mutual fund industry and showed that they are consistent with competition. It
was also found that concentration and barriers to entry are low, actual entry is
common and continuous, pricing exhibits no dominant long-term trend, and market
shares fluctuate significantly. Their study also focused on “effects of competition on
fee” and “pricing anomalies.” They suggested legal interventions are necessary in
setting fee in mutual funds of United States.
Subha and Bharati’s (2007) study is carried out for open-ended mutual fund
schemes and 51 schemes are selected by convenient sampling method. NAVs are
taken for a period of 1 year from 1 October 2004 to 30 September 2005. Out of the
51 funds, as many as 18 schemes earned higher returns than the market return. The
remaining 33 funds however generated lower returns than the market.
Sondhi’s (2007) study analyzes the financial performance of 36 diversified
equity mutual funds in India, in terms of rates of return, comparison with risk-
free return, benchmark comparison, and risk-adjusted returns of diversified equity
funds. Fund size, ownership pattern of AMC, and type of fund are the main factors
considered in this study. The study reveals that private sector is dominating public
sector.
Cheng-Ru Wu et al.’s (2008) study adopts modified Delphi method and the
analytical hierarchy process to design an assessment method for evaluating mutual
fund performance. The most important criteria for mutual fund performance should
be “mutual fund style” followed by “market investment environment.” This result
indicates investor’s focus when they evaluate the mutual fund performance.
Eleni Thanou’s (2008) study examines the risk-adjusted overall performance of
17 Greek Equity Mutual Funds between the years 1997 and 2005. The study
evaluated performance of each fund based on the CAPM performance methodol-
ogy, calculating the Treynor and Sharpe Indexes for the 9-year period as well as for
three sub-periods displaying different market characteristics. The results indicated
that the majority of the funds under examination followed closely the market,
achieved overall satisfactory diversification, and some consistently outperformed
the market, while the results in market timing are mixed, with most funds
displaying negative market timing capabilities.
Kajshmi et al. (2008) studied a sample of schemes in the 8-year period. This study
considers performance evaluation and is restricted to the schemes launched in the
year 1993 when the industry was thrown open to private sector under the regulated
environment by passing the SEBI (Mutual Funds) Regulations 1993. The perfor-
mance of the sample schemes were in line with that of the market as evident from the
positive beta values. All the sample schemes were not well diversified as depicted by
the differences in the Jensen alpha and Sharpe’s differential return.
Massimo Masa and Lei Zhang (2008) found the importance of organizational
structure on Asset Management Company of mutual fund. Their study found that
more hierarchical structures invest less in firms located close to them and deliver
3 An Appraisal of Modeling Dimensions 109
and public sector. The levels of competition and their trends have been obtained for
the periods March 2003–March 2009. This study found overall mutual fund indus-
try is facing a high competitive environment. An increasing trend of competition
was observed within bank institution, private sector foreign, and private sector joint
venture mutual funds.
Mohit Gupta and Aggarwal’s (2009) study focused on the portfolio creation and
industry concentration of 18 ELSS schemes during April 2006 to April 2007.
Mutual fund industry concentration was the variable used in classification or cluster
creation. This exercise was repeated each month for the period under study. Finally
portfolio performance was compared with index fund, portfolio of three randomly
picked funds of the previous month, and the return and risk parameters of ELSS
category as a whole.
Talat Afza and Ali Rauf’s (2009) study aims to provide guidelines to the
managers of open-ended Pakistani mutual funds and benefit small investors by
pointing out the significant variables influencing the fund performance. An effort
has been made to measure the fund performance by using Sharpe ratio with the help
of pooled time-series and cross-sectional data and focusing on different fund
attributes such as fund size, expenses, age, turnover, loads, and liquidity. The
quarterly sample data are collected for all the open-ended mutual funds listed on
Mutual Fund Association of Pakistan (MUFAP), for the years 1999–2006. The
results indicate that among various funds attributes are: lagged return, liquidity and
had significant impact on fund performance.
Amar Ranu and Depali Ranu (2010) critically examined the performance of
equity funds and found out the top 10 best performing funds among 256 equity
mutual fund schemes in this category. They considered three factors for selection:
(a) mutual funds having 5 years of historical performance, (b) fund schemes having
a minimum of Rs.400 crore of assets under management, and (c) funds which have
average return more than 22.47. They found that HDFC TOP 200 (Growth) option
was outperforming among the top 10 best performing equity funds.
Sunil Wahal and Albert Wang (2010) found impact of the entry of new mutual
funds on incumbents using the overlap in their portfolio holdings as a measure of
competitive intensity. Their study revealed that funds with high overlap also
experience quantity competition through lower investor flows, have lower alphas,
and higher attrition rates. These effects only appeared after the late 1990s, at which
point there appears to be endogenous structural shift in the competitive environ-
ment. Their concluding remark is that “the mutual fund market has evolved into one
that displays the hallmark features of a competitive market.”
Sukhwinder Kaur Dhanda et al.’s (2012) study considered the BSE-30 as
a benchmark to study the performance of mutual funds in India. The study period
has been taken from 1 April 2009 to 31 March 2011. The findings of the study
reveal that only three schemes have performed better than benchmark. In the year
2009, HDFC Capital Builder has the top performer. It was 69.18 returns and 26.37
SD and 0.78 beta. HDFC Capital Builder scheme has given the reward for
variability and volatility. HDFC Top 200 Fund and Birla Sun Life Advantage
Funds are on second and third position in terms of return. HDFC Top 200 Fund
3 An Appraisal of Modeling Dimensions 111
has shown better performance than Birla Sun Life Advantage Fund in terms of SD,
beta, Sharpe ratio, and Treynor ratio. Birla Sun Life Advantage Fund has more risk
than the benchmark. Kotak Select Focus Fund has the poorer performer in terms of
risk and return. Except two schemes all other schemes have performed better than
benchmark. Except Kotak Select Focus Fund all other schemes are able to give
reward for variability and volatility.
Given the additional assumption that the capital market is in equilibrium, all three
models yield the following expression for the expected one-period return on any
security (or portfolio) j:
E R j ¼ R F þ bJ ½ E ð R m Þ R F (3.1)
Rp Rf (3.2)
wherein
Rp is actually return on the portfolio, which is monthly average return of fund
and
Rf is monthly average return on treasury bills 91 days.
The overall performance further can be bifurcated into performance due to
selectivity and risk.
112 G.V. Satya Sekhar
Thus,
Rp Rf ¼ Rp Rp bp þ Rp bp Rf (3.3)
Treynor and Mazuy developed a prudent and exclusive model to measure invest-
ment managers’ market timing abilities. This formulation is obtained by adding
squared extra return in the excess return version of the capital asset pricing model as
given below:
2þ
Rpt Rft ¼ a þ bp Rmt Rft þ yp Rmt Rft ept (3.4)
where Rpt is monthly return on the fund, Rft is monthly return on 91 days treasury
bills, Rmt is monthly return on market index, and Ept is error term.
This model involves running a regression with excess investment return as
dependent variable and the excess market return and squared excess market return
as independent variables. The value of coefficient of squared excess return acts as
a measure of market timing abilities that has been tested for significance of using
t-test. Significant and positive values provide evidence in support of the investment
manager’s successful market timing abilities.
Statman measured mutual funds using the following equation (Statman 2000):
eSDAR (excess standard deviation and adjusted return)
¼ Rf þ R p Rf Sm =Sp Rm (3.5)
Elango’s model also compares the performance of public sector funds vs private
sector mutual funds in India. In order to examine the trend in performance of NAV
during the study period, growth rate in NAV was computed. The growth rate was
computed based on the following formula (Elango 2003):
Growth rate: Rg ¼ ðY t Y 0 =Y 0 Þ 100 (3.6)
Equation: Y ¼ A þ b X (3.7)
The pricing model adopted by Jow-Ran Chang, Nao-Wei Hung, and Cheng-Few
Lee is based on competitive equilibrium version of intemporal asset pricing model
derived in Campbell. The dynamic asset pricing model incorporates hedging risk as
well as market. This model uses a log-linear approximation to the budget constraint
to substitute out consumption from a standard intertemporal asset pricing model.
Therefore, asset risk premia are determined by the covariances of asset returns with
the market return and with news about the discounted value of all future market
returns. Formally, the pricing restrictions on asset i imported by the conditional
version of the model are
114 G.V. Satya Sekhar
where
Etri, t + 1, log return on asset; rf, t + 1, log return on riskless asset; Vii denotes Vart (ri,t + 1);
g is the agent’s coefficient of relative risk aversion; Vim denotes Covt (ri, t + 1, rm,t + 1)
and Vih ¼ Covt (ri,t + 1, (Et + 1 Et), _1j ¼ 1 rj rm,t + 1 + j); the parameter,
r ¼ 1 exp(c w); and c w is the mean log consumption to wealth ratio.
This states that the expected excess log return in an asset, adjusted for a Jensen’s
inequality effect, is a weighted average of two covariances: the covariance with the
return from the market portfolio and the covariance with news about future returns
on invested wealth. The intuition in this equation that assets are priced using their
covariances with the return on invested wealth and future returns on invested wealth.
3.3.8 MM Approach
Leah Modigliani and Franco Modigliani are better known as M2 in the investment
literature. This measure is developed adjusting portfolio return. This adjustment is
carried on the uncommitted (cash balances) part of the investment portfolio at the
riskless return so as to enable all portfolio holdings to participate in the return
generation process. This adjustment is needed to bring out the level playing field for
portfolio risk-return and vis-à-vis market return. The effect of this adjustment is
reported below (Modigliani and Modigliani 1997):
M2 ¼ Rp Rm (3.9)
Rp ¼ ðRf ð1 Sdm=SdpÞÞ þ ðRp Sdm=SdpÞ (3.10)
Meijun Qian’s (2009) study reveals about the staleness, which is measured prices
imparts a positive statistical bias and a negative dilution effect on mutual fund
performance. First, evaluating performance with non-synchronous data generates
3 An Appraisal of Modeling Dimensions 115
the dilution effect of arbitrage flows. Thus, assuming that information generated
in time t is not fully incorporated into prices until one period later, the observed
fund return becomes a weighted average of true returns in the current and last
periods:
r t ¼ a þ br mt þ et , (3.11)
where rt denotes the true excess return of the portfolio with mean m and variance s2
and rmt denotes the excess market return with mean mm and variance sm. Both rt and
rmt are i.i.d, and the error term et is independent of rmt. Rt* is the observed excess
return of the portfolio with zero flows, while is the weight on the lagged true
return. That is, the higher the , the staler the prices. Assumedly, arbitrage traders
can earn the return rt*, by trading at the fund’s reported net assets values
(Table 3.1).
3.4 Conclusion
References
Afza, T., & Rauf, A. (2009). Performance Evaluation of Pakistani mutual funds. Pakistan
Economic and Social Review, 47(2), 199–214.
Ammann, M., & Verhofen, M. (2008). The impact of prior performance on the risk-taking of
mutual fund manager. Annals of Finance, (5), 69–90.
Arugaslan, O., Edwards, E., & Samant, A. (2008). Risk-adjusted performance of international
mutual funds. Managerial Finance, 34(1), 5–22.
Baker, M., Litov, L., Wachter, J., & Wurgler, J. (2010). Can mutual fund managers pick stocks?
Evidence from their trades prior to earning announcements. Journal of Financial and Quan-
titative Analysis, 45, 1111–1131.
Baker Kent, H., Haslem, J. A., & Smith, D. M. Performance and Characteristics of Actively
Managed Institutional Equity Mutual Funds [Electronic Copy]. http://ssrn.com/
abstract=1124577
3 An Appraisal of Modeling Dimensions 117
Barua, S. K., & Varma, J. R. (1993). Speculative dynamics: The case of master shares (Advances
in financial planning and forecasting, Vol. 5). Greenwitch: Jai Press.
Bauman, W. S., & Miller, R. E. (1995). Portfolio performance rankings in stock market cycles.
Financial Analysts Journal, 51, 79–87.
Beckmann, D., Lutje, T., & Rebeggiani, L. (2007). Italian asset managers’ behavior: Evidence on
overconfidence, risk taking, and gender (Discussion Paper No. 358). Hannover: Leibniz
Universitat Hannover, Department of Economics, ISSN: 0949-9962.
Bertoni, A., Brtmetti, G., & Cesari, C. (2005). Mutual fund bench marking and market bubbles-a
behavioural approach. Transition Studies Review, 12(1), 36–43.
Bhargava, R., Gallo, J., & Swason, P. T. (2001). The performance, asset allocation and investment
style of international equity manager. Review of Quantitative Finance and Planning, 17,
377–395.
Born, E. K. (1983). International banking in the 19th and 20th centuries. New York: St. Martin’s
Press.
Brown, S. J., & Goetzmann, W. N. (1995). Performance persistence. The Journal of Finance,
50(2), 679–698.
Bu, Q., & Lacey, N. (2008). On understanding mutual fund terminations. Journal of Economics
and Finance, 33, 80–99.
Carhart, M. M. (1997). Persistence in mutual fund performance. Journal of Finance, 52, 57–82.
Chander, R. (2006). Informational efficiency, parameter stationarity and bench mark consis-
tency of investment performance. The ICFAI Journal of Applied Finance. March 2006,
p. 29–45.
Chander, R. (2006b). Investment manager’s market timing abilities: Empirical evidences. The
ICFAI Journal of Applied Finance, 12(8), 15–31.
Chander, S., & Singh, J. (2004). Performance of mutual funds in India: An empirical evidence. The
ICFAI Journal of Applied Finance, 10, 45–63.
Cheng-Ru Wu, Hsin-Yuan Chang, Li-Syuan Wu (2008). A framework of assessable mutual fund
performance. Journal of Modeling in Management, 3(2), 125–139.
Choi, Y. K. (2006). Relative portfolio performance evaluation and incentive structure. Journal of
Business, 79(2), 903–921.
Christofeersen, S. E. (2001). Why do money fund managers voluntarily waive their fee? Journal of
Finance, 74, 117–140.
Coates, J. C., & Hubbard, G. R. (2007). Competition in the mutual fund industry: Evidence and
implications for policy (Discussion Paper No.592). Source: http://ssrn.com/abstract¼1005426
Comer, G. (2006). Hybrid mutual funds and market timing performance. Journal of Business,
79(2), 771–797.
Detzler, L. M. (1999). The performance of global bond mutual funds. Journal of Banking and
Finance, 23, 1195–1217.
Dhanda, S. K., Batra, G. S., & Anjum, B. (2012). Performance evaluation of selected open ended
mutual funds in India. International Journal of Marketing, Financial Services & Management
Research, 1(1), January 2012, ISSN 2277 3622. Online Available at http://indianresearch-
journals.com
Domain, D. L., & Rechenstein, W. (1998). Performance and persistence in money market mutual
fund returns. Financial Services Review, 6, 169–183.
Droms, W. G., & Walker, D. A. (1994). Investment performance of international mutual funds.
Journal of Financial Research, 17, 1–14.
Dunn, P. C., & Theisen, R. D. (1983). How consistently do active managers win? Journal of
Portfolio Management, 9, 47–51.
Elango, R. (2003). Which fund yields more returns? A comparative analysis on the NAV
performance of select public v/s private/foreign open-ended mutual fund schemes in India.
Mutual Funds. The Management Accountant, Vol. 39(4), (2004), p283–290.
Eling, M. (2006). Performance measurement of hedge funds using data envelopment analysis.
Financial Markets and Portfolio Management, 2o, 442–471.
118 G.V. Satya Sekhar
Elton, E. J., Gruber, M. J., & Blake, C. R. (1996). Market timing ability and volatility implied in
investment newsletters’ asset allocation recommendations. Journal of Financial Economics,
42, 397–421.
Eun, C. S., Kolodny, R., & Resnick, B. G. (1991). U.S. based international mutual funds:
A performance evaluation. The Journal of Portfolio Management, 17, 88–94.
Fama, E. F. (1972). Components of investment performance. Journal of Finance, 27, 551–567.
Friend, I., Brown, F. E., Herman, E. S., & Vickers, D. (1962). A study of mutual funds. Washing-
ton, D.C.: U.S. Government Printing Office.
Goetzman, W. N., & Ibbotson, R. G. (1994). Do winners repeat; predicting mutual fund perfor-
mance. The Journal of Portfolio Management, 20, 9–18.
Grinblatt, M., & Titman, S. (1989). Mutual fund performance – An analysis of quarterly holdings.
Journal of Business, 62, 393–416.
Grinblatt, M., & Titman, S. (1992). The persistence of mutual fund performance. Journal of
Finance, 47(5), 1977–1984.
Grinblatt, M., & Titman, S. (1994). Momentum investment strategies – Portfolio performance and
herding: A study of mutual fund behavior. American Economic Review, 85, 1088–1105.
Gupta, M., & Aggarwal, N. (2009). Mutual fund portfolio creation using industry concentration.
The ICFAI Journal of Management Research, Viii(3), 7–20.
Gupta, O. P., & Gupta, A. (2004). Performance evaluation of select Indian mutual fund schemes:
An empirical study. The ICFAI Journal of Applied Finance, 10, 81–97.
Hendricks, D., Patel, J., & Zeckhauser, R. (1993). Hot hands in mutual funds: A short run
persistence of relative performance. Journal of Finance, 48, 93–130.
Jayadeve, M. (1996). Mutual fund performance: An analysis of monthly returns. Finance India,
X(1), 73–84.
Jensen, M. C. (1967). The performance of mutual funds in the period 1945–1964. Journal of
Finance, 23, 389–416.
Jensen, M. C. (1968). The performance of mutual funds in the period 1945–1964. Journal of
Finance, 23, 389–416.
Jow-Ran Chang, Mao-Wei Hung, Cheng-few Lee (2003). An intemporal CAPM approach to evaluate
mutual fund performance. Review of Quantitative Finance and Accounting, 20, 425–433.
Kajshmi, N., Malabika, D., & Murugesan, B. (2008). Performance of the Indian mutual funds;
A study with special reference to growth schemes. Asia-Pacific Business Review, Iv(3), 75–81.
Khurshid, S. M. Z., Rohit, & Sing, G. P. (2009). Level and trends of competition among the mutual
funds in India. Research Journal of Business Management, 3(2), 47–67.
Lintner, J. (1965). The valuation of risk assets and the selection of risky investments in stock
portfolios and capital budgets. The Review of Economics and Statistics, 47(1), 13–37.
Massa, M., & Lei Zhang. (2008). The effects of organizational structure on asset management.
http://ssrn.com/abstract¼1328189
Meijun Qian (2009). Stale prices and the performance evaluation of mutual funds. Journal of
Financial and Quantitative Analysis.
Mc Donald, J. G. (1974). Objectives and performance of mutual funds 1960–1967. Journal of
Financial and Quantitative Analysis, 9, 311–333.
Modigliani, F., & Modigliani, L. (1997). Risk adjusted performance. Journal of Portfolio Man-
agement, 23, 45–54.
Obeid, A. T. (2004). A modified approach for risk-adjusted performance attribution. Financial
Markets and Portfolio Management, 18(3), 285–305.
Philpot, J., Hearth, D., Rimbey, J., & Schulman, C. (1998). Active management, fund size and
bond mutual fund returns. The Financial Review, 33, 115–126.
Ramasamy, B., & Yeung, C. H. M. (2003). Evaluating mutual funds in an emerging market: Factors
that matter to financial advisors. International Journal of Bank Marketing, 21, 122–136.
Ranu, A., & Ranu, D. (2010, January). The best performing equity funds. The Analyst.
Redmand, A. L., Gullett, N. S., & Manakyan, H. (2000). The performance of global and
international mutual funds. Journal of Financial and Strategic Decisions, 13(1), 75–85.
3 An Appraisal of Modeling Dimensions 119
Sadhak, H. (2003). Mutual funds in India: Marketing strategies and investment practices.
New Delhi: Response Books.
Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk.
Journal of Finance, 19, 225–242.
Sharpe, W. F. (1966). Mutual fund performance. Journal of Business, 39, 119–138.
Sidana, G. (2007). Classifying mutual funds in India: Some results from clustering. Indian Journal
of Economics and Business, II(2).
Sisodiya, A. S. (2004, July). Mutual fund industry in India – Coming of age. Chartered Financial
Analyst, pp. 17–22.
Smith, D. M. (2009). The economics of mutual funds. In J. A. Haslem (Ed.), A companion to
mutual funds (Chap. 3). New York: Wiley.
Smith, K. & Tito, D. (1969). Risk-return measures of post-portfolio performance. Journal of
Financial and Quantitative Analysis, 4, 449–471.
Sondhi, H. J. (2007). Financial performance of equity mutual funds in India. New Delhi: Deep &
Deep.
Statman, M. (2000). Socially responsible mutual funds. Financial Analysts Journal, 56, 30–38.
Subha, M. V., & Bharati, J. S. (2007). An empirical study on the performance of select mutual fund
schemes in India. Journal of Contemporary Research in Management, I(1).
Sudhakar, A., & Sasi, K. K. (2005, November). Performance evaluation of mutual funds: A case
study. Southern Economist, pp. 19–23.
Thanou, E. (2008). Mutual fund evaluation during up and down market conditions: The case of
Greek equity mutual funds. International Research Journal of Finance and Economics, (13).
Treynor, J. L. (1965). How to rate management of investment funds. Harvard Business Review, 43,
63–75.
Treynor, J. L., & Mazuy, K. K. (1066). Can mutual funds outguess the markets. Harvard Business
Review, 44, 131–136.
Treynor, J. L., & Mazuy, K. K. (1966). Can mutual funds outguess the markets. Harvard Business
Review, 44, 131–136.
Volkman, D. A., & Wohar, M. E. (1995). Determinants of persistence in relative performance of
mutual funds. Journal of Financial Research, 18, 415–430.
Wahal, S., & Alber (Yan) Wang (2010). Competition among mutual funds. Journal of Financial
Economics. Source: http://ssrn.com/abstract¼1130822
Yuxing Yan (1999). Measuring the timing ability of mutual fund managers. Annals of Operations
Research, 87, 233–243.
Simulation as a Research Tool for
Market Architects 4
Robert A. Schwartz and Bruce W. Weber
Contents
4.1 Studying Market Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
4.2 The Interaction Between Theoretical Modeling, Empirical Analysis,
and Simulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
4.2.1 An Application: Call Auction Trading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123
4.2.2 An Application: The Search for an Equilibrium Price and Quantity . . . . . . . . . . . 125
4.2.3 The Realism of Computer-Generated Data Versus Canned Data . . . . . . . . . . . . . . . 128
4.3 An Equity Market Trading Simulation Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
4.3.1 Informed Orders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
4.3.2 Liquidity Orders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
4.3.3 Momentum Orders . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
4.4 Simulation in Action . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 131
4.4.1 Trading Instructions: Simulation A . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
4.4.2 Trading Instructions: Simulations B-1 and B-2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
4.4.3 Trading Instructions: Simulation C . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
4.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
Modeling Securities Trading . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
Further Details on Simulation Model and Environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 142
Components of the Market Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 121
DOI 10.1007/978-1-4614-7750-1_4,
# Springer Science+Business Media New York 2015
122 R.A. Schwartz and B.W. Weber
Abstract
Financial economists have three primary research tools at their disposal: theoretical
modeling, statistical analysis, and computer simulation. In this chapter, we focus on
using simulation to gain insights into trading and market structure topics, which are
growing in importance for practitioners, policy-makers, and academics. We show
how simulation can be used to gather data on trading decision behavior and to
analyze performance in securities markets under controlled yet competitive condi-
tions. We find that controlled simulations with participants are a flexible and
reliable research tool when it comes to studying issues involving traders and market
architecture. The role of the discrete event simulation model we have developed is
to create a backdrop, or a controlled stochastic environment, for running market
experiments with live subjects. Simulations enable us to gather data on trading
participants’ decision making and to ascertain the ability of incentives and market
structures to influence outcomes. The statistical methods we use include experi-
mental design and careful controls over experimental parameters such as the
instructions given to participants. Furthermore, results are assessed both at the
individual level to understand how participants respond to incentives in a trading
setting and also at the market level to know whether the predicted outcomes are
achieved and how well the market operated.
There are two statistical methods described in the chapter. The first is discrete
event simulation and the model of computer-generated trade order flow that we
describe in Sect. 4.3. To create a realistic, but not ad hoc, market background, we
use draws from a log-normal returns distribution to simulate changes in a stock’s
fundamental value, or P*. The model uses price-dependent Poisson distributions to
generate a realistic flow of computer-generated buy and sell orders whose intensity
and supply-demand balance vary over time. The order flow fluctuations depend on
the difference between the current market price and the P* value. In Sect. 4.4, we
illustrate the second method, which is experimental control to create groupings of
participants in our simulations that have the same trading “assignment.” The result
is the ability to make valid comparisons of traders’ performances in the simulations.
Keywords
Trading simulations • Market microstructure • Order flow models • Random
walk models • Experimental economics • Experimental control
Using good experimental design we can draw statistically valid conclusions from
simulations at both the individual level to understand how participants respond to
incentives in a trading setting and also at the market level to know whether the
predicted outcomes are achieved and how well the market operated. We begin by
considering the interaction between the three tools.
Consider an electronic call auction. At a call, submitted orders are batched together
for simultaneous execution at a single clearing price. For the batching, buy orders
are cumulated from the highest price to the lowest to produce a function that
resembles a downward sloping demand curve, and sell orders are cumulated from
the lowest price to the highest to produce a function that resembles an upward
sloping supply curve. The algorithm generally used for determining the clearing
price at the time that the market is called finds the price which maximizes the
number of shares that execute. In an abstract, theoretical model, the number of
shares that trade is maximized at the price where the downward sloping buy curve
crosses the upward sloping sell curve. Consequently, with continuous order
functions, the clearing price which maximizes the number of shares that trade in
a call auction is uniquely determined by the point where the two curves cross, and
this price, in economics parlance, is an equilibrium value.
124 R.A. Schwartz and B.W. Weber
All told, this call auction procedure has excellent theoretic properties, and one
might expect that designing a well-functioning call would be a straightforward task.
This is not the case, however. In reality, as the saying goes, “the devil is in the details.”
For theoretical modeling, we might for analytic convenience assume a large
enough number of participants so that no one individual has any market power. We
might further assume that the cumulated buy and sell curves are continuous
functions and that the call auction is the only trading facility available. In reality,
however, the buy and sell curves are step functions and some players’ orders will be
large enough to impact a call’s clearing price.
To illustrate, assume an exact match of 40,000 shares to buy and 40,000 shares
to sell at a price of $50 and that, at the next higher price of $50.10, sell orders totaling
50,000 shares and buy orders totaling 30,000 shares exist. A buyer could move the
price up by entering more than 10,000 shares at $50.10 or greater.
Notice also that the real-world buy and sell curves are step functions
(neither price nor quantity is a continuous variable) and thus that the cumulated
buy orders may not exactly match the cumulated sell orders at the price where the
two curves cross. Moreover, many exchange-based call auctions are offered
along with continuous trading in a hybrid environment. These realities of the
marketplace affect participants’ order placement decisions and, consequently,
impact market outcomes for both price and the number of shares that trade.
In response, participants will enter their orders strategically when coming to
the market to trade.
These strategic interactions and the decisions that market participants make
depend on the call auction’s rules of order disclosure (i.e., its transparency) and its
rules of order execution which apply when an exact cross is not obtained. What
guidance do market architects have in dealing with questions such as these other than
their own, hopefully educated, speculation? The questions being raised may be too
context specific for theory to address and, if a new market structure is being consid-
ered, the data required for empirical analysis will not yet exist. This is when
simulation analysis can be used to good advantage.
Regarding transparency, alternative call auction structures include full
transparency (i.e., display the complete set of submitted orders), partial
transparency (e.g., display an indicated clearing price and any order imbalance at
that price), or no transparency at all (i.e., be a dark pool). Regarding the procedure
for dealing with an inexact cross, the alternatives for rationing orders on the
“heavy” side of the market include pro rata execution, the application of time
priority to orders at the clearing price exactly, and the application of time priority
to all orders at the clearing price and better (i.e., to higher priced buys or to lower
priced sells). These and other decisions have been debated in terms of, for
instance, the ability to game or to manipulate an auction, the incentive to enter
large orders, and the incentive to submit orders early in the book building
period before the auction is called. But definitive answers are difficult to
come by. This is when valuable guidance can be (and has been) obtained via the
use of simulation analysis.
4 Simulation as a Research Tool for Market Architects 125
and
where P0 denotes price with respect to the ordinary curve, PR denotes a reservation
price, and N is the number of shares held.
The ordinary curve shows that if, for instance, the price of shares is P01 , the
participant maximizes expected utility by holding N1 shares; if alternatively price is
P02 , the participant maximizes expected utility by holding N2 shares; and so on.
Values given by the reservation curve show that at a quantity N1, the maximum the
participant would pay is PR1 when the alternative is to hold no shares at all; or, at
a quantity N2, the maximum the participant would pay is PR2 when the alternative is to
hold no shares at all; and so on. Identifying the reservation price demand curve
enables us to obtain easily a monetary measure of the gains from trading.
To facilitate the exposition, assume for the moment that the participant initially
holds no shares. Then if, for instance, N1 shares are acquired, we have
Surplus ¼ N1 PR P
where “Surplus” denotes the gains from buying N1 shares and P is the price at which
the N1 shares were bought (note that, for a purchase, we have Surplus > 0 for
P < PR). The participant controls P via the price of his/her order but knows neither the
price at which the order will execute (if price improvement is possible) nor whether
or not the order will, in fact, execute. Because P is not known with certainty, Surplus
is not known with certainty and thus the investor seeks to maximize the expected
value of Surplus. It follows from the derivations cited above that the maximization of
the expected value of Surplus is consistent with the maximization of the expected
utility of the end of the holding period wealth (i.e., at time T2).
With the demand curves to hold shares established, it is straightforward to obtain
linear functions that describe the investor’s propensity to buy and to sell shares in
relation to both ordinary and reservation prices. If the investor’s initial position is
zero shares, the buy curve is the same as the downward sloping demand curve. By
extending the demand curve up and through the price intercept (a) into the negative
quadrant, we see that at prices higher than the intercept, a, the investor will want to
hold a negative number of shares (i.e., establish a short position by selling shares
that he/she does not currently own). By flipping the portion of the demand to hold
curve that is in the negative quadrant into the positive quadrant (and viewing
a negative shareholdings adjustment as a positive sell), we obtain a positively
inclined (vertical) mirror image that is the sell function.
The procedure just described can easily be replicated for any initial share
holdings (either a long or a short position). The individual buy and sell curves
can be aggregated across investors and, following standard economic theory, the
intersection of the aggregate buy curve with the aggregate sell curve establishes
the equilibrium values of share price and the number of shares that trade.
This equilibrium would be achieved if all participants were, simultaneously,
4 Simulation as a Research Tool for Market Architects 127
to submit their complete demand functions to the market, as they presumably would
in a perfect, frictionless environment.
Continue to assume that investors know their continuous, negatively inclined
demand curves to hold shares of the risky asset, but let us now consider how they
might operate in a non-frictionless marketplace where they cannot all simulta-
neously submit their complete and continuous, downward sloping buy functions
and upward sloping sell functions. What specific orders will they send to the
market, and how will these orders interact so as to be turned into trades? Will
equilibrium values for price and quantity be achieved? This highly complex issue
might best be approached by observing participant behavior and, to this end,
simulation can be used as the research tool. Here is how one might go about it.
Let participants compete with each other in a networked, simulated environment.
Each participant is given a demand curve to hold shares of a risky stock, and each is
asked to implement that demand curve by submitting buy or sell orders to the market.
The participants are motivated to place their orders strategically given their demand
curves, the architectural structure of the marketplace, and the objective against which
their performance is assessed – namely, the maximization of expected surplus.
The simulation can be structured as follows. Give all of the participants in
a simulation run the same demand curve,
P ¼ 20 0:5N0
where N0 represents the number of shares initially held. Divide the participants into
two equal groups, A and B, according to the number of shares they are initially
holding, with N0A ¼ 4 for group A players and N0B ¼ 8 for group B players.
Accordingly, the buy curve for each individual in group A is
P ¼ 18 0:5Q
P ¼ 16 þ 0:5Q
where Q is the number of shares bought or sold. The associated reservation curves are
and
From the ordinary (as opposed to the reservation) buy and sell curves, and
recalling that the groups A and B are of equal size, we obtain an equilibrium
price of 17 and an equilibrium quantity traded of 2 (per participant). From the
128 R.A. Schwartz and B.W. Weber
reservation buy and sell curves, we see that if each participant bought or sold two
shares, the surplus for each would be
and
information about a security is reflected in a stock’s market price, then only totally
new, totally unanticipated information can cause the stock’s price to change. But
totally new and thus totally unanticipated information can be either bullish or bearish
with equal probability and thus, with equal probability, can lead to either positive or
negative price changes (returns). Thus, the argument goes, in an informationally efficient
market, returns are not predictable and stock prices follow random walks.
It would be relatively straightforward to structure an equity market simulation
based on machine-driven prices that follow a random walk. One would start
a simulation run with an arbitrarily selected seed price and have that price evolve
as the simulation progresses according to random draws from a (log-normal) returns
distribution with arbitrary variance and a zero mean. Real-world prices do not
evolve in this fashion, however. In a world characterized by trading costs, imperfect
information, and divergent (i.e., nonhomogeneous) expectations based on publicly
available information, prices do not follow simple random walks. Rather, price
changes (returns) in relatively brief intervals of time (e.g., intraday) evolve in
dynamic ways that encompass complex patterns of first-order and higher-order
correlations. Structuring a computer simulation to produce prices that capture this
dynamic property of real-world markets is the objective. In the next session of this
chapter, we set forth the major properties of a machine-driven trading simulation,
TraderEx, which we have formulated so as to achieve this goal.
also been established in the academic microstructure literature that, for an equity
market not to fail, informed traders must interact with liquidity traders, and noise
traders must also be part of the mix.
The second component of the order flow, liquidity orders, is also modeled as
a Poisson arrival process, but with one important difference: at any point of time,
the probability of the newly arriving liquidity order being a buy equals the
probability of its being sell equals 0.5. All liquidity orders are priced, with the
price determined by a draw from a double triangular distribution that is located with
reference to the best posted bid and offer quotes.
A new liquidity order is entered on the book as a limit order if it is a buy
with a price lower than the best posted offer or if it is a sell with a price higher
than the best posted bid. A new liquidity order with a price equal to or more
aggressive than the best posted offer (for a buy) or the best posted bid (for a sell) is
executed immediately as a market order. Liquidity orders can (randomly) cause the
market’s bid and ask quotes to drift away from the equilibrium value, P*. When this
occurs, informed orders that are entered as market orders pull market prices
back towards P*.
4 Simulation as a Research Tool for Market Architects 131
The third component of the order flow is orders entered by noise traders. TraderEx
activity includes just one kind of noise trader – a momentum player – and it operates
as follows: whenever three or more buy orders (or sell orders) arrive sequentially,
the conditional probability is increased that the next arriving order will also be
a buy (or a sell).
As in the microstructure literature, noise traders are needed in the simulation
model to keep participants from too easily identifying price movements that have
been caused by informed orders responding to a change in P*. This is what our
momentum orders achieve. For instance, assume that P* jumps several ticks above
the best posted offer. An accelerated arrival of informed buy orders would be
triggered and prices on the TraderEx book would rise over a sequence of trades,
causing a pattern of positively autocorrelated price movements that can, with
relative ease, be detected by a live participant. But, to obscure this, the momentum
orders create faux price trends that mimic, and therefore obfuscate, the information-
induced trends.
Momentum orders play a further role in the TraderEx simulations. They sys-
tematically cause transaction prices to overshoot P*. Then, as informed orders kick
in, prices in the simulation mean revert back to P*. This mean reversion and its
associated accentuated short-run volatility encourage the placement of limit orders.
This is because overshooting causes limit orders to execute, and limit order placers
profit when price then mean reverts. To see this, assume that the stock is currently
trading at the $23.00 level and that P* jumps from $23.00 to $24.00. As price starts
to tick up to $24.00, momentum orders join the march and carry price beyond
$24.00–$24.20. Assume a limit order to sell is on the book at $24.20 and that it
executes. The limit order placer then benefits from having sold at $24.20 when the
momentum move ends and when a P* of $24.00 exerts its influence and price mean
reverts back towards $24.00.
Since the earliest version of TraderEx was developed in 1995, we have run
hundreds of “live market” simulations with students in our trading and market
microstructure electives and with executive education participants. In addition, we
have developed training modules on trading for new hires at a number of global
banks. We have also run controlled experimental economics studies of trading
decision making and alternative market structures (Schwartz and Weber 1997).
To illustrate the potential for research from using market simulation, we will
examine the data generated by simulation participants’ behavior in a number of
settings. In a January 2011 class session of the “Trading and Financial Market
Structure” elective at London Business School, a simulation was run which covered
132 R.A. Schwartz and B.W. Weber
Eight teams were each given the instruction to sell 1,500 units, and seven teams
were each asked to buy 1,300. Five other teams had the role of either day traders or
proprietary trading desks. Three of these five teams were instructed to buy 900 then
sell 900 and to have a closing position of 0. Two teams were asked to sell 800, then
buy 800, and to close with a flat position. A trial simulation was run, and perfor-
mance metrics were discussed before the simulation began. The teams with a sell
instruction were told they would be assessed on the basis of the highest average
selling price, while buying teams competed on the basis of the lowest average
buying price. The “prop” teams were told that maximizing closing P&L was their
objective but that they should finish flat and have no “overnight risk.” A screen
similar to the one the participants saw is shown in Fig. 4.2.
4
Simulation as a Research Tool for Market Architects
133
Fig. 4.2 End of a trading day in a networked order book simulation from the screen used by the instructor or simulation administrator
134 R.A. Schwartz and B.W. Weber
1,500
Closing
position
1,000 goals: +1300,
-1,500, or 0
500
0
LBS18
LBS14
LBS12
LBS10
LBS16
LBS11
LBS19
LBS13
LBS15
LBS20
LBS21
LBS4
LBS6
LBS2
LBS5
LBS3
LBS1
LBS8
LBS7
LBS9
500
−1,000
−1,500
−2,000
Fig. 4.3 Final positions of 21 trading teams. Teams were supposed to end with holdings of either
1,500 (sellers), +1,300 (buyers), or flat (0, prop-day traders)
One of the first lessons of behavioral studies of trading done via simulation is
that following instructions is not simple for participants. As Fig. 4.3 shows, seven of
the 21 teams did not end the simulation with the position they were instructed to
have. Three of the selling teams sold more than instructed and three of the
proprietary trading teams had short, nonzero positions at the end of the day. Of
course, the noncompliant teams had excuses – “the end of the day came too fast,” or
“there were not enough willing buyers in the afternoon,” or “we didn’t want to pay
more than the VWAP price on the screen.” These complaints could, of course, also
apply in a real equity market.
In the simulation, the market opened at £20.00. The day’s high was £23.60, the
low £18.80, and the last trade of the day was at £21.50. The £4.80 high-low range
(24 %) reflects a volatile day. The day’s volume-weighted average price (VWAP)
was £20.03. Trading volume was 42,224 units and 784 trades took place. Although
the teams were in the same market, and had the same buying or selling instructions
with the same opportunities, there were vast differences in performance. As Fig. 4.4
shows, the best buying team paid £19.68, or £0.35 less than both the worst team and
VWAP, adding nearly 2 % to the investment return. The best selling team received
£0.95 more per share than VWAP and £1.23 per share more than the worst selling
team. This outcome would add almost 5 % to one selling investor’s return relative
to the average. The conclusion is clear: trading performance has a substantial
impact on investors’ returns.
4 Simulation as a Research Tool for Market Architects 135
100% 90% 80% 70% 60% 50% 40% 30% 20% 10% 0%
£19.683
LBS11
Buy 1,300
LBS5
LBS3
LBS19
LBS1
LBS8
LBS13
LBS14
LBS4
LBS6
LBS2
LBS12
LBS10
LBS16
VWAP = £20.03
£19.60 £19.70 £19.80 £19.90 £20.00 £20.10 £20.20
Fig. 4.4 Performance of trading teams as measured by average buying or selling price (scale
shown on bottom of figure reading left to right) and the percentage of teams’ trading done via limit
orders (scale shown on top of figure reading right to left)
Also shown in Fig. 4.4 is the team’s use of limit orders. Again, there is
substantial variation, with the best buying team trading exclusively with limit
orders and the best selling team completing 79 % of its trading with limit orders.
Note that, as the chart shows, a higher use of limit orders did not assure good trading
outcomes.
Note in Fig. 4.4 that the buying teams all matched or improved on VWAP,
while only three of the selling teams were able to for more than VWAP. Teams’
trading was completed with a varied combination of limit orders and market
orders, with the best team, for instance, selling its 1,500 with 77 % limit orders
and 23 % market orders. No significant correlation existed between order choice
and performance.
As Fig. 4.5 shows, among the five proprietary trading teams, the largest loss was
generated by the team (LBS21) that also had the largest risk as measured by average
absolute value of their inventory position during the trading day. The greatest profit
came from LBS15, a team that only took a moderate level of risk. Again, the
simulation reveals substantial variation and behavioral differences across traders.
136 R.A. Schwartz and B.W. Weber
Average Position
-1,345 P&L LBS21
LBS20
LBS15
LBS9
LBS7
Fig. 4.5 Performance of 5-day trading teams. Collectively the proprietary traders lost money, and
only two teams returned to a zero position
A quote-driven market structure was used in two other simulations with the
same graduate students. The underlying order flow and P* model simulated by
the computer is the same as in the order-driven market. The students were put into
22 teams, and two markets (i.e., two separate networks) with 11 teams each
were run. In each of the networks, seven teams were market makers, and the
other teams were given large orders to fill over the course of the day.
In B-1, the market opened at 20.00, and the day’s high and low prices were 21.00
and 18.70, respectively. The last trade was 19.50, and the VWAP was 20.00, with
706 trades generating a volume of 39,875 units. Participants were told that a unit
represented 1,000 shares and that those with buy or sell instructions were handling
institutional-sized orders that could affect prices.
In their trading, four of the market makers generated a positive profit; as
a group they earned 1.4 pence per share, or seven basis points, in
their trading. As Fig. 4.6 shows, only three of the dealers (LBS1, LBS3, LBS2)
ended the day with a flat (either zero or no more than 35 units) closing
position. The chart shows the seven dealers’ average inventory position, closing
profit or loss, and their closing inventory position. LBS4, for instance, had an
average position over the day of +797, a loss of 674, and a closing
position of +1,431.
4 Simulation as a Research Tool for Market Architects 137
LBS1
LBS3
LBS8
LBS4
LBS2
Fig. 4.6 Performance of seven market maker trading teams. The teams that controlled risk by
keeping their average absolute value of their positions below 350 were able to make profits. Teams
LBS4 and LBS2 incurred large losses and held large risk positions
In the B-2 market, the simulation started at 20.00, and the day’s high and low
prices were 22.30 and 19.00, respectively. The last trade was 22.20, and the VWAP
was 20.22, with 579 trades generating volume of 36,161 units. Reflecting
higher volatility, only two market makers generated a positive profit and, as
a group, they lost 2.7 pence per share, or 13 basis points, in their trading.
Figure 4.7 shows that three of the six dealers (LBS31, LBS25, LBS3) had
a flat position at the end of the day. The other dealers had short positions
reflecting the P* (equilibrium price) increases and buying pressure that drove
the prices up during the simulation.
LBS30
LBS25
LBS3
LBS28
Fig. 4.7 Performance of six market maker trading teams. Only two teams, LBS #31 and #26, were
able to make profits, and the teams that allowed their average position to exceed 300 had the
largest losses
In the example below, we provide the buyers’ reservation price function, which starts
at $26.00 but decreases by one for each 1,000 shares bought. If a participant buys 6,000
shares over the course of the simulation, for instance, the reservation price is $20.00.
If they paid on average $18, then the surplus generated is 12,000.
The sellers’ reservation price function is below. It starts at $12.00 but increases
by one for each 1,000 shares bought. If a participant sells 2,000 shares, the
reservation price is $14.00. If the sellers’ average selling price is $18, for instance,
the surplus is 16,000.
Fig. 4.8 Trading experiment with supply and demand functions and surplus calculations based on
a participant trading at the market average price over the day of $19.11. The optimal trading
strategy was to end with a position of 3,500 long (buyer) or short (seller)
Fig. 4.9 Even when provided with reservation price functions, the live participants in the markets traded near the equilibrium and showed little inclination to
“overtrade.” The average position was only 501 units away from the optimal closing position of 3,500
R.A. Schwartz and B.W. Weber
4 Simulation as a Research Tool for Market Architects 141
perfectly liquid; filling large orders, while competing to outperform price bench-
marks, is challenging. Participants often did not complete their instructions even
though completion was part of their assessment. Third, trading can add (or subtract)
value for investment managers. Even in fairly simple simulation games, perfor-
mance varied widely across participant teams. Finally, risk was not well controlled
by the (admittedly inexperienced) trading participants. Although given specific
position limits as guidelines, many participants had large average positions and
suffered substantial losses from adverse price movements.
Market architects today combine different trading systems designs (Zhang et al.
2011). Beyond studies of trading decision making, the live simulations provide
a method for comparing alternative market structures. For instance, implicit trading
costs incurred in handling a large order over a trading day can be contrasted in
markets with and without dealers, and with and without a dark liquidity pool. Live
simulations, by extending what can be learned with analytical modeling and
empirical data analytics, provide a laboratory for examining a broad set of questions
about trading behavior and market structures.
4.5 Conclusion
Appendix
The simulation methodology was chosen for its ability to accommodate critical
institutional features of the market mechanism and off-exchange dealers’ opera-
tions. While higher level abstractions and simplifications could yield an analytically
tractable model, it is not consistent with the goals of this chapter. These complex-
ities included here are the specialist’s role, the use of large and small order sizes,
and the limit and market orders.
142 R.A. Schwartz and B.W. Weber
Earlier market structure research provides useful insights, but missing institu-
tional details prevent it from determining the effect of third markets. Garman’s
(1976) model of an auction market identifies a stochastic order arrival process and
a market structure consistent with negative serial autocorrelations or the tendency
of price changes to reverse themselves. Garman’s model has no specialist role, and
an analytic solution is obtained only for the case with one possible price. He notes
“the practical difficulties of finding analytic solutions in the general case are
considerable, and numerical techniques such as Monte Carlo methods suggest
themselves.” Mendelson (1987) derives analytic results that provide a comparison
of market consolidation or fragmentation on market performance. The work
provides numerous insights into market design trade-offs. As a simplification,
however, all orders from traders are for one unit of the security.
Simulation has yielded useful results in other microstructure research. Garbade
(1978) investigated the implications of interdealer brokerage (IDB) operations in
a competing dealer market with a simulation model and concluded that there
are benefits to investors from IDBs through reduced dispersion of quotes and trans-
action prices closer to the best available in the market. Cohen et al. (1985) used
simulation to analyze market quality under different sets of trading priority rules. They
showed that systems that consolidate orders and that maintain trading priorities by an
order’s time of arrival in the market increase the quality of the market. Hakannson
et al. (1985) studied the market effects of alternative price-setting and own-inventory
trading policies for an NYSE-style specialist dealer using simulation. They found that
pricing rules “independent of the specialist’s inventories break down.”
Our simulation model has been used in experiments with subject-traders to test
hypotheses concerning market structures. The simulation model is dynamic, with
informational changes occurring in a way that creates the possibility of realizing
trading profits from close attention to price changes and careful order handling. In
our human-machine interactive environment, the computer generates orders from
an unlimited number of “machine-resident” traders and investors. This enables us
easily to satisfy the conditions for an active, competitive market.
To be useful and valid, simulation models must reflect real-world dynamics
without being burdened by unnecessary real-world detail. A simulation model also
requires a strong theoretical foundation. The advantage of simulation over theoretical
modeling is that “theorizing” requires abstracting away from some of the very details
of market structure that exchange officials and regulators wish to study. Consequently,
theoretical modeling can give only limited insight into the effects of market design
changes on the behavior of market participants. The advantage of simulation vis-à-vis
empirically testing of new market structures is that the simulated experiments can be
run at much lower cost and across a broader range of alternatives.
The objective of our computer simulation is to provide a backdrop for assessing
the decisions of live participants. Trading in the model is in a single security and is
4 Simulation as a Research Tool for Market Architects 143
the result of “machine-generated” order flow interacting with the order placement
decisions of the live participants. Assumptions are made about the arrival process
of investors’ orders, changes to a price, p*, which is an “equilibrium value” at which
the expected arrival rate of buy orders equals the expected arrival rate of sell orders.
P* follows a random walk jump process. In other words, the equilibrium value
jumps randomly from one level at interarrival times based on sampling an exponential
distribution. After a shift in p*, the orders of the informed traders pull the quotes and
transaction prices up or down, causing them to trend towards the new p* level.
Occasionally, market prices can also trend away from p* because of the orders of
momentum traders or the chance arrival of a string of buy or sell liquidity orders.
However, movements away from p* are unsustainable; eventually the imbalanced
order flow causes a price reversal and market prices gravitate back towards p*.
Little work in experimental economics has used computers to create background
order flow into which participants individually enter orders (Smith and Williams
1992; Friedman 1993). Our test environment does. We use discrete event computer
simulation to do the following:
• Generate a background public order flow that can (1) be placed on a public limit
order book for later execution, or (2) execute as market orders immediately
against the public limit order book.
• Give the live participants the opportunity to trade a quantity of stock. Depending
on the market structure, the live participants can (i) place them in a public limit
order book and wait for a trade to occur or (ii) execute them against the public
limit order book immediately. Variants of the simulation model include market
makers, or call auctions, or dark liquidity pools to facilitate transactions.
• Maintain the screen which displays (i) orders on the public limit order book,
(ii) a time-stamped record of all transaction sizes and prices for each trading
session, and (iii) the users’ position, risk, and profit performance data.
• Capture information concerning (i) the live participants’ decisions and
(ii) market quality measures such as bid-ask spreads.
In summary, with this realistic and theory-based model and the ability in the
simulation to control the level of transparency provided, we have a rigorous
environment to assess trading decisions and the effects of different market rules.
In the simulation model, assumptions are made about the arrival process of
investors’ orders, elasticity of supply and demand, and order placement strategies,
price volatility, and the proportions of market and limit orders.
Order Arrival. Orders arrive according to a price-dependent Poisson function.
Using time-stamped transactions data on six stocks traded on the London Stock
Exchange, a Kolmogorov-Smirnov goodness-of-fit test fails to reject the null
hypothesis of exponential interarrival in 17 out of 22 sample periods at the 0.10
level of significance. We would expect to reject in just over two cases due to
random realizations. The fit is not perfect in part because transactions tend to
144 R.A. Schwartz and B.W. Weber
Price
buy orders will arrive with Stochastic
$33.00
greater intensity than sell equilibrium,
orders p*(t)
$32.00 D(p(t))
$31.00
5.5 6.5 7.5 8.5 9.5
Arrival Rate: Orders per Hour
cluster somewhat more than predicted by the theoretical model (Weber 1991).
Given the shortcoming of using empirical distributions in simulations (Law and
Kelton 1989), the Poisson assumption appears sufficiently justified for capturing the
typical behavior of the order arrival process.
The Poisson interarrival time, T, is exponentially distributed with bt equal to the
mean interarrival time at time t. The mean interarrival time is set at the beginning of
each experiment and assumed to hold constant. A realization at time t is thus
Tt ¼ C(b). The supply and demand structure follows closely those previously
developed in the market microstructure literature (Garbade and Silber 1979), in
which buy and sell arrival rates are step functions of the difference between the
quoted price and the equilibrium value of the security Fig. 4.10.
Garman (1976) termed the intersection of the supply and demand functions
a “stochastic equilibrium.”
Demand/buy orders, D(p):
The constant a1 reflects the proportion of arrivals that are market orders.
The coefficient a2 determines the arrival rate of limit orders with reservation prices.
Limit order traders are sensitive to discrepancies between available prices and the
equilibrium value. The parameter, d, is the range around the equilibrium
value from which limit prices for limit orders are generated. At a price pi lower
than the equilibrium value at the time, pt*, the arrival rate of buy orders will
exceed the rate of sell order arrivals. The resulting market buy orders and limit
order bids will exceed the quantity of sell orders for below-equilibrium values.
The arrival rate discrepancy will cause prices to rise since in expectation,
orders will trade against the lowest offer quotes, and add new, higher priced
bid quotes.
Order Size. Orders are between one and 250 units of the security. This reflects
a convenient normalization that is consistent with the empirically observable range
of order sizes. A unit may represent, for instance, three round lots, or 300 shares.
Beyond 250 units, we assume the trade would be handled as a block trade, and
negotiated outside of the standard market design, or arrive in the market in smaller
broken-up pieces. Large orders can have “market impact,” and can move prices up
for large buyers, and force them down for larger sellers. The functioning of the
market for large orders is consistent with observed trade discounts for large sell
orders and premiums for large buy orders.
Order Placement Strategies. The machine-generated order flow consists of
liquidity, informed, and momentum trading orders. The liquidity orders are either
limited price orders or market (immediately executable) orders. Market orders
execute on arrival but are “priced” to reflect a maximum acceptable premium or
discount to the current bid or offer. If the market order is large enough, its
price impact (the need to hit successive lower priced bids or lift higher priced
offers) will exceed the acceptable discount or premium, and the remaining
order quantity will become a limit order after partially executing against the
limit order book.
Information Generation. Idiosyncratic information events occur that change
the share value, p*, at which buying and selling order arrival rates are
balanced. Information event occurs according to a Poisson arrival process.
When an information innovation occurs, the price will have a random walk jump.
Price Random Walk. Idiosyncratic information events occur that change
the share value, p*, at which the arrival rate of buy orders equals the arrival rate
of sell orders. The time between information change is assumed to be exponentially
distributed with mean, 12 h. Empirical validation is difficult, because informatio-
n affects the share values in unobservable ways. When there is a change in
information that will shift the “balance price,” p* evolves according to a random
walk without return drift. To assure nonnegative prices, the natural log of price is
used, yielding a log-normal distribution for the equilibrium price. The white noise
term, et, is normally distributed with variance linear in the time since the
last observation. This is consistent with the price diffusion models used in
the financial economics literature (Cox and Rubinstein 1985):
146 R.A. Schwartz and B.W. Weber
Eðln pt þ T j pt Þ ¼ ln pt
Empirical validation for the random walk model comes from numerous tests,
whose results “are remarkably consistent in their general finding of randomness . . .
serial correlations are found to be small” (Malkiel 1987).
Information Effects. If the bid and offer quotes straddle p*, there is no informed
order flow and buying and selling order arrival rates will be equal. When p* is
outside of the bid-offer range, additional one-sided market orders will be generated
according to a Poisson process.
References
Clemons, E., & Weber, B. (1997). Information technology and screen-based securities trading:
Pricing the stock and pricing the trade. Management Science, 43(12), 1693–1708.
Cohen, K., Conroy, R., & Maier, S. (1985). Order flow and the quality of the market. In
Y. Amihud, T. Ho, & R. Schwartz (Eds.), Market making and the changing structure of the
securities industry (pp. 93–109). Lanham, MD: Lexington Books.
Cox, J., & Rubinstein, M. (1985). Options markets. Englewood Cliffs: Prentice-Hall.
Francioni, R., Hazarika, S., Schwartz, R., & Reck, M. (2010). Security market microstructure: The
analysis of a non-frictionless market. In C. F. Lee & A. C Lee (Eds.), Handbook of quantitative
finance and risk management (pp. 333–353). Germany: Springer New York & Heidelberg.
Friedman, D. (1993). How trading institutions affect financial market performance: Some labora-
tory evidence. Economic Inquiry, 31, 410–435.
Garbade, K. (1978). The effect of interdealer brokerage on the transactional characteristics of
dealer markets. Journal of Business, 51(3), 477–498.
Garbade, K., & Silber, W. (1979). Structural organization of secondary markets: clearing fre-
quency, dealer activity and liquidity risk. The Journal of Finance, 34, 577–593.
Garman, M. (1976). Market microstructure. Journal of Financial Economics, 3, 257–275.
Greene, W. H. (2011). Econometric analysis (7th ed.). Upper Saddle River, NJ: Prentice Hall.
Hakansson, N., Beja, A., & Kale, J. (1985). On the feasibility of automated market making by
a programmed specialist. Journal of Finance, 40, 1–20.
Ho, T., Schwartz, R., & Whitcomb, D. (1985). The trading decision and market clearing under
transaction price uncertainty. Journal of Finance, 40, 21–42.
Kennedy, P. (2008). A guide to econometrics (6th ed.). Hoboken, New Jersey: Wiley-Blackwell.
Law, A., & Kelton, W. (1989) Simulation modeling and analysis, McGraw-Hill, New York, NY.
Malkiel, B. (1987). Efficient market hypothesis. In B. Malkiel (Ed.), The New Palgrave:
A dictionary of economics (pp. 127–134). London: Macmillan Press.
Mendelson, H. (1987) Consolidation, Fragmentation, and Market Performance, Journal of
Financial and Quantitative Analysis, 22, 189–207.
Parker, C., & Weber, B. (2012). How IT can disrupt markets: A simulation analysis (Working
Paper). London: London Business School.
Schwartz, R., & Weber, B. (1997). Next-generation securities market systems: An experimental
investigation of quote-driven and order-driven trading. Journal of Management Information
Systems, 14, 57–79.
4 Simulation as a Research Tool for Market Architects 147
Smith, V., & Williams, A. (1992, December). Experimental market economics. Scientific
American, pp. 116–121.
Weber, B. (1991) Information technology and securities markets: Feasibility and desirability of
alternative electronic trading systems, Unpublished dissertation, University of Pennsylvania.
Zhang, S. S., Storkenmaier, A., Wagener, M., & Weinhardt, C. (2011). The quality of electronic
markets. In Forty Fourth Hawaii International Conference on System Sciences, Kauai, Hawaii,
pp. 1–10.
Zhang, X. F. (2010). High frequency trading, stock volatility, and price discovery (Working
Paper). New Haven, CT: Yale University.
Motivations for Issuing Putable Debt:
An Empirical Analysis 5
Ivan E. Brick, Oded Palmon, and Dilip K. Patro
Contents
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 150
5.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
5.3 Hypotheses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
5.4 Data and Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
5.5 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 162
5.6 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 174
Appendix 1: Sample of Firms Issuing Putable Bonds and the Put Bond Characteristics . . . . 175
Appendix 2: Sample of Firms Issuing Poison Put Bonds and the Put Bond
Characteristics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 179
Appendix 3: Estimating the Standard Abnormal Returns and the White t-Statistic . . . . . . . . . . 181
Appendix 4: Generalized Method of Moments (GMM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 182
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 184
Abstract
This paper examines the motivations for issuing putable bonds in which the
embedded put option is not contingent upon a company-related event. We find
We would like to thank Sanjay Deshmukh and conference participants at the Seattle FMA
meetings, October 2000, for helpful comments on an earlier version of the paper circulated under
a different title. The views expressed in this paper are strictly that of the authors and not of the
OCC or the Comptroller of the Currency.
I.E. Brick (*)
Department of Finance and Economics, Rutgers, The State University of New Jersey,
Newark/New Brunswick, NJ, USA
e-mail: ibrick@business.rutgers.edu; ibrick@andromeda.rutgers.edu
O. Palmon
Department of Finance and Economics, Rutgers Business School Newark and New Brunswick,
Piscataway, NJ, USA
e-mail: palmon@business.rutgers.edu
D.K. Patro
RAD, Office of the Comptroller of the Currency, Washington, DC, USA
e-mail: dilip.patro@occ.treas.gov
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 149
DOI 10.1007/978-1-4614-7750-1_5,
# Springer Science+Business Media New York 2015
150 I.E. Brick et al.
that the market favorably views the issue announcement of these bonds that we
refer to as bonds with European put options or European putable bonds. This
response is in contrast to the response documented by the literature to other bond
issues (straight, convertible, and most studies examining poison puts) and to the
response documented in the current paper to the issue announcements of poison
put bonds. Our results suggest that the market views issuing European putable
bonds as helping mitigate security mispricing. Our study is an application of
important statistical methods in corporate finance, namely, event studies and the
use of general method of moments for cross-sectional regressions.
Keywords
Agency costs • Asymmetric information • Corporate finance • Capital structure •
Event study methodology • European put • General method of moments •
Management myopia • Management entrenchment • Poison put
5.1 Introduction
This paper examines the motivations for issuing putable bonds in which the
embedded option is not contingent upon company-related events. The option
entitles bondholders to sell the bond back to the firm on the exercise date (usually
3–10 years after the bond is issued) at a predetermined price (usually at par). We
refer to these bonds as bonds with European put options or European putable bonds.
Unlike the poison put bonds (i.e., bonds with event risk covenants) that have been
studied by the literature,1 the exercise of the option in a putable bond is not
contingent upon a company-related event. This distinction is important because
a poison put protects the bondholder from a specific event (e.g., a takeover) and
may be designed to help prevent that event. In contrast, putable debt provides
protection to the bondholder against any deterioration in the value of her claim.
This distinction is important also because the two types of embedded put options
may serve different purposes. Crabbe and Nikoulis (1997) provide a good overview
of the putable bond market.
Corporate managers determine which contingent claims the company issues to
finance its activities. This choice includes the debt-equity mix and the specific
design of the debt. The design of the debt includes its maturity, seniority, collateral,
and the type of embedded options included in the bond contract. The theoretical
corporate finance literature indicates that including convertible, callable, and/or
putable bonds in the capital structure may help mitigate agency costs and reduce
asymmetric information. Haugen and Senbet (1981) theoretically demonstrate that
the optimal combination of embedded call and put options should eliminate the
1
See for example, Crabbe (1991), Bae et al. (1994, 1997), Cook and Easterwood (1994), and Roth
and McDonald (1999), Nash et al. (2003). Billett et al. (2007) find that 5 % of corporate bonds of
their sample have a non-poison putable option.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 151
asset substitution problem. Bodie and Taggart (1978) and Barnea et al. (1980) show
that bonds with call options can mitigate the underinvestment problem. Robbins
and Schatzberg (1986) demonstrate that the increased interest cost of callable bond
can be used to convey the true value of the firm to the market.2
The literature has not empirically examined the motivation and equity valuation
impact of issuing debt with European put options. We fill this gap in the literature in
several ways. First, we examine the stock price reaction around the announcement
dates of the two types of putable bond issues: bonds with European put options and
bonds with poison puts. Further, we consider four alternative motivations for
incorporating either a European or a poison put option in a bond contract and test
the hypotheses that putable debt is issued to reduce security mispricing, agency
costs of debt, management entrenchment, and myopia. The first possible motivation
for issuing putable debt is asymmetric information. Consider a company that is
undervalued by the market. The market should also undervalue its risky straight-
debt issue. However, if this firm were to issue putable bonds, the put option would
be overvalued. (Recall that put options are negatively related to the value of the
underlying asset.) Consequently, the market value of bonds with a put option is less
sensitive to asymmetric information than straight bonds, minimizing market
mispricing of these debt securities. Additionally, if the market overestimates the
risk of the bond, it may overvalue the embedded put option at issuance, thereby
increasing bond proceeds and benefiting the shareholders.3 The second possible
motivation is mitigating agency costs of debt. For example, the existence of a put
option mitigates the advantage to stockholders (and the loss to bondholders) from
risk shifting, thereby reducing the incentive to shift risk. Risk shifting hurts putable
bondholders less than holders of straight bonds because the value of the put option
(which is held by bondholders) is an increasing function of the firm’s risk. The third
possible motivation is the relatively low coupon rate (a myopic view that ignores
the potential liability to the firm due to the put option). The fourth possible
motivation is that the put option may serve to entrench management. This fourth
motivation is most relevant for firms issuing poison puts since these put options are
exercisable contingent upon company-related events that are usually related to
a change in ownership.
We find that the market reacts favorably to the issue announcement of European
put bonds. We also examine the relationship between the abnormal returns around
the put issue announcement date and firm characteristics that proxy for asymmetric
information problems, potential agency costs (i.e., risk-shifting ability and the level
of free cash flow), and management myopia. The empirical evidence is consistent
with the view that the market considers issuing putable bonds as mitigating security
mispricing caused by asymmetric information. The results do not support the idea
2
Other benefits posited by the literature include minimizing tax liabilities. See for example, Brick
and Wallingford (1985), and Brick and Palmon (1993).
3
Brennan and Schwartz (1988) offer a similar argument to explain the benefits of issuing
convertible bonds.
152 I.E. Brick et al.
that putable bonds are issued to obtain lower coupon rates (i.e., management
myopia). Our empirical findings are robust to a number of alternate specifications.
In contrast to European putable bonds, and consistent with the management
entrenchment hypothesis (see Cook and Easterwood (1994) and Roth and
McDonald (1999)), we find that the market reacts unfavorably to the issue
announcement of poison put bonds. However, consistent with Bae et al. (1994)
who argue that poison put bonds are useful in mitigating agency cost problems, we
find that the abnormal returns around the issue announcement of poison put bonds
are positively related to the protection level of the event risk covenant. Thus, our
results are consistent with the view that European put bonds are effective in
mitigating security mispricing problems, but, in contrast, poison put bonds are
related to management entrenchment or mitigating agency cost problems.
The paper’s organization is as follows. In the next section we summarize the
previous literature. Section 5.3 develops the empirical hypotheses. We describe the
data and empirical methodology in Sect. 5.4. The empirical results are summarized
in Sect. 5.5. We offer concluding remarks in Sect. 5.6.
The theoretical corporate finance literature concludes that the firm’s financing
decision may affect its equity value for several reasons. First, issuing debt
increases firm value because it decreases its tax liabilities.4 Second, issuing debt
may be, in part, a signaling mechanism that informs the market of private
information. For example, Ross (1977) and Ravid and Sarig (1991) demonstrate
that the manager of a firm with better prospects than the market perceives has an
incentive to signal her firm’s quality by issuing a greater amount of debt than
issued in a symmetric information environment. Third, prudent level of debt can
reduce the agency costs arising from the conflict of interest between managers and
shareholders as demonstrated by Jensen and Meckling (1976). For example,
Jensen (1986) demonstrates that leverage can minimize the deleterious effect of
free cash flow on the firm.5 However, leverage is also shown to generate other
4
See, for example, Modigliani and Miller (1963), Scott (1976) and Kim (1978). In contrast, Miller
(1977) suggests that the tax benefit of interest is marginal. However, Mackie-Mason (1990)
empirically demonstrates the significant impact of corporate taxes upon the observed finance
choices of firms.
5
Hence, empirically, we would expect that as firms announce increased levels of debt, the stock
price should increase. However, studies by Dann and Mikkelson (1984), Mikkelson and Partch
(1986), Eckbo (1986), and Shyam-Sunder (1991) indicate that there is no systematic relationship
between the announcement of firm’s debt financing and its stock price or that this relationship is
weakly negative. One potential explanation for this result is that the market can predict future debt
offerings as argued by Hansen and Chaplinsky (1993). Another potential explanation, as suggested
by Miller and Rock (1985) and documented by Hansen and Crutchley (1990), is that raising
external capital may indicate a cash shortfall.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 153
6
Myers (1977) demonstrates that shareholders may avoid some profitable net present value pro-
jects because the benefits accrue to the bondholders. Jensen and Meckling (1976) demonstrate that
leverage increases the incentives for managers, acting as agents of the stockholders, to increase the
risk of the firm.
154 I.E. Brick et al.
5.3 Hypotheses
7
We assume that managers of companies that are overvalued have no incentive to resolve security
mispricing. Consequently, managers of undervalued firms could send a credible signal to the
market of the firm’s undervaluation with the inclusion of a put feature. Overvalued firms should
not be able to mimic since the put option represents a potential liability that is greater to these firms
than of the undervalued firms.
8
The bondholders-stockholders agency conflict has been found to be important in security design
by Bodie and Taggart (1978), Barnea et al. (1980), Haugen and Senbet (1981), Jung et al. (1996),
and Lewis et al. (1998).
9
Equivalently, the inclusion of a put option should restrain management from consuming
a suboptimal amount of perquisites or nonpecuniary benefits.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 155
In contrast, giving bondholders the option to put the bond back to the firm at face
value shifts more of the negative price impact of undertaking negative net present
value projects to the stockholders. This negative stock price impact may reduce
management compensation that is tied to equity performance and/or induce tender
offers that will ultimately help replace the current management. These increased
costs to stockholders and management should induce management to refrain from
undertaking negative net present value projects. Thus, we hypothesize that the
benefit to the stockholders from incorporating a put option in the bond contract
should be directly related to the level of free cash flow. Third, the gain in firm value
should be related to the magnitude of the agency cost problem that the firm faces.
The valuation of the put option by the market (assuming market efficiency and
symmetric information) should be positively related to the magnitude of the agency
cost problem faced by the company. Thus, the benefit to the firm should also be
directly related to the aggregate value of the implied put option of the issue (scaled
by firm size). Fourth, for our sample of poison put bonds, the benefit to shareholders
should increase with the strength of the event risk covenant, since the greater the
strength of the event risk covenant, the less likely management will engage in value
decreasing activities. Bae et al. (1994) tested a similar set of hypotheses for
a sample of poison put bonds.10
The third possible motivation is the low coupon rate (compared to the coupon
rates of straight or callable debt issues). This motivation reflects management
myopia as it ignores the potential liability to the firm due to the put option.11 That
is, myopic management may not fully comprehend the increased risk to the firm’s
viability that is posed by the put option. In particular, bondholders would have the
right to force the firm to (prematurely) retire its debt at a time that is most
inconvenient to the firm, which in turn can precipitate a financial crisis. Further,
if the cost of financial distress is significant, given rational markets and myopic
management, issuing putable debt may negatively impact the value of equity.
This possible motivation has the following empirical implications. First, because
management myopia implies that management pursues suboptimal policies, the
issuance of putable debt should be associated with a decline in equity value.12
Second, the decline should be more severe the larger is the aggregate value of
the implied put option. Third, because expected costs of financial distress are
10
Although bonds become due following formal default, hence all bonds in a sense become
putable, bondholders usually recover substantially less than face value following formal default
because equity value has already vanished. In contrast, a formal inclusion of put option may allow
bondholders to recover the face value at the expense of shareholders when financial distress is not
imminent but credit deterioration has occurred.
11
Other studies that have examined managerial myopia include Stein (1988), Meulbroek
et al. (1990), Spiegel and Wilkie (1996), and Wahal and McConnell (2000).
12
Haugen and Senbet (1978, 1988) theoretically demonstrate that the organizational costs of
bankruptcy are economically insignificant. However, if the putable bonds increase the potential
of technical insolvency, then it will increase potential agency costs that are not necessarily
insignificant.
156 I.E. Brick et al.
negatively related to the financial stability of the company, the decline should be
more severe the lower the credit rating of the bond.
The fourth possible motivation is that the use of put options enhances manage-
ment entrenchment. This hypothesis is relevant for firms issuing poison but not
European put bonds. In particular, many bonds with event risk covenants (i.e.,
poison puts) place restrictions on the merger and acquisition activity of the issuing
firms, thereby strengthening the hands of management to resist hostile takeover
bids. This possible motivation has the following empirical implications. First,
issuing bonds with put options should be followed by a decrease in equity value.
Second, in contrast to the mitigating agency costs hypothesis, the abnormal
returns of equity around the issue announcement date should be inversely related
to the level of event-related covenant protection offered to the bondholder.
Our sample of bonds with put options is taken from the Warga Fixed Income
Database. The database includes pricing information on bonds included in the
Lehman Brothers Bond Indices from January 1973. Using this database we select
all bonds with fixed exercise date put option (i.e., not event contingent), issued
between January 1973 and December 1996, excluding issues by public utilities,
multilateral agencies (such as the World Bank), and sovereign issues. In our sample
we keep only those issues for which we find announcement dates in the Dow Jones
News Retrieval Service. This resulted in a sample of 158 bonds of which the earliest
bond is issued in 1979. Our final sample of bonds is selected according to the
following criteria:
(a) The issuing company’s stock returns are available on the CRSP tapes. For
20 companies CRSP data are not available, resulting in 138 issues. To reduce
confounding effects, all repeat issues by the same company within a year of
a sample issue are eliminated. We also eliminated observations for which we
find other contemporaneous corporate events. This further reduced our sample
to 104 issues.
(b) Furthermore, we eliminate from the sample companies that do not have suffi-
cient accounting data, credit rating, issue size, and industry code in either the
Compustat tapes, Moody’s Industrial Manuals, or the Warga’s Fixed Income
database. This reduces our sample size by 13 firms.
(c) We eliminate one more company for which we found no I/B/E/S data, thus
yielding a final sample of 90 firms.
The list of these firms and the characteristics of these bonds are reported in
Appendix 1. The maturity of these put bonds ranges from 5 to 100 years, with an
average initial maturity of 24 years. The put exercise dates ranges from 2 to
30 years, with an average put expiration period of 7 years.13
13
Four bonds have multiple put exercise dates.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 157
For the sake of comparison, we also construct a sample poison put bonds.14 The
issue announcements of poison put bonds are taken from Dow Jones Interactive.
We also searched LexisNexis but did not find any new announcements.15 This
search resulted in 67 observations. Our final sample of bonds with poison put
feature is further refined using criteria (a), (b), and (c) described above. Criterion
(a) reduces the sample size to 57 observations, and applying criterion (b) further
reduces it to 47 observations. The list of these firms and the characteristics of these
bonds are reported in the Appendix 2.
For these issues, we collect CRSP daily returns for our event study.16 We
calculate the abnormal returns (AR) on equity for each firm around the date of the
issue announcement, assuming that the stochastic process of returns is generated
by the market model. We define the announcement date to be the earlier of the
date on which the bond is issued and the date on which the issue information
appears in the news wires, or published in the Wall Street Journal, as depicted by
the Dow Jones Interactive. Our final sample includes only bonds whose issue
announcement explicitly mentions the inclusion of a put option. We estimate the
market model coefficients using the time period that begins 200 trading days
before and ends 31 trading days before the event, employing the CRSP value-
weighted market index as the benchmark portfolio. We use these coefficients to
estimate abnormal returns for days 30 to +30. We calculate the t-statistics
for the significance of the abnormal and cumulative abnormal returns using
the methodology employed by Mikkelson and Partch (1986). See Appendix 3
for details.
We examine the abnormal returns and their determinants for the sample of
issuers of bonds with European put options. First, we test whether issuing these
bonds significantly affects equity values. The stock price should, on average, react
positively to the issue announcement if either reducing debt security mispricing due
to asymmetric information or mitigating agency costs is a major motivation for
issuing putable bonds. On the other hand, the stock price should, on average, react
negatively to the issue announcement if management myopia (i.e., the relatively
low coupon rate compared to straight or callable debt issues) is a major motivation
of management in issuing putable bonds and if costs of financial distress are
significant.
Second, we estimate the following cross-sectional regression equation that
relates the cumulative abnormal return of the issuing firm’s equity to firm
14
Unlike the European put bond sample which we were able to obtain from Warga’s Fixed Income
Database, we obtained our sample of poison puts from Dow Jones Interactive by using keywords
such as poison put, event risk, and covenant ranking. Warga’s database does not include an
identifiable sample of poison put bonds.
15
We were only able to find a sample of poison put bonds with issue announcements using Dow
Jones Interactive and LexisNexis for the period between 1986 until 1991. We did use these news
services for the periods between 1979 and 2000.
16
The event study methodology was pioneered by Fama et al. (1969).
158 I.E. Brick et al.
characteristics that proxy for agency costs, asymmetric information problems, and
managerial myopia:
where CAR3i is the 3-day (i.e., t ¼ 1, 1) cumulative abnormal return for firm I.
The variables, FCF, RISK, SIZE, INTSAVED, and ANALYSTS proxy for agency
costs, information asymmetry, or management myopia. FINS is a dummy variable
that indicates whether the company is a financial institution.
FCF is the level of free cash flow of the firm for the fiscal year prior to the issue
announcement of putable bonds. We construct two alternative measures of FCF
which are similar to the definition employed by Lehn and Poulsen (1989)
and Bae et al. (1994). FCF1 is defined as [Earnings before Interest and
Taxes – Taxes – Interest Expense – Preferred Dividend Payments – Common
Stock Dividend Payments]/Total Assets. FCF2 is defined as [Earnings before
Interest and Taxes – Taxes – Interest Expense]/Total Assets.17 The inclusion of
a put option in the bond contract should help restrain management from misusing its
cash flow resources. We therefore expect b1 to be positive if reducing agency cost is
a major motivation for issuing putable bonds.
RISK is a dummy variable that equals 1 if the putable bond is rated by Standard
and Poor’s as BBB+ or below and zero otherwise.18 The impact of issuing putable
bonds on equity value may be associated with the variable RISK because of two
alternative hypotheses. First, the level of agency costs due to the asset substitution
problem is positively related to the probability of financial distress. We use this
dummy variable as a proxy for the firm’s ability and incentive to shift risk.
According to the agency cost motivation, we expect b2 to be positive. Second,
the greater the probability of financial distress, the more likely are bondholders to
exercise their put option and force the firm to prematurely retire the debt at a time
that is most inconvenient to the firm. Thus, according to the management myopia
hypothesis, we expect b2 to be negative.
SIZE is the natural logarithm of the total asset level of the issuing firm at the end
of the fiscal year preceding the issue announcement date. SIZE may have two
contrasting impacts on CAR3. First, SIZE may be interpreted as a proxy for the
level of asymmetric information. We expect that the degree of asymmetric infor-
mation is inversely related to SIZE. Hence, according to the debt security
mispricing motivation, we expect the putable bond issue announcement to have
a larger positive impact on small firms than on large firms. Consequently, we expect
17
Note that FCF1 is the actual free cash flow of the firm while FCF2 is the (maximum) potential
free cash flow if dividends are not paid. We do not subtract capital expenditures from the free cash
flow variables because they maybe discretionary and maybe allocated suboptimally to satisfy
management’s interests.
18
Had we assigned the value of 1 to only junk bonds, the number of observations with the risk
variable equal to one would be too small for statistical inference.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 159
b3 to be negative. On the other hand, we also expect that the risk of default is
inversely related to firm size. The higher the probability of default, the lower the
probability that the firm will be in existence and be able to pay its obligations
(including the par value of the bond if the put option is exercised) at the exercise
date of these European put options. Thus, size may be an indirect proxy of the
aggregate value of the put option. In this case, we expect b3 to be positive.
INTSAVED is the scaled (per $100 of assets for the fiscal year prior to the issue
announcement date) annual reduction in interest expense due to the incorporation of
a European put option in the bond contract. The annual interest expense reduction is
calculated as the product of the dollar amount of the putable bond issue and the
yield difference between straight and putable bonds. These yield differences are
calculated by subtracting the yields to maturity of the putable bond from the yield to
maturity of an equivalent non-putable bond, also taken from Warga’s Fixed Income
Database. The equivalent bond has similar maturity, credit rating, and call feature
as the putable bond. The equivalent non-putable bond is selected from the issuing
firm if, at the issue date of the putable debt, the issuing firm has an appropriate
(in terms of maturity, credit rating, and call feature) outstanding non-putable bond.
Otherwise, we choose an appropriate equivalent bond from a firm of the same
industry code as given by the Warga’s Fixed Income database. We posit that value
of the embedded put option is directly related to the (present) value of INTSAVED.
If the benefits of mitigating either agency costs or security mispricing are a major
reason for issuing European put bonds, and if these benefits are directly related to
the value of the embedded option, then b4 should be positive. In contrast, if
managerial myopia is a major reason for issuing European put bonds, and if the
expected reduction in equity value due to the cost of financial distress is related to
the value of the option, then b4 should be negative. The b4 coefficient may be
negative also because it proxies for a loss of interest tax shield due issuing putable
rather than straight bonds.
The variable ANALYSTS, defined as the natural logarithm of one plus the number
of analysts who follow the issuing firm for the quarter prior to the putable bond
issue announcement date, is another proxy for the degree of asymmetric informa-
tion.19 We obtain the number of analysts following each firm, for the year prior to
the bond announcement, from the I/B/E/S tapes. We hypothesize that the degree of
asymmetric information is negatively related to ANALYSTS. Thus, if asymmetric
information motivation is a major motivation for issuing putable debt, we expect
b5 to be negative.
We note that the European put and the poison put samples vary in their
proportion of financial service companies. The European put sample contains
25 (out of 90) financial service companies, while the poison put contains only
one. To control for the potential sector impact, we introduce the dummy variable
FINS which equals one if the company is a financial institution and zero otherwise
19
An alternative measure of the degree of asymmetric information is the number of analysts. The
empirical results are robust to this alternative measure.
160 I.E. Brick et al.
Table 5.1 A summary of expected signs of abnormal returns and their determinants for issuers of
bonds with European put options
Agency cost Security mispricing Managerial myopia
Abnormal returns Positive Positive Negative
FCF Positive No prediction No prediction
RISK Positive No prediction Negative
SIZE No prediction Ambiguous No prediction
INTSAVED Positive Positive Negative
ANALYSTS No prediction Negative No prediction
The determinants of abnormal returns are: RISK is a dummy variable equal to one if the bond issue
has an S&P bond rating of BBB+ or below and zero otherwise. FCF is defined in two separate
ways. FCF1 and FCF2 are the two free cash flow measures as described in the text. In particular,
FCF1 is defined as [Earnings before Interest and Taxes – Taxes – Interest Expense – Preferred
Dividend Payments – Common Stock Dividend Payments]/Total Assets. FCF2 is defined as
[Earnings before Interest and Taxes – Taxes – Interest Expense]/Total Assets. INTSAVED mea-
sures the relative amount of aggregate interest expense saved per $1,000 of total assets of the
issuing firm. ANALYSTS is the natural logarithm of one plus the number of analysts who follow the
issuing firm for the year prior to the putable bond issue announcement date. Those predictions that
are confirmed by our empirical study are in bold letters, and those that are significantly different
from zero are also underlined
for our regression analysis of the European put sample.20 Table 5.1 provides
a summary of the expected signs of the regression coefficients for the sample of
European put bonds.
We estimate our regressions using the general method of moments (GMM)
procedure. See Appendix 4. The procedure yields unbiased White t-statistics
estimates that are robust to heteroscedasticity.21 Note that because our instruments
are the regressors themselves, the parameter estimates from OLS and GMM are
identical. This is also discussed in Appendix 3.
We test the robustness and appropriateness of our specification by estimating
alternative specifications that include additional variables. First, we include
a dummy variable for the existence of a call feature to take into account the
effectiveness of the call feature to mitigate agency costs. Second, we include
a dummy variable to indicate that the bond’s put option expires within 5 years
because effectiveness of the put bond in mitigating agency costs may be related to
20
In the next section, we report the regression results when we exclude financial service companies
from our European put sample. Essentially, the basic results of our paper are not affected by the
inclusion or exclusion of financial services company.
21
The violation of the homoscedastic assumption for OLS does not lead to biased regression
coefficients estimators but potentially biases the computed t-statistic. The GMM procedure provides
an asymptotically unbiased estimation of the t-statistics without specifying the heteroscedastic
structure of the regression equation. The t-statistics obtained using GMM are identical to those
obtained using ordinary least squares (OLS) in the absence of heteroscedasticity.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 161
time to expiration of the put option. Third, we also include interaction variables
between these dummy variables and FCF, SPRISK, INTSAVE, and ANALYSTS.
In alternative specifications, we include interest rate volatility measures to take
into account the potential sensitivity of the value of the put option to interest rate
volatility. It should be noted that the value of the put option depends on the
volatility of the corporate bond yield. This volatility may be due to factors specific
to the company (such as corporate mismanagement and asymmetric information
about firm’s value) and macroeconomic factors relating to the stability and level of
market-wide default-free interest rates. Our independent variables in Eqs. 5.1 and
5.2 control for firm-specific yield volatility. To incorporate the possible impact of
market-wide interest rate volatility, we include several alternative variables. We
measure interest rate volatility as the standard deviation of the monthly 5-year
Fama-Bliss discount rate from CRSP. The standard deviation is alternatively
measured during a period of 60 months and 24 months immediately prior and
after the announcement date. Alternatively, we use the level of the Fama-Bliss
discount rate as a proxy for interest rate volatility.
We repeat the analysis for the sample of issuers of poison put bonds. As with the
sample of bonds with an embedded European put option, we first examine the
abnormal returns and their determinants for the sample of issuers of poison putable
bonds. In essence, we test whether issuing these bonds significantly affects equity
value. The stock price should, on average, react positively to the issue announce-
ment if either mitigating agency costs or reducing debt security mispricing due to
asymmetric information is a major motivation for issuing poison bonds. On the
other hand, the stock price should, on average, react negatively to the issue
announcement if management myopia (i.e., the relatively low coupon rate com-
pared to straight or callable debt issues) or management entrenchment is a major
motivation of management in issuing putable bonds.
We use two alternative specifications for the cross-sectional study that relates
the abnormal returns of the poison putable bond sample to firm characteristics.
The first is described by Eq. 5.1 The second includes a variable, COVRANK, that
equals the S&P event risk ranking on a scale of 1–5. S&P event risk ranking of one
implies that the embedded put option provides the most protection to bondholders
against credit downgrade events. Event risk ranking of five offers the least
protection to bondholders against credit downgrade events. Thus, the second
specification is:
If mitigating agency costs was a major motivation for issuing poison put bonds
as suggested by Bae et al. (1994), then we would expect regression coefficients
b1, b2, and b4 to be positive. We also expect b6 to be negative because
COVRANK is negatively related to extent of bondholder protection, and this
protection should deter management from asset substitution activities.
162 I.E. Brick et al.
In contrast, if, as Cook and Easterwood (1994) and Roth and McDonald (1999)
argue, management entrenchment is a major motivation behind for issuing
putable bonds, then we expect b1 to be negative since we expect the potential
loss of value due to management entrenchment to be positively related to free
cash flow and b6 to be positive.22
This section presents the empirical results for tests of our hypotheses discussed
above. Panel A of Table 5.2 provides summary statistics of issuers of bonds with
a European put option. The average annual sales, long-term debt, and total assets for
the fiscal year prior to the issue announcement date are $14.66 billion, $4.21 billion,
and $30.79 billion, respectively. The mean leverage ratio, defined as the long-term
debt to total assets, is 18.6 %. The average issue size of the putable bond is $190
million and represents, on average, 2.7 % of the firm’s assets. As measured by
RISK, 24 % of putable bonds are rated below A-. The average free cash flow of the
firm as a percentage of the firm’s total assets is below 4 %. The average amount of
interest saved (INTSAVED) due to the inclusion of a put option feature in the bond
issue is approximately 0.03 % of the total assets of the issuing firm. The maximum
interest expense saved is as high as $1.23 per $100 of total assets.23 However,
INTSAVED is negative for ten observations, which may be due to the lack of
a closely matched straight bond. To guard against this error-in-variable problem,
we estimate our regression Eq. 5.1 using two alternative samples: the entire sample
and a sample comprising of positive INTSAVED observations. The number of
analysts following a company ranges from 1 to 41. Panel B of Table 5.2 provides
the corresponding summary statistics for the sample of issuers of poison put bonds.
We note that these firms are smaller than the issuers of bonds with European put
options. Additionally, issuers of poison puts have a smaller number of following
analysts, and the bonds tend to be somewhat riskier than the issuers of bonds with
European put options.
Panel A of Table 5.3 presents the average daily abnormal performance of the
equity of issuers of bonds with European put options in our sample for t ¼ 30 to
t ¼ 30. These abnormal returns are obtained from a market model. Please note that
the t-statistics presented in Table 5.3 are based on standardized abnormal returns. In
an efficient market, we expect the market to impound the economic informational
impact of the new bond issue on the day of the announcement (t ¼ 0). The average
abnormal return at t ¼ 0 is almost 0.33 % and is significantly positive at the 1 % level.
22
If a major motivation for issuing putable bonds is to enhance management entrenchment, then
we expect that the value of entrenchment is directly related to the level of free cash flow which
management can misappropriate.
23
Crabbe (1991) demonstrates that bonds with poison puts reduce the cost of borrowing for firms.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 163
This suggests that the market views favorably the announcement of putable bonds.
This may be due to mitigating agency costs or resolving asymmetric information.
However, for the 3-day window (t ¼ 1, 1), the abnormal return is 0.19 % but is not
statistically significant from zero.
164 I.E. Brick et al.
Table 5.3 The average daily abnormal returns for firms issuing putable bonds from 30 days prior
to the putable bond issuance announcement to 30 days after the announcement. The t-statistics are
based on standardized abnormal returns. CAR is the cumulative abnormal return
Event day Abnormal return t-statistic CAR
Panel A: Sample of 90 issuers of bonds with European put options
30 0.0006 0.1988 0.0006
29 0.0004 0.2305 0.0010
28 0.0011 0.8169 0.0021
27 0.0003 0.1248 0.0017
26 0.0002 0.2028 0.0019
25 0.0003 0.0263 0.0022
24 0.0003 0.2174 0.0025
23 0.0002 0.1306 0.0023
22 0.0007 0.2573 0.0030
21 0.0003 0.0367 0.0026
20 0.0024 1.6700 0.0003
19 0.0000 0.1348 0.0002
18 0.0025 1.6932 0.0022
17 0.0029 1.7141 0.0007
16 0.0005 0.0154 0.0002
15 0.0022 1.4378 0.0024
14 0.0022 1.4701 0.0046
13 0.0002 0.1052 0.0043
12 0.0007 0.4783 0.0050
11 0.0014 1.2803 0.0036
10 0.0003 0.1904 0.0039
9 0.0014 0.5100 0.0025
8 0.0006 0.6052 0.0019
7 0.0001 0.4020 0.0021
6 0.0005 0.1268 0.0026
5 0.0027 1.3209 0.0053
4 0.0001 0.0050 0.0052
3 0.0020 1.4407 0.0071
2 0.0028 1.8590 0.0043
1 0.0021 1.4323 0.0022
0 0.0033 2.6396 0.0055
1 0.0007 0.3534 0.0061
2 0.0007 0.3177 0.0068
3 0.0002 0.2628 0.0070
4 0.0010 0.7364 0.0060
5 0.0000 0.2821 0.0059
6 0.0036 2.4342 0.0023
7 0.0002 0.4355 0.0022
8 0.0004 0.4253 0.0018
(continued)
5 Motivations for Issuing Putable Debt: An Empirical Analysis 165
0.0050
0.0000
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
−9
−8
−7
−6
−5
−4
−3
−2
−1
−30
−29
−28
−27
−26
−25
−24
−23
−22
−21
−20
−19
−18
−17
−16
−15
−14
−13
−12
−11
−10
CAR
−0.0050
−0.0100
Motivations for Issuing Putable Debt: An Empirical Analysis
−0.0150
Event Date
Fig. 5.1 Cumulative abnormal returns from t ¼ 30 to t ¼ +30 around the issue announcement of bonds with European or poison put features
167
168 I.E. Brick et al.
Table 5.4 Cross-sectional regression results for the European put sample: base case
Sample of 80 bonds with positive
Full sample of 90 firms INTSAVED
Coef. t-ratio p-value Coef. t-ratio p-value
Intercept 0.0388 1.69 0.095 0.03657 1.58 0.1193
FCF1 0.0016 0.09 0.9289 0.00392 0.22 0.8277
RISK 0.0016 0.25 0.8026 0.00245 0.37 0.7117
SIZE 0.0097 2.93 0.0044 0.010637 3.03 0.0034
INTSAVED 0.0269 2.59 0.0115 0.023819 2.27 0.0259
ANALYSTS 0.0153 2.48 0.0152 0.01848 2.82 0.0062
FINS 0.0203 2.79 0.0065 0.02062 2.75 0.0075
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼b6¼0 26.38 0.0002 29.19 0.000
Adj. R2 0.1185 0.1314
Coef. t-ratio p-value Coef. t-ratio p-value
Intercept 0.0390 1.71 0.0915 0.03642 1.58 0.1191
FCF2 0.0004 0.02 0.9808 0.00392 0.21 0.8309
RISK 0.0016 0.24 0.8075 0.00247 0.37 0.7094
SIZE 0.0097 2.93 0.0043 0.010633 3.03 0.0034
INTSAVED 0.0270 2.59 0.0112 0.023751 2.26 0.0266
ANALYSTS 0.0152 2.47 0.0157 0.01849 2.8 0.0065
FINS 0.0202 2.73 0.0078 0.02066 2.7 0.0085
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼b6¼0 26.73 0.0002 29.25 0.000
Adj. R2 0.1184 0.1314
This table reports the regression coefficients and their t-statistics of the following regression
equation:
CAR3i ¼ b0 + b1FCFi + b2RISKi + b3SIZEi + b4INTSAVEDi + b5ANALYSTSi + b6FINS + ei
CAR3 is the cumulative abnormal return measured from the day before the announced issue to the
day after the announced issue; FCF1 and FCF2 are the two free cash flow measures as described in
the text; RISK is a dummy variable equal to one if the bond issue has an S&P bond rating of BBB+
or below and zero otherwise; SIZE is the natural logarithm of the total assets of the issuing firm at
the end of the fiscal year prior to the issue announcement; INTSAVED measures the relative
amount of aggregate interest expense saved per $100 of total assets of the issuing firm; ANALYSTS
is the natural logarithm of one plus the number of analysts following the firm; and FINS is equal to
one if the parent company is a financial institution and is equal to zero otherwise. The p-values
assume a two-tail test. All the t-ratios are heteroscedasticity consistent
Panel B of Table 5.3 presents the corresponding results for the issuers of poison put
bonds. Note that the abnormal return at t ¼ 0 is 0.28 % with a t-statistic of 1.38.
Furthermore, during the 3-day (i.e., t ¼ 1, 1) window, the abnormal return is
negative (0.52 %) and is significantly different than zero (t-statistic of 2.03).
This negative abnormal return is consistent with the view that poison put bonds help
entrench the current management. The cumulative abnormal returns of the two sample
types of embedded put option are also depicted in Fig. 5.1. The patterns of CARs
5 Motivations for Issuing Putable Debt: An Empirical Analysis 169
around the issue announcement dates clearly show the positive trend of CARs
for bonds with European puts versus the negative trend of CARs for bonds with
poison puts.
Next, to test our hypotheses, we examine what determines the cross-sectional
variations in the CARs for the sample of European and poison put bonds. Table 5.4
reports the parameter and t-statistic estimates for regression Eq. 5.1 for the sample
of issuers of bonds with European put options. The four regressions reported in
Table 5.4 differ by the sample that is used and by the specification used for the free
cash flow variable. The regression in the top left panel uses the entire sample of
90 observations and the variable FCF1. The regression in the top right panel reports
the results using only the 80 observations in which the variable INTSAVED is
positive. The regressions reported in the bottom panels use FCF2 instead of FCF1.
First note that our estimates are robust to these alternative specifications. The
Wald test of the hypothesis that all five independent variables are jointly zero is
rejected at the 1 % significance level for all four regression specifications. The
regression coefficients on the ANALYSTS variable, b5, are negative and significantly
different from zero for all four regressions. These estimates are consistent with the
reduction in security mispricing motivation for issuing putable bonds. In contrast,
the regression coefficients for FCF and RISK are not significantly different from
zero, indicating a lack of empirical support for the mitigating agency cost hypoth-
esis. The regression coefficients for the INTSAVED variables are positive and are
significantly different from zero in all four regressions. This result is again consis-
tent with the security mispricing motivation and inconsistent with the management
myopia motivation. Additionally, the regression coefficient b3 for the size variable
is also positive and significantly different from zero, a result more consistent with
size being related to the probability of survivorship of the firm. In summary,
announcement of bonds with European puts is associated with positive abnormal
returns which are related to our proxies for potential benefits from mitigating
security mispricing.
Panel A of Table 5.5 presents the corresponding estimates for the sample of
issuers of poison put bonds, while panel B of Table 5.5 presents the estimates of
Eq. 5.2 for the same sample. The estimates reported in Table 5.5 are different from
those in Table 5.4. The coefficients of the ANALYSTS variable, b5, are significantly
positive for the full sample. The coefficients for Size are significantly negative for
the full sample. All the other variables are not significantly different than zero. The
estimates of Eq. 5.2 indicate that the coefficients of the variable COVRANK are
negative and significantly different from 0.
Our results so far indicate that the equity abnormal returns around the issue
announcement dates of poison put bonds are negative, consistent with the mana-
gerial entrenchment evidence in Cook and Easterwood (1994) and Roth and
McDonald (1999). Additionally, the positive and significant coefficient of the
ANALYSTS variable may also be consistent with the management entrenchment
hypothesis because it appears that the market negative response to the issuance of
poison putable bonds arises from the less followed firms, where the management
170 I.E. Brick et al.
Table 5.5 Cross-sectional regression results for the poison put sample
Sample of 28 bonds with
Full sample of 47 firms positive INTSAVED
Coef. t-ratio p-value Coef. t-ratio p-value
Panel A: Sample of issuers of bonds with poison put options
Intercept 0.015 0.530 0.602 0.030 0.680 0.506
FCF1 0.170 1.330 0.191 0.082 0.510 0.617
RISK 0.010 1.550 0.129 0.001 0.090 0.932
SIZE 0.007 2.250 0.030 0.006 1.090 0.288
INTSAVED 0.022 0.370 0.714 0.024 0.280 0.782
ANALYSTS 0.014 2.100 0.042 0.006 0.370 0.717
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼0 11.927 0.036 1.906 0.861
Adj. R2 0.015 0.181
Intercept 0.018 0.730 0.472 0.026 0.640 0.532
FCF2 0.153 1.280 0.208 0.045 0.270 0.788
RISK 0.009 1.400 0.168 0.000 0.030 0.978
SIZE 0.007 2.480 0.017 0.005 1.020 0.318
INTSAVED 0.027 0.460 0.649 0.016 0.180 0.858
ANALYSTS 0.014 1.980 0.054 0.006 0.330 0.744
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼0 11.538 0.042 1.857 0.868
Adj. R2 0.007 0.192
Sample of 23 bonds with
Full sample of 40 firms positive INTSAVED
Coef. t-ratio p-value Coef. t-ratio p-value
Panel B: Sample of issuers of bonds with poison put options with COVRANK
Intercept 0.052 1.200 0.239 0.136 2.550 0.021
FCF1 0.158 1.100 0.281 0.271 1.330 0.201
RISK 0.011 1.140 0.264 0.006 0.370 0.720
SIZE 0.008 1.800 0.081 0.014 2.370 0.031
INTSAVED 0.029 0.430 0.670 0.071 0.720 0.484
COVRANK 0.008 3.900 0.000 0.012 3.520 0.003
ANALYSTS 0.013 2.090 0.045 0.009 0.700 0.497
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼b6¼0 48.723 0.000 20.513 0.002
Adj. R2 0.114 0.058
Intercept 0.053 1.440 0.160 0.117 3.030 0.008
FCF2 0.132 1.000 0.327 0.159 0.890 0.386
RISK 0.010 1.110 0.275 0.000 0.010 0.995
SIZE 0.009 2.030 0.051 0.013 2.590 0.020
INTSAVED 0.032 0.510 0.616 0.030 0.240 0.810
COVRANK 0.008 4.100 0.000 0.010 3.610 0.002
(continued)
5 Motivations for Issuing Putable Debt: An Empirical Analysis 171
strategy may not be as well known prior to the bond issuance. However, these
returns are negatively related to the event risk covenant ranking consistent with the
agency cost evidence in Bae et al. (1994). The negative abnormal returns experi-
enced by issuers of poison puts and the different coefficients on SIZE, INTSAVED,
and ANALYSTS indicate that the European put bonds are viewed differently than
poison put bonds.
In summary, the results reported in Tables 5.2–5.5 are consistent with the view
that European put bonds, where the put option does not depend on a specific
company event, are more effective in mitigating problems that are associated
with security mispricing. Furthermore, there is no empirical support for the hypoth-
esis that European put bonds are used as a vehicle for management entrenchment. In
contrast, the evidence for poison put bonds is consistent with both management
entrenchment and mitigating agency costs.
We now discuss several robustness checks of our basic results. Note that, as
reported in Table 5.4, the regression coefficient of the financial industry dummy
variable, b6, is negative and significantly different from zero. To verify that the
differing results between the European put bond sample (where 25 out of 90 firms
172 I.E. Brick et al.
Table 5.6 Cross-sectional regression results for the European put sample excluding financial
service companies
Sample of 61 bonds with positive
Full sample of 65 firms INTSAVED
Coef. t-ratio p-value Coef. t-ratio p-value
Intercept 0.0410 1.27 0.2087 0.0416 1.25 0.2182
FCF1 0.0005 0.04 0.9715 0.0016 0.10 0.9200
RISK 0.0028 0.41 0.6821 0.0033 0.49 0.6279
SIZE 0.0108 2.79 0.0071 0.0114 2.80 0.0071
INTSAVED 0.0248 1.85 0.0694 0.0242 1.79 0.0785
ANALYSTS 0.0178 2.54 0.0138 0.0191 2.70 0.0093
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼0 32.66 0.0000 32.10 0.0000
Adj. R2 0.1450 0.1524
Intercept 0.0409 1.28 0.2073 0.0414 1.25 0.2181
FCF2 0.0000 0.00 0.9992 0.0031 0.17 0.8676
RISK 0.0028 0.41 0.6805 0.0034 0.50 0.6206
SIZE 0.0108 2.78 0.0072 0.0114 2.79 0.0071
INTSAVED 0.0248 1.85 0.0686 0.0241 1.79 0.0784
ANALYSTS 0.0178 2.54 0.0137 0.0191 2.70 0.0091
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼0 32.66 0.0000 32.10 0.0000
Adj. R2 0.1450 0.1525
This table reports the regression coefficients and their t-statistics of the following regression
equation:
CAR3i ¼ b0 + b1FCFi + b2RISKi + b3SIZEi + b4INTSAVEDi + b5ANALYSTSi + ei
CAR3 is the cumulative abnormal return measured from the day before the announced issue to the
day after the announced issue; FCF1 and FCF2 are the two free cash flow measures as described in
the text; RISK is a dummy variable equal to one if the bond issue has a S&P bond rating of BBB+
or below, and zero otherwise; SIZE is the natural logarithm of the total assets of the issuing firm at
the end of the fiscal year prior to the issue announcement; INTSAVED measures the relative
amount of aggregate interest expense saved per $100 of total assets of the issuing firm; ANALYSTS
is the natural logarithm of one plus the number of analysts following the firm. The p-values assume
a two-tail test. All the t-ratios are heteroscedasticity consistent
are from the financial service sector) and the poison put sample (where only one
firm is from the financial service sector), we replicate the regressions of Tables 5.4
and 5.5 excluding all financial service companies. The estimates for the subsample
of European put bonds are reported in Table 5.6. Note that the coefficients and
significance levels are very similar to those reported in Table 5.4. Because only one
poison put company is from the financial sector and because the estimates of the
corresponding subsample of poison put bonds are very similar to those reported in
Table 5.5, we do not report these estimates. Thus, we conclude that the differences
between poison and European put bonds are not due to the different sector compo-
sition of our samples.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 173
Table 5.7 The cross-sectional regression results for the European put sample: impact of time to
expiration
Sample of 80 bonds with
Full sample of 90 firms positive INTSAVED
Coef. t-ratio p-value Coef. t-ratio p-value
Intercept 0.0288 1.27 0.2063 0.0282 1.22 0.2279
FCF1 0.0037 0.24 0.8147 0.0007 0.05 0.9627
RISK 0.0010 0.16 0.8703 0.0021 0.32 0.7527
SIZE 0.0088 2.74 0.0075 0.0098 2.88 0.0052
INTSAVED 0.0143 1.35 0.1814 0.0131 1.22 0.2249
ANALYSTS 0.0165 2.68 0.0090 0.0193 2.95 0.0043
FINS 0.0185 2.60 0.0112 0.0194 2.61 0.0109
EXPLT5 0.0097 1.91 0.0601 0.0083 1.37 0.1739
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼b6¼ b7¼0 34.82 0.0000 37.88 0.000
Adj. R2 0.1413 0.1420
Intercept 0.0291 1.3 0.1981 0.0283 1.22 0.2248
FCF2 0.0051 0.33 0.7397 0.0010 0.06 0.9506
RISK 0.0010 0.15 0.8794 0.0021 0.31 0.7542
SIZE 0.0088 2.74 0.0075 0.0098 2.88 0.0052
INTSAVED 0.0144 1.36 0.1784 0.0132 1.22 0.2257
ANALYSTS 0.0164 2.67 0.0092 0.0193 2.93 0.0045
FINS 0.0184 2.53 0.0133 0.0194 2.57 0.0122
EXPLT5 0.0097 1.91 0.0593 0.0083 1.38 0.1733
w2 p-value w2 p-value
H0:b1¼b2¼b3¼b4¼b5¼b6¼b7¼0 34.85 0.0000 37.82 0.000
Adj. R2 0.1415 0.1420
This table reports the regression coefficients and their t-statistics of the following regression equation:
CAR3i ¼ b0 + b1FCFi + b2RISKi + b3SIZEi + b4INTSAVEDi + b5ANALYSTSi + b6FINS + b7EXPLT5 + ei
CAR3 is the cumulative abnormal return measured from the day before the announced issue to the
day after the announced issue; FCF1 and FCF2 are the two free cash flow measures as described in
the text; RISK is a dummy variable equal to one if the bond issue has a S&P bond rating of BBB+
or below and zero otherwise; SIZE is the natural logarithm of the total assets of the issuing firm at
the end of the fiscal year prior to the issue announcement; INTSAVED measures the relative
amount of aggregate interest expense saved per $100 of total assets of the issuing firm; ANALYSTS
is the natural logarithm of one plus the number of analysts following the firm; FINS is equal to one
if the parent company is a financial institution and is equal to zero otherwise; and EXPLT5 is equal
to one if the expiration of the embedded option is less than 5 years from the issue date and is equal
to zero otherwise. The p-values assume a two-tail test. All the t-ratios are heteroscedasticity
consistent
Next, we examine whether the term to expiration of European put bonds affects
our estimates.24 Table 5.7 repeats the regressions of Table 5.4 when we introduce
an additional explanatory dummy variable, EXPLT5, that is equal to one if the
24
Recall that by definition the option of poison put bonds does not have an expiration date.
174 I.E. Brick et al.
expiration date of the embedded European put option is less than 5 years from
the issue date and zero otherwise. We find that INTSAVE is still positive but is
no longer significant and the regression coefficient for EXPLT5 is positive and
is significantly different from zero at the 10 % level for the full sample.
The estimates and the significance of the other coefficients are largely
unaffected. We conclude that these results still confirm the security mispricing
hypothesis because the regression coefficients for ANALYSTS are significantly
negative.
Our results are robust to alternative specifications we described at the end of the
previous section. In particular, including a dummy variable for the existence of
a call feature, interaction variables and alternative measures of interest rate vola-
tility do not affect our results. These results are not reported here but are available
upon request.
Finally, we test for multicollinearity by examining the eigenvalues of
the correlation matrix for (non-dummy) independent variables. For orthogonal
data the eigenvalue, l, for each variable should equal 1, and S1/l ¼ the
number of regressors (i.e., five for our study). For our sample of bonds with an
embedded European put option, this sum is 7.14 when FCF1 is used, and
this sum ¼ 7.19 when FCF2 is used, indicating a lack of significant
multicollinearity. Similar results are obtained for the poison put sample. There-
fore, our findings are robust to alternate specifications and are not driven by
multicollinearity.
This paper examines the motivations and equity valuation impact of issuing Euro-
pean putable bonds, a bond containing an embedded European put option held by
bondholders. The option entitles them to sell the bond back to the firm on the
exercise date at a predetermined price. Unlike a poison put bond which has been
studied by the literature, the exercise of the put option in a European putable bond is
not contingent upon a company-related event.
We find that the market reacts favorably to the issue announcement of such
putable bonds. We consider three alternative motivations for incorporating
a European put option in a bond contract: reducing the security mispricing
impact of asymmetric information, mitigating agency costs, and the relatively
low coupon rate (a myopic view that ignores the potential liability to the
firm due to the put option). We test these hypotheses by conducting a
cross-sectional empirical study of the impact of putable debt issue announce-
ments on the equity value of the issuing companies. Our results indicate that the
market favorably views putable bonds as a means to reduce security mispricing.
We also find that the market reaction to poison put announcements differs
from the market reaction to issue announcements for bonds with European
put options.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 175
Put S&P
Amount Issue Maturity expiration YTM credit
Name Coupon ($000’s) date (years) (years) Callability (%) rating
Air Products 7.34 100,000 06/15/ 30 12 N 7.20 7
and Chemicals 1996
Inc
American 8.50 150,000 06/09/ 10 5 N 8.56 5
Express 1989
Anadarko 7.25 100,000 03/17/ 30 5 N 6.90 9
Petroleum 1995
Corp
Anadarko 7.73 100,000 09/19/ 100 30 N 7.21 9
Petroleum 1996
Corp
Bankamerica 7.65 150,000 04/26/ 10 2 C 12.23 10
Corp 1984
Bausch & 6.56 100,000 08/12/ 30 5 N 6.71 8
Lomb Inc 1996
Baxter 8.88 100,000 06/14/ 30 5 C 8.96 9
International 1988
Inc
Burlington 7.29 200,000 05/31/ 40 12 N 7.39 10
Northern 1996
Santa Fe
Champion 15.63 100,000 07/22/ 12 3 C 15.68 9
International 1982
Corp
Champion 6.40 200,000 02/15/ 30 10 N 6.69 10
International 1996
Corp
Chase 7.55 250,000 06/12/ 12 5 C 8.87 5
Manhattan 1985
Corp – old
Chrysler Corp 12.75 200,000 11/01/ 15 5 N 12.75 10
1984
Chrysler Corp 9.65 300,000 07/26/ 20 5 C 9.65 10
1988
Chrysler Corp 9.63 200,000 09/06/ 20 2 C 9.65 11
1988
Circus Circus 6.70 150,000 11/15/ 100 7 N 6.66 9
Enterprise Inc 1996
Citicorp 9.40 250,000 12/06/ 12 2 C 11.05 4
1983
Citicorp 10.25 300,000 12/19/ 10 2 C 10.31 4
1984
(continued)
176 I.E. Brick et al.
Put S&P
Amount Issue Maturity expiration YTM credit
Name Coupon ($000’s) date (years) (years) Callability (%) rating
Citicorp 8.75 250,000 12/12/ 15 3 C 8.84 4
1985
Coca-Cola 7.00 300,000 09/27/ 30 10 N 7.00 5
Enterprises 1996
Commercial 8.50 100,000 02/05/ 10 5 N 8.84 8
Credit 1988
Commercial 8.70 150,000 06/09/ 20 10 N 9.06 8
Credit 1989
Commercial 8.70 100,000 06/11/ 20 3 N 7.54 7
Credit 1990
Commercial 7.88 200,000 02/01/ 30 10 N 7.30 6
Credit 1995
Conagra Inc 7.13 400,000 10/02/ 30 10 N 6.95 9
1996
Corning Inc 7.63 100,000 07/31/ 30 10 N 7.81 6
1994
Deere & Co 8.95 199,000 06/08/ 30 10 C 8.99 7
1989
Diamond 7.65 100,000 06/25/ 30 10 N 7.37 10
Shamrock Inc 1996
Dow 8.48 150,000 08/22/ 30 14 C 9.35 7
Chemical 1985
Eastman 7.25 125,000 04/08/ 10 5 N 8.48 8
Kodak Co 1987
Eaton Corp 8.00 100,000 08/18/ 20 10 N 7.90 7
1986
Eaton Corp 8.88 38,000 06/14/ 30 15 N 9.00 7
1989
Eaton Corp 6.50 150,000 06/09/ 30 10 N 6.64 7
1995
Enron Corp 9.65 100,000 05/17/ 12 7 N 9.50 10
1989
First Chicago 8.50 98,932 05/13/ 12 7 N 8.70 7
Corp 1986
First Interstate 7.35 150,000 08/23/ 15 6 C 13.70 4
Bancorp 1984
First Interstate 9.70 100,000 07/08/ 15 5 C 9.91 4
Bancorp 1985
First Union 7.50 250,000 04/25/ 40 10 N 7.96 8
Corp (NC) 1995
First Union 6.82 300,000 08/01/ 30 10 N 7.43 8
Corp (NC) 1996
Ford Motor Co 7.50 250,000 10/30/ 15 3 N 9.63 6
1985
(continued)
5 Motivations for Issuing Putable Debt: An Empirical Analysis 177
Put S&P
Amount Issue Maturity expiration YTM credit
Name Coupon ($000’s) date (years) (years) Callability (%) rating
Ford Motor Co 9.95 300,000 02/10/ 40 3 N 8.56 6
1992
General 6.75 250,000 11/06/ 25 5 C 6.80 2
Electric Co 1986
General 8.25 500,000 04/26/ 30 3 C 8.41 2
Electric Co 1988
General 8.38 200,000 04/30/ 10 5 N 8.38 5
Motors Corp 1987
General 8.63 400,000 06/09/ 10 5 N 8.59 5
Motors Corp 1989
General 8.88 500,000 05/31/ 20 5 N 8.80 5
Motors Corp 1990
Harris Corp 6.65 100,000 08/01/ 10 5 N 6.90 8
1996
Ingersoll- 6.48 150,000 06/01/ 30 10 N 6.64 7
Rand Co 1995
Intl Business 13.75 125,000 03/09/ 12 3 C 14.05 2
Machines 1982
Corp
ITT Industries 8.50 100,000 01/20/ 10 5 N 8.50 7
Inc 1988
ITT Industries 8.55 100,000 06/12/ 20 9 N 9.20 7
Inc 1989
ITT Industries 3.98 100,000 02/15/ 15 3 N 8.60 7
Inc 1990
Johnson 7.70 125,000 02/28/ 20 10 N 7.83 7
Controls Inc 1995
K N Energy 7.35 125,000 07/25/ 30 10 N 7.47 9
Inc 1996
Litton 6.98 100,000 03/15/ 40 10 N 7.01 9
Industries Inc 1996
Lockheed 7.20 300,000 05/01/ 40 12 N 7.29 9
Martin Corp 1996
Marriott Corp 9.38 250,000 06/11/ 20 10 N 9.74 8
1987
Merrill Lynch 11.13 250,000 03/26/ 15 5 C 13.36 4
& Co 1984
Merrill Lynch 9.38 125,000 06/04/ 12 6 C 9.48 4
& Co 1985
Merrill Lynch 8.40 200,000 10/25/ 30 5 N 8.67 6
& Co 1989
Motorola Inc 8.40 200,000 08/15/ 40 10 N 8.36 4
1991
Motorola Inc 6.50 400,000 08/31/ 30 10 N 6.55 4
1995
(continued)
178 I.E. Brick et al.
Put S&P
Amount Issue Maturity expiration YTM credit
Name Coupon ($000’s) date (years) (years) Callability (%) rating
Occidental 9.25 300,000 08/03/ 30 15 N 9.37 10
Petroleum 1989
Corp
Penney 6.90 200,000 08/16/ 30 7 N 7.07 7
(JC) Co 1996
Philip Morris 7.00 150,000 07/15/ 5 2 N 7.30 7
Cos Inc 1986
Philip Morris 9.00 350,000 05/09/ 10 6 N 9.39 7
Cos Inc 1988
Philip Morris 6.95 500,000 06/01/ 10 5 N 6.91 7
Cos Inc 1996
Pitney Bowes 8.63 100,000 02/10/ 20 10 N 8.70 4
Inc 1988
Pitney Bowes 8.55 150,000 09/15/ 20 10 Y 8.64 4
Inc 1989
Regions 7.75 100,000 09/15/ 30 10 N 8.07 7
Financial Corp 1994
Ryder System 9.50 125,000 07/01/ 15 2 C 9.69 7
Inc 1985
Seagram Co 4.42 250,000 08/03/ 30 15 N 9.90 7
Ltd 1988
Security 12.50 150,000 09/27/ 12 6 C 12.62 3
Pacific Corp 1984
Security 7.50 150,000 04/03/ 15 3 C 7.54 3
Pacific Corp 1986
Service Corp 7.00 300,000 05/26/ 20 7 N 6.80 9
International 1995
Southtrust 1.62 100,000 05/09/ 30 10 N 7.28 8
Corp 1995
State Street 7.35 150,000 06/15/ 30 10 N 7.23 5
Corp 1996
Suntrust 6.00 200,000 02/15/ 30 10 N 6.33 7
Banks Inc 1996
Triad Systems 14.00 71,500 08/09/ 8 3 C 13.99 16
Corp 1989
TRW Inc 9.35 100,000 05/31/ 30 10 N 9.28 7
1990
Union Carbide 6.79 250,000 06/01/ 30 10 N 6.83 10
Corp 1995
United 8.50 150,000 09/22/ 30 10 N 8.60 9
Dominion 1994
Realty Trust
Westinghouse 11.88 100,000 03/19/ 12 3 C 13.38 6
Electric 1984
(continued)
5 Motivations for Issuing Putable Debt: An Empirical Analysis 179
Put S&P
Amount Issue Maturity expiration YTM credit
Name Coupon ($000’s) date (years) (years) Callability (%) rating
Westinghouse 8.88 150,000 05/31/ 24 4 N 8.84 13
Electric 1990
Whitman Corp 7.29 100,000 09/19/ 30 8 N 7.26 9
1996
WMX 8.75 250,000 04/29/ 30 5 C 8.79 4
Technology 1988
WMX 7.65 150,000 03/15/ 20 3 N 7.71 6
Technology 1991
WMX 6.22 150,000 05/09/ 10 3 N 7.41 6
Technology 1994
WMX 6.65 200,000 05/16/ 10 5 N 6.43 6
Technology 1995
WMX 7.10 450,000 07/31/ 30 7 N 7.16 7
Technology 1996
Xerox Corp 11.25 100,000 08/25/ 15 3 C 11.44 5
1983
S&P
Amount Years to credit
Name Coupon ($million) Issue date maturity Callability YTM (%) rating
Aar Corp 9.500 65 10/27/89 12 N 9.425 10
AMR 9.750 200 03/15/90 10 N 9.85 7
Anheuser- 8.750 250 12/01/89 10 N 8.804 5
Busch
Armstrong 9.750 125 08/18/89 19 N 9.5 5
World
Ashland Oil 11.125 200 10/08/87 30 C 10.896 7
Becton 9.950 100 03/13/89 10 N 10 6
Dickinson
Bowater Inc 9.000 300 08/02/89 20 N 9.331 7
Chrysler 10.300 300 06/15/90 2 N 10.43 11
Financial
Coastal 10.250 200 12/06/89 15 N 10.02 11
Corporation
Consolidated 9.125 150 08/17/89 10 N 9.202 6
Freightways
Corning Inc 8.750 100 07/13/89 10 N 8.655 7
CPC Intl 7.780 200 12/15/89 15 N 7.780 8
(continued)
180 I.E. Brick et al.
S&P
Amount Years to credit
Name Coupon ($million) Issue date maturity Callability YTM (%) rating
Cummins 9.750 100 03/21/86 30 C 10.276 9
Engine
Cyprus 10.125 150 04/11/90 12 N 10.172 10
Minerals
Dresser 9.373 68.8 03/10/89 11 C 9.650 8
Industries
Eaton Corp 9.000 100 03/18/86 30 C 9.323 7
Federal 9.200 150 11/15/89 5 N 9.206 9
Express
General 10.125 115 03/22/90 12 N 10.269 8
American
Transportation
Georgia Pacific 10.000 300 06/13/90 7 N 10.053 9
Grumman Corp 10.375 200 01/05/89 10 C 10.375 9
Harris Corp 10.375 150 11/29/88 30 C 10.321 8
Harsco 8.750 100 05/15/91 5 N 8.924 8
International 9.700 150 03/21/90 10 N 9.823 8
Paper
Kerr-Mcgee 9.750 100 04/01/86 30 C 9.459 8
Knight-Rydder 9.875 200 04/21/89 20 N 10.05 5
Lockhee Corp 9.375 300 10/15/89 10 N 9.329 7
Maytag 8.875 175 07/10/89 10 N 9.1 8
Monsanto 8.875 100 12/15/89 20 N 8.956 7
Morton 9.250 200 06/01/90 30 N 9.358 5
International
Parker- 9.750 100 02/11/91 30 C 9.837 7
Hannifin
Penn Central 9.750 200 08/03/89 10 N 9.358 11
Penn Central 10.875 150 05/01/91 20 N 11.016 11
Potlatch 9.125 100 12/01/89 20 N 9.206 8
Questar 9.875 50 06/11/90 30 C 9.930 6
Ralston Purina 9.250 200 10/15/89 20 N 9.45 8
Rite-Aid 9.625 65 09/25/89 27 C 9.99 6
Rohm And 9.373 100 11/15/89 30 C 9.618 7
Haas
Safety-Kleen 9.250 100 09/11/89 10 N 9.678 9
Sequa Corp 9.625 150 10/15/89 10 N 9.574 11
Stanley Works 8.250 75 04/02/86 10 C 8.174 7
Strawbridge 8.750 50 10/24/89 7 C 9.374 8
And Clothier
Union Camp 10.000 100 04/28/89 30 C 10.185 6
Unisys 10.300 300 05/29/90 7 N 10.794 10
(continued)
5 Motivations for Issuing Putable Debt: An Empirical Analysis 181
S&P
Amount Years to credit
Name Coupon ($million) Issue date maturity Callability YTM (%) rating
United Airlines 12.500 150 06/03/88 7 C 12.500 15
United 8.875 300 11/13/89 30 N 9.052 5
Technologies
VF Corp 9.500 100 10/15/89 10 C 9.500 7
Weyerhaeuser 9.250 200 11/15/90 5 N 9.073 6
The S&P ratings are based on a scale from 1 (AAA+) to 24 (Not rated). In the Callability column,
C denotes callable bond and N denotes non-callable bond
Fama et al. (1969) introduced the event study methodology when they analyzed
the impact of stock dividend announcements upon stock prices. Essentially,
they used the Market Model to estimate the stochastic relationship between
stock returns and the market portfolio. In particular, we estimate the following
regression:
We estimate the market model coefficients using the time period that begins
200 trading days before and ends 31 trading days before the event, employing the
CRSP value-weighted market index as the benchmark portfolio. We use these
coefficients to estimate abnormal returns for days 30 to +30. The abnormal return
is defined as
X
n
ARt ¼ ARi, t =n: (5.5)
i¼1
In order to conduct tests of significance, we must ensure that ARi,t has identical
standard deviation. Assuming that the forecast values are normally distributed, we
can scale ARi,t by the standard deviation of the prediction, Sit, given by Eq. 5.4 In
particular,
n h 2 2 io1=2
Sit ¼ s2i þ ð1=EDÞ þ Rmt Rm =S Rmi Rm (5.6)
182 I.E. Brick et al.
where s2i is the variance of the error term of Eq. 5.3, ED is the number of days
to estimate the market model for firm I, Rmt is the return of the market portfolio, and
Rm is the mean market return in the estimation period.
Hence the standardized abnormal return, SARi,t, is equal to ARi,t/Sit. SARi,t is
distributed normally with a mean of zero and a standard deviation equal to one.
The mean standardized abnormal return for time t, SARt, is the sum of the SARi,t
divided by n. SARt is normally distributed with a mean of zero and
a standard deviation of the square root of l/n. The cumulative abnormal returns for
days 1 to k, CARk, is the sum of mean abnormal returns for t ¼ 1 to k.
The standardized cumulative mean excess returns for the k days after month
t ¼ 0, SCARt, is equal to the sum of SARt for days 1 to k. SCART is normally
distributed with a standard deviation of square root of k/n. Please see Campbell
et al. (1996) for a detailed discussion of these tests as well as event studies in general.
In our cross-sectional regressions where we regress the CARs on firm-specific
variables, we estimate the model using GMM. However, since we are not
interested in conditional estimates, our regressors are the instruments. Therefore
our parameters estimates are the same that would be obtained by OLS. However,
we use the White (1980) heteroscedastic-consistent estimator. In the standard
regression model,
Y ¼ Xb þ e (5.7)
When the errors are homoscedastic, O ¼ s2I and the variance reduces to s2(X0 X)1.
However when O is unknown as shown by White (1980), Eq. 5.8 can be used
using a consistent estimator of O. White showed that can be done using the residual
from Eq. 5.7.
So when the heteroscedasticity is of the unknown form,
1 1
Using ðX0 XÞ X0 diag e2 X ðX0 XÞ (5.9)
The law of large numbers ensures that the RHS of above equation is the same as
The sample GMM estimator of the parameters may be written as (see Hansen
1982)
h X i0 X
Y ¼ arg min 1=T f ðYt ; yÞ WT 1=T f ðYt ; yÞ: (5.13)
So essentially GMM finds the values of the parameters so that the sample
moment conditions are satisfied as closely as possible. In our case for the regression
model,
yt ¼ Xt 0 b þ et : (5.14)
References
Bae, S. C., Klein, D. P., & Padmaraj, R. (1994). Event risk bond covenants, agency costs of debt
and equity, and stockholder wealth. Financial Management, 23, 28–41.
Bae, S. C., Klein, D. P., & Padmaraj, R. (1997). Firm characteristics and the presence of event risk
covenants in bond indentures. Journal of Financial Research, 20, 373–388.
Barnea, A., Haugen, R. A., & Senbet, L. W. (1980). A rationale for debt maturity structure and call
provisions in the agency theoretic framework. Journal of Finance, 25, 1223–1234.
Billett, M. T., King, T. D., & Mauer, D. C. (2007). Growth opportunities and the choice of
leverage, debt maturity, and covenants. Journal of Finance, 62, 697–730.
Bodie, Z., & Taggart, R. A. (1978). Future investment opportunities and the value of the call
provision on a bond. Journal of Finance, 33, 1187–1200.
Brennan, M. J., & Schwartz, E. S. (1988). The case for convertibles. Journal of Applied Corporate
Finance, 1, 55–64.
Brick, I. E., & Palmon, O. (1993). The tax advantages of refunding debt by calling, repurchasing,
and putting. Financial Management, 22, 96–106.
Brick, I. E., & Wallingford, B. A. (1985). The relative tax benefits of alternative call features in
corporate debt. Journal of Financial and Quantitative Analysis, 20, 95–106.
Campbell, J., Lo, A. W., & MacKinlay, C. (1996). The econometrics of financial markets.
Princeton: Princeton University Press.
Chatfield, R. E., & Moyer, R. C. (1986). Putting away bond risk: An empirical examination of the
value of the put option on bonds. Financial Management, 15(2), 26–33.
Cook, D. O., & Easterwood, J. C. (1994). Poison put bonds: An analysis of their economic role.
Journal of Finance, 49, 1905–1920.
Crabbe, L. (1991). Event risk: An analysis of losses to bondholders and ‘super poison put’ bond
covenants. Journal of Finance, 46, 689–706.
Crabbe, L., & Nikoulis, P. (1997). The putable bond market: Structure, historical experience, and
strategies. The Journal of Fixed Income, 7, 47–60.
Dann, L. Y., & Mikkelson, W. H. (1984). Convertible debt issuance, capital structure change and
financing-related information: Some new evidence. Journal of Financial Economics, 13,
157–186.
David, A. (2001). Pricing the strategic value of putable securities in liquidity crises. Journal of
Financial Economics, 59, 63–99.
Eckbo, B. E. (1986). Valuation effects of corporate debt offerings. Journal of Financial Econom-
ics, 5, 119–152.
Fama, E., Fisher, L., Jensen, M., & Roll, R. (1969). The adjustment of stock prices to new
information. International Economic Review, 10, 1–21.
Green, R. (1984). Investment incentives, debt, and warrants. Journal of Financial Economics, 13,
115–136.
Greene, W. (2008). Econometric analysis. Boston: Pearson.
Hansen, L. P. (1982). Large sample properties of generalized method of moments estimators.
Econometrica, 50, 1029–1054.
Hansen, R. S., & Chaplinsky, S. (1993). Partial anticipation, the flow of information and the
economic impact of corporate debt sales. Review of Financial Studies, 6, 709–732.
Hansen, R. S., & Crutchley, C. (1990). Corporate earnings and financings: An empirical analysis.
Journal of Business, 63, 347–371.
Haugen, R. A., & Senbet, L. W. (1978). The insignificance of bankruptcy costs to the theory of
optimal capital structure. Journal of Finance, 33, 383–393.
Haugen, R. A., & Senbet, L. W. (1981). Resolving the agency problems of external capital through
options. Journal of Finance, 26, 629–647.
Haugen, R. A., & Senbet, L. W. (1988). Bankruptcy and agency costs: Their significance
to the theory of optimal capital structure. Journal of Financial and Quantitative Analysis,
23, 27–38.
5 Motivations for Issuing Putable Debt: An Empirical Analysis 185
Jensen, M. C. (1986). Agency costs of free cash flow, corporate finance, and takeovers. American
Economic Review, 76, 323–329.
Jensen, M. C., & Meckling, W. H. (1976). Theory of the firm: Managerial behavior, agency costs
and ownership structure. Journal of Financial Economics, 3, 305–360.
Jung, K., Kim, Y., & Stulz, R. M. (1996). Timing, investment opportunities, managerial discretion,
and the security issue decision. Journal of Financial Economics, 42, 159–185.
Kim, E. H. (1978). A mean-variance theory of optimal capital structure and corporate debt
capacity. Journal of Finance, 33, 45–63.
Lehn, K., & Poulsen, J. (1989). Free cash flow and stockholders gain in going private transactions.
Journal of Finance, 44, 771–788.
Lewis, C., Rogalski, R., & Seward, J. K. (1998). Industry conditions, growth opportunities and the
market reactions to convertible debt financing decisions. (Unpublished mimeo, Vanderbilt
University).
Mackie-Mason, J. K. (1990). Do taxes affect corporate financing decisions? Journal of Finance,
45, 1471–1493.
Meulbroek, L. K., Mitchell, M. L., Mulherin, J. H., Netter, J. M., & Poulsen, A. B. (1990). Shark
repellants and managerial myopia: An empirical test. Journal of Political Economy, 98,
1108–1117.
Mikkelson, W. H., & Partch, M. M. (1986). Valuation effects of security offerings and the issuance
process. Journal of Financial Economics, 15, 31–60.
Miller, M. H. (1977). Debt and taxes. Journal of Finance, 32, 261–275.
Miller, M. H., & Rock, K. (1985). Dividend policy under asymmetric information. Journal of
Finance, 40, 1031–1051.
Modigliani, F., & Miller, M. H. (1963). Corporate income taxes and the cost of capital:
A correction. American Economic Review, 63, 433–443.
Myers, S. (1977). Determinants of corporate borrowing. Journal of Financial Economics, 5,
147–175.
Nash, R. C., Netter, J. M., & Poulsen, A. B. (2003). Determinants of contractual relations between
shareholders and bondholders: Investment opportunities and restrictive covenants. Journal of
Corporate Finance, 9, 201–232.
Ravid, S. A., & Sarig, O. H. (1991). Financial signaling by committing to cash outflows. Journal of
Financial and Quantitative Analysis, 26, 165–180.
Robbins, E. H., & Schatzberg, J. D. (1986). Callable bonds: A risk-reducing signaling mechanism.
Journal of Finance, 41, 935–950.
Ross, S. (1977). The determination of financial structure: The incentive signaling approach. Bell
Journal of Economics, 8, 23–40.
Roth, G., & McDonald, C. G. (1999). Shareholder-management conflict and event risk covenants.
Journal of Financial Research, 22, 207–225.
Scott, J. (1976). A theory of optimal capital structure. Bell Journal of Economics, 7, 33–54.
Shyam-Sunder, L. (1991). The stock price effect of risky versus safe debt. Journal of Financial
and Quantitative Analysis, 26, 549–558.
Spiegel, Y., & Wilkie, S. (1996). Investment in a new technology as a signal of firm value under
regulatory opportunism. Journal of Economics and Management Strategy, 5, 251–276.
Stein, J. C. (1988). Takeover threats and managerial myopia. Journal of Political Economy, 96,
61–80.
Tewari, M., & Ramanlal, P. (2010). Is the put option in U.S. structured bonds good news for both
bondholders and stockholders? International Research Journal of Finance and Economics, 52,
50–59.
Wahal, S., & McConnell, J. J. (2000). Do institutional investors exacerbate managerial
myopia. Journal of Corporate Finance: Contracting, Governance and Organization, 6,
307–329.
White, H. (1980). A heteroskedasticity-consistent covariance matrix estimator and a direct test for
heteroskedasticity. Econometrica, 48, 817–838.
Multi-Risk Premia Model of US Bank
Returns: An Integration of CAPM and APT 6
Suresh Srivastava and Ken Hung
Contents
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 188
6.2 Multiple Regression Model of Bank Return . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 189
6.2.1 Pricing of Interest Rate Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 190
6.3 Multi-Risk Premia Asset-Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 192
6.4 Data Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 193
6.5 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 195
6.5.1 Two-Variable Regression Model and Pricing of Interest Rate Risk . . . . . . . . . . . . 195
6.5.2 Multi-risk Premia Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 197
6.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 200
Appendix 1: An Integration of CAPM and APT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
Principal Component Factor Specification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201
Appendix 2: Interest Rate Innovations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 203
Orthogonalization Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
Univariate ARMA Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 204
Vector ARMA Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 205
Abstract
Interest rate sensitivity of bank stock returns has been studied using an
augmented CAPM, a multiple regression model with market returns and interest
rate as independent variables. In this paper, we test an asset-pricing model in
which the CAPM is augmented by three orthogonal factors which are proxies for
the innovations in inflation, maturity risk, and default risk. The model proposed
S. Srivastava (*)
University of Alaska Anchorage, Anchorage, AK, USA
e-mail: afscs@uaa.alaska.edu
K. Hung
Division of International Banking & Finance Studies, Texas A&M International University,
Laredo, TX, USA
e-mail: ken.hung@tamiu.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 187
DOI 10.1007/978-1-4614-7750-1_6,
# Springer Science+Business Media New York 2015
188 S. Srivastava and K. Hung
is an integration of CAPM and APT. The results of the two models are compared
to shed light on sources of interest rate risk.
Our results using the integrated model indicate the inflation beta to be
statistically significant. Hence, innovations in short-term interest rates contain
valuable information regarding inflation premium; as a result the interest rate risk
is priced with respect to the short-term interest rates. Further, it also indicates that
innovations in long-term interest rates contain valuable information regarding
maturity premium. Consequently the interest rate risk is priced with respect to
the long-term interest rates. Using the traditional augmented CAPM, our investi-
gation of the pricing of the interest rate risk is inconclusive. It shows that interest
rate risk was priced from 1979 to 1984 irrespective of the choice of interest rate
variable. However, during the periods 1974–1978 and 1985–1990, bank stock
returns were sensitive only to the innovations in the long-term interest rates.
Keywords
CAPM • APT • Bank stock return • Interest rate risk • Orthogonal factors •
Multiple regression
6.1 Introduction
Interest rate sensitivity of commercial bank stock returns has been the subject of
considerable academic research. Stone (1974) proposed a multiple regression
model incorporating both the market return and interest rate variables as return-
generating independent variables. While some studies have found the interest rate
variable to be an important determinant of common stock returns of banks (Fama
and Schwert 1977; Lynge and Zumwalt 1980; Christie 1981; Flannery and James
1984; Booth and Officer 1985), others have found the returns to be insensitive
(Chance and Lane 1980) or only marginally explained by the interest rate factor
(Lloyd and Shick 1977). A review of the early literature can be found in Unal and
Kane (1988). Sweeney and Warga (1986) used the APT framework and concluded
that the interest rate risk premium exists but varies over time. Flannery et al. (1997)
tested a two-factor model for a broad class of security returns and found the effect
of interest rate risk on security returns to be rather weak. Bae (1990) examined the
interest rate sensitivity of depository and nondepository firms using three different
maturity interest rate indices. His results indicate that depository institutions’ stocks
are sensitive to actual and unexpected interest rate changes, and the sensitivity
increases for longer-maturity interest rate variables. Song (1994) examined the
two-factor model using time-varying betas. His results show that both market
beta and interest rate beta varied over the period 1977–1987. Yourougou (1990)
found the interest rate risk to be high during a period of great interest rate volatility
(post-October 1979) but low during a period of stable interest rates (pre-October
1979). Choi et al. (1992) tested a three-factor model of bank stock returns using
market, interest, and exchange rate variables. Their findings about interest rate risk
are consistent with the observations of Yourougou (1990).
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 189
The issue of interest rate sensitivity remains empirically unresolved. Most of the
studies use a variety of short-term and long-term bond returns as the interest rate
factor without providing any rationale for their use. The choice of bond market
index seems to affect the pricing of the interest rate risk. Yet, there is no consensus
on the choice of the interest rate factor that should be used in testing the two-factor
model. In this paper, we provide a plausible explanation of why pricing of interest
rate risk differs with the choice of interest rate variable. We also suggest a hybrid
return-generating model for bank stock returns in which the CAPM is augmented
by three APT-type factors to account for unexpected changes in the inflation
premium, the maturity-risk premium, and the default-risk premium. The use of
three additional factors provides a better understanding of the interest rate sensi-
tivity and offers a plausible explanation for the time-varying interest rate risk
observed by other investigators. Our empirical investigation covers three distinc-
tion economic and bank regulatory environments: 1974–1978, a period of increas-
ing but only moderately volatile interest rates in a highly regulated banking
environment; (2) 1979–1984, a period characterized by high level of interest rates
with high volatility, in which there was gradual deregulation of the banking
industry; and (3) 1985–1990, a low interest rate and low-volatility period during
which many regulatory changes were made in response to enormous bank loan
losses and bankruptcies. The results of the multifactor asset-pricing model are
compared with those from the two-factor model in order to explain the time-
varying interest rate risk.
The rest of this paper is divided into five sections. In Sect. 6.2, we
describe the two-factor model of the bank stock return and the pricing of the
interest rate risk. The multi-risk premia model and the specification of the
factors are discussed in Sect. 6.3. The data for this analysis is described in
Sect. 6.4. Section 6.5 presents empirical results, and Sect. 6.6 concludes the
paper.
where Rjt is the bank common stock return, Rmt is the market return, and RIt is the
innovation in the interest rate variable. Coefficients aj and b1j are analogous to the
alpha and beta coefficients of the market model, and b2j represents interest rate risk.
Since then, numerous researchers have studied the pricing of interest rate risk with
varying results. While Stone (1974) and others did not place an a priori restriction
on the sign of b2j, the nominal contracting hypothesis implies that it should
be positive. This is because the maturity of bank assets is typically longer than
190 S. Srivastava and K. Hung
that of liabilities.1 Support for this hypothesis was found by Flannery and James
(1984) but not by French et al. (1983).
An important issue in the empirical investigation of the two-factor model is the
specification of an appropriate interest rate factor. Theoretical consideration of
factor analysis requires that two factors, Rmt and RIt, be orthogonal whereby choice
of the second factor (RIt) would not influence the first factor loading (blj). The
resolution of this constraint requires a robust technique for determining the unex-
pected changes in the interest rate that is uncorrelated with the market return. There
are three approaches to specify the interest rate factor (Appendix 2). In the first
approach, the expected change in the interest rate is estimated using the high
correlation between the observed interest rate and the market rate. The residual –
difference between observed and estimated rates – is used as the interest rate factor.
The second approach is to identify and estimate a univariate ARMA model for the
interest rate variable and use the residuals from the ARMA model as the second
factor. In the third approach, the interest rate variable (RIt) and the market return
(Rmt) are treated as the components of a bivariate vector, which is modeled as
a vector ARMA process. The estimated model provides the unanticipated change in
interest rate variable to be used as the second factor in the augmented CAPM,
Eq. 6.1. Srivastava et al. (1999) discuss the alternate ways of specifying the
innovations in the interest rate variable and its influence on the pricing of the
interest rate risk. In this paper, the error term from the regression of interest rates
on market returns is used as the orthogonal interest rate factor in Eq. 6.1.
where Rrf is the real risk-free rate plus an inflation premium, DRP is the default-risk
premium, MRP is the maturity-risk premium, and LP is the liquidity-risk premium.
1
The sign of b2j is negative when changes in bond yields and not the bond market return are used as
the interest rate variable (see Sweeney and Warga 1986).
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 191
where the debt security within the brackets identifies the choice of the bond market
index. A number of researchers have indicated that bank stock returns are sensitive
to unexpected changes in long-term but not short-term interest rates. This observa-
tion is consistent with the expectations expressed in Eq. 6.3. However, we are
unable to isolate and identify which component of the interest rate risk is being
2
Federal Reserve’s intervention changes the short-term interest rates. These rate changes take
place after considerable deliberation and are often anticipated by financial markets. Hence, it
neither generates interest rate innovations nor produces interest rate risk.
3
Liquidity premium is ignored in our discussion and subsequent model construction because our
sample does not include financially distressed banks, so there are insufficient variations in the
liquidity premium.
192 S. Srivastava and K. Hung
where DPt, DMRt, and DDRt are the proxies for the innovations in inflation,
maturity risk, and default risk, respectively (specification of these principal
component factors consistent with APT is discussed in the Appendix). Coeffi-
cients b1, b2P, b2M, and b2D are the measures of market risk, inflation risk,
maturity risk, and default risk, respectively. The expected return is given by
CAPM:
E Rj ¼ aj þ b1j EðRmÞ (6.5)
4
An appropriate interest rate variable that should be used to examine the pricing of the interest rate
risk in the two-variable framework is not easily available. However, one could construct an index
composed of long-term US bonds and corporate bonds with duration equal to the net duration of
bank assets and appropriate default risk. This interest rate variable will identify the true pricing of
the interest rate risk.
5
This model’s conceptual framework is provided by Chen et al. (1986). However, their factors are
not orthogonal.
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 193
The three APT-type factors have managerial implications regarding bank’s asset
and liability management. The first factor, DPt, has implications pertaining to the
management of the treasury securities portfolio held by the bank. The value of this
portfolio is sensitive to the changes in the short-term interest rate. In addition, these
changes impact the short-maturity funding gap. The null hypothesis b2P ¼ 0 implies
that management has correctly anticipated future changes in short-term inflation
and has taken steps to correctly hedge the pricing risk through the use of derivative
contracts and minimize the short-maturity funding gap.
The second factor, DMRt, has implications regarding the bank’s long-term assets
and liabilities. The market value of a bank’s net worth is very sensitive to changes
in the slope of the term structure. The null hypothesis b2M ¼ 0 conjectures that
management has correctly anticipated future changes in the slope of the term
structure and has immunized the institution’s net worth by a sensible allocation
assets in the loan portfolio, sale of loans to the secondary markets, securitization of
loans, and the use of derivative contracts. The third factor, DDRt, relates to the
management of loan losses and overall default-risk rating of bank assets. The null
hypothesis b2D ¼ 0 infers that management has correctly anticipated future loan
losses due to changes in exogenous conditions, and subsequent loan losses,
however large, will not adversely affect the stock returns.
Month-end yield for 3-month T-bill, 6-month T-bill, 1-year T-note, 7-year T-note,
and BAA corporate bonds, for the period January 1974 to December 1990, were
obtained, and monthly returns were calculated.6 Return on the CRSP equally
weighted index of NYSE stocks was used as the market return. Month-end closing
prices and dividends for a sample of 88 banks were obtained from Compustat’s
Price, Dividend, and Earning (PDE) data tape, and monthly returns were calculated.
Availability of continuous data was the sole sample selection criteria. This selection
criterion does introduce a survivorship bias. However, it was correctly pointed out
by Elyasiani and Iqbal (1998) that the magnitude of this bias could be small and not
affect the pricing of interest risk. Equally weighted portfolios of bank returns were
calculated for this study. The total observation period is divided into three
contrasting economic and bank regulatory periods: (1) an increasing but moderately
volatile interest rate period from January 1974 to December 1978 in a highly
regulated environment; (2) a high interest rate and high-volatility period from
January 1979 to December 1984, during which there was gradual deregulation of
the industry; and (3) a low interest rate and low-volatility period from January 1985
to December 1990, during which many regulatory changes were made in response
to banks loan loss problems. The descriptive statistics of the sample are
6
The return on a bond index is calculated from the yield series as RIt ¼ (YIt YI,t1)/YIt
where YIt is the bond index yield at time t.
194 S. Srivastava and K. Hung
Table 6.1 Summary statistics of portfolio returns and interest rate yields
1974–1978 1979–1984 1985–1990
Bank portfolio return 9.66%a 22.28% 6.48%
(5.95%)b (4.93%) (6.17%)
Portfolio betac 0.7091 0.7259 1.0102
CRSP market return 22.05% 22.05% 8.58%
(7.63%) (5.39%) (5.36%)
3-month T-bill yield 6.21% 10.71% 6.92%
(1.27%) (2.41%) (1.00%)
6-month T-bill yield 6.48% 10.80% 7.02%
(1.27%) (2.21%) (0.95%)
1-year T-note yield 7.05% 11.70% 7.62%
(1.25%) (2.27%) (1.00%)
7-year T-note yield 7.23% 11.91% 8.67%
(0.54%) (1.79%) (1.09%)
BAA bond yield 9.66% 14.04% 10.84%
(0.68%) (1.97%) (0.98%)
Monthly inflationd 0.78% 0.46% 0.25%
(0.69%) (0.50%) (0.23%)
Maturity-risk premiume 0.68% 0.20% 1.05%
(0.88%) (1.32%) (0.72)
Default-risk premiumf 1.94% 2.13% 2.17%
(0.77%) (0.76%) (0.46%)
a
Average returns and yields are reported on annualized basis
b
Standard deviation is in parenthesis
c
Estimated using single-index market model
d
Measured by the relative change in consumer price index
e
Yield differential between 7-year T-note and 1-year T-note
f
Yield differential between BAA-rated corporate bond and 7-year T-note
Table 6.2 Two-variable model of the bank stock returns: Rjt ¼ aj + b1jRmt + b2jRIt + ejt
Interest rate variablea
3-month 6-month 1-year 7-year BAA
Period Estimates T-bill T-bill T-note T-note Bond
1974–1978 Constant, a 0.0046 0.0045 0.0043 0.0033 0.0042
(0.18)b (0.17) (0.16) (0.14) (0.16)
Market beta, b1 0.6921 0.6886 0.6897 0.6845 0.6987
(15.05) (14.67) (15.17) (16.47) (15.76)
Interest rate beta, b2 0.0566 0.0608 0.0700 0.2893 0.2248
(1.01) (1.07) (1.24) (2.77) (0.93)
R2 0.828 0.828 0.830 0.845 0.828
F-statistic 137.3 137.7 138.8 156.3 136.9
1979–1984 Constant, a 0.0052 0.0052 0.0052 0.0052 0.0052
(1.49) (1.53) (1.57) (1.70) (1.62)
Market beta, b1 0.7261 0.7258 0.7262 0.7262 0.7264
(11.73) (12.01) (12.37) (13.33) (12.74)
Interest rate beta, b2 0.1113 0.1425 0.1714 0.3582 0.6083
(3.47) (3.99) (4.41) (5.97) (5.15)
R2 0.684 0.699 0.716 0.755 0.732
F-statistic 74.8 80.1 87.2 106.6 94.4
1985–1990 Constant, a 0.0018 0.0018 0.0018 0.0018 0.0018
(0.51) (0.51) (0.52) (0.53) (0.52)
Market beta, b1 1.010 1.010 1.010 1.010 1.010
(15.31) (15.30) (15.45) (15.70) (15.57)
Interest rate beta, b2 0.0508 0.0374 0.1036 0.1752 0.2741
(0.51) (0.40) (1.22) (1.95) (1.62)
R2 0.773 0.772 0.777 0.794 0.780
F-statistic 117.3 117.1 120.0 125.1 122.5
a
The error term from the regression of interest rate on market return is used as the appropriate
orthogonal interest rate variable
b
t-statistics are in the parenthesis
7
Test of two-variable model for the period 1991–2007 indicated that interest rate risk is not priced.
Tables can be provided to interested readers.
196 S. Srivastava and K. Hung
the interest rate beta was statistically significant with all the interest factors, and
its magnitude varied substantially. This period was characterized by a relatively
flat but volatile yield curve. For the period 1985–1990, the interest rate beta was
statistically significant only with the 7-year T-note returns as the interest rate
factor. In general, the estimated value of b2j for the periods 1974–1978 and
1985–1990 is smaller and less statistically significant than the value for the period
1979–1984.
The magnitude of interest rate beta, whether significant or not, varies with the
choice of the interest rate variable. The size of b2j increases with the choice of
securities in the following order: 3-month T-bill, 6-month T-bill, 1-year T-note,
7-year T-note, and BAA-rated bond (except b2j estimated using BAA bond return
for the period 1974–1978 and the 6-month T-bill return in 1985–1990). This
observation validates inequality (6.3) in all the periods and suggests the expanding
nature of the investment opportunity set with increased horizon (Merton 1973). As
stated earlier, the difference between b2j estimated using 7-year T-note returns and
that using 3-month T-bill returns measures the effect on the bank stock returns due
to the unexpected changes in the maturity-risk premium. Further, difference
between b2j estimated using BAA bond returns and that using 7-year T-note returns
measures the effect on the bank stock returns due to the unexpected changes in the
default-risk premium.
The fact that bank stock returns are more sensitive to the long-term interest rates
than to short-term interest rates is consistent with our expectation about the size of
b2 expressed in inequalities (6.3). Similar results were reported by other researchers
(such as Unal and Kane (1988) and Chen et al. (1986)). A shift of focus from short-
term to long-term inflation expectation could explain this result. An alternative
explanation is that bank balance sheet returns are better approximated by long-term
than by short-term bond returns. To the extent that balance maturity mismatches
occur, they should be related to long-term bond returns. The reason is simply that
long-term bond returns include the present value of more future period returns than
do short-term bond returns. That is, long-term bond returns include price changes
not included in short-term bond returns. The price changes in the long-term bond
returns represent price changes in long-term bank contracts. The most representa-
tive term equals the term of assets or liabilities, whichever is longer. If the maturity
mismatch is a net asset (liability) position, then the long-term bond maturity would
reflect the asset (liability) maturity. The estimate of a large, positive coefficient (b2j)
for 7-year T-notes and BAA bonds implies that banks mismatch in favor of assets
with relatively long maturities.
The plausible causes for the change in the interest risk from period to period are
(1) changes in real returns, as reported by Fama and Gibbons (1982), (2) unexpected
changes in short-term and long-term inflation expectation, (3) shift of focus from
short-term to long-term inflation expectation, (4) unexpected changes in risk aver-
sion, (5) unexpected changes in the default risk of bank’s nominal contracts, and
(6) structural instability of the systematic interest rate equation used to extract
interest rate innovations. The estimated coefficients of multi-risk premia model will
shed some light on this issue.
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 197
The estimates of the first systematic interest rate Eq. 6.11, which specifies innovations
in the short-term default-free bond returns, are reported in Table 6.3. The ordinary
least square (OLS) estimates were rejected because of the presence of serial auto-
correlation as indicated by the Durbin-Watson statistic (see Woolridge 2009).
A Durbin-Watson statistic equal to 2 indicates the absence of any serial autocorre-
lation. The generalized least square (GLS) estimates seemed more appropriate
because the Durbin-Watson statistic was approximately equal to 2 and the value of
R2 was higher. Results reported in Table 6.3 exhibit a significant correlation between
market return and short-term interest rate variable in the period 1974–1978, but not in
the periods 1979–1984 and 1985–1990. However, a low value of R2 indicates that the
relationship expressed in Eq. 6.4 is not as robust as one would have preferred. This
approach of extracting changes in short-term inflation was used by Fama and Schwert
(1977) and French et al. (1983). Their estimation period overlapped our estimation
period 1974–1978 but not the later ones. In Table 6.4 we present the results of the
second systematic interest rate Eqs. 6.12a and 6.12b. The estimated coefficient, y1,
determined the average slope of the yield curve during the estimation period. The OLS
estimates were rejected because of the presence of serial correlation indicated by the
Durbin-Watson statistics. The term structure coefficient y1 was significant in all the
periods for Specifications a and b. However, coefficients y2 of Specification b was
Table 6.3 Estimating a proxy for ex-post unexpected inflation: RST,t ¼ d0 + d1Rmt + et
Estimated coefficients
1974–1978 1979–1984 1985–1990
Ordinary least square
Constant, d0 0.0068 0.0070 0.0037
(0.87)a (0.45) (0.89)
Market linkage coefficient, d1b 0.2987 0.5791 0.0891
(2.97) (2.19) (1.13)
R2 0.132 0.076 0.018
Durbin-Watson statistic 1.90 1.19 1.12
Generalized least square
Constant, d0 0.0075 0.0013 0.0044
(0.82) (0.08) (0.83)
Market linkage coefficient, d1b 0.3173 0.3136 0.0806
(3.16) (1.44) (1.15)
R2 0.145 0.306 0.242
Durbin-Watson statistic 1.99 1.97 1.92
Ordinary Least Square estimates were rejected because of the presence of serial correlation
indicated by Durbin-Watson statistic
RST,t and Rmt are 3-month T-bill and CRSP equally-weighted market returns
a
t-statistics are in the parenthesis
b
Coefficient d1 measures the stock–bond market linkage
198 S. Srivastava and K. Hung
Table 6.4 Estimating alternate proxies for maturity-risk premia: Specification a : RLT, t ¼ y0 þ
y1 RST, t þ et , Specification b : RLT, t ¼ y0 þ y1 RST, t þ y2 Rmt þ et
1974–1978 1979–1984 1985–1990
Specification a: generalized least square
Constant, y0 0.0028 0.0053 0.0017
(1.02)a (2.11) (0.66)
Term structure coefficient, y1b 0.3988 0.5201 0.8183
(10.30) (16.29) (14.96)
R2 0.675 0.846 0.820
Durbin-Watson statistic 1.98 1.83 1.84
Specification b: generalized least square
Constant, y0 0.0024 0.0064 0.0011
(0.84) (2.47) (0.48)
Term structure coefficient, y1b 0.4070 0.5122 0.8105
(10.02) (15.78) (15.34)
Market linkage coefficient, y2c 0.0216 0.0576 0.0944
(0.68) (1.07) (2.56)
R2 0.677 0.849 0.835
Durbin-Watson statistic 1.98 1.82 1.80
Ordinary least square estimates were rejected because of the presence of serial correlation
indicated by Durbin-Watson statistic
RLT,t and RST,t are 7-year T-note and 1-year T-note returns
a
t-statistics are in the parenthesis
b
Coefficient y1 measures the average slope of the yield curve
c
Coefficient y2 accounts for the stock–bond market linkage
significant only during 1985–1990. The errors from Specification a for periods
1974–1978 and 1979–1984 and from Specification b for the period 1985–1990 were
used to specify the second risk factor DMRt. Estimates of the third systematic interest
rate Eqs. 6.13a and 6.13b are presented in Table 6.5. As before, the GLS estimates
were deemed appropriate. The errors from Specification a for the periods 1974–1978
and 1979–1984 and from Specification b for the period 1985–1990 were used to
specify the unexpected change in default risk, DDRt.
Results of the multi-risk premia model are presented in Table 6.6. The inflation
beta is statistically significant in the periods 1979–1984 and 1985–1990 but not in
the period 1974–1978. It was shown in Table 6.3 that quasi-differenced short-term
interest rates were correlated with the market return for the period 1974–1978 but
not for the periods 1979–1984 and 1985–1990. Hence, one could argue that when
short-term interest rates are correlated with the market return (i.e., 1974–1978), the
error term from Eq. 6.8 contains no systematic information. This results in the
inflation beta being insignificant and the interest rate risk not priced with respect to
the short-term rates within the context of the two-factor model. A corollary is that,
when short-term interest rates are uncorrelated with the market return (i.e.,
1979–1984 and 1985–1990), the error term from Eq. 6.11 contains valuable infor-
mation leading to the inflation beta being significant. The maturity beta was found
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 199
Table 6.5 Estimating alternate proxies for the default-risk premia: Specification a : RBAA, t ¼
j0 þ j1 RLT, t þ et , Specification b : RBAA, t ¼ j0 þ j1 RLT, t þ j2 Rmt þ et
1974–1978 1979–1984 1985–1990
Specification a: generalized least square
Constant, j0 0.0021 0.0039 0.0010
(0.60)a (1.12) (0.66)
Default-risk coefficient, j1b 0.0842 0.4129 0.4734
(2.27) (12.27) (13.47)
R2 0.553 0.800 0.772
Durbin-Watson statistic 1.88 1.97 1.99
Specification a: generalized least square
Constant, j0 0.0019 0.0051 0.0006
(0.54) (1.41) (0.48)
Term structure coefficient, j1b 0.0861 0.3976 0.4637
(2.30) (11.35) (13.83)
Market linkage coefficient, j2c 0.0066 0.0491 0.0577
(0.49) (1.49) (2.39)
R2 0.555 0.806 0.785
Durbin-Watson statistic 1.87 1.97 2.00
Ordinary least square estimates were rejected because of the presence of serial correlation
indicated by Durbin-Watson statistic
RBAA,t and RLT,t are BAA corporate bond and 7-year T-note returns
a
t-statistics are in the parenthesis
b
Coefficient j1 measures the average yield differential between RBAA,t and RLT,t
c
Coefficient j2 accounts for the stock–bond market linkage
to be statistically significant in the periods 1974–1978 and 1979–1984 but not for
the period 1985–1990. Results in Table 6.4 (Specification b) showed that long-term
interest rates were correlated with the market return for the period 1985–1990 but
not for the periods 1974–1978 and 1979–1984. Hence, we posit that when long-
term interest rates are correlated with the market return (i.e., 1985–1990), the error
term from Eq. 6.12b contains no systematic information. This results in the maturity
beta being insignificant. A corollary is that, when long-term interest rates are
uncorrelated with the market return (i.e., 1974–1978 and 1979–1984), the error
term from Eq. 6.12b contains valuable information producing a significant maturity
beta and the interest rate risk is priced with respect to the long-term rates within the
context of the two-factor model. The default beta was found to be statistically
significant in the period 1985–1990 but not for the periods 1974–1978 and
1979–1984. The economic factors that lead to significant correlation between
market returns and long-term interest rates (Eq. 6.12b) or between market returns
and BAA-rated bond returns (Eq. 6.13b) caused the interest rate risk to be priced
with respect to the long-term rates within the context of the two-factor model
(1985–1990). Since the correlation between market return and interest rate changes
over time, the interest rate risk also changes over time.
200 S. Srivastava and K. Hung
Table 6.6 Multi-risk premia model of the bank stock returns: Rjt ¼ a0j + b1jRmt + b2QjDPt +
b2MjDMRt + b2DjDDRt + ejt
Estimated coefficientsa
1974–1978 1979–1984 1985–1990
Constant, a0 0.0056 0.0041 0.0047
(1.84)b (1.04) (1.31)
Market beta, b1 0.7397 0.6833 0.7185
(16.94) (10.55) (17.55)
Inflation beta, b2P 0.0698 0.1292 0.2157
(1.27) (3.45) (2.29)
Maturity beta, b2M 0.5350 0.4881 0.2476
(2.87) (2.93) (1.38)
Default beta, b2D 0.1973 0.1153 0.5696
(0.56) (0.43) (1.94)
R2 0.849 0.754 0.851
Durbin-Watson statistic 2.00 2.04 1.98
Ordinary least square estimates were rejected because of the presence of serial correlation
indicated by Durbin-Watson statistic
Portfolio beta estimated using single-index model are 0.7091, 0.7259, and 1.0102 for the periods
1974–1978, 1979–1984, and 1985–1990, respectively
a
Generalized least square estimates
b
t-statistics are in the parenthesis
Negative inflation and maturity betas for the period 1985–1990 need some
explanation because it is contrary to a priori expectation. For the period
1985–1990, the bank portfolio and market return both dropped dramatically to
6.48 % and 8.58 %, respectively. However, the estimated portfolio beta increased
from about 0.7 in the two earlier periods to about 1.0 in this period (beta estimated
independently by the single-index model). Consequently, the market factor alone
will overestimate the bank portfolio’s expected return. The negative values of
inflation beta and maturity beta (though insignificant) correct the overestimation.
Economic factors and regulatory changes that fuelled M&A activities during this
period must have been such that they increased the portfolio beta without increasing
the ex-post portfolio return. Some of these unidentifiable factors are negatively
correlated with the interest rates. One of the shortcomings of the factor analytic
approach is that factors are at times unidentifiable. In spite of difficulties in
explaining some of the results for the period 1985–1990, the multi-risk premia
model does provide greater insight into the pricing of interest rate risk.
6.6 Conclusions
In this paper, we examine the interest rate sensitivity of commercial bank returns
covering three distinct economic and regulatory environments. First, we investigate
the pricing of the interest rate risk within the framework of the two-factor model.
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 201
Our results indicate that interest rate risk was priced during 1979–1984 irrespective
of the choice of interest rate variable. However, during the periods 1974–1978 and
1985–1990, bank stock returns were sensitive only to the unexpected changes in the
long-term interest rates. Next, we tested to an asset-pricing model in which the
traditional CAPM is augmented by three additional factors to account for unex-
pected changes in the inflation, the maturity premium, and default premium. Our
results show that the inflation beta was significant for the periods 1979–1984 and
1985–1990, but not for the period 1974–1978; the maturity beta was significant for
the periods 1974–1978 and 1979–1984 but not for the period 1985–1990; and the
default beta was significant for the period 1985–1990 but not for the periods
1974–1978 and 1978–1984.
We can infer that when short-term interest rates are correlated with the
market return, the innovations in short-term interest rate are indeed white
noise. However, innovations in short-term interest rates contain valuable
information when short-term interest rates are uncorrelated with the market
return. This will lead to a significant inflation beta and the interest rate risk
will be priced with respect to the short-term rates within the context of the
two-factor model. We can also infer that when long-term interest rates are
correlated with the market return, the innovations in long-term interest rate
are indeed white noise. However, innovations in long-term interest rates
contain valuable information when long-term interest rates are uncorrelated
with the market return. This results in a significant maturity beta and priced
interest rate risk with respect to the long-term rates within the context of the
two-variable model.
Rjt ¼ aj þ b1j Rmt þ b2Pj DPt þ b2Mj DMRt þ b2Dj DDRt þ ejt (6.7)
where DPt, DMRt, and DDRt are the proxies for the innovations in inflation,
maturity risk, and default risk, respectively. Coefficients b1, b2P, b2M, and b2D are
the measures of systematic market risk, inflation risk, maturity risk, and default risk,
respectively, and are consistent with APT.
In addition, there are no common market shocks that may influence any of the
risk factors. So the following conditions
EðRmt :DPt Þ ¼ 0
EðRmt :DMRt Þ ¼ 0 (6.10)
EðRmt :DDRt Þ ¼ 0
where Rmt is the return on the stock market index and RST,t is the return on the short-
term default-free bond return. The error term in Eq. 6.11, eIt, specifies the unex-
pected change in inflation and is the generated regressor which serves as a proxy for
factor DPt. Fama and Schwert (1977) and French et al. (1983) employ similar
approaches to specify a proxy for changes in inflation.
The yield differential between short-term and long-term default-free bond
represents the slope of the yield curve. The relationship used to construct the
maturity-risk factor is given by the second systematic interest rate equation:
where y1 measures the average slope of the yield curve. Alternately, we will also
test the following systematic interest rate equation:
Inclusion of the independent variable, Rmt, will control for the stock–bond
market linkage. The estimated coefficient y1 measures the slope of the yield
curve, and it determines the maturity-risk premium on long-term assets and liabil-
ities. The error term in Eqs. 6.12a or 6.12b, eMt, specifies the unexpected change in
maturity risk and will be used as the generated regressor which serves as a proxy for
factor DMRt.
The portfolio of commercial banks used for this study consisted of money
center and large national and regional banks. Most of these banks were well
capitalized and had acquired a balanced portfolio of assets. The average
default-risk rating of the portfolio of banks should be close to the default-
risk rating of a high-grade corporate bond. Hence, the yield differential
between BAA corporate bond and long-term treasury security is used to
construct the default-risk factor. Our third systematic interest rate equation is
written as
where j1 measures the average default risk on BAA corporate bond. Inclusion of the
independent variable, Rmt, will control for the stock–bond market linkage. The error
term in Eqs. 6.13a or 6.13b, eDt, specifies the unexpected change in default risk
and serves as the generated regressor which will be used as the proxy for
factor DDRt.
Orthogonalization Procedure
Economic factors producing changes in the market return also induced term
structure movements. This leads to a high correlation between the market factor
and the interest rate factor. Hence, market return can be used as an instrument
variable to forecast the expected changes in the interest rates. To find the unex-
pected component of interest rates, the expected interest rate is purged by
regressing RIt on Rmt and using the residuals. The systematic interest rate risk
equation is
The residuals, eit, are the unsystematic interest rates and are used to replace RIt
in Eq. 6.14. The validity of this approach has been questioned on methodological
grounds. It is pointed out that this orthogonalization procedure produces biased
estimates of coefficients (intercept and b1j in Eq. 6.14) and that the deficiency of b2j
is not improved. On the other hand, use of an unorthogonal interest rate factor leads
to the errors-in-variable problem, i.e., the estimated coefficient b2j also captures
some of the effects responsible for changing the market factor, and hence, it is not
a true measure of the interest rate risk. Another problem using an unorthogonal
interest rate factor stems from the fact that interest rate (RIt) is usually
autoregressive. Therefore, residuals, eit, from Eq. 6.14 are autocorrelated unless
GLS parameter estimation procedure is employed. To use the GLS procedure,
a variance-covariance matrix has to be specified, which is not an easy task.
The second approach is to identify and estimate an ARMA model for the interest
rate variable, RIt (Flannery and James (1984)). The unanticipated change in interest
rate from the estimated model (i.e., residuals) is used to replace RIt in Eq. 6.14. In
general, the ARMA model of order (p, q) for the univariate time series, RIt, is
written as
where eIt, eI,t1, . . . are identically and independently distributed random errors
with mean zero. The ARMA procedure for the modeling of time series data is
outlined in Box and Jenkins (1976). The modeling is usually done in three steps.
First, a tentative parsimonious model is identified. Second, the parameters are
estimated, and diagnostic (Box-Pierce Q) statistics and residual auto correlation
plots are examined. The model is acceptable if the time series of residuals is white
noise and the Box-Pierce Q statistics are significant.
6 Multi-Risk Premia Model of US Bank Returns: An Integration of CAPM and APT 205
This approach leads to unbiased estimates of all the coefficients in Eq. 6.14. The
shortcoming of the univariate ARMA approach is that valuable information
contained in the stock and bond market linkage is ignored. Consequently, coeffi-
cient b2j captures some of the effects of economic factors producing stock market
changes.
YðBÞ ¼ I y1 B y2 B2 . . . yq Bq (6.18)
where B is the back shift operator, I is kxk unit matrix, and et is a sequence of an
independent k-dimensional vector with zero mean and positive definite covariance
matrix. The interest rate variable RIt and the market return Rmt are treated as the
components of a bivariate vector. Then vector (RIt, Rmt)T is modeled as a vector AR
process. The estimated model provides the unanticipated change in interest rate
variable to be used as the second factor in the augmented CAPM model.
References
Bae, S. C. (1990). Interest rate changes and common stock returns of financial institutions:
Revised. Journal of Financial Research, 13, 71–79.
Booth, J. R., & Officer, D. T. (1985). Expectations, interest rates, and commercial bank stocks.
Journal of Financial Research, 8, 51–58.
Box, G. E. P., & Jenkins, G. M. (1976). Time series analysis, forecasting and control. San Fran:
Holden-Day.
Chance, D. M., & Lane, W. R. (1980). A Re-examination of interest rate sensitivity in the common
stock of financial institutions. Journal of Financial Research, 3, 49–55.
Chen, N., Roll, R., & Ross, S. (1986). Economic forces and the stock market. Journal of Business,
59, 383–404.
Choi, J. J., Elyasiani, E., & Kopecky, K. (1992). The sensitivity of bank stock returns to
market, interest, and exchange rate risks. Journal of Banking and Finance, 16, 983–1004.
Christie, A. A. (1981). The stochastic behavior of common stocks variances: Value, leverage and
interest rate effects. Journal of Financial Economics, 10, 407–432.
206 S. Srivastava and K. Hung
Cox, J. C., Ingersoll, J. E., & Ross, S. A. (1985). An intertemporal general equilibrium model of
asset prices. Econometrica, 53, 363–84.
Elyasiani, E., & Iqbal, M. (1998). Sensitivity of the bank stock returns distribution to changes in
the level and volatility of interest rate: A GARCH-M model. Journal of Banking & Finance,
22, 535–563.
Fama, E. F., & Gibbons, M. R. (1982). Inflation, real returns and capital investment. Journal of
Monetary Economics, 9, 297–323.
Fama, E. F., & Schwert, W. G. (1977). Asset returns and inflation. Journal of Financial
Economics, 4, 115–146.
Flannery, M. J., & James, C. M. (1984). The effect of interest rate changes on the common stock
returns of financial institutions. Journal of Finance, 39, 1141–1153.
Flannery, M. J., Hameed, A. S., & Harjes, R. H. (1997). Asset pricing time-varying risk premia and
interest rate risk. Journal of Banking and Finance, 21, 315–335.
French, K. R., Ruback, R. C., & Schwert, G. W. (1983). Effects of nominal contracting hypothesis
on stock returns. Journal of Political Economy, 91, 70–96.
Johnson, R., & Wichern, D. (2007). Applied multivariate statistical analysis (6th ed.). Upper
Saddle River: Pearson Education.
Llyod, W. P., & Shick, R. (1977). A test of Stone’s two index models of returns. Journal of
Financial and Quantitative Analysis, 12, 363–376.
Lynge, M., & Zumwalt, J. (1980). An empirical study of interest rate sensitivity of commercial
bank returns: A multi index approach. Journal of Financial and Quantitative Analysis, 15,
731–742.
Merton, R. C. (1973). An intertemporal capital asset pricing model. Econometrica, 41, 867–887.
Song, F. (1994). A two factor ARCH model for deposit-institution stock returns. Journal of Money,
Credit, and Banking, 26, 323–340.
Srivastava, S. C., Hamid, S., & Choudhury, A. H. (1999). Stock and bond market linkage in the
empirical study of interest rate sensitivity of bank returns. Journal of Applied Business
Research, 15, 47–58.
Stone, B. (1974). Systematic interest rate risk in a two – index model of returns. Journal of
Financial and Quantitative Analysis, 9, 709–721.
Sweeney, R. J., & Warga, A. D. (1986). The pricing of interest-rate risk: Evidence from the stock
market. Journal of Finance, 41, 393–410.
Unal, H., & Kane, E. J. (1988). Two approaches to assessing the interest-rate sensitivity
of deposit-institutions’ equity returns. In C. Andrew (Ed.), Research in finance
(Vol. 7, pp. 113–132). Greenwich Conn.: Jai Press.
Woolridge, J. (2009). Introductory econometrics: A modern approach (4th ed.). Mason: Cengage
Learning.
Yourougou, P. (1990). Interest rate and the pricing of depository financial intermediary common
stock: Empirical evidence. Journal of Banking and Finance, 14, 803–320.
Nonparametric Bounds for European
Option Prices 7
Hsuan-Chu Lin, Ren-Raw Chen, and Oded Palmon
Contents
7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 208
7.2 The Bounds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 210
7.3 Comparisons . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 214
7.4 Extensions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 218
7.5 Empirical Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 222
7.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 228
Appendix 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 229
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 230
Abstract
There is much research whose efforts have been devoted to discovering the
distributional defects in the Black-Scholes model, which are known to cause severe
biases. However, with a free specification for the distribution, one can only find
upper and lower bounds for option prices. In this paper, we derive a new nonpara-
metric lower bound and provide an alternative interpretation of Ritchken’s (1985)
upper bound to the price of the European option. In a series of numerical examples,
our new lower bound is substantially tighter than previous lower bounds.
The financial support of National Science Council, Taiwan, Republic of China (NSC 96-2416-H-
006-039-), is gratefully acknowledged.
H.-C. Lin (*)
Graduate Institute of Finance and Banking, National Cheng-Kung University, Tainan, Taiwan
e-mail: hsuanchu@mail.ncku.edu.tw
R.-R. Chen
Graduate School of Business Administration, Fordham University, New York, NY, USA
e-mail: rchen@fordham.edu
O. Palmon
Department of Finance and Economics, Rutgers Business School – Newark and New Brunswick,
Piscataway, NJ, USA
e-mail: palmon@business.rutgers.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 207
DOI 10.1007/978-1-4614-7750-1_7,
# Springer Science+Business Media New York 2015
208 H.-C. Lin et al.
This is prevalent especially for out-of-the-money (OTM) options where the previ-
ous lower bounds perform badly. Moreover, we present that our bounds can be
derived from histograms which are completely nonparametric in an empirical
study. We first construct histograms from realizations of S&P 500 index returns
following Chen, Lin, and Palmon (2006); calculate the dollar beta of the option and
expected payoffs of the index and the option; and eventually obtain our bounds. We
discover violations in our lower bound and show that those violations present
arbitrage profits. In particular, our empirical results show that out-of-the-money
calls are substantially overpriced (violate the lower bound).
Keywords
Option bounds • Nonparametric • Black-Scholes model • European option • S&P
500 index • Arbitrage • Distribution of underlying asset • Lower bound • Out-of-
the-money • Kernel pricing
7.1 Introduction
In a seminal paper, Merton (1973) presents for the first time the no-arbitrage bounds
of European call and put options. These bounds are nonparametric and do not rely on
any assumption.1 Exact pricing formulas such as the Black and Scholes (1973) model
and its variants, on the other hand, rely on strong assumptions on the asset price
process and continuous trading. Due to the discreteness of actual trading opportuni-
ties, Perrakis and Ryan (1984) point out that option analyses in continuous time limit
the accuracy and applicability of the Black-Scholes and related formulas. Relying on
Rubinstein’s (1976) approach, the single-price law, and arbitrage arguments, they
derive upper and lower bounds for option prices with both a general price distribution
and discrete trading opportunities. Their lower bound is tighter than that of Merton.
Levy (1985) applies stochastic dominance rules with borrowing and lending at the
risk-free interest rate to derive upper and lower option bounds for all unconstrained
utility functions and alternatively for concave utility functions. The derivation of these
bounds can be applied to any kinds of stock price distribution as long as the stock is
“nonnegative beta,” which is identical to the assumption of Perrakis and Ryan (1984).
Moreover, Levy claims that Perrakis and Ryan’s bounds can be obtained by applying
the second-degree stochastic dominance rule. However, Perrakis and Ryan do not
cover all possible combinations of the risky asset with the riskless asset, and their
bounds are therefore wider than those of Levy. Levy also applies the first-degree
stochastic dominance rule (FSDR) with riskless assets to prove that Merton’s bounds
are in fact FSDR bounds and applies the second-degree stochastic dominance rule to
strengthen Merton’s bounds on the option value. At the same time, Ritchken (1985)
uses a linear programming methodology to derive option bounds based on primitive
prices in incomplete markets and claims that his bounds are tighter than those of
Perrakis and Ryan (1984).
1
The only assumption is that both option and its underlying stock are traded securities.
7 Nonparametric Bounds for European Option Prices 209
With an additional restriction that the range of the distribution of the one-period
returns per dollar invested in the optioned stock is finite and has a strictly positive
lower limit, Perrakis (1986) extends Perrakis and Ryan (1984) to provide bounds for
American options. Instead of assuming that no opportunities exist to revise positions
prior to expiration in Levy (1985) and Ritchken (Ritchken 1985), Ritchken and Kuo
(1988) obtain tighter bounds on option prices under an incomplete market by
allowing for a finite number of opportunities to revise position before expiration
and making more restrictive assumptions on probabilities and preferences. The
single-period linear programming option model is extended to handle multiple
periods, and the stock price is assumed to follow a multiplicative multinomial
process. Their results show that the upper bounds are identical to those of Perrakis,
while the lower bounds are tighter. Later, Ritchken and Kuo (1989) also add suitable
constraints to a linear programming problem to derive option bounds under higher
orders of stochastic dominance preferences. Their results show that while the upper
bounds remain unchanged beyond the second-degree stochastic dominance, the
lower bounds become sharper as the order of stochastic dominance increases.2
Claiming that Perrakis and Ryan (1984), Levy (1985), Ritchken (1985), and
Perrakis (1986) are all parametric models, Lo (1987) derives semi-parametric
upper bounds for the expected payoff of call and put options. These upper bounds
are semi-parametric because they depend on the mean and variance of the stock price
at maturity but not on its entire distribution. In addition, the derivation of
corresponding semi-parametric upper bounds for option prices is shown by adopting
the risk-neutral pricing approach of Cox and Ross (1976).3 To continue the work of
Lo (1987), Zhang (1994) and De La Pena et al. (2004), both of which assume that the
underlying asset price must be continuously distributed, sharpen the upper option
bounds of Lo (1987). Boyle and Lin (1997) extend the results of Lo (1987) to
contingent claims on multiple underlying assets. Under an intertemporal setting,
Constantinides and Zariphopoulou (2001) derive bounds for derivative prices with
proportional transaction costs and multiple securities. Frey and Sin (1999) examine
the sufficient conditions of Merton’s bounds on European option prices under random
volatility. More recently, Gotoh and Konno (2002) use the semi-definite programming
and a cutting plane algorithm to study upper and lower bounds of European call option
prices. Rodriguez (2003) uses a nonparametric method to derive lower and upper
2
To further explore the research work of Ritchken and Kuo (1989) under the decreasing absolute
risk aversion dominance rule, Basso and Pianca (1997) obtain efficient lower and upper option
pricing bounds by solving nonlinear optimization problem. Unfortunately, neither model provides
enough information of their numerical examples for us to compare our model with. The Ritchken-
Kuo model provides no Black-Scholes comparison, and the Basso-Pianca model provides only
some partial information on the Black-Scholes model (we find the Black-Scholes model under 0.2
volatility to be 13.2670 and under the 0.4 volatility to be 20.3185, which are different from what
are reported in their paper (12.993 and 20.098, respectively)) which is insufficient for us to provide
any comparison.
3
Inspired by Lo (1987), Grundy (1991) derives semi-parametric upper bounds on the moments of
the true, other than risk-neutral, distribution of underlying assets and obtains lower bounds by
using observed option prices.
210 H.-C. Lin et al.
bounds and contributes a new tighter lower bound than previous work. Huang (Huang
2004) puts restrictions on the representative investor’s relative risk aversion and
produces a tighter call option bound than that of Perrakis and Ryan (1984). Hobson
et al. (2005) derive arbitrage-free upper bounds for the prices of basket options. Peña
et al. (2010) conduct static-arbitrage lower bounds on the prices of basket options via
linear programming. Broadie and Cao (2008) introduce new and improved methods
based on simulation to obtain tighter lower and upper bounds for pricing American
options. Lately, Chung et al. (2010) also use an exponential function to approximate
the early exercise boundary to obtain tighter bounds on American option prices.
Chuang et al. (2011) provide a more complete review and comparison of theoretical
and empirical development on option bounds.4
In this paper we derive a new and tighter lower bound for European option prices
under a nonparametric framework. We show that Ritchken’s (1985) upper bound is
consistent with our nonparametric framework. Both bounds are nonparametric
because the price distribution of underlying asset is totally flexible, can be arbitrarily
chosen, and is consistent with any utility preference.5 We compare our lower bound
with those in previous studies and show that ours dominate those models by a wide
margin. We also present the lower bound result on the model with random volatility
and random interest rates (Bakshi et al. 1997; Scott 1997) to demonstrate how easily
our model can be made consistent with any parametric structure.6 Finally, we
present how our bounds can be derived from histograms which are nonparametric
in an empirical study. We discover violations of our lower bound and show that those
violations present arbitrage profits.7 In particular, our empirical results show that
out-of-the-money calls are substantially overpriced (violate the lower bound).
where Mt,T is the marginal rate of substitution, also known as the pricing kernel
that discounts the future cash flow at time T; Et[.] is the conditional expectation
4
Since our paper only provides a nonparametric method on examining European option bounds,
our literature review is much limited. For a more complete review and comparison on prior studies
of option bounds, please see Chuang et al. (2011).
5
Christoffersen et al. (2010) provide results for the valuation of European-style contingent claims
for a large class of specifications of the underlying asset returns.
6
Given that our upper bound turns out to be identical to Ritchken’s (1985), we do not compare with
those upper bound models that dominate Ritchken (e.g., Huang (2004), Zhang (1994) and De La
Pena et al. (2004)). Also, we do not compare our model with those models that require further
assumptions to carry out exact results (e.g., Huang (2004) and Frey and Sin (1999)), since it is
technically difficult to do.
7
For the related empirical studies of S&P 500 index options, see Constantinides et al. (2009, 2011).
7 Nonparametric Bounds for European Option Prices 211
under the physical measure ℙ taken at time t; and St is the value of an arbitrary
asset at time t. The standard kernel pricing theory (e.g., Ingersoll (1989))
demonstrates that
Dt, T ¼ Et Mt, T , (7.2)
where Dt,T is the risk-free discount factor that gives the present value of $1 over the
period (t,T). The usual separation theorem gives rise to the well-known risk-neutral
pricing result:
St ¼ Et Mt, T St
ðT Þ
¼ Et Mt, T E^t ½ST , (7.3)
ðT Þ
¼ Dt, T E^t ½ST
If the risk-free interest rate is stochastic, then E^t(T)[.] is the conditional expec-
tation under the T-forward measure P ^ ðT Þ . When the risk-free rate is non-stochastic,
then the forward measure reduces to the risk-neutral measure P ^ and will not depend
ðT Þ ^
upon maturity time, i.e., Et ½ ! E t ½. 8
Note that Eqs. 7.1 and 7.3 can be applied to both the stock and the option prices.
This leads to the following theorem which is the main result of this paper.
Theorem 7.1 The following formula provides a lower bound for the European call
option Ct:
Ct ¼ Dt, T Et ½CT þ bC St Dt, T Et ½ST (7.4)
cov½CT ; ST
where bC ¼ var½ST :
Proof By Eq. 7.1, the option price must follow Ct ¼ Et[Mt,T CT], and hence
Ct ¼ Et Mt, T CT
¼ Et Mt, T Et ½CT þ cov Mt, T , CT (7.5)
¼ Dt, T Et ½CT þ cov Mt, T , CT
or
Ct Dt, T Et ½CT ¼ cov Mt, T ; CT : (7.6)
Similarly,
St Dt, T Et ½ST ¼ cov Mt, T ; ST : (7.7)
8
Without loss of generality and for the ease of exposition, we take non-stochastic interest rates and
proceed with the risk-neutral measure P ^ for the rest of the paper.
212 H.-C. Lin et al.
where bP ¼ bC 1. Note that bP < 0 < bC. This also implies that bc<1. Finally, it is
straightforward to show that bP ¼ covvar½P½ST T;ST using the put-call parity.
Therefore, for both call and put options, since the relationship between the
pricing kernel M and stock price S is convex, the theorem provides a lower
bound.9 Merton (1973) shows that the stock price will be the upper bound for
the call option and the strike price should be the upper bound for the put
option; otherwise, arbitrage should occur. Ritchken (1985) provides a tighter
upper bound10 than that of Merton, which is stated in the following theorem,
although we provide an alternative proof to Ritchken that is consistent with our
derivation of the lower bound.
10
Perrakis and Ryan (1984) and Ritchken (1985) obtain the identical upper bound.
7 Nonparametric Bounds for European Option Prices 213
Theorem 7.2 The following formulas provide upper bounds for the European call
and put options (Ritchken (1985))11:
St
Ct ¼ Et ½CT
Et ½ST (7.12)
St St
Pt ¼ Et ½PT þ Dt, T K:
Et ½ST E t ½ ST
Proof Similar to the proof of the lower bound, the upper bound of the call option is
provided as follows:
St Et Mt, T ST
Et ½CT ¼ Et ½CT
Et ½ST E t ½ ST
Et Mt, T ST CT cov Mt, T ST , CT
¼
Et ½ST
Et Mt, T ST CT (7.13)
>
Et ½ S T
ðCÞ
Et ½ST
¼ Ct
Et ½ S T
> Ct :
The third line of the above equation is a result from the fact that cov[Mt,TST,
CT] < 0. The fourth line of the above equation is a change of measure with the call
option being the numeraire. The last line of the above equation is a result based upon
t [ST]/Et[ST] > 1.
E(C) 12
By the put-call parity, we can show that the upper bound of the put option
requires an additional term:
St
Pt ¼ Ct þ Dt, T K St ¼ Et ½PT þ ST K þ Dt, T K St
Et ½ S T
(7.14)
St St
¼ Et ½PT þ Dt, T K:
Et ½ S T Et ½ S T
The lower and upper bounds we show in this paper have two important advantages
over the existing bounds. The bounds will converge to the true value of the option if:
• The expected stock return, EtS½St T , approaches the risk-free rate.
• The correlation between the stock and the call or put option (rSC or rSP)
approaches 1 or 1.
11
This is same as Proposition 3-i (Eq. 7.26) in Ritchken (1985).
By the definition of measure change, we have Et[CTST] ¼ Et[CT]E(C)
12 (C)
t [ST] which implies Et [ST]/
E[ST] ¼ Et[CTST]/{Et[CT]Et[ST]} > 1.
214 H.-C. Lin et al.
These advantages help us identify when the bounds are tight and when they are
not. The first advantage indicates that the bounds are tight for low-risk stocks and
not tight for high-risk stocks. The second advantage indicates that the bounds are
tighter for in-the-money options than out-of-the-money options.
7.3 Comparisons
The main purpose of this section is to compare our lower bound to several
lower bounds in previous studies, namely, Merton (1973), Perrakis and Ryan
(1984), Ritchken (1985), Ritchken and Kuo (1988), Gotoh and Konno (2002),
and Rodriguez (2003) using the Black-Scholes model as the benchmark for its
true option value. We also compare Ritchken’s upper bound (which is also our
upper bound) with more recent works by Gotoh and Konno (2002) and
Rodriguez (2003).
The Black-Scholes model has five variables: stock price, strike price, volatility
(standard deviation), risk-free rate (constant), and time to maturity. In addition to
the five variables, the lower bound models need the physical expected stock return.
The following is the base case for the comparison:
Current stock S0 50
Strike K 50
Volatility s 0.2
Risk-free rate r 0.1
Time to maturity T 1
Stock expected return m 0.2
In the Black-Scholes model, stock price (S) evolution follows a log normal
process:
where instantaneous expected rate of stock return m and volatility of stock price s
are assumed to be constants and where dW is a wiener process. The call option price
is computed as
where
ln S0 ln K þ ðr 0:5s2 ÞT
d ¼ pffiffiffi :
s T
7 Nonparametric Bounds for European Option Prices 215
The lower bound is computed by simulating the stock price using Eq. 7.15 via
a binomial distribution approximation:
Sj ¼ S0 uj dnj : (7.17)
where
erDt d
p^ ¼ ,
ud
pffiffiffiffi
u ¼ es Dt
,
pffiffiffiffi
d ¼ es Dt
,
and Dt ¼ T/n represents the length of the partition. Under the actual measure, the
formula will change to
n j
Pr½j ¼ p ð1 pÞnj (7.19)
j
where
emDt d
p¼ :
ud
b ½j rT
Pr
M0T ½j ¼ e : (7.20)
Pr½j
In our results, we let n be great enough so that the binomial model price is 4-digit
accurate to the Black-Scholes model. We hence set n to be 1,000. The results are
reported in Table 7.1. The first panel presents the results for various moneyness
levels, the second panel presents the results for various volatility levels, the third
panel presents the results for various interest rates, the fourth panel presents the
results for various maturity times, and the last presents the results for various stock
216 H.-C. Lin et al.
expected returns. In general, the lower bounds are tighter (for all models) when the
moneyness is high (in-the-money), volatility is high, risk-free rate is high, time to
maturity is short, and the expected return of stock is low.
This table presents comparisons between our lower bound and existing lower
bounds by Merton (1973), Perrakis and Ryan (1984), and Ritchken (1985). The
base case parameter values are:
• Stock price ¼ 50
• Strike price ¼ 50
• Volatility ¼ 0.2
• Risk-free rate ¼ 10 %
• Time to maturity ¼ 1 year
• Stock expected return (m) ¼ 20 %
The highlighted rows represent the base case.
As we can easily see, universally our model for the lower bound is tighter than any
of the comparative models. One result particularly worth mentioning is that our lower
bound performs better than the other lower bound models in out-of-the-money
options. For example, our lower bound is much better than Ritchken’s (1985)
lower bound when the option is 20 % out-of-the-money and continues to show
value when Ritchken’s lower bound returns to 0 (see the first panel of Table 7.1).
While Ritchken and Kuo (1988) claim to obtain tighter lower bounds than
Perrakis (1986) and Perrakis and Ryan (1984), they do not show direct comparisons
in their paper. Rather, they present through a convergence plot (Figure 3 on page
308 in Ritchken and Kuo (1988)) of a Black-Scholes example with the true value
being $5.4532 and the lower bound approaching roughly $5.2. The same parameter
values with our lower bound show a lower bound of $5.4427, which demonstrates
a substantial improvement over the Ritchken and Kuo model.
The comparisons with more recent studies of Gotoh and Konno (2002) and
Rodriguez (2003) are given in Tables 7.2 and 7.3.13 Gotoh and Konno use
semi-definite programming and a cutting plane algorithm to study upper and lower
bounds of European call option prices. Rodriguez uses a nonparametric method to
derive lower and upper bounds. As we can see in Tables 7.2 and 7.3, except for very
few upper bound cases, none of the bounds under the Gotoh and Konno’s model and
Rodriguez’s model are very tight, compared to our model. Furthermore, note that our
model requires no moments of the underlying distribution.14
The base case parameter values are:
• Stock price ¼ 40
• Risk-free rate ¼ 6 %
• Stock expected return (m) ¼ 20 %
13
We also compare with the upper bound by Zhang (1994), which is an improved upper bound by
Lo (1987), and show overwhelming dominance of our upper bound. The results (comparison to
Tables 7.1, 7.2, and 7.3 in Zhang) are available upon request.
14
The upper bounds by the Gotoh and Konno model perform well in only in-the-money, short
maturity, and low volatility scenarios, and these scenarios are where the option prices are close to
their intrinsic values, and hence the percentage errors are small.
7 Nonparametric Bounds for European Option Prices 217
(continued)
218 H.-C. Lin et al.
Note: S is the stock price; vol is the volatility; rate is the risk-free rate; Mu (m) is the expected rate
of return of stock; Blk-Sch is the Black-Scholes (1973) solution; Merton is the Merton (1973)
model; PR is Perrakis and Ryan (1984) model; Ritch. is the Ritchken (1985) model; Our is our
model; $error is error in dollar; %error is error in percentage
7.4 Extensions
15
The term “dollar beta” is originally from Page 173 of Black (1976). Here we mean bc and br.
7 Nonparametric Bounds for European Option Prices 219
Table 7.2 Comparison of upper and lower bounds with the Gotoh and Konno (2002) model
Lower bound Upper bound
Stk Our GK Blk-Sch Our GK
S ¼ 40; rate ¼ 6 %; vol ¼ 0.2; t ¼ 1 week
30 10.0346 10.0346 10.0346 10.1152 10.0349
35 5.0404 5.0404 5.0404 5.1344 5.0428
40 0.4628 0.3425 0.4658 0.5225 0.5771
45 0.0000 0.0000 0.0000 0.0000 0.0027
50 0.0000 0.0000 0.0000 0.0000 0.0003
S ¼ 40; rate ¼ 6 %; vol ¼ 0.8; t ¼ 1 week
30 10.0400 10.0346 10.0401 10.1202 10.1028
35 5.2644 5.0404 5.2663 5.3483 5.4127
40 1.7876 1.2810 1.7916 1.8428 2.2268
45 0.3533 0.0015 0.3548 0.3717 0.5566
50 0.0412 0.0000 0.0419 0.0444 0.1021
S ¼ 40; rate ¼ 6 %; vol ¼ 0.8; t ¼ 12 week
30 11.9661 10.4125 12.0278 12.7229 12.8578
35 8.7345 6.2980 8.8246 9.4774 9.7658
40 6.2141 3.8290 6.3321 6.8984 7.5165
45 4.3432 2.5271 4.4689 4.9421 6.8726
50 2.9948 1.5722 3.1168 3.4990 4.5786
Note: S is the stock price; Stk is the strike price; vol is the volatility; rate is the risk-free rate; Blk-
Sch is the Black-Scholes (1973) solution; GK is the Gotoh and Konno (2002) model; Our is our
model; $error is error in dollar; %error is error in percentage
specifications. However, here, given that there is no closed form solution to the
covariance our model requires, we shall use Monte Carlo to simulate the lower
bound. In order to be consistent with the lower bound, we must use the same Monte
Carlo paths for the valuation of the option. For the ease of exposition and simplicity,
we assume the following joint stochastic processes of stock price S, interest rate r,
and volatility V under the actual measure, respectively:
pffiffiffiffi
dS ¼ mSdt þ V SdW ^2
^2
dr ¼ aðy r Þdt þ vd W (7.21)
dV ¼ Vd W^2
Table 7.3 Comparison of upper and lower bounds with the Rodriguez (2003) model
Lower bound Upper bound
S Our Rodriguez Blk-Sch Our Rodriguez
30 0.0221 0 0.0538 0.1001 0.1806
32 0.0725 0.0000 0.1284 0.2244 0.3793
34 0.1828 0.0171 0.2692 0.4451 0.7090
36 0.3878 0.1158 0.5072 0.7973 1.2044
38 0.7224 0.3598 0.8735 1.3100 1.8900
40 1.2177 0.7965 1.3950 2.0044 2.7767
42 1.8982 1.4521 2.0902 2.8927 3.8619
44 2.7711 2.3329 2.9676 3.9697 5.1315
46 3.8319 3.4286 4.0255 5.2211 6.5640
48 5.0709 4.7177 5.2535 6.6302 8.1337
50 6.4703 6.1724 6.6348 8.1753 9.8149
52 8.0072 7.7635 8.1494 9.8327 11.5835
54 9.6574 9.4631 9.7758 11.5794 13.4187
56 11.3974 11.2462 11.4933 13.3950 15.3032
58 13.2067 13.0922 13.2832 15.2621 17.2235
60 15.0683 14.9845 15.1292 17.1674 19.1693
62 16.9716 16.9101 17.0179 19.1024 21.1328
64 18.9035 18.8593 18.9384 21.0573 23.1086
66 20.8559 20.8250 20.8822 23.0262 25.0926
68 22.8239 22.8020 22.8429 25.0056 27.0822
70 24.8018 24.7867 24.8157 26.9915 29.0754
Note: S is the stock price; Blk-Sch is the Black-Scholes (1973) solution; Rodriguez is the
Rodriguez (2003) model; Our is our model; $error is error in dollar; %error is error in percentage
The no-arbitrage price of the call option is computed under the risk-neutral
measure as
ðt
^
C ¼ E exp r u du max S t K, 0 : (7.24)
0
7 Nonparametric Bounds for European Option Prices 221
Table 7.4 Lower bound under the random volatility and random interest rate model
S BCC/Scott Our $error %error
25 1.9073 1.1360 0.7713 40.44
30 3.2384 2.4131 0.8253 25.49
35 5.2058 4.1891 1.0167 19.53
40 7.5902 6.5864 1.0039 13.23
45 10.2962 9.5332 0.7630 7.41
50 13.6180 12.8670 0.7510 5.52
55 17.1579 16.5908 0.5671 3.30
60 21.1082 20.5539 0.5543 2.63
65 25.0292 24.7251 0.3040 1.21
70 29.3226 29.0742 0.2484 0.85
Note: S is the stock price; BCC/Scott are Bakshi et al. (1997) and Scott (1997) models; Our is our
model; $error is error in dollar; %error is error in percentage
The bounds are computed under the actual measure. For example,
Given that dWidWj ¼ 0, we can simulate the interest rates and volatility
separately and then simulate the stock price. That is, conditional on known interest
rates and volatility, under independence, the stock price is log normally distributed.
We perform our simulations using 10,000 paths over 52 weekly periods. The
parameters are given as follows:
Strike K 50
Time to maturity T 1
Stock expected return m 0.2
Reverting speed a 0.5
Reverting level y 0.1
Interest rate volatility u 0.03
Initial interest rate r0 0.1
Initial variance V0 0.0416
Volatility on variance 0.2
Note that implicitly we assume the price of risk for both interest rate process and
volatility process to be 0, for simplicity and without loss of generality. The results
are shown in Table 7.4. Compared to the model of the Black-Scholes (i.e., the first
panel of Table 7.4), the lower bound performs similarly in the random volatility
and random interest rate model. Take the base case as an example where the
Black-Scholes price is 6.6322, the Bakshi-Cao-Chen/Scott price is 13.6180 as
a result of extra uncertainty in the stock price due to random volatility and interest
16
This is so because the initial volatility is 0.2.
222 H.-C. Lin et al.
rates. The error of the lower bound of our model is 0.7510 in the Bakshi-Cao-Chen/
Scott case as opposed to 0.4398 in the Black-Scholes case.17 The percentage error is
5.52 % in the Bakshi-Cao-Chen/Scott case versus 6.63 % in the Black-Scholes case.
The in-the-money options have larger percentage errors than those of the out-of-
the-money options.
The parameters are given as follows:
Strike price 50
Time to maturity 1
Stock expected return 0.2
Reverting speed 0.5
Reverting level 0.1
Interest rate volatility 0.03
Price of risk 0
Initial interest rate 0.1
Initial volatility 0.2
Volatility on volatility 0.2
The Monte Carlo paths are 10,000. The stock price, volatility, and interest rate processes are
assumed to be independent
In this section, we test the lower and upper bounds against data. Charles Cao has
generously provided us with the approximated prices of S&P 500 index call option
contracts, matched levels of S&P 500 index, and approximated risk-free 90-day
T-Bill rates for the period of June 2, 1988 through December 31, 1991.18 For each
day, the approximated option prices are calculated as the average of the last bid and
ask quotes. Index returns are computed using daily closing levels for the S&P
500 index that are collected and confirmed using data obtained from Standard and
Poor’s, CBOE, Yahoo, and Bloomberg.19
The dataset contains 46,540 observations over 901 days (from June 2, 1988, to
December 31, 1991). Hence, on average there are over 50 options for various
maturities and strikes. The shortest maturity of the dataset is 7 days and the longest
is 367 days. 15 % of the data are less than 30 days to maturity, 32 % are between
17
This Black-Scholes case is from the highlighted row in the first panel of Table 7.1.
18
The data are used in Bakshi et al. (1997).
19
The (ex-dividend) S&P 500 index we use is the index that serves as an underlying asset for the option.
For option evaluation, realized returns of this index need not be adjusted for dividends unless the timing
of the evaluated option contract is correlated with lumpy dividends. Because we use monthly
observations, we think that such correlation is not a problem. Furthermore, in any case, this should
not affect the comparison of the volatility smile between our model and the Black-Scholes model.
7 Nonparametric Bounds for European Option Prices 223
30 and 60 days, 30 % are between 60 and 180 days, and 24 % are more than
180 days to maturity. Hence, these data do not have maturity bias.
The deepest out-of-the-money option is 18.33 %, and the deepest
in-the-money option is 47.30 %. Roughly half of the data are at-the-money options
(46 % of the data are within 5 % in-the-money and out-of-the-money). 10 % are
deep-in-the-money (more than 15 % in-the-money), but less than 1 % of the data are
deep-out-of-the-money (more than 15 % out-of-the-money). Hence, the data have
disproportional fraction of in-the-money options. This is clearly a reflection of the
bull market in the sample period.
The best way to test the lower and upper bounds derived in this paper is to use
a nonparametric, distribution-free model. Note that the lower bound in Theorem 7.1
requires only the expected return of the underlying stock and the covariance
between the stock and the option. There is no further requirement for the lower
and the upper bounds. In other words, our lower bound model can permit any
arbitrary distribution of the underlying stock and any parametric specification of the
underlying stock such as random volatility, random interest rates, and jumps.
Hence, to best test the bounds with a parsimonious empirical design, we adopt
the histogram method introduced by Chen et al. (2006) where the underlying asset
is modelled by past realizations, i.e., histogram.
We construct histograms from realizations of S&P 500 index (SPX) returns. We
calculate the price on day t of an option that settles on day T using a histogram of
S&P 500 index returns for a holding period of T–t, taken from a 5-year window
immediately preceding time t.20 For example, an x-calendar-day
option price on
any date is evaluated using a histogram of round 252365 x -trading-day holding period
returns where round [.] is rounding the nearest integer.21 The index levels used to
calculate these returns are taken from a window that starts on the 1260th ( 5
252) trading day before the option trading date and ends 1 day before the trading
date. Thus, this histogram contains 1, 260 round 252 365 x -trading-day return reali-
zations. Formally, we compute histogram of the (unannualized) returns by the
following equation:
where each i is an observation in time and t is the last i. For example, if t is 1988/06/
02 and x is 15 calendar days (or ten business days). We further choose our
histogram horizon to be 5 years or 1,260 business days. Fifteen business days
after 1988/06/02 is 1988/06/17. To estimate a distribution of the stock return for
1988/06/17, we look back a series of 10-business-day returns. Since we choose
a 5-year historical window, or 1,260-business-day window, the histogram will contain
20
We use three alternative time windows, 2-year, 10-year, and 30-year, to check the robustness of
our procedure and results.
21
The conversion is needed because we use trading-day intervals to identify the appropriate return
histograms and calendar-day intervals to calculate the appropriate discount factor.
224 H.-C. Lin et al.
1,260 observations. The first return in the histogram, R1933/06/02,1933/06/07,1, is the differ-
ence between the log of the stock price on 1988/06/01, ln St–1, and the log of the stock
price 15 calendar days (ten business days) earlier on 1988/05/17, ln St–1–a. The second
observation in the histogram, R1933/06/02/1933/06/07,2 is computed as ln St–2–ln St–2–a.
After, we complete the histogram of returns, we then convert it to the histogram of
prices by multiplying every observation in the histogram by the current stock price:
The expected option payoff is calculated as the average payoff where all the
realizations in the histogram are given equal weights. Thus, Et[CT,T,K] and Et[ST]
are calculated as
8
> 1 XN
< Et CT , T , K ¼ max S T , i K, 0
N i¼1
> XN , (7.28)
: Et ½ S T ¼ 1 S T , i
N i¼1
where N is the total number of realized returns and Ct,T,K is the price observed at time t,
of an option that expires at time T with strike price K. Substituting the results in Eq.
7.28 in the approximation pricing formula of Eq. 7.4, we obtain our empirical model:
C t ¼ P t , T E t C T , T , K þ bC S t P t , T E t ½ S T
1 XN 1 XN
¼ Pt , T max ST , i K, 0 þ bC St Pt, T S
i¼1 T , i
(7.29)
N i¼1 N
where the dollar beta is defined as bC ¼ covvar½C½ST T;ST as defined in Eq. 7.4.
Note that option prices should be based upon projected future volatility levels
rather than historical estimates. We assume that investors believe that the distribu-
tion of index returns over the time to maturity follows the histogram of a particular
horizon with a projected volatility. In practice, traders obtain this projected vola-
tility by calibrating the model to the market price. We incorporate the projected
volatility, n*t,T,K, into the histogram by adjusting its returns:
v
t, T , K
R
t, T , K , i ¼ Rt, T , i Rt, T þ Rt, T ::, i ¼ 1, , N, (7.30)
vt, T
where the historical volatility nt,T is calculated as the standard deviation of the
historical returns as follows:
1 XN
2
v2t, T ¼ R t , T , i R t, T (7.31)
N1 i¼1
where Rt,T,i ¼ ST,i/St and Rt, T ¼ N1 SNi¼1 Rt, T , i is the mean return.
7 Nonparametric Bounds for European Option Prices 225
perform a simple buy and hold trading strategy. If the lower bound is violated, we will
buy the option and hold it till maturity. For the 4,233 (out of 46,540) violations, the
buy and hold strategy generated $22,809 or an average of $5.39 per contract.
Given that the buy and hold strategy can be profitable simply due to the bull market,
we compute those that had no violation of the lower bound. For the 42,307 cases that
violated no lower bound, the average profit is $1.83. Hence, the options that violated
the lower bound imply a trading profit 200 % above the average.
7 Nonparametric Bounds for European Option Prices 227
a 30-day Maturity
14 actual ATM
lbdd
12 ubdd
10
0
02/27/88 09/14/88 04/02/89 10/19/89 05/07/90 11/23/90 06/11/91 12/28/91 07/15/92
b 60-day Maturity
18 actual ATM
lbdd
16
ubdd
14
12
10
8
6
4
2
0
02/27/88 09/14/88 04/02/89 10/19/89 05/07/90 11/23/90 06/11/91 12/28/91 07/15/92
91-day Maturity
c 25 ATM
actual
lbdd
ubdd
20
15
10
0
02/27/88 09/14/88 04/02/89 10/19/89 05/07/90 11/23/90 06/11/91 12/28/91 07/15/92
182-day Maturity
d 40 ATM
actual
lbdd
35 ubdd
30
25
20
15
10
0
02/27/88 09/14/88 04/02/89 10/19/89 05/07/90 11/23/90 06/11/91 12/28/91 07/15/92
Fig. 7.1 (a) Plot of 30-day maturity actual ATM option prices and its upper and lower bound values.
(b) Plot of 60-day maturity actual ATM option prices and its upper and lower bound values. (c) Plot
of 91-day maturity actual ATM option prices and its upper and lower bound values. (d) Plot of
182-day maturity actual ATM option prices and its upper and lower bound values
7.6 Conclusion
In this paper, we derive a new and tighter lower bound for European option
prices comparing with those of previous studies. We further reinterpret
Ritchken’s (1985) upper bound under a nonparametric framework. Our
model contributes to the literature in two different ways. First, our bounds
require no parametric assumption of the underlying stock or the moments of
the distribution. Furthermore, our bounds require no assumptions on interest
rates or volatility. The only requirements of our model are the dollar beta
of the option and expected payoffs of the stock and the option. Hence, our
bounds can be applied to any model such as the random volatility and
random interest rate model by Bakshi et al. (1997) and Scott (1997).
Second, despite of much looser and flexible assumptions, our bounds are
significantly tighter than the existing upper and lower bound models. Most
importantly, our bounds are tighter for the out-of-the-money options that
cannot be bounded efficiently by previous models. Finally, we apply our model to
real data using histograms of the realized stock returns. The results show that
nearly 10 % of the observations violate the lower bound. These violations are
shown to generate significant arbitrage profits, after correction of the bull market in
the sample period.
7 Nonparametric Bounds for European Option Prices 229
Appendix 1
Probability M S C ¼ max{S K, 0}
p2 M+x S+y>K S+yK
2p(1 p) M Se<K 0
(1 p)2 Mx Sy<K 0
When e ¼ 0, the relationship between the pricing kernel M and stock price S is
linear, and we obtain equality. When e > 0, the relationship is convex. We first
calculate the mean values:
E½M ¼ M x þ 2px
E½S ¼ S y þ 2pðy
e þ 2p e
2
E½C ¼ p ðS þ y K :
2
where
z ¼ y2 þ e2 1 2p þ 2p2 þ 2eyð2p 1Þ > 0:
As a result, it is straightforward to show that
e ð 1 p Þðy eÞ
¼ 2p ð1 pÞxðS þ y K 1 þ
2
z
2p ð1 pÞxðS þ y K ¼: cov½M, C:
2
230 H.-C. Lin et al.
The fourth line is obtained because cov[M,C] < 0 and 1 þ eð1pzÞðyeÞ > 1. Note that
the result is independent of p since all it needs is 0 < p < 1 for 1 þ eð1pzÞðyeÞ to be
greater than 1. Also note that when e ¼ 0 the equality holds.
References
Bakshi, G., Cao, C., & Chen, Z. (1997). Empirical performance of alternative option pricing
models. Journal of Finance, 52, 2003–2049.
Basso, A., & Pianca, P. (1997). Decreasing absolute risk aversion and option pricing bounds.
Management Science, 43, 206–216.
Black, F. (1976). The pricing of commodity contacts. Journal of Financial Economics, 3, 167–179.
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of
Political Economy, 81, 637–654.
Boyle, P., & Lin, X. (1997). Bounds on contingent claims based on several assets. Journal of
Financial Economics, 46, 383–400.
Broadie, M., & Cao, M. (2008). Improved lower and upper bound algorithms for pricing American
options by simulation. Quantitative Finance, 8, 845–861.
Chen, R. R., Lin, H. C., & Palmon, O. (2006). Explaining the volatility smile: Reduced form versus
structural option models. 2006 FMA Annual Meeting. Salt Lake City, UT.
Christoffersen, P., Elkamhi, R., Feunou, B., & Jacobs, K. (2010). Option valuation
with conditional heteroskedasticity and nonnormality. Review of Financial Studies,
23, 2139–2189.
Chuang, H., Lee, C. F. & Zhong, Z. K. (2011). Option bounds: A review and comparison
(Working paper).
Chung, S. L., Hung, M. W., & Wang, J. Y. (2010). Tight bounds on American option prices.
Journal of Banking and Finance, 34, 77–89.
Constantinides, G., & Zariphopoulou, T. (2001). Bounds on derivative prices in an intertemporal
setting with proportional transaction costs and multiple securities. Mathematical Finance,
11, 331–346.
Constantinides, G. M., Jackwerth, J. C., & Perrakis, S. (2009). Mispricing of S&P 500 index
options. Review of Financial Studies, 22, 1247–1277.
Constantinides, G. M., Czerwonko, M., Jackwerth, J. C., & Perrakis, S. (2011). Are options on
index futures profitable for risk-averse investors? Empirical evidence. Journal of Finance, 66,
1407–1437.
Cox, J., & Ross, S. (1976). The valuation of options for alternative stochastic process. Journal of
Finance Economics, 3, 145–166.
De La Pena, V., Ibragimov, R., & Jordan, S. (2004). Option bounds. Journal of Applied Proba-
bility, 41, 145–156.
Frey, R., & Sin, C. (1999). Bounds on European option prices under stochastic volatility.
Mathematical Finance, 9, 97–116.
Gotoh, J. Y., & Konno, H. (2002). Bounding option prices by semidefinite programming: A cutting
plane algorithm. Management Science, 48, 665–678.
Grundy, B. (1991). Option prices and the underlying assets return distribution. Journal of Finance,
46, 1045–1069.
Hobson, D., Laurence, P., & Wang, T. H. (2005). Static-arbitrage upper bounds for the prices of
basket options. Quantitative Finance, 5, 329–342.
Huang, J. (2004). Option pricing bounds and the elasticity of the pricing kernel. Review of
Derivatives Research, 7, 25–51.
Ingersoll, J. (1989). Theory of financial decision making. Totowa, New Jersey: Rowman &
Littlefield.
7 Nonparametric Bounds for European Option Prices 231
Levy, H. (1985). Upper and lower bounds of put and call option value: Stochastic dominance
approach. Journal of Finance, 40, 1197–1217.
Lo, A. (1987). Semi-parametric upper bounds for option prices and expected payoff. Journal of
Financial Economics, 19, 373–388.
Merton, R. (1973). The theory of rational option pricing. The Bell Journal of Economics and
Management Science, 4, 141–183.
Peña, J., Vera, J. C., & Zuluaga, L. F. (2010). Static-arbitrage lower bounds on the prices of basket
options via linear programming. Quantitative Finance, 10, 819–827.
Perrakis, S. (1986). Option pricing bounds in discrete time: Extensions and the pricing of the
American put. Journal of Business, 59, 119–141.
Perrakis, S., & Ryan, J. (1984). Option pricing in discrete time. Journal of Finance, 39, 519–525.
Ritchken, P. (1985). On option pricing bounds. Journal of Finance, 40, 1219–1233.
Ritchken, P., & Kuo, S. (1988). Option bounds with finite revision opportunities. Journal of
Finance, 43, 301–308.
Ritchken, P., & Kuo, S. (1989). On stochastic dominance and decreasing absolute risk averse
option pricing bounds. Management Science, 35, 51–59.
Rodriguez, R. (2003). Option pricing bounds: Synthesis and extension. Journal of Financial
Research, 26, 149–164.
Rubinstein, M. (1976). The valuation of uncertain income stream and the pricing of options. The
Bell Journal of Economics, 7, 407–425.
Scott, L. (1997). Pricing stock options in a jump diffusion model with stochastic volatility and
interest rates: Applications of Fourier inversion methods. Mathematical Finance, 7, 413–424.
Zhang, P. (1994). Bounds for option prices and expected payoffs. Review of Quantitative Finance
and Accounting, 4, 179–197.
Can Time-Varying Copulas Improve the
Mean-Variance Portfolio? 8
Chin-Wen Huang, Chun-Pin Hsu, and Wan-Jiun Paul Chiou
Contents
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 234
8.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 235
8.3 Empirical Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236
8.3.1 Copulas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 236
8.3.2 Copula Specifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237
8.3.3 Gaussian Copula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237
8.3.4 Student’s t-Copula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 237
8.3.5 Archimedean Copulas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 238
8.3.6 Portfolio Construction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 239
8.4 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240
8.5 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240
8.5.1 Dependence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 240
8.5.2 Average Portfolio Returns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 241
8.5.3 Testing Significance of Return Difference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 243
8.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246
Appendix 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 247
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 250
C.-W. Huang
Department of Finance, Western Connecticut State University, Danbury, CT, USA
e-mail: huangc@wcsu.edu
C.-P. Hsu (*)
Department of Accounting and Finance, York College, The City University of New York,
Jamaica, NY, USA
e-mail: chsu@york.cuny.edu
W.-J.P. Chiou
Department of Finance and Law College of Business Administration, Central Michigan
University, Mount Pleasant, MI, USA
e-mail: Chiou1P@cmich.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 233
DOI 10.1007/978-1-4614-7750-1_8,
# Springer Science+Business Media New York 2015
234 C.-W. Huang et al.
Abstract
Research in structuring asset return dependence has become an indispensable
element of wealth management, particularly after the experience of the recent
financial crises. In this paper, we evaluate whether constructing a portfolio using
time-varying copulas yields superior returns under various weight updating
strategies. Specifically, minimum-risk portfolios are constructed based on vari-
ous copulas and the Pearson correlation, and a 250-day rolling window tech-
nique is adopted to derive a sequence of time-varied dependencies for each
dependence model. Using daily data of the G7 countries, our empirical findings
suggest that portfolios using time-varying copulas, particularly the Clayton
dependence, outperform those constructed using Pearson correlations. The
above results still hold under different weight updating strategies and portfolio
rebalancing frequencies.
Keywords
Copulas • Time-varying dependence • Portfolio optimization • Bootstrap • Out-
of-sample return • Performance evaluation • GARCH • Gaussian copula •
Student’s t-copula • Gumbel copula • Clayton copula
8.1 Introduction
The return of a portfolio depends heavily on its asset dependence structure. Over the
past decade, copula modeling has become a popular alternative to Pearson corre-
lation modeling when describing data with an asymptotic dependence structure and
a non-normal distribution.1 However, several critical issues attached to the appli-
cations of copulas emerge: Do portfolios using time-varying copulas outperform
those constructed with Pearson correlations? How does the risk return of copula-
based portfolios change over the business cycle? The estimation of parameters has
become particularly critical for finance academics and professionals on the heels of
the recent financial crises. In this chapter, we model the time-varying dependence
of an international equity portfolio using several copula functions and the Pearson
correlation. We investigate whether a portfolio constructed with copula dependence
yields superior returns as compared to a portfolio constructed using a Pearson
correlation under various weight updating strategies.
This paper extends the existing literature in two ways. First, we estimate our
time-varying copulas using a rolling window of the latest 250 trading days. It is well
accepted that the dependencies between asset returns are time varying (Kroner and
Ng 1998; Ang and Bekaert 2002). Differing from the regime-switching type used in
Rodriguez (2007) and Okimoto (2008) or the time-evolving type GARCH model
used in Patton (2006a), we estimate time-varying copulas via a rolling window
1
See Chan et al. (1999), Dowd (2005), Patton (2006a), Engle and Sheppardy (2008), Chollete
et al. (2009), Bauer and Vorkink (2011).
8 Can Time-Varying Copulas Improve the Mean-Variance Portfolio? 235
based on daily data from the previous year. The rolling window method has several
benefits. First, it is a method frequently adopted by practitioners. Second, the rolling
window method considers only the past year’s information when forming depen-
dencies, thus avoiding disturbances that may have existed in the distant past.
Several studies have applied this technique, such as Aussenegg and Cech (2011).
However, Aussenegg and Cech (2011) considered only the daily Gaussian and
Student’s t-copulas in constructing models, and it is reasonable to consider monthly
and quarterly frequencies, given that portfolio managers do not adjust their portfo-
lios on a daily basis. Our research also extends Aussenegg and Cech’s (2011) study
by including the Archimedean copulas to govern the strength of dependence.
Second, our study investigates how the choice of copula functions affects
portfolio performance during periods of economic expansion and recession. The
expansion and recession periods we define are based on announcements from the
National Bureau of Economic Research (NBER). While the use of copula functions
in financial studies has grown enormously, little work has been done in comparing
copula dependencies under different economic states.
Using daily US dollar-denominated Morgan Stanley Capital International
(MSCI) indices of the G7 countries, our empirical results suggest that the
copula-dependence portfolios outperform the Pearson-correlation portfolios. The
Clayton-dependence portfolios, for most scenarios studied, deliver the highest
portfolio returns, indicating the importance of lower-tail dependence in building
an international equity portfolio. Moreover, the choice of weight updating fre-
quency matters. As we increase the weight updating frequency from quarterly to
monthly, the portfolio returns for the full sample and recession periods also
increase, regardless of the choice of dependence measure. Our finding supports
the value of active portfolio reconstruction during recession periods.
This paper is organized as follows. Section 8.2 reviews the literature on copula
applications in portfolio modeling. Section 8.3 describes the empirical models.
Section 8.4 presents the data used. The main empirical results are reported in
Sect. 8.5. Section 8.6 concludes.
8.3.1 Copulas
2
For detailed derivations, please refer to Cherubini et al. (2004), Demarta and McNeil (2005),
Embrechts et al. (2003), Embrechts et al. (2005), Franke et al. (2008), Nelson (2006), and
Patton (2009).
8 Can Time-Varying Copulas Improve the Mean-Variance Portfolio? 237
In this paper, we consider four copula functions: the Gaussian, Student’s t, Gumbel,
and Clayton. The Gaussian copula focuses on the center of the distribution and
assumes no tail dependence. The Student’s t-copula stresses both the center of the
distribution and symmetric tail behaviors. The Clayton copula emphasizes lower-
tail dependence, while the Gumbel copula focuses on upper-tail dependence.
Table 8.1 summarizes the characteristics of each copula in detail.
The Gaussian copula is frequently seen in the finance literature due to its close
relationship to the Pearson correlation. It represents the dependence structure of two
normal marginal distributions. According to Nelson (2006), the bivariate Gaussian
copula can be expressed as
where F denotes the univariate standard normal distribution function and Fr is the
joint distribution function of the bivariate standard normal distribution with correla-
tion coefficient –1 r 1. The Gaussian copula has no tail dependence unless r ¼ 1.
Unlike the Gaussian copula, which fails to capture tail behaviors, the Student’s t-copula
depicts the dependence in the center of the distribution as well as in the tails. The
Student’s t-copula is defined using the multivariate t distribution and can be written as
238 C.-W. Huang et al.
ð t1 ð 1 vþ2
v ðxÞ tv ðyÞ 1 s2 2rst þ t2 2
where t2v, r indicates the bivariate joint t distribution, tv1 is the inverse of
the distribution of a univariate t distribution, v is the degrees of freedom, and r
is the correlation coefficient of the bivariate t distribution when v > 2.
The upper-tail dependence presents the probability that Y exceeds its qth
quantile given that X exceeds its qth quantile, considering the limit as q goes to
its infinity. If the limit lu∈[0,1] exists, then X and Y are said to show upper-tail
dependence. In the same manner, the coefficient of lower-tail dependence (ll) of
X and Y is described as
h
ll ðx; yÞ ¼ limq0þ P Y Fy ðqÞX Fx ðqÞ : (8.6)
Since both FU1 and FU2 are continuous density functions, the upper-tail
dependence can be presented as
P Y > Fy ðqÞX > Fx ðqÞ
lu ¼ lim : (8.7)
q0 P X > Fx ð qÞ
where wi is the weight of asset i, and V is the covariance matrix of asset returns.
Because dependence is a time-varying parameter, the data from a subset of
250 trading days prior to the given sample date t is used to derive the dependence
for date t. With 1,780 daily data points in our sample, we calculate a total of 1,531
dependencies for each copula method and the Pearson correlation. With these
dependencies, optimal portfolio weightings can be obtained by solving
a quadratic function subject to specified constraints. The optimal weightings for
time t are used to calculate the realized portfolio returns for (t + 1).4
In practice, portfolio managers periodically reexamine and update the optimal
weights of their portfolios. If the asset allocation of an existing portfolio has
deviated from the target allocation to a certain degree and if the benefit of updating
exceeds its costs, a portfolio reconstruction action is executed. In this paper,
we construct a comprehensive study of portfolio returns by varying the state of
3
Short selling usually involves other service fees, which vary depending on the creditability of the
investors. Because the focus of this study is on the effect of the dependence structure on portfolio
performance, we assume that short selling is not allowed to simplify the comparison.
4
For example, we use return data from t1 to t250 to calculate the optimal portfolio weights with
dependencies estimated from the copulas and the Pearson correlation. The optimal portfolio
weights are applied to the return data at t251 to calculate the realized portfolio returns.
240 C.-W. Huang et al.
the economy (i.e., expansion or recession), the dependence structure, and the
frequency of weight updating, i.e., quarterly, monthly, and daily. Quarterly
updating allows investors to update the optimal weights on the first trading days
of March, June, September, and December; monthly updating allows investors to
update the optimal weights on the first trading days of each month. Under daily
updating, investors update the optimal weights every trading day.
8.4 Data
The data are the US dollar-denominated daily returns of the Morgan Stanley Capital
International (MSCI) indices for the G7 countries, including Canada, France,
Germany, Italy, Japan, the UK, and the USA. The sample period covers from the
first business day in June 2002 to the last business day in June 2009, for a total of
1,780 daily observations. Based on the definitions provided by the National Bureau
of Economic Research, we separate the data into an expansion period from June
2002 to November 2007 and a recession period from December 2007 to June 2009.
Table 8.2 presents the descriptive statistics. Among the G7 countries, Canada
had the highest daily returns, while the USA had the lowest. Germany, however,
experienced the most volatile returns. All return series exhibit high kurtosis,
suggesting fat tails on return distributions. The results of the Jarque-Bera test reject
the assumption that the G7 indices have normal distributions.
8.5.1 Dependence
Using 1,780 daily data points from the G7 countries, for each dependence model,
we estimate 21 dependence pairs, each containing a sequence of 1,531 dependen-
cies. The parameters for the Gaussian, Student’s t, Gumbel, and Clayton copula
functions are estimated using the two-stage inference for the margins (IFM) method
proposed by Joe and Xu (1996) and Joe (1997).
8 Can Time-Varying Copulas Improve the Mean-Variance Portfolio? 241
The dependencies from the Pearson correlation are calculated using the standard
method. Appendix 1 shows the maximum and the minimum of the 21 dependence
pairs of each dependence model.
The graphs in Fig. 8.1 show the dependencies between the USA and other
countries as estimated via the various copulas and the Pearson correlation. In
general, the Gaussian copula estimation is similar to that of the corresponding
Pearson correlation, but the Student’s t-copulas show significant jumps over time.
For our sample period, Japan shows a low dependence with the US market as
compared to other economies.
Table 8.3 presents the average portfolio returns for the full sample period,
the expansion period, and the recession period for the quarterly, monthly, and
daily weight updating strategies. Under the quarterly weight updating, the
Clayton-dependence portfolios have the highest average returns at 6.07 % in the
expansions and 12.52 % in the recessions, and the Pearson-correlation portfolios
have the lowest average returns, 5.48 % in the expansions and 14.25 % in the
recessions. The order of portfolio performance, in the form of its dependence
model regardless of the state of economy, is as follows: the Clayton copula, the
Gumbel copula, the Student’s t-copula, the Gaussian copula, and the Pearson
correlation. Because both the Clayton and Gumbel copulas highlight the tail
dependence between assets, the empirical evidence suggests that with
a quarterly weighting strategy, tail dependence, particularly lower-tail depen-
dence, is important for obtaining superior average portfolio returns across differ-
ent states of the economy.
As we increase the updating frequency from quarterly to monthly, similar empir-
ical results are observed. That is, the Clayton-copula portfolios yield the highest
average returns, while the Pearson-correlation portfolios provide the lowest average
returns. In the expansion periods, the order of portfolio performance, in the form of its
dependence model, is as follows: the Clayton copula, the Student’s t-copula, the
Gaussian copula, the Gumbel copula, and the Pearson correlation. In recession
periods, the order of portfolio performance, in the form of its dependence model, is
as follows: the Clayton copula, the Gumbel copula, the Student’s t- copula, the
Gaussian copula, and the Pearson correlation. According to Kole et al. (2007), the
Gaussian copula, which does not consider lower-tail dependence, tends to be overly
optimistic on the portfolio’s diversification benefits, and the Gumbel copula, which
focuses on the upper tail and pays no attention to the center of the distribution, tends
to be overly pessimistic on the portfolio’s diversification benefits. We verify this
argument by observing that the Gumbel-copula portfolio outperforms only the
Pearson-correlation portfolio in the expansion periods, while the Gaussian-copula-
dependence portfolio outperforms the Pearson-correlation portfolio only in recession
periods. Interestingly, as we increase the weight updating frequency from quarterly to
monthly, the average portfolio returns for the full sample and recession periods also
242 C.-W. Huang et al.
The Dependence using Different Copulas and Pearson Correlation: US vs. Canada
0.90
a
0.80
0.70
0.60
0.50
0.40
0.30 Clayton
Gaussian
0.20 Gumbel
0.10 t
Pearson
0.00
Jan-03 May-04 Sep-05 Jan-07 May-08 Sep-09
The Dependence using Different Copulas and Pearson Correlation: US vs. France
0.800
0.700
b
0.600
0.500
0.400
0.300 Clayton
0.200 Gaussian
Gumbel
0.100 t
Pearson
0.000
Jan-03 May-04 Sep-05 Jan-07 May-08 Sep-09
The Dependence using Different Copulas and Pearson Correlation: US vs. Germany
0.80
c
0.70
0.60
0.50
0.40
0.30 Clayton
0.20 Gaussian
Gumbel
0.10 t
Pearson
0.00
Jan-03 May-04 Sep-05 Jan-07 May-08 Sep-09
The Dependence using Different Copulas and Pearson Correlation: US vs. Italy
0.08
0.70
d
0.60
0.50
0.40
0.30 Clayton
0.20 Gaussian
Gumbel
0.10 t
0.00 Pearson
Jan-03 May-04 Sep-05 Jan-07 May-08 Sep-09
The Dependence using Different Copulas and Pearson Correlation: US vs. Japan
0.40 Clayton
e Gaussian
0.30 Gumbel
t
0.20 Pearson
0.10
0.00
Jan-03 May-04 Sep-05 Jan-07 May-08 Sep-09
−0.10
−0.20
The Dependence using Different Copulas and Pearson Correlation: US vs. UK
0.80
0.70
f
0.60
0.50
0.40
0.30 Clayton
0.20 Gaussian
Gumbel
0.10 t
Pearson
0.00
Jan-03 May-04 Sep-05 Jan-07 May-08 Sep-09
Fig. 8.1 The dependence using different copulas and Pearson correlation. Panel (a) USA versus
Canada. Panel (b) USA versus France. Panel (c) USA versus Germany. Panel (d) USA versus
Italy. Panel (e) USA versus Japan. Panel (f) USA versus UK
increase, regardless of the choice of dependence measures. Thus, the empirical results
seem to support the need for active portfolio reconstruction during recessions.
As the weight updating frequency increases to daily, the Clayton copula delivers
only the highest average portfolio returns during the expansion period. The
Student’s t-copula, by contrast, generates the highest portfolio average returns
for the full sample and recession periods. The influence of the lower-tail
dependence seems to diminish under daily weight reconstruction. The Gaussian-
copula portfolio delivers the worst portfolio performance in both expansion and
recession periods.
The results reported in the previous section show the average portfolio returns for
different dependencies and weight updating frequencies. One issue with average
returns is that if extreme values exist over the examined period, the empirical
results may be biased and relatively high-standard deviations will be reported.
Previous methods of examining the robustness of portfolio performance usually
build on the data normality assumption (Jobson and Korkie 1981; Memmel
2003), which goes against the empirical facts.
244 C.-W. Huang et al.
To cope with this problem, Ledoit and Wolf (2008) proposed an alternative
testing method using the inferential studentized time-series bootstrap. Ledoit and
Wolf’s (2008) method is as follows.5 Let a and b be two investment strategies, and
let rat and rbt be the portfolio returns for strategies a and b, respectively, at time t,
where t ranges from 1 to i. The mean vector m and the covariance matrix S for the
return pairs (ra1,rb1)’,. . .,(rat,rbt)’ are denoted by
X
s2 s
ma
m¼ and ¼ a ab
: (8.12)
mb sab s2b
5
For detailed derivations and computer codes, please refer to Ledoit and Wolf (2008).
8 Can Time-Varying Copulas Improve the Mean-Variance Portfolio? 245
and
where D is the difference between the two Sharpe ratios, and Sa and Sb are the
Sharpe ratios for strategies a and b, respectively.
Let the second moments of the returns from strategies a and b be denoted by ga
and gb. Then ga ¼ E(gat2) and gb ¼ E(gbt2). Let u and ^u be (ma, mb, ga, gb)’ and
^ can be expressed as
ðm^a , m^b , g^a , g^a Þ0 , respectively. Then D and D
^ ¼ f ð^u Þ
D ¼ f ðuÞ and D (8.15)
ma mb
pffi d
where f ðuÞ ¼ pffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffiffiffiffi
ffi and ið^u uÞ! N ð0; CÞ:
ga ma
2 gb mb
2
yt ¼ r ta m^a , r tb m^b , r 2 ^ 2 ^
ta ga , r tb gb t ¼ 1, . . . , i: (8.16)
6
Ledoit and Wolf (2008) suggested that 5,000 iterations guarantee a sufficient sample. We adopt
a higher standard of 10,000 iterations to strengthen our testing results.
246 C.-W. Huang et al.
8.6 Conclusions
Appendix 1
CA FR DE IT JP UK USA
Panel A: Gaussian dependence
CA
Max
Min
FR
Max 0.7064
Min 0.3914
DE
Max 0.6487 0.9686
Min 0.2016 0.7856
IT
Max 0.6320 0.9407 0.9274
Min 0.2143 0.2303 0.7001
JP
Max 0.1814 0.2761 0.2478 0.4023
Min 0.2115 0.2238 0.2229 0.0119
UK
Max 0.6393 0.9100 0.8665 0.8575 0.4664
Min 0.1945 0.2384 0.2008 0.2050 0.0329
(continued)
248 C.-W. Huang et al.
CA FR DE IT JP UK USA
USA
Max 0.7221 0.5031 0.5231 0.4661 0.5831 0.5671
Min 0.1864 0.2127 0.2087 0.2414 0.2241 0.1997
Panel B: Student’s t dependence
CA
Max
Min
FR
Max 0.7509
Min 0.3921
DE
Max 0.4578 0.9810
Min 0.1070 0.7476
IT
Max 0.4367 0.9810 0.9586
Min 0.1089 0.7476 0.6937
JP
Max 0.1093 0.1679 0.1522 0.6846
Min 0.1284 0.1511 0.1574 0.0093
UK
Max 0.4478 0.8055 0.7380 0.7316 0.7335
Min 0.1094 0.1546 0.1359 0.1230 0.0284
USA
Max 0.5733 0.8055 0.3457 0.3176 0.4375 0.6614
Min 0.1107 0.1546 0.1343 0.1360 0.1519 0.2687
Panel C: Gumbel dependence
CA
Max
Min
FR
Max 0.5947
Min 0.3220
DE
Max 0.3744 0.9063
Min 0.0000 0.5928
IT
Max 0.3666 0.7286 0.8544
Min 0.0000 0.0000 0.5516
JP
Max 0.0961 0.1384 0.1222 0.5356
Min 0.0000 0.0000 0.0000 0.0200
(continued)
8 Can Time-Varying Copulas Improve the Mean-Variance Portfolio? 249
CA FR DE IT JP UK USA
UK
Max 0.3650 0.6946 0.6156 0.6086 0.5855
Min 0.0000 0.0000 0.0000 0.0000 0.0234
USA
Max 0.4340 0.2816 0.2952 0.2649 0.3261 0.5967
Min 0.0000 0.0000 0.0000 0.0000 0.0000 0.2493
Panel D: Clayton dependence
CA
Max
Min
FR
Max 0.6763
Min 0.2899
DE
Max 0.3585 0.9327
Min 0.0000 0.6696
IT
Max 0.3463 0.7635 0.9004
Min 0.0000 0.0000 0.6006
JP
Max 0.0038 0.0408 0.0287 0.6476
Min 0.0000 0.0000 0.0000 0.0000
UK
Max 0.3657 0.7193 0.6567 0.6506 0.6794
Min 0.0000 0.0000 0.0000 0.0000 0.0000
USA
Max 0.5248 0.2450 0.2476 0.1987 0.3472 0.6622
Min 0.0000 0.0000 0.0000 0.0000 0.0000 0.1982
Panel E: Pearson correlation
CA
Max
Min
FR
Max 0.7002
Min 0.3966
DE
Max 0.6944 0.9726
Min 0.3645 0.7890
IT
Max 0.6833 0.9596 0.9477
Min 0.3877 0.8204 0.6965
(continued)
250 C.-W. Huang et al.
CA FR DE IT JP UK USA
JP
Max 0.3549 0.4594 0.4702 0.4073
Min 0.0411 0.0251 0.0170 0.0072
UK
Max 0.7080 0.9573 0.9298 0.9181 0.4612
Min 0.3676 0.7791 0.6572 0.6990 0.0154
USA
Max 0.7586 0.6096 0.7443 0.5871 0.2078 0.5480
Min 0.3764 0.2647 0.2921 0.2481 0.1562 0.1913
References
Aas, K., Czado, C., Frigessi, A., & Bakken, H. (2007). Pair-copula constructions of multiple
dependence. Insurance: Mathematics and Economics, 2, 1–25.
Ang, A., & Bekaert, G. (2002). International asset allocation with regime shifts. Review of
Financial Studies, 15, 1137–1187.
Ang, A., & Chen, J. (2002). Asymmetric correlations of equity portfolios. Journal of Financial
Economics, 63, 443–494.
Aussenegg, W., & Cech, C. (2011, forthcoming). Simple time-varying copula estimation. In
A. S. Barczak, E. Dziwok (Ed.), Mathematical econometrical and computational methods in
finance and insurance. Katowice: University of Economics in Katowice.
Bauer, G. H., & Vorkink, K. (2011). Forecasting multivariate realized stock market volatility.
Journal of Econometrics, 160, 93–101.
Chan, L. K., Karceski, J., & Lakonishok, J. (1999). On portfolio optimization: Forecasting
covariances and choosing the risk model. Review of Financial Studies, 12, 937–974.
Cherubini, U., Luciano, E., & Vecchiato, W. (2004). Copula methods in finance. Hoboken: Wiley.
Chollete, L., Heinen, A., & Valdesogo, A. (2009). Modeling international financial
returns with a multivariate regime-switching copula. Journal of Financial Econometrics, 7,
437–480.
Chollete, L., Peña, V., & Lu, C. (2011). International diversification: A copula approach. Journal
of Banking and Finance, 35, 403–417.
Demarta, S., & McNeil, A. (2005). The t copula and related copulas. International Statistical
Review, 73, 111–129.
Dowd, K. (2005). Copulas and coherence. Journal of Portfolio Management, Fall, 32(1), 123–127.
Embrechts, P., Frey, R., & McNeil, A. (2005). Quantitative risk management. Princeton: Princeton
University Press.
Embrechts, P., Lindskog, F., & McNeil, A. (2003). Modelling dependence with copulas and
applications to risk management. In S. T. Rachev (Ed.), Handbook of heavy tailed distributions
in finance (pp. 329–384). Amsterdam: Elsevier/North-Holland.
Engle, R. F., & Sheppardy K. (2008). Evaluating the specification of covariance models for large
portfolios (Working Paper). New York University.
Franke, J., Hardle, W., & Hafner, C. (2010). Statistics of financial markets: An introduction. New
York: Springer.
Hu, L. (2006). Dependence patterns across financial markets: A mixed copula approach. Applied
Financial Economics, 16, 717–729.
8 Can Time-Varying Copulas Improve the Mean-Variance Portfolio? 251
Jobson, J., & Korkie, B. (1981). Performance hypothesis testing with the Sharpe and Treynor
measures. Journal of Finance, 36, 889–908.
Joe, H. (1997). Multivariate models and dependence concepts. New York: Chapman & Hall/CRC.
Joe, H., & Xu, J. J. (1996). The estimation method of inference functions for margins for
multivariate models (Technical Report 166). Department of Statistics, University of British
Columbia.
Kole, E., Koedijk, K., & Verbeek, M. (2007). Selecting copulas for risk management. Journal of
Banking and Finance, 31, 2405–2423.
Kroner, K. E., & Ng, V. (1998). Modeling asymmetric comovements of asset returns. Review of
Financial Studies, 11, 817–844.
Ledoit, O., & Wolf, M. (2008). Robust performance hypothesis testing with the Sharpe ratio.
Journal of Empirical Finance, 15, 850–859.
Longin, F., & Solnik, B. (2001). Extreme correlation of international equity markets. Journal of
Finance, 56, 649–676.
Markowitz, H. (1952). Portfolio selection. Journal of Finance, 7, 77–91.
Memmel, C. (2003). Performance hypothesis testing with the Sharpe ratio. Finance Research
Letters, 1, 21–23.
Nelson, R. (2006). An introduction to copulas. New York: Springer.
Okimoto, T. (2008). New evidence of asymmetric dependence structures in international equity
markets. Journal of Financial and Quantitative Analysis, 43, 787–816.
Patton, A. (2006a). Modelling asymmetric exchange rate dependence. International Economic
Review, 47, 527–556.
Patton, A. (2006b). Estimation of multivariate models for time series of possibly different lengths.
Journal of Applied Econometrics, 21, 147–173.
Patton, A. (2009). Copula-based models for financial time series. In T. Mikosch, J. P. Kreiss,
T. G. Andersen, R. A. Davis, J. P. Kreiss, & T. Mikosch (Eds.), Handbook of financial time
series. New York: Springer.
Rodriguez, J. C. (2007). Measuring financial contagion: A copula approach. Journal of Empirical
Finance, 14, 401–423.
Sklar, A. (1959) “Fonctions de Répartition à n Dimensions et Leurs Marges,” Publications de
l’Institut de Statistique de l’Université de Paris 8, 229–231.
Determinations of Corporate Earnings
Forecast Accuracy: Taiwan Market 9
Experience
Contents
9.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 254
9.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255
9.3 Testable Hypotheses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258
9.3.1 Firm Size . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258
9.3.2 Volatility of Market . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 258
9.3.3 Volume Turnover . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259
9.3.4 Corporate Earnings Variance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 259
9.3.5 Type of Industry . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 260
9.3.6 Forecasting Experience . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 260
9.4 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 260
9.4.1 Comparison of Management and Analyst’s Earnings Forecast . . . . . . . . . . . . . . . . . 260
9.4.2 Factors Influencing the Absolute Errors of Earnings Forecast . . . . . . . . . . . . . . . . . . 263
9.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 266
Methodology Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 269
Sample Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 269
Variable Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 270
Market Volatility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 272
Regression Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 274
Wilcoxon Two-Sample Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 276
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 276
K. Hung (*)
Division of International Banking & Finance Studies, Texas A&M International University,
Laredo, TX, USA
e-mail: ken.hung@tamiu.edu
K.-H. Lee
Department of Finance, College of Business, Bloomsburg University of Pennsylvania,
Bloomsburg, PA, USA
e-mail: klee@bloomu.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 253
DOI 10.1007/978-1-4614-7750-1_9,
# Springer Science+Business Media New York 2015
254 K. Hung and K.-H. Lee
Abstract
Individual investors are actively involved in stock market and are making
investment decision based on publicly available and nonproprietary information,
such as corporate earnings forecasts from the management and the financial
analyst. Also, the management forecast is another important index investors
might use.
To examine the accuracy of the earnings forecasts, the following test meth-
odology have been conducted. Multiple regression models are used to examine
the effect of six factors: firm size, market volatility, trading volume turnover,
corporate earnings variances, type of industry, and experience. If the two-sample
groups are related, Wilcoxon two-sample test will be used to determine the
relative earnings forecast accuracy.
The results indicate that firm size has no effect on management forecast,
voluntary management forecast, mandatory management forecast, and analysts’
forecast. There are some indications that forecasting accuracy is affected by
market ups and downs. The results also reveal that relative accuracy of earnings
forecasts is not a function of trading volume turnover. However, management’s
earnings forecast and analysts’ forecasts are sensitive to earnings variances.
Readers are well advised and referred to the chapter appendix for methodo-
logical issues such as sample selection, variable definition, regression model,
and Wilcoxon two-sample test.
Keywords
Multiple regression • Wilcoxon two-sample test • Corporate earnings • Forecast
accuracy • Management earnings • Firm size • Corporation regulation •
Volatility • Trade turnover • Industry
9.1 Introduction
In recent times, individual investors are actively involved in stock market and are
making investment decision based on publicly available and nonproprietary infor-
mation. Corporate earnings forecasts are an important investment tool for investors.
Corporate earnings forecasts come from two sources: the company management
and financial analyst. As an insider, the management has the advantage of
possessing more information and hence provides a more accurate earnings forecast.
However, because of the existing relationship of the company with its key investor
group, the management may have a tendency to take an optimistic view and
overestimate its future earnings. In contrast, the financial analysts are less informed
about the company and often rely on management briefings. They have more
experiences in the overall market and economies and are expected to analyze
companies objectively. Hence, analysts should provide reliable and more accurate
earnings forecast.
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 255
Whether investors should rely on the earnings forecast made by the management
or by the analyst is a debatable issue. Many researchers have examined the accuracy
of such earnings forecasts. Because there are differences in methodologies, sample
selections and time horizons, the findings and conclusions from the previous studies
are conflicting and inconclusive. This motivated us to do a new analysis by using
a different methodology.
According to Regulations for Publishing Corporate Earnings Forecast imposed
by the Department of Treasury,1 publicly traded Taiwanese companies have to
publish their earnings forecasts under the following situations:
1. To issue new stocks or acquired liabilities;
2. When more than one third of the board has been changed;
3. When one of the situations as listed in section 185 of the corporation regulations
happens;
4. Merger and acquisitions;
5. Profit gain/loss up to one third of its annual revenue due to an unexpected event;
6. Revenue loss over 30 % compared to last year;
7. Voluntarily publish its earnings forecast
Since management earnings forecasts are mandatory or voluntary, the focus of
this research is to examine the accuracy of management’s overall earnings forecast,
management voluntary earnings forecast, management mandatory earnings fore-
cast, and financial analyst earnings forecast.
Jaggi (1980) examined the impact of company size on forecast accuracy using
management’s earnings forecasts from the Wall Street Journal and analysts’
earnings forecasts from Value Line Investment Service from 1971 to 1974. He
argued that because a larger company has strong financial and human capital
resources, its management’s earnings forecast would be more accurate than the
analyst’s. The sample data were classified into six categories based on the size of
the firms’ total revenue to examine the factors that attribute to the accuracy of
management’s earnings forecast with the analyst’s. The result of his research did
not support his hypothesis that management’s forecast is more accurate than the
analyst’s.
Bhushan (1989) assumed that it is more profitable trading large companies’
stocks because large companies have better liquidity than the small ones. Therefore,
the availability of information is related to company size. His research results
support his hypothesis that the larger the company size, the more information is
available to financial analysts and the more accurate their earnings forecasts are.
Kross and Schreoder (1990) proposed that brokerage firm’s characteristics influ-
ence analysts’ earnings forecasting accuracy. In their analysis, sample analysts’
1
Securities Regulation Committee, Department of Treasury, series 00588, volume #6, 1997
256 K. Hung and K.-H. Lee
earnings forecasts from 1980 to 1981 were obtained from Value Line Investment,
and the market value of a firm is used as the size of the firm. The results of this study
on analysts’ earnings forecasts did not find a positive relation between the company
size and the analyst’s forecast accuracy. Xu (1990) used the data range from 1986 to
1990 and the logarithm of average total revenue as a proxy of company earnings
and examined factors associated with the accuracy of analysts’ earnings forecast.
The hypothesis that the larger the firm size is, the more accurate the analysts’
earnings forecast would be was supported.
Su (1996) focuses on comparison of relative accuracy of management and
analysts’ earnings forecasts by using cross-sectional design method. Samples
selection includes forecast data during the time period from 1991 to 1994. The
company’s “book value” of its total assets is used as a proxy for the size of the
company in the regression analysis. The author believes that analysts are more
attracted to larger companies, and there are more incentives for them to follow large
companies than small companies in their forecasting. Therefore, the size of
a company will affect the relative accuracy of analyst earnings forecast. On the
other hand, large companies possess excessive human and financial resources and
information which analysts have no access to, to allow managers to anticipate the
corporate future earnings with high accuracy. The study results show that analyst
and voluntary earnings forecast accuracy for larger companies are higher than
forecast accuracy for small companies.
Yie (1996) examines the factors influencing management and financial analyst
earnings forecasting accuracy. Data used in this study are the earnings forecasts
during the years 1991–1995. She uses the company’s total assets and market value
of the company’s equity as proxies for company size. The finding of this research
reveals that the relative earnings forecast accuracy (management, voluntary man-
agement, mandatory management, and analyst) is not affected by the size of the
company when the company’s total assets are used as the proxy of company size.
The result also indicates that mandatory management’s earnings forecast and
analysts’ earnings forecasts are influenced by company size if market value of
company’s equity is used.
Xu (1990) examines the relative accuracy of analysts’ earnings forecasts,
a hypothesis that market volatility is one of factors that influence the relative
accuracy of analyst earnings forecast. In upmarket situation, a vast amount of
information regarding corporate earnings and overwhelming trading activities
may hinder the analyst from getting realistic and objective information; thus,
overoptimistic forecast might be a result. In contrast, when market is experiencing
a downturn, individual investors are less speculative and more rational; thus,
information about corporate earnings tends to be more accurate. Under these
circumstances, the analyst tends to provide earnings forecasts with a higher level
of accuracy. The results of this study support the author’s hypothesis. Jiang (1993)
examines the relative accuracy between management’s earnings forecast and ana-
lyst earnings forecast. He hypothesizes that analysts’ earnings forecast has a higher
degree of accuracy in down market compared to upmarket situation. He uses
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 257
samples forecast data from years 1991 to 1993 in his analysis and finds that the
result of this research supports his argument.
Das et al. (1998) used a cross-sectional approach to study the optimistic
behavior of financial analysts. Especially, they focused on the predicative accu-
racy of past information analysts’ earnings forecast associated with magnetite of
the bias in analysts’ earnings forecasts. The sample selection covers the time
period from 1989 to 1993 with 274 companies’ earnings forecasts information.
A regression method was used in this research. The term “optimistic behavior” is
referred to as the optimistic earnings forecasts made by financial analysts. The
authors hypothesize the following scenario: there is higher demand for nonpublic
information for firms whose earnings are more difficult to predict than for firms
whose earnings can be accurately forecasted using public information. Their
finding supports the hypothesis that analysts will make more optimistic forecasts
for low-predictability firms with an assumption that optimistic forecast facilitates
access to management’s nonpublic information. Clement (1999) studies the rela-
tion between the quality and forecast accuracy of analysts’ reports. It also identifies
systematic and time persistence in analysts’ earnings forecast accuracy and exam-
ines the factors associated with degree of accuracy. Using the I/B/E/S database, the
author has found that earnings forecast accuracy is positively related with analysts’
experience (a surrogate for analyst ability and skill) and employer size (a surrogate
for resources available) and inversely related with the number of firms and
industries followed by the analyst. The sample selection covers the time horizon
from 1983 to 1994 with earnings forecasts of 9,500 companies and 7,500 analysts.
The author believes that as the analyst’s experience increases, his earnings forecast
accuracy will increase, which implies that the analyst has a better understanding of
the idiosyncrasies of a particular firm’s reporting practices or he might establish
a better relationship with insiders and therefore gain better access to the managers’
private information. An analyst’s portfolio complexity is also believed to have
association with his earnings forecast accuracy. He hypothesizes that forecast
accuracy would decrease with the number of industries/firms followed. The effect
of available resources impacts analyst’s earnings forecast in such a way that
analysts employed by a larger broker firm supply more accurate forecasts than
smaller ones. The rationale behind this hypothesis is that the analyst hired
by a large brokerage firm has better access to the private information of managers
at the companies he follows. Large firms have more advanced networks that
allow the firms to better disseminate their analyst’s recommendations into the
capital markets. The results of this research support the hypothesis made by the
author.
Xiu (1992) studies the relative accuracy of management and analysts’ earn-
ings forecasts using Taiwan database covering the period 1986–1990. The
management and analyst’s earnings forecasts used in the study are from Busi-
ness News, Finance News, Central News Paper, and The United Newspaper. The
research methodology is to examine management’s earnings forecast accuracy
with prior and posterior analyst’s earnings forecasts. The result reveals that
258 K. Hung and K.-H. Lee
The size of a firm is believed to have influence on the accuracy of analyst’s and
management’s earnings forecast. Jaggi (1980), Bhushan (1989), and Clement (1999)
found that the larger the company is, the more accurate the earnings forecast will
be. They believe that holding other factors constant, larger companies have more
financial and human resources available that allow the management to draw more
precise earnings forecast than smaller companies. Thus, forecasts and recommenda-
tions supplied by larger firms are more valuable and accurate than the smaller firms:
H1: The accuracy of management’s earnings forecast increases with the size of
the firm.
H2: The accuracy of management’s voluntary earnings forecast increases with the
size of the firm.
H3: The accuracy of management’s mandatory earnings forecast increases with the
size of the firm.
H4: The accuracy of analysts’ earnings forecast increases with the size of the firm.
period, investors tend to be rational about the next movement of the market; there is
less biased information regarding corporate earnings among the general investors;
thus, the information accessed by analysts will allow them to be more objective in
the earnings forecast. In contrast, the management has the insights on what is really
happening in the aspects of operation, finance, top management changes, and
profitability of the business. Even if they are less vulnerable regardless of what
the market situation is, voluntary management’s earnings forecast might be affected
by market volatility to some extent:
H5: Management’s earnings forecast will not be affected by volatility of market.
H6: Voluntary management’s earnings forecast is a function of the of market
volatility.
H7: Mandatory management’s earnings forecast is not affected by market volatility.
H8: Accuracy of analysts’ earnings forecast is affected by market volatility.
There may exist a relationship between the type of industry and earnings forecast
accuracy. They hypothesize that the difference between different industries may
result in different levels of accuracy on earnings forecast. Some analysts may not
possess adequate knowledge necessary in the forecasting in a particular industry;
therefore, their forecast may not be as accurate as management’s earnings forecast.
Hence, the following hypotheses can be tested:
H17: Type of industries influences the accuracy of management‘s earnings forecast.
H18: Type of industries affects the accuracy of voluntary management’s earnings
forecast.
H19: Type of industries affects the accuracy of mandatory management’s earnings
forecast.
H20: Type of industries affects the accuracy of analysts’ earnings forecast.
Table 9.1 Descriptive statistics of management and analyst’s earnings forecast errors
Management’s forecast error Analyst’s forecast error
Year Sample size Mean Standard error Maximum value Minimum value Mean Standard error Maximum value Minimum value
1999 402 1.15 3.64 38.55 0.0006 1.07 3.18 35.76 0.0006
1998 360 1.61 5.80 63.48 0.0017 2.10 7.10 75.81 0.0037
1997 317 0.74 2.20 29.87 0.0040 0.78 3.17 53.23 0.0002
1996 267 1.12 5.41 55.45 0.0007 1.33 6.49 71.34 0.0009
1995 226 0.86 2.93 33.27 0.0001 0.87 2.31 28.42 0.0018
1994 178 0.62 2.17 18.8 0.0015 0.62 2.17 23.86 0.0025
1993 157 12.71 149.1 1869.1 0.0011 13.80 164.1 2057.4 0.0003
1992 126 0.65 1.48 11.58 0.0013 0.73 1.48 8.14 0.0005
1991 144 0.84 2.99 32.93 0.0014 0.65 1.10 5.67 0.0079
1990 120 5.98 32.60 334.94 0.0017 4.86 20.92 187.3 0.0075
1989 116 1.11 2.28 18.66 0.0009 0.91 1.59 10.56 0.0056
1988 96 1.25 3.29 19.57 0.0001 1.08 2.74 15.43 0.0006
1987 78 0.66 1.88 15.73 0.0171 0.57 1.78 14.62 0.0025
1993–1999 1,907 2.031 0.983 1869.1 0.0001 2.236 1.083 2057. 0.0002
1987–1993 680 1.796 0.535 334.93 0.0001 1.503 0.34 187.368 0.0005
1987–1999 2,587 1.969 37.57 1869.1 0.0001 2.043 40.88 2057.44 0.0002
K. Hung and K.-H. Lee
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 263
Table 9.2 Wilcoxon signed-rank test for earnings forecast accuracy of management and analyst’s
forecasts errors
Year Sample size Negative ranks Positive ranks Ties Z-value Sig.
1999 402 150 230 22 3.119 0.002***
1998 360 209 140 11 5.009 0***
1997 317 147 140 30 0.5 0.96
1996 267 143 119 5 1.567 0.17
1995 226 117 89 20 2.857 0.004***
1994 178 82 90 6 0.339 0.734
1993 157 74 79 4 0.502 0.616
1992 126 60 62 4 1.136 0.256
1991 144 69 69 6 0.407 0.684
1990 120 43 76 1 2.941 0.003***
1989 116 46 67 3 2.7 0.007***
1988 96 28 60 8 3.503 0***
1987 78 27 51 0 3.315 0.001***
1993–1999 1,299 626 607 66 1.592 0.111
1987–1993 1,288 569 665 54 2.138 0.033**
1987–1999 2,587 1,195 1,272 120 0.346 0.73
Negative ranks: absolute error of management’s earnings forecast < absolute error of analyst
earnings forecast
Positive ranks: absolute error of management’s earnings forecast > absolute error of analyst
earnings forecast
Tie: absolute error of management’s earnings forecast ¼ absolute error of analyst earnings forecast
*
Significant level ¼ 0.1, ** significant level ¼ 0.05, *** significant level ¼ 0.01
Table 9.4 Regression model for the absolute errors of management earnings forecasts dating
from 1987 to 1999
Independent variable Correlation coefficient t-statistic p-value of t-statistic
Intercept 11.59 0.52 0.604
Size 0.69 0.71 0.478
I1: Market volatility 2.86 1.55 0.122
TR: Rate of trading volume turnover 0.09 0.10 0.923
CV: Corporate earnings variances 2.74 3.50 0.00***
E: Forecasting experience 1.00 1.25 0.212
I2: Cement 0.71 0.11 0.196
I3: Food 0.38 0.07 0.943
I4: Plastic 1.13 0.20 0.844
I5: Textile 0.87 0.19 0.853
I6: Electrical machinery 0.85 0.14 0.888
I7: Electronic equipment and cable 0.78 0.13 0.895
I8: Chemical 14.83 2.70 0.007***
I9: Glass and ceramic 1.36 0.17 0.867
I10: Paper manufacturing 0.84 0.11 0.912
I11: Steel 0.79 0.14 0.891
I12: Rubber 0.46 0.68 0.946
I13: Auto 2.74 0.28 0.782
I14: Electronics 0.82 0.17 0.865
I15: Construction 1.01 0.18 0.856
I16: Transportation 0.53 0.08 0.933
I17: Travel 0.76 0.10 0.924
I18: Insurance 0.92 0.14 0.886
I19: Grocery 0.65 0.09 0.925
R-square 0.016
I2–I19: dummy variables for industry
*
Significant level ¼ 0.10, ** significant level ¼ 0.05, *** significant level ¼ 0.01
increases. This supports the hypothesis H12 that corporate earnings variance has an
effect on management earnings forecast accuracy. From the examination for
Tables 9.5 (1993–1999) and 9.6 (1987–1992), management earnings forecast
accuracy is affected by corporate earnings variances during the recent years
(1993–1999).
Table 9.5 Regression model for the absolute errors of management earnings forecasts dating
from 1993 to 1999
Independent variable Correlation coefficient t-statistic p-value of t-statistic
Intercept 18.16 0.59 0.552
Size 1.00 0.76 0.450
I1: Market volatility 2.54 1.02 0.310
TR: Rate of trading volume turnover 0.04 0.03 0.976
CV: Corporate earnings variances 3.48 3.42 0.001***
E: Forecasting experience 1.26 1.23 0.218
I2: Cement 1.94 0.21 0.836
I3: Food 0.96 0.15 0.885
I4: Plastic 2.08 0.27 0.789
I5: Textile 1.39 0.23 0.815
I6: Electrical machinery 1.20 0.14 0.875
I7: Electronic equipment and cable 1.42 0.18 0.859
I8: Chemical 20.36 2.84 0.005***
I9: Glass and ceramic 1.91 0.17 0.854
I10: Paper manufacturing 2.22 0.19 0.851
I11: Steel 1.13 0.16 0.875
I12: Rubber 0.74 0.08 0.935
I13: Auto 3.29 0.26 0.798
I14: Electronics 3.01 0.50 0.617
I15: Construction 1.55 0.23 0.822
I16: Transportation 0.37 0.05 0.964
I17: Travel 0.36 0.03 0.974
I18: Insurance 1.99 0.23 0.815
I19: Grocery 0.70 0.08 0.939
R-square 0.016
I2–I19:dummy variables for industry
*
Significant level ¼ 0.10, ** significant level ¼ 0.05, *** significant level ¼ 0.01
9.5 Conclusions
The results of our research indicate that company size has no effect on any of the
following: management forecast, voluntary management forecast, mandatory
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 267
Table 9.6 Regression model for the absolute errors of management earnings forecasts dating
from 1978 to 1992
Independent variable Correlation coefficient t-statistic p-value of t-statistic
Intercept 15.55 0.87 0.386
Size 0.68 0.87 0.383
I1: Market volatility 2.57 1.78 0.075*
TR: Rate of trading volume turn over) 0.25 0.47 0.636
CV: Corporate earnings variances 0.39 0.59 0.555
E: Forecasting experience 0.57 0.76 0.450
I2: Cement 0.59 0.11 0.914
I3: Food 0.54 0.11 0.913
I4: Plastic 0.16 0.03 0.974
I5: Textile 1.17 0.26 0.799
I6: Electrical machinery 0.68 0.12 0.905
I7: Electronic equipment and cable 0.40 0.08 0.936
I8: Chemical 2.05 0.41 0.680
I9: Glass and ceramic 0.40 0.06 0.955
I10: Paper manufacturing 2.05 0.37 0.709
I11: Steel 0.56 0.10 0.923
I12: Rubber 0.14 0.02 0.981
I13: Auto 1.94 0.24 0.814
I14: Electronics 7.98 1.64 0.102
I15: Construction 0.04 0.01 0.994
I16: Transportation 0.36 0.06 0.949
I17: Travel 0.96 0.16 0.876
I18: Insurance 1.63 0.29 0.770
I19: Grocery 3.78 0.66 0.509
R-square 0.035
I2–I19: dummy variables for industry
*
Significant level ¼ 0.10, ** significant level ¼ 0.05, *** significant level ¼ 0.01
management forecast, and analysts’ forecast. This result agrees with Jaggi
(1980) and Kross and Schreoder (1990) that analyst’s earnings forecast accuracy
is not related with the size of company, but differs from the results suggested by
Bhushan (1989), Das et al. (1998), Clement (1999), Xu (1990), and Jiang (1993)
that company size does influence the relative precision of management or ana-
lysts’ earnings forecasts.
It can be seen that the relative accuracy of management’s earnings forecast and
analyst’s earnings forecast is not affected by market situation across the entire
range of sampled forecasts. There are some indications that forecasting accuracy
is affected by market ups and downs. For instance, the relative accuracy of
voluntary management’s earnings forecast during the entire time period, accuracy
of management’s forecast, and analysts’ earnings forecasts during the years 1978
through 1992 are more accurate when market is up and are less accurate during the
268 K. Hung and K.-H. Lee
down market. This result agrees with what Su has suggested – earnings forecast
accuracy is affected by market volatility, but in different ways. We believe that
due to the fact that more individual investors who are most likely to chase the
market when it is up saturate the Taiwan stock market. They examine corporate
earnings with more caution, so their expectations for companies in general are
more realistic and rational. Therefore, overall earnings forecast accuracy is
increased, and vice versa.
The results of this study reveal that relative accuracy of all four kinds of earnings
forecasts is not the functions of trading volume turnover. This agrees with the
results obtained by Chai, but it disagrees with the results of Jiang (1993). Results of
regression analysis indicate that management’s earnings forecast and analysts’
forecasts are sensitive to the corporate earnings variances. This conclusion proves
the hypothesis supports H13 through H16 and supports the theories of Kross and
Schreoder (1990) and Jiang (1993). We postulate that corporate earnings variance
of earnings surprises is an important indicator for a company’s profitability and its
earnings in the future. The management and analysts use past year’s earnings
surprises to forecast that future earnings, with an assumption of higher forecast
inaccuracy, are a result of a high degree of earnings deviation. Therefore, they will
need to exercise their highest ability in making earnings forecast more accurate. But
we found that the higher the corporate earnings variance is, the lower the forecast
accuracy will be for both the management and for analysts. Corporate earnings
variances should not be used as an important indicator as how a company is
operating, but it represents a complicated business environment it operates
in. The higher the complicity of the business environment is, the less accurate the
prediction/forecast will be.
Analyst earnings forecast and management’s earnings forecast are biased for the
chemical industry over the entire time period of sampled forecast; voluntary
management’s earnings forecasts for textile, electrical machinery, and paper
manufacturing industries are inaccurate during 1993–1999. Mandatory manage-
ment’s earnings forecasts are very inaccurate for the food industry, textile industry,
and travel industry in the time period of 1993–1999. This supports Kross and
Schreoder (1990) who concluded that analyst’s earnings forecast is affected by
the type of industry he/she follows.
The results reveal that the relative accuracy of management’s earnings forecast
and analyst’s earnings forecast do not respond to the differences of forecasters’
previous experiences. But, the relative accuracy of mandatory management’s
earnings forecast for forecasters affect the entire time period and the subsampled
voluntary management’s forecast previous earnings forecast experiences. This
conclusion agrees with Clement’s (1999) finding.
We rationalize that forecast accuracy is positive related to the forecasting
experiences as hypotheses H21 through H24 state. The results from this study
indicate otherwise. The forecast accuracy of mandatory and voluntary manage-
ment’s earnings forecast has a negative relationship with previous forecasting
experiences. We argue that it is because of (1) mis-quantifying variable as
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 269
a proxy of forecasting experiences and (2) only using the past mandatory man-
agement’s earnings forecasts as the base of future focusing without paying
attention to ways on how to reduce the forecasting errors in those forecasts.
Therefore, the more forecasting experiences the forecaster has, the less accurate
the forecast will be.
Methodology Appendix
Sample Selection
Table 9.7 Sample of earnings forecasts by management and analysts from Taiwan database
selected for the study
Year Management voluntary Management mandatory Management Analysts
1999 407 236 319 479
1998 408 238 335 430
1997 384 227 288 376
1996 328 201 223 322
1995 281 193 218 279
1994 219 135 112 247
1993 172 137 84 225
1992 139 78 84 200
1991 156 154 16 175
1990 125 125 NAa 154
1989 123 123 NA 130
1988 112 112 NA 106
1987 87 87 NA 87
Total 2,941 2,046 1,679 3,210
a
Mandatory forecast requirement was introduced in 1991
Variable Definition
1 Xn
MFm, i, t ¼ FEm, i, j, t
N j¼1
1 Xn
MF1m, i, t ¼ FE1m, i, j, t
N j¼1
1 Xn
MF2m, i, t ¼ FE2m, i, j, t
N j¼1
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 271
AFEm, i, t ¼ MFm, i, t AEi, t =AEi, t
AFE1m, i, t ¼ MF1m, i, t AEi, t =AEi, t
AFE2m, i, t ¼ MF2m, i, t AEi, t =AEi, t
where,
MFm,i,t: mean management’s pretax earnings forecast for company i at time t
MF1m,i,t: mean management’s voluntary pretax earnings forecast for company
i at time t
MF2m,i,t: mean management’s mandatory pretax earnings forecast for company
i at time t
FEm,i,j,t: management’s jth pretax earnings forecast for company i at time t
FE1m,i,j,t: voluntary management’s jth pretax earnings forecast for company i at
time t
FE2m,i,j,t: mandatory management’s jth pretax earnings forecast for company i in
year t
AFEm,i,t: absolute error of management’s pretax earnings forecast for company
i at time t
AFE1m,i,t: absolute error of voluntary management’s pretax earnings forecast for
company i at time t
AFE2m,i,t: absolute error of mandatory management’s pretax earnings forecast
for company i at time t
AEi,t: actual pretax EPS for company i at time t
Analysts forecast errors are calculated as follows:
1 Xn
MFf, i, t ¼ FEf, i, j, t
n j¼1
AFEf, i, t ¼ MFf, i, t AEi, t =AEi, t
where
MFf,i,t: mean analysts’ pretax earnings forecast for company at time t
FEf,i,j,t: analysts’ jth pre-tax earnings forecast for company i at time t;
AFEf,i,t: absolute error of analyst’s pre-tax earnings forecast for company i at
time t;
AEi,t: actual pre-tax EPS for company i at time t.
272 K. Hung and K.-H. Lee
Company Size
Unlike other previous researchers who used market value of company’s equity
as a indication of size of a company, in this study, a company’s last year’s
total revenue is used as the size of the company. The reason is because
Taiwanese market is not efficient and investors are not informed fully with
information they need during their investment decision-making process; specu-
lations among individual investors are the main cause of the stock market
volatility; thus, market value of company’s equity cannot fully represent
a company’s real size.In order to better control the company size for our
regression analysis, a logarithm of company’s last year’s total revenue is used
as the following:
SIZEi, t ¼ ln TAi, t1
where
SIZEi,t: the size of company i at time t;
TAi,t1: total revenue of company i at time t1.
Market Volatility
This study adapts what Pettengill, Glenn N, Sundaram, Sridhar, and Mathur
and Ike used in their research to measure market volatility. Market volatility
is measured as upmarket or down market by using market-adjusted return.
This return is calculated as RmRf, in which Rm is the last month’s
market return minus the last first month’s market return divided by the first
month’s market return in a given year. Rf is the risk-free interest rate in the
same year:
ReturnðMarketadjustedÞ ¼ Rm Rf
where
Upmarket if Return (Market-adjusted) >0
Down market if Return (Market-adjusted) <0
A dummy variable is used to identify market volatility. Market volatility is set to
1 if a year’s Return (Market – adjusted) is greater than 0 and set to 0 otherwise. Table 9.3
reports the market volatility of Taiwan market.
1X n
AVi, t ¼ V i, j, t
N j¼1
AVi, t1
TRi, t ¼ ln 1, 000
CSi, t1
where
Vi,j,t: daily trading volume in day j at time t for company i
AVi,t: mean daily trading volume at time t for company i
CSi,t1: number of shares outstanding at time t1 for company i
TRi,t: rate of trading volume turnover at time t for company i
!
sðXÞ
CVi, t ¼ LN
X
vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
uX n 2
u
u Xt X
t
t¼1
sðXÞ ¼
n1
1X 3
X¼ Yt
3 t¼1
where
CVi,t: corporate earnings variance at time t for company i
s(x): actual corporate earnings variance for company i
Xt: actual earnings at time t for company i
X: mean EPS (before tax) for company i
Yt: actual EPS at time t for company i
Type of Industry
There are two major ways to classify industries:
(i) “Industry classification of Republic of China” by State Council in 1987
(ii) Industry classification used by stock exchange house
274 K. Hung and K.-H. Lee
In this research, we use the latter one to classify industries, and a variable Ij is set
to represent nineteen different industries: cement, food, plastic, textile, electrical
machinery, electronic equipment and cable, chemical, glass and ceramic, paper
manufacturing, steel, rubber, auto, electronics, construction, transportation, travel,
insurance, and others.
X
n
Em , i , t ¼ Em, i, j, t1
j¼1
X
n
E1m, i, t ¼ E1m, i, j, t1
j¼1
X
n
E2m, i, t ¼ E2m, i, j, t1
j¼1
X
n
Ef , i , t ¼ Ef , i, j, t1
j¼1
where
Em,i,t: total number of times of management’s earnings forecasting experience at
time t for company i
E1m,i,t: total number of times of voluntary management’s earnings forecasting
experience at time t for company i
E2m,i,t: total number of times of mandatory management’s earnings forecasting
experience at time t for company i
Ef,i,t: total number of times of analysts’ earnings forecasting experience at time
t for company i
Regression Model
A multiple regression model is used to examine the effect of six factors: firm size,
market volatility, trading volume turnover, corporate earnings variances, type of
industry, and experience.
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 275
where
AFEm,i: absolute error percentage of management’s forecast for company i
AFE1m,i: absolute error percentage of voluntary management’s forecast for
company i
AFE2m,i: absolute error percentage of mandatory management’s forecast for
company i
AFEf,i: absolute error percentage of analysts’ forecast for company i
SIZEi: size of company i
Ii,1: market volatility (1 if market is upmarket, 0 if market is down market)
276 K. Hung and K.-H. Lee
If the two-sample groups are related, Wilcoxon two-sample test will be used to
determine the relative earnings forecast accuracy:
W Eð W Þ
Z ¼ pffiffiffiffiffiffiffiffiffiffiffiffi
V ðW Þ
where
W1: rank sum of absolute error percentage for management’s earnings forecasts
W2: rank sum of absolute error percentage for analysts’ earnings forecasts
W: smaller value between W1 and W2
E(W): expected values of W-distribution
pffiffiffiffiffiffiffiffiffiffiffiffi
VðWÞ: deviation of W-distribution
References
Bartley, J. W., & Cameron, A. B. (1991). Long-run earnings forecasts by managers and financial
analysts. Journal of Business Finance & Accounting, 18, 21–41.
Bhushan, R. (1989). Firm characteristics and analyst following. Journal of Accounting & Eco-
nomics, 11, 255–274.
Chia, Yuxuan (1994). Examination of mandatory management earnings forecasts and the rate
of trading turnover. Unpublished thesis, Accounting Department, National Chengchi
University.
Clement, B. M. (1999). Analyst forecast accuracy: Do ability, resources, and portfolio complexity
matter? Journal of Accounting & Economics, 27, 285–303.
Das, S., Carolyn, B., Levine, K. S. (1998). Earnings predictability and bias in analysts’ earnings
forecasts. Accounting Review, 73, 277–294.
Imhoff, E. A., & Pare, P. (1982). Analysis and comparison of earnings forecasts agents. Journal of
Accounting Research, 20, 429–439.
Jaggi, B. (1980). Further evidence on the accuracy of management forecasts vis-à-vis analyst’
forecasts. The Accounting Review, 55, 96–101.
Jiang, Jianquan (1993). Factors associated with analyst’s earnings forecast accuracy
under different assumptions in Taiwan. Unpublished thesis, School of Business, Taiwan
University.
Kross, W., Ro, B., & Schreoder, D. (1990). Earnings expectations: The analysts information
advantage. The Accounting Review, 65, 461–476.
9 Determinations of Corporate Earnings Forecast Accuracy: Taiwan Market Experience 277
Pettengill, G. N., Sundaram, S., & Mathur, I. (1995). The conditional relation between beta and
returns, Journal of Financial and Quantitative Analysis, 30(1), 101–116
Su, Yongru (1996). Comparison of relative accuracy of management and analysts earnings
forecast. Unpublished thesis, Accounting Department, Soochow University.
Xu, Xiubin (1990). A examination of analysts earnings forecasts accuracy. Unpublished thesis,
Accounting Department, National Chengchi University.
Xu, Jinjuan (1992). The effect of management earnings forecast on portfolio management.
Unpublished thesis, Accounting Department, National Chengchi University, Taiwan.
Yie, Meiling (1996). Examination on factors in management and financial analyst earnings
forecasting accuracy. Unpublished thesis, Accounting Department, National Chung Cheng
University.
Market-Based Accounting Research
(MBAR) Models: A Test of ARIMAX 10
Modeling
Anastasia Maggina
Contents
10.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 280
10.2 Review of the Literature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281
10.3 Methodology and Models Used . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 283
10.4 Sample Selection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 285
10.5 Empirical Findings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 286
10.5.1 Unit Root Tests: Testing for Stationarity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 292
10.5.2 Forecasting Dependent Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 294
10.6 Conclusions and Suggestions for Further Future Research . . . . . . . . . . . . . . . . . . . . . . . . . . . . 296
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 296
Abstract
The purpose of this study is to provide evidence drawn from publicly traded
companies in Greece as far as some of the standard models of accounting earnings
and returns relations mainly collected through the literature. Standard models such
as earnings level and earnings changes have been investigated in this study. Models
that fit better to the data drawn from companies listed on the Athens Stock
Exchange have been selected employing autoregressive integrated moving average
with exogenous variables (ARIMAX) models. Models I (price on earnings model),
II (returns on change in earnings divided by beginning-of-period price and prior
period), V (returns on change in earnings over opening market value),VII (returns
deflated by lag of 2 years on earnings over opening market value), and IX
(differenced-price model) have statistically significant coefficients of explanatory
variables. In addition, model II (returns on change in earnings divided by
beginning-of-period price and prior period with MSE (minimum squared error)
A. Maggina
Business Consultant/Research Scientist, Avlona, Attikis, Greece
e-mail: anastasiamaggina@yahoo.gr; a.maggina@yahoo.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 279
DOI 10.1007/978-1-4614-7750-1_10,
# Springer Science+Business Media New York 2015
280 A. Maggina
loss function in ARIMAX (2,0,2)) is prevalent. These models take place with
backward-looking information instead of forward-looking information that recent
literature is assessed. Application of generalized autoregressive conditional
heteroscedasticity (GARCH) models is suggested for further future research.
Keywords
Market-based accounting research • Price on earnings model • Earnings level •
Earnings change • Return models • Autoregressive-moving average with exog-
enous variables • Minimum value of squared residuals (MSE loss function) •
Unit root tests • Stationarity • Dickey-Fuller test
10.1 Introduction
The relationship between accounting earnings and stock prices has been discussed both
in the accounting and financial literature. In a consideration of market-based account-
ing research, the association between returns and earnings has been very low due to
(i) poor specification of the estimating equation, (ii) poor informational properties of
reported earnings, (iii) inappropriate choice of the assumed proxy for expected earn-
ings, and (iv) the availability of more timely sources of the value-relevant information
in earnings statement (Strong and Walker 1993). This has been resolved by allowing
for time-series and cross-sectional variation in the regression parameters, by including
an earnings yield and partitioning all-inclusive earnings into pre-exceptional, excep-
tional, and extraordinary components. Yet, empirical evidence on returns-earnings
association in certain applications is very strong (Easton and Harris 1991; Kothari
and Zimmerman 1995). Standard empirical models, that is, price-earnings association,
and returns-earnings association (earnings level and earnings changes) have been
investigated in the USA and UK with no empirical evidence from at least the rest of
Europe. The purpose of this paper is to develop an empirical background by investi-
gating whether some of the standard models that have been collected through the
literature would be relevant for evaluating accounting earnings/returns associations in
a stock market being in transition from an emerging to a matured one. In other words,
we purport to select those models that better fit to available data. The rationale of the
Greek stock market response is one major objective, while the other is to make some
comparisons with other findings presented in the literature.
The Athens stock exchange has been established on 1876. It was an emerging
market some years ago and has run up to a mature one. It is a normal market with no
surprises, and investments are not only a place for corporations to get financing;
more importantly it is a place to create wealth. Registered companies are young,
growth-oriented to long-established enterprises. The Greek stock market has played
a great role in the economic development of the country in the last half of the
twentieth century. Facts that have influenced the ASE is the inclusion of the country
in the Economic and Monetary Union and the crash of 1999 that affected the life of
many Greek families. The ASE operates in a country which has been in deep
recession since 2008.
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 281
Common interest of both accounting and finance scholars and users of financial
statements (primarily, financial analysts and investors) is to obtain concrete and
increased knowledge of the association between accounting earnings and stock
prices for a more or less predictive ability explanation. An evaluation of the
informational content of, at least, the most basic accounting numbers contributes
to the improvement of models. Among the various treatments of the relation
between accounting earnings and stock prices, the seminal work of Ball and
Brown (1968) is precedent. Their work indicates that accounting earnings and
some of its components exhibit an information content drawn through stock prices.
Later on, Brown et al. (1985) worked on the relative ability of the current innova-
tion in annual earnings and the revision in next year’s earnings forecasts to explain
changes in stock prices. However, it has empirically been tested (Fama et al. 1969;
Fama 1970). Successive changes in individual common stocks are very nearly
independent because of an efficient market in which adjustments to new informa-
tion are rapidly made (Mandelbrot 1966; Samuelson 1965). Beaver (1970) and
Rosenberg and Marathe (1975) worked to discover financial statement components
that are related to risk and, thus, predict expected stock returns. They maintain
that financial statements constitute an assured source of information (and can be
prepared in a short time period) and, consequently, the direction of 1-year-ahead
earnings changes affects stock prices. Their results indicate that the summary
measure robustly predicts future stock returns.
Beaver et al. (1980) developed a model known as BLM theory. They followed
the traditional price-earnings relation and tested for the information content of
prices with respect to future earnings. BLM regress percentage change in price
which may contain information about future earnings not reflected in the current
earnings. Lev and Ohlson (1982) describe price and return models as complemen-
tary. Given that stock prices are associated with accounting earnings, analysts
forecast the reported accounting numbers, and as Brown et al. (1985) maintain,
analysts use their 1-year-ahead forecasts to convey their expectations about perma-
nent earnings. In all the above treatments of the subject matter referring to account-
ing earnings and stock prices, the more and more prevailing role of managerialism
could be more or less emphatic. Trueman (1986) explained why managers would be
willing to disclose forecasts of higher or lower than expected earnings. Undoubt-
edly, the disclosure of favorable accounting numbers or the disclosure of forecasts
of favorable expected earnings contributes to an increase of the market value of the
firm. Hirst et al. (2008) assert that managers often issue earnings forecasts to correct
information asymmetry problems and, thus, influence their firm’s stock price.
282 A. Maggina
Lipe and Kormendi (1987) found that stock returns were considered as a function of
the revisions in expectations of earnings. They show that the stock return reaction to
earnings is a function of (1) the time-series properties of earnings, (2) the interest
rate used to discount expected future earnings, and (3) the relative ability of
earnings. Furthermore, Freeman (1987) maintains that the relation between security
prices and firm-specific information is associated with the market value of a firm’s
common equity and concludes that the magnitude of the abnormal returns related to
earnings is a decreasing function of firm size. Of much interest is a hypothesis
tested by Freeman (1987) which implies that (i) the abnormal security returns
related to accounting earnings occur earlier for large than for small firms and
(ii) abnormal returns are lower for large firms and higher for small firms. The
information content theorists consider size as an important conditioning variable
when testing the information content of prices with respect to future earnings and
contemporaneous price changes.
Beaver et al. (1987), Collins et al. (1987), Collins and Kothari (1989), and many
others show that unexpected earnings for a year are correlated with returns
from a prior year. Christie (1987) concludes that while return and price models
are economically equivalent, return models are econometrically less
problematic. Meanwhile, Holthausen and Verrecchia (1988) worked on price
changes when information is announced. Landsman and Magliolo (1988) argue
that price models are superior to return models for certain applications. According
to Cornell and Landsman (1989), stock prices respond to earnings announcements.
Alternatively stated, unexpected increases in earnings are associated with a rise in
stock prices, and unexpected decreases in earnings are associated with a fall in stock
prices. Relating unexpected accounting earnings and security prices aims to assess
the information content of the latter (Collins and Kothari 1989). Conservative
proponents support the view that financial statement ratios are the basic tools for
evaluating and predicting accounting earnings and, consequently, security prices.
Ou and Penman (1989) maintain that the relationship of financial statement char-
acteristics to value is not apparent. Easton and Zmijewski (1989) and that of Board
and Walker (1990) analyze a coefficient that measures the response of stock prices
to accounting earnings coefficient. They measure the response of stock prices to
accounting earnings announcements, and they empirically show that the higher this
coefficient, the smaller the stock price changes.
Lipe (1990) worked on the relation between stock returns and accounting earnings,
assuming that market observes current-period information other than earnings. In the
process of relating stock prices to accounting earnings, (i) a coefficient which mea-
sures the stock-return response to a one dollar earnings changes as a function of both
“predictability” and “persistence” of earnings, and (ii) the variance of stock price
changes during the reporting of earnings has been tested in the literature (Cho and
Jung 1991; etc.). Easton and Harris (1991) presented models relating earnings vari-
ables and security returns, concluding that both current earnings level and the earnings
change variables play a role in the security valuation. They are in fact correlated.
Strong and Walker (1993) used a panel regression approach to examine the association
between annual stock price returns and reported earnings figures of industrial
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 283
Literature concerning accounting earnings and stock prices has been formulated
in the framework of an earnings-based valuation model expanded with the
dividend irrelevance proposition. These standard models which have also been
tested by Easton and Harris (1991) and Zimmerman and Kothari (1995) are as
follows:
I. Price on earnings
Ai, j =Pi, j1 ¼ ai, j þ bi, j Ai, j Ai, j1 =Pi, j1 þ Ai, j1 =Pi, j1 (10.2)
284 A. Maggina
VI. Returns on change in earnings over opening market value and on earnings over
opening market value
Pi, j Pi, j1 þ di, j =Pi, j1 ¼ a þ b1i, j Ai, j Ai, j1 =Pi, j1
þ b2i, j Ai, j =Pi, j1 þ ei, j (10.6)
VII. Returns (deflated by lag of 2 years) on earnings over opening market value
Pi, j Pi, j1 þ di, j =Pi, j2 ¼ ai, j þ bi, j Ai, j =Pi, j1 þ ei, j (10.7)
where
Pi,j ¼ stock price (per share) of firm i in period j
Ai,j ¼ earnings per share of firm i in period j
di,j ¼ dividend per share of firm i in period j
a ¼ a constant in a linear relationship(intercept parameter)
b1,b2 ¼ a slope parameter or a coefficient in a linear regression
i ¼ cross-selection item, j ¼ time-series item
To be familiar and consistent with the existing literature, some requirements are
stressed. For example, earnings per share divided by price at the beginning of the
return period (Ai,j/Pi,j 1) refers to current earnings level variable. Change in
earnings divided by beginning-of-period price refers to earnings change variable
[(Ai,j Ai,j 1)/Pi,j 1]. Thus far, the models that have been selected to be tested
express the following:
Model I: Expresses price as a multiple of earnings.
Model II: Expresses historical price-earnings ratios with an earnings change versus
earnings levels explanation form.
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 285
But in general
Exchange Annual Yearbook, the annual statistical bulletin, and the Internet.
Total number of companies refers to the time period 1974 up to 2005 (the most
recently available data when writing the paper). The full sample (1974–2005) is
separated in two samples (1974–1999 and 2000–2005) which correspond to
two-time periods, that is, before the Euro currency and the Euro era. As in
Kothari and Zimmerman (1995), to avoid any undue influence of extreme
observations, the largest and the lowest 1 % of observations is excluded from
the sample. EPS take positive or zero values. All firms have a December fiscal
year-end. Annual earnings include those from discontinued operations and
extraordinary items.
For each year, a nonconstant number of companies is available. We define the total
number of companies at each year t as nt. We would like to estimate the following
nine models for the period from 1974 to 2005:
Model I
Pt ¼ a þ bAt þ et , for et N 0; s2 , (10.15)
X
nt X
nt
where Pt ¼ nt 1 Pi, t and At ¼ nt 1 Ai, t:
i¼1 i¼1
Model II
At At At1 At1
¼ a þ b1 þ b2 þ et , for et N 0; s2 , (10.16)
Pt1 Pt1 Pt1
At Xnt
Ai, t At At1 Xnt
Ai, t Ai, t1 At1
where ¼ nt 1 , ¼ nt 1 , and ¼ nt 1
Pt1 i¼1
P i , t1 P t1 i¼1
P i , t1 P t1
Xnt
Ai, t1
.
i¼1
Pi, t1
Model III
Pt Pt1 þ dt At
¼ a þ b1 þ et , for et N 0; s2 , (10.17)
Pt1 Pt1
Pt Pt1 þ dt Xnt
Pi, t Pi, t1 þ di, t At Xnt
Ai, t
where ¼ nt 1 and ¼ nt 1 .
Pt1 i¼1
P i, t1 P t1 i¼1
P i, t1
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 287
Model IV
Pt Pt1 þ dt At1
¼ a þ b1 þ et , for et N 0; s2 , (10.18)
Pt1 Pt1
Pt Pt1 þ dt Xnt
Pi, t Pi, t1 þ di, t Xnt
Ai, t1
where ¼ nt 1 and APt1 ¼ n t
1
.
Pt1 i¼1
P i, t1
t1
i¼1
Pi, t1
Model V
Pt Pt1 þ dt At At1
¼ a þ b1 þ et , for et N 0; s2 , (10.19)
Pt1 Pt1
Pt Pt1 þ dt Xnt
Pi, t Pi, t1 þ di, t
where ¼ nt 1 and
Pt1 i¼1
Pi, t1
At At1 X Ai, t Ai, t1
nt
¼ nt 1 .
Pt1 i¼1
Pi, t1
Model VI
Pt Pt1 þ dt At At1 At
¼ a þ b1 þ b2 þ et , for et N 0; s2 , (10.20)
Pt1 Pt1 Pt1
Pt Pt1 þ dt X nt
Pi, t Pi, t1 þ di, t
where ¼ nt 1 ,
Pt1 i¼1
Pi, t1
At At1 Xnt
Ai, t Ai, t1 At Xnt
Ai , t
¼ nt 1 , and ¼ nt 1 .
Pt1 i¼1
P i, t1 P t1 i¼1
P i, t1
Model VII
Pt Pt1 þ dt At
¼ a þ b1 þ et , for et N 0; s2 , (10.21)
Pt2 Pt1
Pt Pt1 þ dt Xnt
Pi, t Pi, t1 þ di, t At Xnt
Ai, t
where ¼ nt 1 and ¼ nt 1 .
Pt2 i¼1
Pi, t2 Pt1 i¼1
Pi, t1
Model VIII
Pt At
¼ a þ b1 þ et , for et N 0; s2 , (10.22)
Pt1 Pt1
Pt X Pi , t
nt
At Xnt
Ai , t
where ¼ nt 1 and ¼ nt 1 .
Pt1 i¼1
Pi, t1 Pt1 i¼1
Pi, t1
288 A. Maggina
Table 10.1 The ARIMAX (p,d,q) specification for each of the nine models and the relative MSE
loss functions
Model I ARIMAX(1,0,1) 2.766371
Model II ARIMAX(2,0,2) 0.003763
Model III ARIMAX(2,0,2) 0.741482
Model IV ARIMAX(2,0,2) 0.719591
Model V ARIMAX(0,1,1) 0.732002
Model VI ARIMAX(2,0,2) 0.584564
Model VII ARIMAX(2,1,1) 0.888239
Model VIII ARIMAX(2,0,2) 0.762231
Model IX ARIMAX(2,1,2) 2.245854
Model IX
ðPt Pt1 Þ ¼ a þ b1 ðAt At1 Þ þ et , for et N 0; s2 , (10.23)
X
nt
where ðPt Pt1 Þ ¼ nt 1 Pi, t Pi, t1 and
i¼1
X
nt
ðAt At1 Þ ¼ nt 1 Ai, t Ai, t1 .
i¼1
The nine aforementioned models are characterized by autocorrelated and
heteroscedastic residuals. Thus, the models are expanded in the ARIMAX
(autoregressive integrated moving average) framework (for details about ARIMAX
modeling, the interested reader is referred to Box and Jenkins (1976)) in order to
model the autocorrelated residuals. Moreover, we take into consideration White’s
(1980) heteroscedasticity-consistent covariance matrix estimator which provides
correct estimates of the coefficient covariances in the presence of heteroscedasticity
of unknown form.
The ARIMAX (p,d,q) models are estimated in the following form:
ð1 L!Þd yt ¼ Xt b þ et !
X
p X q
1 ci L et ¼ 1 þ
i
d i L et
i (10.24)
i¼1 i¼1
et N ð0; s2 Þ,
Model II
At At At1 At1
¼ 0:004 þ 0:99 þ 1:04 þ et
Pt1 Pt1 Pt1 (10.26)
et ¼ 1:5et1 0:84et2 þ et 1:14et1 þ 0:52et2
et N ð0; s2 Þ:
Model III
Pt Pt1 þ d t At
¼ 0:28 0:21 þ et
Pt1 Pt1 (10.27)
et ¼ 0:32et1 þ 0:46et2 þ et þ 0:92et1 1:16et2
et N ð0; s2 Þ:
Model IV
Pt Pt1 þ d t At1
¼ 0:47 þ 2:48 þ et
Pt1 Pt1 (10.28)
et ¼ 0:21et1 þ 0:36et2 þ et 0:68et1 1:42et2
et N ð0; s2 Þ:
Model V
Pt Pt1 þ d t At At1
ð 1 LÞ ¼ 0:055 0:5 þ et
Pt1 Pt1 (10.29)
et ¼ et 1:52et1
et N ð0; s2 Þ:
Model VI
Pt Pt1 þ d t At At1 At
¼ 0:33 2:6 þ 2:4 þ et
Pt1 Pt1 Pt1 (10.30)
et ¼ 0:4et1 þ 0:3et2 þ et 0:37et1 1:9et2
et N ð0; s2 Þ:
Model VII
Pt Pt1 þ d t At
ð1 LÞ ¼ 0:2 0:76 þ et
Pt2 Pt1 (10.31)
et ¼ 0:1et1 þ 0:3et2 þ et 1:7et1
et N ð0; s2 Þ:
290 A. Maggina
Model VIII
Pt At
¼ 1:3 0:22 þ et
Pt1 Pt1 (10.32)
et ¼ 0:31et1 þ 0:46et2 þ et þ 0:94et1 1:1et2
et N ð0; s2 Þ:
Model IX
In the sequel, for the models being estimated, the estimates of the coefficients,
their standard errors, the t-ratios, and their p-values are presented. According to the
t-statistics, computed as the ratio of the coefficients to their standard errors, the
coefficients of the explanatory variables are statistically significant in models
I (price on earnings), II (returns on change in earnings divided by beginning-of-
period price and prior period earnings divided by the price at the beginning of the
return period), V (returns on change in earnings over opening market value), VII
(returns (deflated by lag of 2 years) on earnings over opening market value, and IX
(differenced-price model).
Thus, we conclude that these models explain the relationship between dependent
and explanatory variables, whereas the models III (returns on earnings over opening
market value), IV (returns on prior earnings model over opening market value), VI
(returns on change in earnings over opening market value, and on earnings over
opening market value), and VIII (return model regressed on earnings over opening
market value) fail to explain any strong relationship for the variables under
investigation.
Model I
Model II
Model III
Model IV
Model V
Model VI
Model VII
Model VIII
Model IX
We estimate the augmented Dickey and Fuller (1979) test in order to investigate the
null hypothesis of a unit root or nonstationarity of the time series under investiga-
tion, i.e., yt. The Dickey-Fuller test is carried out by estimating:
X
l
Dyt ¼ a0 þ a1 yt1 þ a2 t þ bi Dyti þ et (10.34)
i¼1
where et N(0,s2), and D is the difference operator, or D ¼ (1 L). The lag order
Xl
l of bi Dyti is selected based on the Schwarz (1978) information criterion.
i¼1
The null and the alternative hypotheses may be written as
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 293
H 0 : a1 ¼ 0
(10.35)
H A : a1 6¼ 0:
The test statistics and the relative p-values from testing 21 for the Eq. 10.20 for
a2 ¼ 0 are the following:
The test statistics and the relative p-values from testing 21 for the Eq. 10.20 for
a0 6¼ 0 and a2 6¼ 0 are the following:
Variable Test statistic p-value
Pt 3.565939 0.0497
At 5.698735 0.0003
At
Pt1
5.667882 0.0004
At At1
Pt1
6.213167 0.0001
At1
Pt1
5.673596 0.0003
Pt Pt1 þ d t
Pt1
4.332965 0.0092
Pt Pt1 þ d t
Pt2
5.931949 0.0002
Pt
Pt1
4.290153 0.0102
(Pt Pt1) 6.875067 0.00
(At At1) 7.024534 0.00
In all the cases, the p-values are less than a 5 % level of significance. Therefore,
the null hypothesis is rejected at any case. Hence, the series are defined, by the
augmented Dickey-Fuller tests, to be stationary.
For the following models, the dependent variable is predicted for the year 2006,
after adding in the dataset the values of the explanatory variables for the year 2006:
Model I
Pt ¼ 8:5 5:1At þ et
et ¼ 1:05et1 þ et 1:46et1 (10.36)
et N ð0; s2 Þ:
Model II
At At At1 At1
¼ 0:004 þ 0:99 þ 1:04 þ et
Pt1 Pt1 Pt1 (10.37)
et ¼ 1:5et1 0:84et2 þ et 1:14et1 þ 0:52et2
et N ð0; s2 Þ:
Model III
Pt Pt1 þ d t At
¼ 0:28 0:21 þ et
Pt1 Pt1 (10.38)
et ¼ 0:32et1 þ 0:46et2 þ et þ 0:92et1 1:16et2
et N ð0; s2 Þ:
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 295
Table 10.2 The 1-year-ahead forecasts (year 2006) of the dependent variables for the nine
models
Number of model Best model Forecast of the dependent variable
Model I ARIMAX(1,0,1) 1.066527
Model II ARIMAX(2,0,2) 0.010287
Model III ARIMAX(2,0,2) 0.094187
Model IV ARIMAX(2,0,2) 0.236392
Model V ARIMAX(0,1,1) 0.715401
Model VI ARIMAX(2,0,2) 0.059651
Model VII ARIMAX(2,1,1) 0.372343
Model VIII ARIMAX(2,0,2) 0.833744
Model IX ARIMAX(2,1,2) 4.755135
Model IV
Pt Pt1 þ d t At1
¼ 0:47 þ 2:48 þ et
Pt1 Pt1 (10.39)
et ¼ 0:21et1 þ 0:36et2 þ et 0:68et1 1:42et2
et N ð0; s2 Þ:
Model V
Pt Pt1 þ d t At At1
ð 1 LÞ ¼ 0:055 0:5 þ et
Pt1 Pt1 (10.40)
et ¼ et 1:52et1
et N ð0; s2 Þ:
Model VI
Pt Pt1 þ d t At At1 At
¼ 0:33 2:6 þ 2:4 þ et
Pt1 Pt1 Pt1 (10.41)
et ¼ 0:4et1 þ 0:3et2 þ et 0:37et1 1:9et2
et N ð0; s2 Þ:
Model VII
Pt Pt1 þ d t At
ð1 LÞ ¼ 0:2 0:76 þ et
Pt2 Pt1 (10.42)
et ¼ 0:1et1 þ 0:3et2 þ et 1:7et1
et N ð0; s2 Þ:
Model VIII
Pt At
¼ 1:3 0:22 þ et
Pt1 Pt1 (10.43)
et ¼ 0:31et1 þ 0:46et2 þ et þ 0:94et1 1:1et2
et N ð0; s2 Þ:
296 A. Maggina
Model IX
References
Ball, R., & Brown, P. (1968). An empirical evaluation of accounting income numbers. Journal of
Accounting Research, 6, 159–178.
Beaver, W. (1970). The time series behavior of earnings variables, empirical research in
accounting. Selected studies. Supplement to Journal of Accounting Research, 62–99.
10 Market-Based Accounting Research (MBAR) Models: A Test of ARIMAX Modeling 297
Beaver, W., Lambert, R., & Morse, D. (1980). The information content of security prices. Journal
of Accounting and Economics, 2, 3–28.
Board, J.L.G., & Walker, M. (1990). Intertemporal and cross sectional variation in the association
between unexpected accounting rates of return and abnormal returns. Journal of Accounting
Research, 182–192.
Box, G. E. P., & Jenkins, G. M. (1976). Time series analysis: Forecasting and control, revised
edition. Oakland: Holden-Day.
Brown, P., Foster, G., & Noreen, E. (1985). Security analyst multiyear earnings forecasts and the
capital market. Monograph, American Accounting Association, 11, 1–45.
Chen, C., Chen, S., & Su, X. (2001). Is accounting information value-relevant in the emerging
Chinese stock market? Journal of International Accounting, Auditing and Taxation, 10, 1–22.
Chen, G., Firth, M., & Ning Gao, D. (2011). The information content of earnings components:
Evidence from the Chinese stock market. European Accounting Review, 20, 669–692.
Cho, J. Y., & Jung, K. (1991). Earnings response coefficients: A synthesis of theory and empirical
evidence. Journal of Accounting Literature, 10, 85–116.
Christie, A. (1987). On cross-sectional analysis in accounting research. Journal of Accounting and
Economics, 9, 231–258.
Collins, D., & Kothari, S. (1989). An analysis of international and cross-sectional determinants of
earnings response coefficients. Journal of Accounting and Economics, 11, 143–181.
Collins, D., Kothari, S., & Rayburn, J. (1987). Firm size and the information content of prices with
respect to earnings. Journal of Accounting and Economics, 9, 111–138.
Cornell, B., & Landsman, W. (1989). Security price response to quarterly earnings announcements
and analyst forecast revisions. The Accounting Review, 64, 680–692.
Dickey, D. A., & Fuller, W. A. (1979). Distribution of the estimators for autoregressive time series
with a unit root. Journal of the American Statistical Association, 74, 427–431.
Easton, P., & Harris, T. (1991). Earnings as an explanatory variable for returns. Journal of
Accounting Research, 29, 19–36.
Easton, P., & Zmijewski, M. (1989). Cross-sectional variation in the stock market response to
accounting earnings announcements. Journal of Accounting and Economics, 29, 19–36.
Fama, E. (1970). Efficient capital markets: A review of theory and empirical work. Journal of
Finance, 25(2), 383–417.
Fama, E., Fisher, L., Jensen M., & Roll, R. (1969). The adjustment of stock prices to new
information. International Economic Review, 10(1), 1–21.
Freeman, R. (1987). The association between accounting earnings and stock returns for large and
small firms. Journal of Accounting and Economics, 9, 195–228.
Greene, W.H. (2007). Econometric analysis. Prentice Hall, New Jersey.
Hirst, E., Koonce, L., & Venkataraman, S. (2008). Management earnings forecasts: A review and
framework. Accounting Horizons, 22, 315–338.
Holthausen, R. W., & Verrecchia, R. E. (1988). The effect of informedness and consensus on price
and volume behavior. The Accounting Review, 63, 191–208.
Kahya, E., Maggina, A., & Theodossiou, P. (1993). The information content of accounting earnings in
Greece: An empirical investigation. Journal of Multinational Financial Management, 3, 143–156.
Kormendi, R., & Lipe, R. (1987). Earnings innovations, earnings persistence and stock returns.
Journal of Business, 60, 323–345.
Kothari, S. P., & Zimmerman, J. L. (1995). Price and return models. Journal of Accounting and
Economics, 20, 155–192.
Landsman, W. R., & Magliolo, J. (1988). Cross-sectional capital market research and model
specification. The Accounting Review, 64, 586–604.
Lev, B., & Ohlson, J. A. (1982). Market based empirical research in accounting:
A review, interpretations, and extensions. Supplement to Journal of Accounting Research, 20,
249–322.
Lipe, R. (1990). The relation between stock returns and accounting earnings given alternative
information. The Accounting Review, 65, 49–71.
298 A. Maggina
Lundholm, R., & Myers, L. (2002). Bringing the future forward: The effect of disclosure on the
returns-earnings relation. Journal of Accounting Research, 40, 809–839.
Mandelbrot, B. (1966). Forecasts of future prices, unbiased markets and martingale models.
Journal of Business, 39, 242–255.
Ou, J., & Penman, S. (1989). Financial statement analysis and the prediction of stock returns.
Journal of Accounting and Economics, 11, 295–329.
Rosenberg, B., & Marathe, V. (1975). The prediction of investment risk: Systematic and residual
risk. In Proceedings of the seminar on the analysis of security prices (pp. 185–226). Chicago:
Center for Research in Security Prices, University of Chicago.
Samuelson, P. (1965). Proof that properly anticipated prices fluctuate randomly. Industrial Man-
agement Review, 31, 41–49.
Schwarz, G. (1978). Estimating the dimension of a model. Annals of Statistics, 6, 461–464.
Strong, N., & Walker, M. (1993). The explanatory power of earnings for stock returns. The
Accounting Review, 68, 385–399.
Trueman, B. (1986). Why do managers voluntarily release earnings forecasts. Journal of Account-
ing and Economics, 8, 53–71.
Wei, W.W. (2006). Time series analysis: Univariate and multivariate methods. Pearson Education.
White, H. (1980). A heteroskedasticity-consistent covariance matrix and a direct test for
heteroskedasticity. Econometrica, 48, 817–838.
An Assessment of Copula Functions
Approach in Conjunction with 11
Factor Model in Portfolio Credit Risk
Management
Contents
11.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 300
11.2 Dependence Structure in Static Factor Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 302
11.3 Dependence Structure in Dynamic Factor Model Using Intensity Function . . . . . . . . . . 303
11.4 Contagious Model: Mutually Exciting Intensity Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 304
11.5 Copula Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 305
11.6 Simulation Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 306
11.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 310
Appendix 1: Point Process and Its Intensity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 311
Appendix 2: Copula Functions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 312
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 315
Abstract
In credit risk modeling, factor models, either static or dynamic, are often used to
account for correlated defaults among a set of financial assets. Within the realm of
factor models, default dependence is due to a set of common systemic factors.
Conditional on these common factors, defaults are independent. The benefit of
a factor model is straightforward coupling with a copula function to give an
analytic formulation of the joint distribution of default times. However, factor
models fail to account for the contagion mechanism of defaults in which a firm’s
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 299
DOI 10.1007/978-1-4614-7750-1_11,
# Springer Science+Business Media New York 2015
300 L.-J. Kao et al.
Keywords
Static factor model • Dynamic factor model • Correlated defaults • Contagious
effect • Hawkes process-Monte • Carlo simulation • Normal copula • t-copula •
Clayton copula • Frank copula • Gumbel copula
11.1 Introduction
dynamic factor model in a point process. Section 11.4 introduces the contagious
effect model based on mutually exciting point processes. The theoretical founda-
tions of the copula function approach are given in Sect. 11.5. A simulation study is
given in Sect. 11.6. Section 11.7 concludes.
X
K
Xi ¼ rij Y j þ riðKþ1Þ ei
j¼1
where rij determines the relative importance of asset i to the common factor Yj and
ri(K+1) the relative weight of idiosyncratic factor ei for asset i. To the extent that the
creditworthiness indices are related to the common systemic factors Y1, . . ., YK, the
likelihood of joint default events across assets varies accordingly. Here the common
factors Y1, . . ., YK and the idiosyncratic factors e1,. . ., eI are independent. In the
following, for simplicity, we will consider a single factor Y and the ith asset’s
creditworthiness index:
qffiffiffiffiffiffiffiffiffiffiffiffiffi
Xi ¼ ri Y þ 1 r2i ei : (11.1)
If assuming the latent common factors Y and idiosyncratic factors e1, . . ., eI are
standard normally distributed, then the creditworthiness indices Xi are also standard
normally distributed. Therefore, the conditional probability
!
F1 ðFi ðti ÞÞ ri y
PðXi ui jyÞ ¼ F pffiffiffiffiffiffiffiffiffiffiffiffiffi
1 r2i
And the joint distribution F(t1, . . ., tI) of default times t1, . . ., tI is linked to
a one-factor normal copula function in the form
ð" !#
I F1 ðFi ðti ÞÞ ri y
Fðt1 ; . . . ; tI Þ ¼ PF pffiffiffiffiffiffiffiffiffiffiffiffiffi fðyÞdy (11.2)
i¼1 1 r2i
where f is the standard normal density function. Alternatively, one can assume that
the latent common factor Y is Gamma distributed with parameter 1/y and let
PðXi ui jyÞ ¼ exp y’1 ðFi ðti ÞÞ
where ’1(s) ¼ sy 1 and ’ is the generator. Then the joint distribution F(t1, . . ., tI)
is linked to a one-factor Clayton copula function in the form
ð
I 1
Fðt1 ; . . . ; tI Þ ¼ P exp y’ ðFi ðti ÞÞ fðyÞdy: (11.3)
i¼1
The Clayton copula function and its generator ’ are given in Definition 2.7 of
Appendix 2.
induce correlated defaults among assets (Duffee 1999; Driessen 2005; Elizalde
2005; Yu 2005). In contrast to the static factor model, the evolution of these
common factors over time is modeled, and conditional on the evolution of these
common factors Xt, defaults are independent. The formulation of the default
intensity as such is referred as a doubly Poisson point process (Lando 1998). In
Definition 1.3, the properties of a point process and a (doubly) Poisson process are
given. An example of a dynamic factor model is given in Duffee (1999), in which
the default intensity of asset i is
li , t ¼ a0, i þ a1, i X1, t X1 þ a2, i X2, t X2 þ li, t c
where a0,i, a1,i, and a2,i are constants and X1,t and X2,t are two latent factors
interpreted as the slope and level of the default-free yield curve, i.e., the risk-free
interest rate
r t ¼ ar þ X1, t þ X2, t :
The asset’s specific intensity li;t , independent across assets, is assumed to obey
a mean-reversion process
qffiffiffiffiffiffi
dli,t ¼ ki yi li,t dt þ sli li,t dW i,t (11.4)
where W1,t, . . ., WI,t are independent Brownian motions. To introduce more default
correlation, Duffie and Singleton (1999b) incorporate joint as well as idiosyncratic
jumps in the default intensity li,t.
In Das et al. (2007), the hypothesis whether default events can be modeled as
a doubly Poisson point process that solely depends on “exogenous” factors is tested.
Based on a time series of US corporate defaults, they strongly rejected the hypoth-
esis that defaults can be modeled as a doubly Poisson point process. Instead of
a factor model, in the following, the contagious model in which default status of
other firms will affect the default intensity of the underlying asset will be
considered.
Examples of contagious models include the infectious default model by Davis and
Lo (2001) in which the intensity function follows a piecewise deterministic Markov
process and the propensity model by Jarrow and Yu (2001). In the propensity
model, firms are divided into K primary and I-K-1 secondary firms: the default
intensities l1,t, . . ., lK,t of primary firms are determined by some exogenously
11 An Assessment of Copula Functions Approach 305
common factors, while those of the secondary firms depend on the status (default or
not) of the primary firms but not the other way around (asymmetric dependence).
Specifically, the default intensity of the ith secondary firm is given by
X
K
li , t ¼ l i , t þ aji, t 1ftj tg (11.5)
j¼1
where K + 1 i I. Here li,t represents the part of secondary firm i’s hazard rate
independent of the default status of other firms.
Recently, Hawkes process, a class of counting processes which allow intensities
to depend on the timing of previous events (Hawkes 1971), had been adopted to
model the aggregate default intensity of a portfolio (Azizpour and Giesecke 2008;
Errais et al. 2010). In Errais et al. (2010), the cumulated default intensity of n firms
is specified by
X
n
lðtÞ ¼ l0 þ aebðttj Þ 1ftj tg : (11.6)
j¼1
In Lando and Nielsen (2010), a total of 2,557 firms with an average of 1,142 and
a minimum of 1,007 firms at any time throughout the sample period from January
1982 to December 2005 are used to calibrate the goodness-of-fit of an extended
dynamic factor model that includes exogenously determined common factors as
well as a contagion effect in terms of a Hawkes process. During the sampling
period, 370 firms default and a contagious effect is concluded.
In this study, we simulate the default times of two firms using a mixture model
based on a dynamic factor model of Duffee (1999) together with a contagious effect
in the specification of a Hawkes process. Various commonly used copula functions
are calibrated against simulated default times to demonstrate the copula functions in
the modeling of credit default risk.
Copulas, introduced by Sklar (1959), have been extensively applied in areas such as
actuarial science using survival data. It was adopted by Li (2000) and Gregory and
Laurent (2005) to the application in modeling the joint distribution of the default
times t1, . . ., tI of a set of I firms. According to Sklar (1959), for any continuous
joint distribution F(t1, . . ., tI), there exists a uniquely determined I-dimensional
copula CF(u1, . . ., uI), where ui ¼ F(ti), 1 i I, such that for all (t1, . . ., tI) in RI,
In Definition 2.1, the formal definition of a copula is given. The theorem by Sklar
(1959) is given in Appendix 2.
306 L.-J. Kao et al.
where WX1;t and WX2;t are two independent Wiener processes. To account for coun-
terparty default risk, however, a modification is made on the default intensity in
Eq. 11.8 by including the default status of the other firm(s). For the two-firm case,
the default intensities are, respectively,
11 An Assessment of Copula Functions Approach 307
l1, t ¼ a0,1 þ a1,1 X1, t X1 þ a2,1 X2,t X2
(11.9)
þ l1,t þ aebðtt2 Þ 1ft2 tg
l2, t ¼ a0, 2 þ a1, 2 X1, t X1 þ a2, 2 X2, t X2
(11.10)
þ l2, t þ aebðtt1 Þ 1ft1 tg
where t1 and t2 are the default times of the 1st and 2nd firm, respectively. The
firm’s specific default intensities l1;t and l2;t are assumed to follow the mean-
reversion processes in Eq. 11.4.
The values of the parameters ki, yi, j, mj, a0,i, a1,i, a2,i, sXj , 1i2, 1j2 are
from Duffee (1999), in which month-end prices of noncallable corporate bonds
from January 1985 to December 1995 across 161 firms, with majority of investment
grade bonds, are used to calibrate default intensity process. In Table 11.1, the
parameter values used are given. Two cases with different contagious effects are
considered in the study. In the first case, the parameter a that specifies the conta-
gious effect is set to 0.1, while the a value is 0.25 in the second case. For each case,
10,000 pairs of simulated default times for the two firms are generated.
Five different copula functions are considered to fit the empirical joint distribu-
tion of the default times t1 and t2. They are the normal, t-, Clayton, Frank, and
Gumbel copula, respectively. For the normal and t-copula, the parameter r is set to
the empirical correlation coefficient of the simulated default times. Minimum mean
squares error is used to obtain the estimated parameter values of the degree of
freedom v in t-copula and y in Clayton, Frank, and Gumbel copula, respectively.
In Table 11.2, the estimated parameters for the five copula functions and the
corresponding sum squares of errors (SSE) are given. As can be seen in Table 11.2,
308 L.-J. Kao et al.
the normal copula is outperformed by the t-, Frank, and Gumbel copulas when
a ¼ 0.10, while outperformed by the t- and Gumbel copulas only when a ¼ 0.25. It
can also be seen that as the contagious effect parameter a increases from 0.10 to
0.25, except the Gumbel copula, the SSEs also increase significantly for the normal,
t-, Clayton, Frank, and Gumbel copulas. In both cases, i.e., a ¼ 0.10 and a ¼ 0.25,
the Gumbel copula performs the best while the Clayton copula performs the worst
among the five copula approximations.
In Figs. 11.1 and 11.2, the contour plots of the differences between the empirical
and copula-based joint distribution functions when the contagious effect parameter
a ¼ 0.10 and a ¼ 0.25, respectively, are given. For the normal copula, as can be seen in
Fig. 11.1 (panel a), larger deviations from the empirical joint distribution occur when
the default times t1 and t2 are in the range from 5 to 10 years when a ¼ 0.10. When
a ¼ 0.25, as illustrated in Fig. 11.2 (panel a), larger deviations occur in the range from
5 to 15 years. For the t-copula, larger deviations from the empirical joint distribution
occur when the default times t1 and t2 are in the range from 5 to 10 years when
a ¼ 0.10. However, in panel b of Fig. 11.2 where a ¼ 0.25, in addition to the range
between 5 and 10 years, the t-copula fails to approximate the empirical joint distri-
bution well for small default times t1 and t2. This implies that as the contagious effect
increases, the t-copula does not explain the lower tail dependence well. In panel d of
Figs. 11.1 and 11.2, when using the Gumbel copula, larger deviations occur when the
default times t1 and t2 are in the range from 2 to 7 years. This indicates that the
Gumbel copula does not approximate well for smaller default times.
Taken together, the goodness-of-fit of the normal, t-, Clayton, Frank, and Gumbel
copula, respectively, decreases as the contagious effect parameter a increases.
This phenomenon is more apparent for the Frank copula. The contour plots of the
11 An Assessment of Copula Functions Approach 309
20 20
a: normal copula b: t-copula
15 15
τ2
τ2
10 10
5 5
0 0
0 5 10 15 20 0 5 10 15 20
τ1 τ1
20 20
c: Clayton copula d: Gumbel copula
15 15
τ2
10
τ2
10
5 5
0 0
0 5 10 15 20 0 5 10 15 20
τ1 τ1
Fig. 11.1 Contours of Differences between Empirical Joint Distribution and Copula Approxi-
mation. Note Panels a-d show the contours of differences with absolute values exceeding 0.01 for
normal, t-Clayton, and Gumbel copula, respectively. The contagious effect parameter a is 0.1
20 20
a: normal copula b: t-copula
15 15
τ2
10 10
τ2
5 5
0 0
0 5 10 15 20 0 5 10 15 20
τ1 τ1
20 20
c: Clayton copula d: Gumbel copula
15 15
τ2
10
τ2
10
5 5
0 0
0 5 10 15 20 0 5 10 15 20
τ1 τ1
Fig. 11.2 Contours of Differences between Empirical Joint Distribution and its Copula Approx-
imation. Note Panels a-d show the contours of differences with absolute values exceeding 0.01 for
normal, t-Clayton, and Gumbel copula, respectively. The contagious effect parameter a is 0.25
differences between the empirical and Frank copula-based joint distribution functions
when the contagious effect parameter a ¼ 0.10 and a ¼ 0.25, respectively, are given
in Panel a–b of Fig. 11.3. As can be seen in Fig. 11.3, when a ¼ 0.10, the Frank
copula fails to approximate the empirical distribution well only for small default
310 L.-J. Kao et al.
20
a: contagious effect = 0.1
15
τ2
10
0
0 2 4 6 8 10 12 14 16 18 20
τ1
20
b: contagious effect = 0.25
15
10
τ2
0
0 2 4 6 8 10 12 14 16 18 20
τ1
Fig. 11.3 Contours of Differences between Empirical Joint Distribution and its Frank Copula Approx-
imation. Note Panels a–d show the contours of differences with absolute values exceeding 0.01 for
normal, t-Clayton, and Gumbel copula, respectively. The contagious effect parameter a is 0.1 and 0.25
times t1 and t2. As the contagious effect parameter a ¼ 0.25, however, the Frank
copula fails to approximate the empirical distribution well not only for small default
times t1 and t2 but also larger t1 and t2. This can also be seen in Table 11.2, where
the SSE increases from 0.1695 to 0.4525 as a increases from 0.10 to 0.25.
11.7 Conclusion
0T 1 0T 1
ð ð
E@ HdN A ¼ E@ HdAA:
0 0
ðt
AðtÞ ¼ lðsÞds:
0
Then l(t) is called the intensity process of the point process N. The Ht-predictable
intensity has the x dt, the probability one of the stopping time tk that occurs during
(t-dt,t] is
Basically, a copula function C links the joint distribution F(y1, . . ., yI) of I multivariate
random variables Y1, . . ., YI to their univariate marginal distributions F1(y1), . . .,
FI(yI). A formal definition of a copula function is given as follows.
Definition 2.1 A function C: [0, 1]N![0, 1] is a copula if there are uniform random
variables U1, . . ., UN taking values in [0, 1] such that C is their joint distribution
function that satisfies:
1. C(1, . . ., 1, ui, 1, . . ., 1) ¼ ui for all i ¼ 1, . . ., N and ui2[0, 1].
2. For all u2[0, 1]N, C(u) ¼ 0 if at least one coordinate ui ¼ 0, i ¼ 1, . . ., N.
3. For all u2[0, 1]N, v2[0, 1]N with ui vi, i ¼ 1, . . ., N, the C volume of the
hypercube
i1 þ...þiN
X 2 X2
... ð1Þ Cðwi1 ; . . . ; wiN Þ 0
i1 ¼1 i¼1
for all (t1, . . ., tI) in RI, where ui ¼ F(ti), 1iI. Moreover, if each Fi is continuous,
then the copula C is unique.
In order to calibrate a copula, indices that measure the dependence structure
introduced by the choice of the copula function are used. We focus on three
dependence measures that depend only on the copula function, not in the marginal
distributions. Among the three dependence measures, Kendall’s tau is the measure
of global dependence, while upper/lower tail dependence coefficients are two local
measures of dependence.
Definition 2.3 Let F be the joint distribution of two random variables X1 and X2
with marginal F1 and F2, respectively. Kendall’s tau is the probability of con-
0
cordance minus the probability of discordance. Specifically, if (X1, X2) and (X1 ,
0
X2 ) are two realizations of F, then Kendall’s tau is defined as
h 0
0
i h 0
0
i
P X 1 X 1 X2 X 2 > 0 P X 1 X 1 X2 X2 < 0 (11.11)
ð1 ð1
tau ¼ 4 Cðu; vÞdCðu; vÞ 1:
0 0
Definition 2.4 Let F(X1, X2) be the joint distribution of two random variables X1 and
X2, with marginal F1 and F2. The coefficients of upper and lower tail dependence
are defined, respectively, as
lU ¼ lim Pr X1 > F1 1
1 ð uÞ X 2 > F2 ð uÞ
u"1
lL ¼ lim Pr X1 < F1 1
1 ðuÞ X2 < F2 ðuÞ :
u#0
where z 0 ¼ (F1(u1), . . ., F1(uI)), and III is the unity matrix. For a normal
copula, the relationship between the linear correlation coefficient r and Kendall’s
tau is
On the other hand, the coefficients of upper and lower tail dependence are both
zero, i.e., a normal copula is tail independent.
Definition 2.6 (t-Copula) The I-dimensional t-copula is expressed as
Cðu1 ; . . . ; uI Þ ¼ tIn, S t1 1
n ðu1 Þ, . . . , tn ðuI Þ
where tn1 denotes the inverse of the distribution function of a univariate t-student
random variable with n degrees of freedom and tIn;S denotes the distribution function
of a multivariate t-distribution with n degrees of freedom and positive-definite
dispersion matrix S. Its density function is
nI1 nþI 0 1 2
nþI
G 2 G 2 1 þ z Sn z
ðnþ1 Þ
nþ1I 12 z2i 2
G 2 jSj I
1þ
i¼1 n
314 L.-J. Kao et al.
uy 1
fðuÞ ¼ ,y 0
y
then Eq. 11.13 is called a Clayton copula. When the generator
yu
e 1
fðuÞ ¼ ln y , y 6¼ 0
e 1
Equation 11.13 is called a Frank copula. When the generator f(u) ¼ (lnu)y,
y 1, Eq. 11.13 is a Gumbel copula. In particular, the density function of a Clayton
copula is
!I1=y
X
I I
cðu1 ; ...; uI Þ ¼ 1 I þ uy
i P uiy1 ðði 1Þy þ 1Þ : (11.14)
i¼1
i¼1
For a Clayton copula, Kendall’s tau and tail dependency measures are
ð1
uy 1 y
tau ¼ 1 þ 4 y1
du ¼
yu y2
0
0
f ð2vÞ
lU ¼ 2 2 lim 0 ¼0
v!0 f ðvÞ
0
f ð2vÞ
lL ¼ 2 lim 0 ¼2
v!0 f ðvÞ
References
Azizpour, S., & Giesecke, K. (2008). Self-exciting corporate defaults: Contagion vs. Frailty.
Working paper, Stanford University.
Credit Suisse Financial Products. (1997). CreditRisk+: A credit risk management framework.
New York.
CreditMetrics. (1997). The benchmark for understanding credit risk. Technical document.
New York: JP Morgan.
Das, S. R., Freed, L., Geng, G., & Kapadia, N. (2006). Correlated default risk. Journal of Fixed
Income, 16(2), 7–32.
Das, S., Duffie, D., Kapadia, N., & Saita, L. (2007). Common failings: How corporate defaults are
correlated. Journal of Finance, 62, 93–117.
Davis, M., & Lo, V. (2001). Infectious defaults. Quantitative Finance, 1, 382–387.
Driessen, J. (2005). Is default event risk priced in corporate bonds? Review of Financial Studies,
18, 165–195.
Duffee, G. R. (1999). Estimating the price of default risk. Review of Financial Studies, 12, 197–226.
Duffie, D., & Singleton, K. J. (1999a). Modeling term structures of defaultable bonds. Review of
Financial Studies Special, 12(4), 687–720.
Duffie, D., Singleton, K. J. (1999b). Simulating correlated defaults. Working paper, Graduate
School of Business, Stanford University.
Elizalde, A. (2005). Do we need to worry about credit risk correlation? Journal of Fixed Income,
15(3), 42–59.
Errais, E., Giesecke, K., & Goldberg, L. R. (2010). Affine point processes and portfolio credit risk.
SIAM Journal on Financial Mathematics, 1, 642–665.
Finger, C. (2000). A comparison of stochastic default rate models. Risk Metrics Journal, 1(2), 49–73.
Frey, R., McNeil, A., & Nyfeler, M. (2001). Copulas and credit models. RISK, 111–114.
Gregory, J., & Laurent, J. P. (2005). Basket default swaps, CDO’s and factor copulas. Journal of
Risk, 7(4), 103–22.
Gupton, G., Finger, C., & Bhatia, M. (1997). Credit metrics technical document. New York:
Morgan Guaranty Trust.
Hawkes, A. G. (1971). Spectra of some self-exciting and mutually exciting point processes.
Biometrika, 58(1), 83–90.
Hull, J., & White, A. (2001). Valuing credit default swaps II: Modeling default correlations.
Journal of Derivatives, 8(3), 12–22.
Jarrow, R., & Turnbull, S. (1995). Pricing derivatives on financial securities subject to credit risk.
Journal of Finance, 50, 53–86.
Jarrow, R., & Yu, F. (2001). Counterparty risk and the pricing of defaultable securities. The
Journal of Finance, 56, 1765–1799.
Kealhofer, S. (1995). Managing default risk in portfolios of derivatives. In: David M. Rowe (ed.)
Derivative Credit Risk: Advances in Measurement and Management. London: Risk Publications.
Koyuoglu, H. U., & Hickman, A. (1998). Reconcilable differences. Risk, 11(10), 56–62.
Lando, D. (1998). On cox processes and credit risky securities. Review of Derivative Research, 2,
99–120.
Lando, D., & Nielsen, M. S. (2010). Correlation in corporate default: Contagion or conditional
independence? Journal of Financial Intermediation, 19(3), 355–372.
Li, D. X. (1999). On default correlation: A copula function approach. Working paper 99–07,
The Risk Metrics Group.
Li, D. X. (2000). On default correlation: A copula approach. Journal of Fixed Income, 9, 43–54.
Lucas, D. (1995). Default correlation and credit analysis. Journal of Fixed Income, 9(4), 76–87.
Merton, R. (1974). On the pricing of corporate debt: The risk structure of interest rates. Journal of
Finance, 29, 449–470.
Nagpal, K., & Bahar, R. (2001). Measuring default correlation. Risk, 14(3), 129–132.
316 L.-J. Kao et al.
Contents
12.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 319
12.2 Literature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320
12.2.1 Panel Data Estimation Procedures: Time Series Versus Cross-Sectional
Effects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 320
12.2.2 Financial Architecture, Transactions Costs, and Risks . . . . . . . . . . . . . . . . . . . . . . . 321
12.2.3 Equity Premium as Measure of Equity Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 323
12.3 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 324
12.3.1 Estimating the Equity Premium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 325
12.3.2 Estimating the Cross-Border Determinants of Financial Architecture . . . . . . . 326
12.3.3 Estimation Methodologies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327
12.3.4 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 327
12.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 328
12.4.1 Descriptive Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 328
12.4.2 Results of Baltagi-Li EC2SLS Modeling and Blundell-Bond Modeling . . . . 328
12.4.3 Additional Econometric Details . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334
12.4.4 Discussion of Initial Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 334
12.4.5 Robustness Checks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 337
12.4.6 Discussion of Robustness Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338
12.4.7 Discussion of Differences in the Results of the Two
Different Modelings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 338
12.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 340
Appendix 1: Pearson Correlation Coefficients . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 341
Appendix 2: Estimation Procedure of Baltagi-Li EC2SLS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 342
Appendix 3: Estimation Procedure of Blundell and Bond (1998) System
GMM Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 344
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 347
R. Aggarwal (*)
University of Akron, Akron, OH, USA
e-mail: aggarwa@uakron.edu
J.W. Goodell
College of Business Administration, University of Akron, Akron, OH, USA
e-mail: johngoo@uakron.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 317
DOI 10.1007/978-1-4614-7750-1_12,
# Springer Science+Business Media New York 2015
318 R. Aggarwal and J.W. Goodell
Abstract
In the study of economic and financial panel data, it is often important to
differentiate between time series and cross-sectional effects. We present
two estimation procedures that can do so and illustrate their application by
examining international variations in expected equity premia and financial archi-
tecture where a number of variables vary across time but not cross-sectionally,
while other variables vary cross-sectionally but not across time. Using two
different estimation procedures, we find a preference for market financing to be
negatively associated with the size of expected premia. However, we also find that
US corporate bond spreads negatively determine financial architecture according
to the first procedure but not according to the second estimation as US corporate
bond spreads change value each year but have the same value across countries.
Similarly some measures that change across countries but do not change across
time, such as cultural dimensions as well as the index of measures against self-
dealing, are significant determinants of financial architecture according second
estimation but not according to the first estimation. Our results show that using
these two estimation procedures together can assess time series versus cross-
sectional variations in panel data. This research should be of considerable interest
to empirical researchers.
We illustrate with simultaneous-equation modeling. Following a Hausman
test to determine whether to report fixed or random-effects estimates, we first
report random-effects estimates based on the estimation procedure of Baltagi
(Baltagi 1981; Baltagi and Li 1995; Baltagi and Li 1994). We consider that the
error component two-stage least squares (EC2SLS) estimator of Baltagi and Li
(1995) is more efficient than the generalized two-stage least squares (G2SLS)
estimator of Balestra and Varadharajan-Krishnakumar (1987). For our second
estimation procedure, for comparative purposes we use the dynamic panel
modeling estimates recommended by Blundell and Bond (1998). We employ
the model of Blundell and Bond (1998), as these authors argue that their
estimator is more appropriate than the Arellano and Bond (1991) model for
smaller time periods relative to the size of the panels. We also use this
two-step procedure and use as an independent variable the first lag of the
dependent variable, reporting robust standard errors of Windmeijer (2005).
Thus, our two different panel estimation techniques place differing emphases
on cross-sectional and time series effects, with the Baltagi-Li estimator
emphasizing cross-sectional effects and the Blundell-Bond estimator
emphasizing time series effects.
Keywords
Panel data estimates • Time series and cross-sectional effects • Econometrics •
Financial institutions • Banks • Financial markets • Comparative financial
systems • Legal traditions • Uncertainty avoidance • Trust • Property rights •
Error component two-stage least squares (EC2SLS) • Generalized two-stage
least squares (G2SLS)
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 319
12.1 Introduction
Recently, there has been a veritable explosion of studies in a wide variety of areas
that employ panel data econometric methodology. Analysis of panel data allows
independent variables to vary both cross-sectionally and across time, while
panel data econometrics correct for cross-correlations between time series and
cross-sectional error terms. In recent years there has been the introduction of
a number of new refinements in analyzing panel data (Petersen 2009; Wooldridge
2010), all maintaining the goal of accounting for cross-correlation between time
series and cross-sectional error terms when assessing coefficient significance.
However, in the study of economic and financial panel data, it is often important to
assess the differential impact of time series versus cross-sectional effects; but panel data
techniques are unclear how this may be accomplished. In other words, panel data
methodologies typically do not inform us fully regarding which effect (time series or
cross-sectional) is more important or more dominant within particular data sets or
contexts. In this chapter we employ two contrasting estimation procedures, which,
respectively, emphasize cross-sectional versus time series differences, to clarify the
impacts of these two influences. We undertake this comparison of econometric methods
within a finance-related context which takes into account possible endogeneity.
We illustrate with simultaneous-equation modeling (outlined below). Following
a Hausman test to determine whether to report fixed or random-effects estimates,
we first report random-effects estimates based on the estimation procedure of
Baltagi (1981; Baltagi and Li 1994, 1995). We consider that the error component
two-stage least squares (EC2SLS) estimator of Baltagi and Li (1995) is more
efficient than the generalized two-stage least squares (G2SLS) estimator of Balestra
and Varadharajan-Krishnakumar (1987).
For our second estimation procedure, for comparative purposes we use the
dynamic panel modeling estimates recommended by Blundell and Bond (1998).
We employ the model of Blundell and Bond (1998), as these authors argue that their
estimator is more appropriate than the Arellano and Bond (1991) model for smaller
time periods relative to the size of the panels. We also use the two-step procedure,
reporting robust standard errors of Windmeijer (2005). Within this modeling we use
as an independent variable the first lag of the dependent variable. Thus, our two
different panel estimation techniques place differing emphases on cross-sectional
and time series effects, with the Baltagi-Li estimator emphasizing cross-sectional
effects and the Blundell-Bond estimator emphasizing time series effects.
Regarding the context of this econometric study, recent research suggests that both
national variations in the structures of financial intermediation and equity premia are
determined by many similar socioeconomic factors so that national variations in equity
premia can be expected to influence national variations in the structure of financial
intermediation and vice versa. But, until very recently prior literatures in these two
areas have ignored each other and the resulting possible endogeneity problems. Recent
exceptions include Aggarwal and Goodell (2011a, b), both of which examine the role
of financial architecture in determining nations’ equity premia. In this chapter we use
this economic context to illustrate two differing estimation procedures.
320 R. Aggarwal and J.W. Goodell
12.2 Literature
In recent years, the development of panel data econometrics has facilitated a large
increase in scholarship where panel data models are applicable. This is particularly
the case in international finance where data can be described across countries and time
as well as across industries and time. The development of panel data methods has
followed from the introduction by Hansen (1982) of Generalized Method of Moments
(GMM). GMM, including the use of instrumental variables, allows the implementa-
tion of consistent estimations based on conditional expectations which are inconsis-
tent with the use of earlier methods such as ordinary least squares regression.
Dynamic effects can render the fixed-effects estimator of panel models biased and
inconsistent, especially for data covering finite and short time periods. Among
alternative estimators that control for persistence is the system Generalized
Method of Moments (GMM) estimator proposed by Blundell and Bond (1998).
This procedure addresses econometric problems such as regressor endogeneity,
measurement error, and weak instruments while controlling for time-invariant,
country-specific effects such as distance or common language. Arellano and Bond
(1991) suggest transforming the model, either in the first differences or in orthogonal
deviations, to eliminate the fixed effects and to estimate it by using the two-step
GMM estimator. The second and higher lags of the endogenous variable in levels are
suitable instruments to overcome the estimation problem. However, when data are
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 321
highly persistent, Blundell and Bond (1998) argue that this procedure can be
improved by using the system – GMM estimation – which supplements the equations
in the first differences with equations in levels. For the equations in the first differ-
ences, the instruments used are the lagged levels, and for the equations in levels, the
instruments are the lagged differences.
In addition, the use of simultaneous-equation modeling on the same data
set allows us to assess the results using the Blundell and Bond estimation procedure
against an alternative. Following a Hausman test to determine whether to report
fixed or random-effects estimates, we report random-effects estimates based on the
estimation procedure of Baltagi (1981; Baltagi and Li 1994, 1995). We believe that
the error component two-stage least squares (EC2SLS) estimator of Baltagi and Li
(1995) is more efficient than the generalized two-stage least squares (G2SLS)
estimator of Balestra and Varadharajan-Krishnakumar (1987) because of
a broader set of transformations of the instruments.
It is very useful to examine the differences obtained with these two estimation
procedures. These two panel data estimation procedures have different emphases
with regard to cross-sectional versus time series effects. Because of the nature of
their construction, while the Baltagi-Li estimator emphasizes cross-sectional
effects, the Blundell-Bond estimator emphasizes time series effects.
1
For instance, Shirai (2004) reports that, because of improvements in official oversight for the
period 1997–2001, Indian capital markets improved significantly in being able to differentiate
high-quality firms from low-quality firms.
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 323
collateral can also lead to such an informational advantage for banks such that they
can charge excessively high interest rates which can weaken economic develop-
ment. Underlying Rajan and Zingales (1998) is the notion that when reliable
information about firms is too difficult or costly for the general public, banks
provide delegated monitoring (see Diamond 1984).
As this brief review of the relevant literature indicates, the international
determinants of financial architecture must include national characteristics such
as quality of investor protection, cultural and legal variables, and the equity
premium. However, there may be an endogeneity problem as financial architecture
seems to depend on a number of similar variables and the two variables, financial
architecture and equity premia, may influence each other. Our research design
includes a resolution of this issue.
We can expect variations in the ex ante equity premia across countries. Equity
premia can be expected to reflect the price of risk in equity investments. Depending
on national characteristics such as the nature of their institutional structures and
their levels of financial development, countries may differ with regard to both the
risk involved in equity investments and in the price of such risk. One way to think
about this extra risk and its price is to think about how the supply and demand for
equity investments may differ across countries, especially as most countries have
less than perfect capital markets.
Ibbotson et al. (2006) suggest that because of many obstacles and limitations, the
supply and demand for equity in markets may not respond to market forces as
would be expected from a theoretical view of efficient markets.2 For example, the
supply of equity may be restricted as bureaucratic rules and regulations may deter the
formation and market listing of corporate shares. Similarly, the demand for equity
may be limited due to uncertain property rights and the unreliability of public
information on potential investments. As the equity premium is the price of equity
risk, it is determined by the balance between supply and demand for equity. In order
to understand the nature and size of equity premia, it is important to account for the
nature of equity demand and supply in actual, imperfect, markets.
While there are no perfect measures of national supply of equity, a number
of variables could be used as indirect indicators. For example, stock market
capitalization as a ratio of GDP is one such widely used measure. But this ratio
by itself is an inadequate proxy because this measure will rise with a rise in
valuation (as well as with new listings) and valuation is itself affected by supply.
Other arguable measures of equity supply would include the number of shares
2
Ibbotson et al. (2006) actually start with somewhat of an alternative view to the commonly held
notion that prices in capital markets are set by the supply and demand for capital. Instead, they
focus on the viewpoint of the supplier of capital (an investor) and suggest that there is a supply and
demand for returns and that it is returns that are priced in the marketplace.
324 R. Aggarwal and J.W. Goodell
listed per unit of population or the number of new shares listed. A more extensive
evaluation of how developed an equity market is might include R2, the degree
to which individual stocks move synchronously with the overall equity market in
that country (Morck et al. 2000), or the degree to which market capitalization
is concentrated in a few firms. Turner (2003) associates the development of
nations’ private bond markets with the quality of their local investor bases.
Regulatory restrictions and lack of accounting standards can inhibit bond trading
by institutional investors. So if the quality of local bond markets in some measure
reflects the narrowness of the local investor base, then the quality of the bond
market might partially and indirectly determine the nonpecuniary (e.g., for control
rights) demand for equity.
Similarly, the demand for equity returns is likely to be influenced by a great
variety of factors that influence the risk level of equity and society’s perceptions,
tolerance, or price of equity risk. The nature of legal protection for investors,
disclosure requirements, the level of social trust that a particular society
believes can be placed in strangers, and the political stability of a country certainly
are some factors that come to mind. However, it is also reasonable to suppose that
there is correlation among many of the social, cultural, legal, and governance factors
that might affect the demand for equity. Ibbotson et al. (2006) suggest that the
demand for equity return is potentially also affected by concern for real returns as
opposed to nominal returns. Further, Moerman and van Dijk (2010) document
evidence that inflation risk is priced in international asset returns – so our investiga-
tion of the demand for equity returns also ought to control for inflation variability.
The supply and demand for equity in a country are also likely to be affected by
its financial architecture, i.e., the relative importance of the banking sector versus
financial markets in a country. It is now well recognized that some countries like
Japan and Germany are bank oriented, while other countries, like the Anglo-Saxon
countries, depend more on financial markets. For example, it can be expected that
equity premia are likely to be lower in countries with well-developed financial and
equity markets with a less restricted, and so greater, supply of equity than in bank-
oriented countries.
12.3 Methodology
As discussed briefly above, it seems that both national variations in the structures of
financial intermediation and equity premia are determined by many similar socio-
economic factors so that national variations in equity premia can be expected to
influence national variations in the structure of financial intermediation and vice
versa. But, until very recently, prior literatures in these two areas have ignored each
other and the resulting possible endogeneity problem. Recent exceptions include
Aggarwal and Goodell (2011a, b), both of which papers examine the role of financial
architecture in partial determining nations’ equity premia. This chapter overcomes
this limitation of prior research and examines the determinants of international
variations in financial architecture, accounting for the relationship between financing
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 325
architecture and ex ante equity risk premia with different estimation techniques.
One estimation technique uses simultaneous-equation estimates that include the
equity premium as an instrumental variable along with a number of other relevant
institutional, governance, and cultural factors. Another estimation technique uses
a dynamic panel data modeling which includes as an independent variable the ex ante
equity premium as a predetermined variable. Specifically, this chapter examines the
nature of international variations in national financial architecture with a special
emphasis on exploring the role of ex ante equity premia.
This chapter uses the abnormal earnings growth (AEG) model to estimate the ex
ante equity premium. This model is chosen not just because of its efficacy at
predicting ex post values but rather also because it is regarded as a fundamental
model upon which other previously used models are based (Penman 2005). In
addition, we use AEG estimations rather than alternative estimations in order to
avoid assumptions of future payout ratios. Further, the AEG model avoids the
problem with some estimation procedures when composite market-to-book ratios
for particular country/years are less than one.
Our estimates of ex ante equity premiums are based on data from the
period 1996–2006. Consensus of all available individual earnings forecasts
are obtained from Institutional Brokers Estimate System (I/B/E/S) of Thomson
Financial, as well as data for actual earnings per share, dividends per share, share
prices, and the number of shares outstanding. In the interests of consistency, forecasts
are collected as of April each year. Most firms end their fiscal years in December,
and, typically, values are reported within 90 days of the fiscal year end. If a firm has
I/B/E/S earnings forecasts for +1 and +2 years and a 5-year growth forecast, it is
retained in the sample. Firm-level data are then aggregated for each year. This paper
uses national treasury bill rates as a measure of the risk-free rate. This data is obtained
from International Financial Statistics of the International Monetary Fund (IMF).
This chapter estimates that abnormal earnings grow at a constant rate after
5 years at a rate equal to the rate of expected nominal GDP growth. Expected
GDP growth is modeled according to the following equation:
where eps1 is the next-period forecasted earnings per share (when all firms aggre-
gated, this becomes next year’s forecasted earnings), p0 is the price which when all
firms are aggregated becomes aggregate market values, d1 are the dividends per
share which become aggregate dividends, glt is the long-term growth rate, and gst is
the short-term growth rate. We proxy the short-term rate as the geometric average
of growth in earnings forecasts up to the fifth forecast. The long-term growth rate
is proxied as the rate of expected nominal GDP growth. We do not incorporate
the 5-year growth forecast from I/B/E/S in the estimate of the short-term growth
rate because of concerns that, by using an average of the 5-year growth
forecast from I/B/E/S, a bias will be introduced as a consequence of equal
weighting across years.
Besides the use of expected equity premia as an instrumental variable, we also use
a number of other independent variables. We control for inflation variability
(INFLATION_VOLATILITY) using the variance of the preceding 5 years as
reported by IFS. To capture time variation in risk appetite, we include the annual
difference from Moody’s Baa corporate bond yield and 10-year Treasury yield as
reported by the St Louis Federal Reserve for the middle of April for each year.
STOCK_VOLATILITY is the annualized standard deviation of monthly equity
returns of the respective Morgan Stanley Country Index. To account for relative
firm size and equity market concentration, we establish a Herfindahl index for each
country for each year (CONCENTRATION). A value of CONCENTRATION
close to 1 would suggest that most market capitalization for a particular country
in a given year is due to a small number of firms. Market capitalization data is
obtained from I/B/E/S. Based on Aggarwal and Goodell (2009), we speculate in
advance of empirical findings that country/years with high market capitalization
concentration will be more market based.
We control for regional differences by including as an independent variable
a dummy variable and REGION_EUROPE that receive “1” if the country is Europe.
We consider that the European region has historically had a unique relationship with
banking (see Rajan and Zingales 2003). We also control for cross-national differ-
ences in wealth and wealth inequality by including real GDP and the Gini coefficient
from World Development Indicators.
In order to assess the cultural impact on financial architecture, we include
as independent variables four cultural dimensions of Hofstede (2001):
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 327
In order to use the estimator of Baltagi and Li (1995), we design a set of simulta-
neous equations. With regard to our modeling of simultaneous equations, we
suggest that the risk factors and costs generally involved with resolving asymmetric
information in markets is reflected in the equity premium. Therefore we model
financial architecture as partially determined by the respective country/year equity
premium as well as political factors. Our empirical estimation models are based on
this set of equations:
X
FIN_ARCHit ¼ ait þ b1it EQ_PREMit þ b2it X2it þ eit (12.3)
X
EQ_PREMit ¼ ait þ b1it FIN_ARCHit þ b2it X1it þ eit (12.4)
12.3.4 Data
We analyze data for 41 countries over a recent 11-year period using panel data
methods. Specifically, we assess data for the 11-year period, 1996–2006, from
328 R. Aggarwal and J.W. Goodell
12.4 Results
Table 12.3 presents the means, and standard deviations of the variables used in our
models. Together these independent variables reflect the factors we described
earlier that may affect cross-national differences in financing predilections. Column
4 shows the standard deviation divided by the mean. This column suggests that the
dummy variable for European region, and our Herfindahl for market concentration
are the most variable of the independent variables (1.06, 0.96), while FIN_ARCH,
and EQUITY PREMIUM are similarly variable (0.73, 0.40) (Table 12.3).
Table 12.4 shows the results of panel regressions using the system of Eqs. 12.3 and
12.4, with our estimate of financial architecture as a dependent variable and our
estimate of the equity premium as an instrumented variable. The estimation is carried
out using the random-effects EC2SLS estimator proposed by Baltagi and Li (1992,
1995), and Baltagi (2001). Baltagi (2001) suggests the EC2SLS estimator is more
3
We begin our period of study in 1996 in order to include our measure of market concentration
which is a Herfindahl index we construct using data from I/B/E/S. There is insufficient data from
I/B/E/S for many countries prior to 1996. Another important reason, however, is that the measures
we use for political stability, control of corruption, and regulatory from Kaufmann et al. (2008) are
only available from 1996. Generally, our sample is restricted to those country/years which have
sufficient data reported by I/B/E/S and are included by Kaufmann et al. (2008). We stop in 2006 to
avoid the effects of the global financial crises and recession that started in 2007.
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 329
Where eps1 is the next-period forecasted earnings per share (when all firms aggregated, this
becomes next year’s forecasted earnings), p0 is the price which when all firms are aggregated
becomes aggregate market values, d1 are the dividends per share which becomes aggregate
dividends, glt is the long-term growth rate, and gst is short-term growth rate. We proxy the
short-term rate as the geometric average of growth in earnings forecasts up to the fifth forecast.
The long-term growth rate is proxied as rate of expected nominal GDP growth
Tables 12.4 and 12.5. SPREAD is significantly negative in Table 12.5 while not
significant in Table 12.4. Three cultural variables, PDI, MAS, and IDV, are
significant in Table 12.4 while not significant in Table 12.5. In Table 12.4, UAI and
MAS are negatively significant at 1 % and 5 %, respectively. In Table 12.4,
IDV and PDI are both positively significant at 1 %. However, UAI, MAS, IDV,
and PDI are not significant in Table 12.5. In other results, both tables show
LN_GDP positively significant, suggesting equity premia are larger in wealthier
countries.
Model 5 adds to the independent variables of Model 4, the comprehensive
index of anti-self-dealing (ANTI_SELF_DEAL) of Djankov et al. (2008). These
result, in Table 12.4, with regulatory efficiency, as reflected in measures against
self-dealing being positively significant. In Table 12.5, however, this variable is
not significant.
Table 12.4 Cross-national determinants of financial architecture: static panel estimation
332
Table 12.6 Cross-national determinants of financial architecture: dynamic panel estimation controlling for three lags
Dependent variable: financial architecture Model
1 2 3 4 5
INTERCEPT 0.23** (0.041) 0.36*** (0.001) 0.86* (0.093) 1.71 (0.946) –2.76 (0.655)
LAG1_FIN_ARCH 1.34*** (0.000) 1.36*** (0.000) 1.31*** (0.000) 1.20*** (0.000) 1.11*** (0.000)
LAG2_FIN_ARCH –0.80** (0.019) –0.76** (0.012) –0.70*** (0.000) –0.65*** (0.000) –0.62*** (0.000)
LAG3_FIN_ARCH 0.26 (0.121) 0.25* (0.069) 0.26*** (0.000) 0.23*** (0.002) 0.19** (0.030)
EQUITY_PREMIUM –0.00** (0.014) –0.00 (0.147) –0.00*** (0.002) –0.00*** (0.000) –0.00*** (0.000)
INFLATION_VOLATILITY 0.00 (0.684) 0.00 (0.708) 0.00 (0.437) 0.00 (0.723)
STOCK_VOLATILITY –0.01*** (0.000) –0.01*** (0.000) –0.01* (0.052) –0.01 (0.180)
CONCENTRATION 0.91*** (0.000) 0.97*** (0.001) 1.28*** (0.000)
REGION_EUROPE 0.08 (0.629) 0.33 (0.490) –0.31 (0.766)
LN_GDP –0.06 (0.248) 0.04 (0.834) 0.45 (0.366)
UAI 0.00 (0.989) –0.00 (0.935)
PDI 0.01 (0.637) 0.02 (0.454)
MAS –0.01 (0.300) –0.02 (0.348)
IDV –0.00 (0.905) –0.00 (0.983)
SPREAD –0.12** (0.013) –0.09** (0.025)
(continued)
Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 335
336
significant (at 10 %) in Model 5 in Table 12.4 but is not significant in Table 12.5.
Many of the cultural dimensions of Hofstede (2001) are significant in Table 12.4.
For instance, PDI, IDV, and MAS are significant in every model in which they are
present, while UAI is significant in Model 4. However, none of these cultural
variables are significant in Table 12.5.
Summarizing these differences, both static and dynamic estimation methods find
evidence of a negative association of financial architecture with the expected equity
premium. This suggests that investors demand less return for holding equity in more
market-oriented countries. Both static and dynamic estimation methods find evidence of
a negative association of financial architecture with stock volatility. This suggests that
less volatile markets are associated with being more market-oriented. Both static and
dynamic estimation methods find evidence of a positive association of financial archi-
tecture with equity market concentration. This suggests that, despite the widely dispersed,
market orientation of the United States, overall, more market-oriented countries have
greater concentration of equity ownership into fewer firms. Both static and dynamic
estimation methods also find evidence of a positive association of financial architecture
with national wealth, suggesting that wealthier countries are more market oriented.
While the two estimation procedures used in Tables 12.4 and 12.5 yield very
similar results with respect to a number of independent variables, some differences
also exist. For instance, SPREAD is significantly negative in Table 12.5 but not
significant in Table 12.4. SPREAD changes value each year but is the same across
countries. We consider that the difference between Tables 12.4 and 12.5 with
respect to SPREAD may be due to a difference in emphasis between these
tables with regard to cross-sectional versus time series effects, with the Baltagi-Li
estimator emphasizing cross-sectional effects and the Blundell-Bond estimator
emphasizing time series effects. Similarly the cultural dimensions of Hofstede
(2001) – as well as ANTI_SELF_DEALING – do not change over our time period.
However these variables vary widely across countries. And so the result that these
variables are significant in Table 12.4 but not significant in Table 12.5 is consistent
with the notion that Table 12.4 emphasizes cross-sectional differences, while
Table 12.5 emphasizes time series differences.
As noted above, because the dynamic panel estimator of Blundell and Bond (1998)
assumes that there is no autocorrelation in the idiosyncratic errors, we also report in
Table 12.5 Arellano-Bond test statistics. The Arellano-Bond test for zero correla-
tion in first-differenced errors is significant for differences of one lag. However,
because the first difference of white noise is necessarily autocorrelated, it is only
necessary to be concerned with second and higher autocorrelations that are higher
than 1 for the models in Table 12.5. However, examining the Arellano-Bond test
across the models in Table 12.5, the values for lags 2 and 4 are also consistently
significant. Therefore, in Table 12.6, we also control for the first three lags of our
dependent variable.
338 R. Aggarwal and J.W. Goodell
We then include the residuals from Eqs. 12.5 and 12.6 as substitute independent
variables for IDV and PDI, respectively. We refer to these variables as RESID_IDV
and RESID_PDI, respectively. This procedure lowers the correlation between IDV
(RESID_IDV) and PDI (RESID_PDI) to zero. The correlation between
RESID_IDV and RESID_PDI is now lowered to –0.45.
The models in Table 12.6 are the same as those in Tables 12.4 and 12.5, with the
exception of including three lags of the dependent variable. The results of
Table 12.6 substantially corroborate the results of Tables 12.5 and 12.4.
EQUITY__PREMIUM is negatively significant in every model. CONCENTRA-
TION is positively significant. STOCK_VOLATILITY is generally negatively
significant, as in Tables 12.4 and 12.5. However, unlike the previous tables this
variable is not significant in Model 5. Like Table 12.5, SPREAD is negatively
significant. Overall the results of Table 12.6 suggest that the results of Table 12.5
are not driven by serial correlation in higher lags and/or excessive multicollinearity.
Table 12.7 shows the results of the most comprehensive model (Model 5 in
Tables 12.4, 12.5, and 12.6) using three lags of the dependent variable and replacing
IDV and PDI with RESID_IDV and RESID_PDI. As seen in Table 12.7, controlling
for correlation between PDI and IDV by orthogonalizing both to LN_GDP results in
little change to the estimates for our most comprehensive model. Model 1 in
Table 12.7 is very similar to Model 5 in Table 12.4. Similarly, Model 2 in Table 12.7
is very similar to Model 5 in Table 12.5.
As noted above, while the two estimation procedures used in this study yield
very similar results with respect to a number of independent variables, some
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 339
12.5 Conclusions
1 2 3 4 5 6 7 8 9 10 11 12 13
1 FIN_ARCH 1
2 EQUITY_PREMIUM 0.03 1
3 INFLATION_VOLATILITY 0.07 0.16 1
4 MARKET_VOLATILITY 0.05 0.10 0.11 1
5 CONCENTRATION 0.03 0.00 0.04 0.30 1
6 REGION_EUROPE 0.19 0.12 0.10 0.14 0.18 1
7 LN_GDP 0.08 0.00 0.09 0.42 0.02 0.38 1
8 UAI 0.32 0.07 0.06 0.13 0.15 0.18 0.12 1
9 PDI 0.04 0.08 0.07 0.27 0.18 0.35 0.63 0.18 1
10 MAS 0.10 0.07 0.01 0.10 0.21 0.16 0.06 0.15 0.06 1
(continued)
342 R. Aggarwal and J.W. Goodell
1 2 3 4 5 6 7 8 9 10 11 12 13
11 IDV 0.10 0.05 0.07 0.37 0.00 0.44 0.51 0.17 0.67 0.12 1
12 SPREAD 0.05 0.14 0.08 0.06 0.10 0.02 0.00 0.02 0.01 0.01 0.01 1
13 ANTI_SELF_DEAL 0.29 0.05 0.09 0.09 0.20 0.48 0.05 0.49 0.02 0.00 0.07 0.01 1
yj ¼ Y j aj þ Xj bj þ uj ¼ Z j dj þ uj , j ¼ 1, 2, . . . , M,
yj ¼ NT 1; Y j ¼ NT Mj 1 ; Xj ¼ NT K j ; (12.7)
0 0 0
Zj ¼ Y j Xj ; dj ¼ aj ; bj ; nj ¼ Mj þ K j 1
uj ¼ Z u uj þ Z l lj þ vj , j ¼ 1, 2, . . . , M, (12.8)
N N
where Zu ¼ IN eT, Zl ¼ eN IT
0 1 2 2 3
uj sujl I N 0 0
6 7
E@ l j A u 1 l 1 v 1 ¼ 4 0
0 0 0
s2ljl I T 0 5 (12.9)
vj 0 0 svjl I TN
2
for j and l ¼ 1,2,. . . .M. As noted by Baltagi (1981), this implies that the covariance
matrix between jth and lth structural equation is
0
Sjl ¼ E uj ul ¼ s2ujl A þ s2ljl B þ s2vjl I NT , (12.10)
N N
where A ¼ IN eTe0 T, B ¼ eNe0 N IT
It follows that
J NT A J NT B J NT
Sij ¼ s23jl þ s1jl
2
þ s2jl
2
þ s2vjl Q (12.11)
NT T NT N NT
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 343
where
0
Q ¼ I NT AT NB þ JNT
NT
and JNT ¼ eNTeNT
also
s23jl ¼ s2vjl þ Ns2ljl þ Ts2ujl and s21jl ¼ s2vjl þ Ts2ujl as well as
s2vjl , s21jl , s22jl , s21jl are the characteristic roots of ∑ij of multiplicity (N1)(T1),
N1, T1, and 1, respectively (Baltagi 1981; Nerlove 1971).
If the rank condition K K1 + M1 1 holds for Eq. 12.7, then we can apply
transformation Qh to Eq. 12.7 to get
0 0 0
Xh yh1 ¼ Xh Zh1 d1 þ Xh uh1 (12.13)
ðhÞ2
^ 11 from Eq. 12.15 into Eq. 12.17 yields
Substituting the estimate of s
1
3
ðhÞ0 ðhÞ2 3
ðhÞ0 ðhÞ2
^
d 1, EC2SLS ¼ S
ðhÞ
Z 1 Pwh Z 1 =^s 11 S
ðhÞ
Z 1 Pwh y1 =^s 11
h¼1 h¼1
Additionally
It then follows that the following (T – 1)(T – 2)/2 linear moments conditions are
valid:
E yit2 Duit t ¼ 3, . . . : , T, (12.22)
where
yit2 ¼ ðyi1 ; yi2 ; . . . ; yit2 Þ0 and Duit ¼ uit uit1 ¼ Dyit aDyit1
12 Assessing Importance of Time-Series Versus Cross-Sectional Changes in Panel Data 345
Defining
2 3
2 3 ui3
yi1 0 0 ... 0 ... 0 6 ui4 7
6 7
6 0 yi1 yi2 . . . 0 . . . 0 7 7 6 : 7
Z di ¼ 6
4: : ; Du ¼ 6 7
: ... : ... : 5 i 6 : 7
6 7
0 0 0 . . . yi1 . . . yiT2 4 : 5
DuiT
i
yi1 ¼ þ ei (12.25)
1a
where
Defining
2 3
2 3 ui3
Dyi2 0 0 ... 0 ... 0 6 ui4 7
6 7
6 0 Dyi2 Dyi3 ... 0 ... 0 7 6 7
Z di ¼ 6 7; Dui ¼ 6 : 7
4: : : ... : ... : 5 6 : 7
6 7
0 0 0 . . . Dyi2 . . . DyiT1 4 : 5
DuiT
The above moment conditions can be written as
0
E Z li ui ¼ 0 (12.27)
al is a GMM estimator in the levels model. This is referred to as the levels’ GMM
estimator (Bun and Windmeijer 2010).
The full set of linear moments under assumptions (Eqs. 12.19, 12.20, 12.21, and
12.24) is
E yit2 Duit ¼ 0 (12.28)
E uit Dyi, t1 ¼ 0
for t ¼ 3, . . ., T
Following Bun and Windmeijer (2010), this can be resolved as
0
E Z si pi ¼ 0, (12.29)
where
2 3
Zdi 0 ... 0
60 Dyi2 ... 0 7
Z si ¼ 6 7; pi ¼ Dui
4: : : : 5 ui
0 0 . . . DyiT
Acknowledgments Authors are grateful to John Hunter, Maurice Peat, Min Qi, an anonymous
reviewer, and others for their kind advice on earlier versions but remain responsible for the
contents including any remaining errors.
References
Aggarwal, R., & Goodell, J. W. (2009). Markets and institutions in financial intermediation:
National characteristics as determinants. Journal of Banking and Finance, 33, 1770–1780.
Aggarwal, R., & Goodell, J. W. (2011a). International variations in expected equity premia: Role
of financial architecture and governance. Journal of Banking and Finance, 35, 3090–3100.
Aggarwal, R., & Goodell, J. W. (2011b). Variations in emerging market equity premia: Impact of
financial architecture. Journal of the Asia Pacific Economy, 16, 313–330.
Aggarwal, R., & Zhao, S. (2009). The diversification discount puzzle: Empirical evidence for
a transactions cost resolution. Financial Review, 44, 113–135.
Arellano, M., & Bond, S. (1991). Some tests of specification for panel data: Monte Carlo evidence
and an application to employment equations. Review of Economic Studies, 58, 277–297.
Arellano, M., & Bover, O. (1995). Another look at the instrumental variable estimation of error-
components models. Journal of Econometrics, 68, 29–51.
Balestra, P., & Varadharajan-Krishnakumar, J. (1987). Full information estimations of a system of
simultaneous equations with error components structure. Econometric Theory, 3, 223–246.
Baltagi, B. H. (1981). Simultaneous equations with error components. Journal of Econometrics,
17, 189–200.
Baltagi, B. H. (2001). Econometric analysis of panel data. New York: Wiley.
Baltagi, B. H., & Li, Q. (1992). A note on the estimation of simultaneous equations with error
components. Econometric Theory, 8, 113–119.
Baltagi, B. H., & Li, Q. (1994). Estimating error component models with general MA
(q) disturbances. Econometric Theory, 10, 396–408.
Baltagi, B. H., & Li, Q. (1995). Testing AR(1) against MA(1) disturbances in an error component
model. Journal of Econometrics, 68, 133–151.
Baltagi, B. H., & Wu, P. X. (1999). Unequally spaced panel data regressions with Ar(1) distur-
bances. Economic Theory, 15, 814–823.
Beck, T., Demirguc-Kunt, A., & Levine, R. (2000). A new database on financial development and
structure. World Bank Economic Review, 14, 597–605.
Bhattacharya, S., & Thakor, A. V. (1993). Contemporary banking theory. Journal of Financial
Intermediation, 3, 2–50.
Blundell, R., & Bond, S. (1998). Initial conditions and moment restrictions in dynamic panel data
models. Journal of Econometrics, 87, 115–143.
Bun, M. J. G., & Windmeijer, F. (2010). The weak instrument problem of the system GMM
estimator in dynamic panel data models. Econometrics Journal, 13, 95–126.
Coase, R. H. (1960). The problem of social cost. Journal of Law and Economics, 3, 1–44.
Diamond, D. W. (1984). Financial intermediation and delegated monitoring. Review of Economic
Studies, 51, 393–414.
Djankov, S., La Porta, R., Lopez-de-Silvanes, F., & Shleifer, A. (2008). The law and economics of
self-dealing. Journal of Financial Economics, 88, 430–465.
Ergungor, O. E. (2004). Market- vs. bank-based financial systems: Do rights and regulations really
matter? Journal of Banking and Finance, 28, 2869–2887.
Fligstein, N. (2001). The architecture of markets: An economic sociology of twenty-first-century
capitalist societies. Princeton: Princeton University Press.
Hansen, L. P. (1982). Large sample properties of generalized method of moments estimators.
Econometrica, 50, 1029–1054.
Hart, O. (1995). Firms, contracts, and financial structure. New York: Oxford University Press.
Hart, O. (2001). Financial contracting. Journal of Economic Literature, 39, 1079–1100.
348 R. Aggarwal and J.W. Goodell
Contents
13.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 350
13.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 352
13.3 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355
13.3.1 Instrumental Variable for Capital . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 355
13.3.2 Generalized Method of Moments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 356
13.4 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 357
13.4.1 Overall Observations of BHC Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 359
13.4.2 Instrumental Variable . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 360
13.4.3 Measures of Bank Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
13.5 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 361
13.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 376
Appendix 1: Stochastic Frontier Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 377
Appendix 2: Generalized Method of Moments . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 379
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 380
Abstract
In this chapter, we thoroughly analyze the relationship between capital and bank
risk-taking. We collect cross section of bank holding company data from 1993 to
2008. To deal with the endogeneity between risk and capital, we employ
stochastic frontier analysis to create a new type of instrumental variable. The
unrestricted frontier model determines the highest possible profitability based
solely on the book value of assets employed. We develop a second frontier based
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 349
DOI 10.1007/978-1-4614-7750-1_13,
# Springer Science+Business Media New York 2015
350 W.-J.P. Chiou and R.L. Porter
on the level of bank holding company capital as well as the amount of assets. The
implication of using the unrestricted model is that we are measuring the uncon-
ditional inefficiency of the banking organization.
We further apply generalized method of moments (GMM) regression to avoid
the problem caused by departure from normality. To control for the impact of
size on a bank’s risk-taking behavior, the book value of assets is considered in
the model. The relationship between the variables specifying bank behavior and
the use of equity is analyzed by GMM regression. Our results support the theory
that banks respond to higher capital ratios by increasing the risk in their earning
asset portfolios and off-balance-sheet activity. This perverse result suggests that
bank regulation should be thoroughly reexamined and alternative tools devel-
oped to ensure a stable financial system.
Keywords
Bank capital • Generalized method of moments • Stochastic frontier analysis •
Bank risks • Bank holding companies • Endogeneity of variables
13.1 Introduction
1
For detailed discussion, please see Berger et al. (1995); Gatev et al. (2009); Hovakimian and Kane
(2000); Shrieves and Dahl (1992); and VanHoose (2007).
13 Does Banking Capital Reduce Risk? 351
It has been argued that excessively high capital requirements can produce social
costs through lower levels of intermediation. In addition, there can be unintended
consequences of high capital requirements such as risk arbitrage (increasing risk to
offset the increase in capital and thereby maintain the same return on capital),
increased securitization, and increased off-balance-sheet activity, all of which
could mitigate the benefits of increased capital standards. See Berger et al. (1995)
and Santos (2001). The extent to which these unintended consequences played
a role in our recent crisis is yet to be determined.
Moral hazard is high on the list of problems receiving attention in this post-
financial crisis environment. The presence of a federal safety net creates moral
hazard because bank management does not have to worry about monitoring by
depositors (see Merton 1977; Buser et al. 1981; Laeven and Levine 2009). Absent
depositor monitoring, banks are free to increase risk. If, however, deposit insurance
and other elements of a federal safety net are reasons for increases in bank risk, why
do they continue to exist? The answer lies in the contemporary theory of financial
intermediation. It has been well established in the literature that there is need for
both demand deposit contracts and the possibility of bank runs (Diamond and
Dybvig 1983; Calomiris and Kahn 1991; Diamond and Rajan 2000; and Santos
2001). If the possibility of bank runs is needed, and bank runs are harmful, then
government deposit insurance is an optimal solution. There is a related issue. Banks
have a unique ability to resolve information asymmetries associated with risky
loans. As a result, bank failures can produce a serious contraction in credit avail-
ability, especially among borrowers without access to public capital markets. The
federal safety net is needed to avoid this credit contraction. Likewise, if a bank is
considered “too big to fail,” then the government will always bail the bank out, and
there is no reason for bank management to limit risk.
13 Does Banking Capital Reduce Risk? 353
It needs to be noted that not everyone is in agreement that the use of capital
requirements is the best way to reduce risk in banking. Marcus (1984) and Keeley
(1990) argue that a bank’s charter value mitigates against increased risk. Banks
operate in a regulated environment, and therefore, a charter to operate contains market
power. Excessive risk increases the cost of financial distress, and this can cause a loss
of charter value. Kim and Santomero (1988) argue that a simple capital ratio cannot be
effective, and any ratio would need to have exactly correct risk weights in a risk-based
system. Gorton and Pennacchi (1992) discuss “narrow banking” and propose splitting
the deposit services of banks from the credit services. In other words, the financial
system would include money market accounts and finance companies. The money
market accounts would only invest in short-term high-quality assets and leave the
lending to the finance companies that would not take in any deposits.
In Prescott (1997), he reviews the precommitment approach to risk management.
Briefly, banks commit to a level of capital, and if that level proves to be insufficient,
the bank is fined. This is used currently in the area of capital in support of a trading
portfolio but cannot be used for overall capital ratios since a fine against a failed
bank is not effective. Esty (1998) studies the impact of contingent liability of
stockholders on risk. In the late nineteenth and early twentieth century, bank
stockholders were subject to a call or an assessment for more money if needed to
meet the claims on a bank. There was a negative relation between increases in risk
and the possible call on bank stockholders. Calomiris (1999) makes a strong case
for requiring the use of subordinated debt in bank capital structures. The need to
issue unguaranteed debt and the associated market discipline would act as an
effective limit to the amount of risk a bank would be able to assume. John
et al. (2000) argue that a regulatory emphasis on capital ratios may not be effective
in controlling risk. Since all banks will have a different investment opportunity set,
an efficient allocation of funds must incorporate different risk-taking for different
investment schedules. These authors go on to argue that senior bank management
compensation contracts may be a more promising avenue to control risk using
incentive-compatible contracts to achieve the optimal level of risk.
Marcus and Shaked (1984) show how Merton’s (1977) put option pricing formula
can be made operational and then used the results to estimate appropriate deposit
insurance premium rates. The results of their empirical analysis indicated that the then
current FDIC premiums were higher than was warranted by the ex ante default risk of
the sample banks. This implies that banks are not transferring excessive risk to the
deposit insurance safety net and capital regulation is effectively working.
Duan et al. (1992) address the question of the impact of fixed-rate versus risk-
based deposit insurance premiums directly. The authors tested for specific risk-
shifting behavior by banks. If banks were able to increase the risk-adjusted value of
the deposit insurance premiums, then they had appropriated wealth from the
FDIC. This is because the FDIC, at the time, could not increase the insurance
premium even though risk had increased. Their empirical findings were that only
20 % of their sample banks were successful in risk-shifting behavior and therefore
the problem was not widespread. This also implies that capital management has
been effective.
354 W.-J.P. Chiou and R.L. Porter
banking industry and the related increase in risk exposure. They suggest that
increased diversification may have been offset by the increased risk of the newly
permissible activities. As a result, it was counterparty risk that was the driving force
for higher capital ratios.
13.3 Methodology
We differ from previous studies that deal with the endogeneity between risk and
capital using traditional methods such as a simultaneous equation approach or two-
or three-stage regression analysis.2 In this study, we follow the method and concept
of Hughes et al. (2001, 2003); and others and use stochastic frontier analysis to
estimate the inefficiency of our sample of bank holding companies. See Jondrow
et al. (1982) for a discussion of fitting production frontier models. We then create
a unique instrumental variable for bank capital to be used in regressions of capital
and risk. The question we ask is: “How efficient is a bank holding company in
converting the resources with which it has to work into profit?” The frontier
developed is exogenous to any specific bank since it is based on the results of all
banks in the sample. From this frontier, we measure the inefficiency of each bank as
the distance between the frontier and that specific bank’s pretax income. This
measure, however, must be adjusted for those elements that are beyond the control
of the bank.
Our unrestricted frontier model determines the highest possible profitability
based solely on the book value of assets employed. The unrestricted model is
specified as
where PTI is pretax income, BVA is of book value of assets, x is statistical noise, B is
systematic shortfall (under management control), and B 0. A quadratic specifi-
cation is used to allow for a nonlinear relation between the pretax income and the
book value of assets.
Our next step is to develop a second frontier based on the level of bank holding
company capital as well as the amount of assets. The implication of using the
unrestricted model is that we are measuring the unconditional inefficiency of the
banking organization. By also conditioning the model on capital, we can develop
a measure of the incremental efficiency or inefficiency of an organization due to its
2
In Appendix 1, we explain the execution of stochastic frontier in this chapter.
356 W.-J.P. Chiou and R.L. Porter
capital level. It is this incremental inefficiency due to a bank’s capital level that we
propose to use as an instrument for capital in a regression of risk on capital.
Specifically, our restricted model, again in a quadratic form, is as follows:
where BVC is the book value of capital, v is statistical noise, and u denotes the
inefficiency of a bank considering its use of both assets and capital.
The two assessments of inefficiency allow us to measure the difference in
profitability due to the use of capital by calculating the difference in the inefficiency
between the restricted and unrestricted model. Specifically,
d¼uB (13.3)
This becomes our instrumental variable for capital. While any measure of
profitability endogenously includes risk, our instrument, the difference between
two measures of profitability conditioned only on capital, is related to capital but
not to risk which is included in both models.
where yk,t is one of the measures of risk or behavior (e.g., total equity/total asset)
for bank i in year t; c is a constant; bk,t is the coefficient of instrumental
3
In Appendix 2, we provide the derivation of the GMM model.
13 Does Banking Capital Reduce Risk? 357
variable of capital, di,k, for k’s regression in year t; gk,t is the coefficient of natural
logarithm of bank’s book value; and k,t is the error term.
13.4 Data
We obtain our data on bank holding companies from Federal Reserve reports FR
Y-9C for the years 1993–2008. Data on risk-weighted assets, tier 1 capital, and tier
2 capital were not included with the FR Y-9C reports from 1993 to 1996. We were
graciously provided this missing information by the authors of Berger et al. (2008).
Table 13.1 displays the descriptive statistics of the sample BHCs in our analysis.
The total of 24,973 bank-year observations ranges from 2,256 in 2005 to 678 in
2008. From 2005 to 2006, there is an especially large drop in the number of BHCs
included in our data. This is primarily due to a change in the reporting criteria for
the FR Y-9C report. Starting in 2006, the threshold for required reporting by a BHC
was increased from BHCs with $150 million in total assets to BHCs with $500
million in total assets. Note that in spite of the 57 % drop in the number of BHCs
reporting in 2006 compared with 2005, the total assets represented in the sample for
these 2 years decreased by only 14 %.
Our data start in 1993 because 1992 was the final year in which capital ratios were
still adjusting in order to conform to the Basle I Capital Accord. As a result, 1993
represents the first year that does not include any mandated changes in the capital ratios.
The entire period of 1993–2008 contains a number of significant events affecting the
banking industry. For instance, the Riegle-Neal Interstate Banking and Branching Act
was passed in 1994 eliminating geographic restrictions on bank expansion. In 1999 the
Gramm-Leach-Bliley Financial Services Modernization Act was passed effectively
repealing the Glass-Steagall Act. Together these two acts overturned 65 years of
legislation and regulation intended to keep banks financially sound.
From an economic point of view, the early portion of our time period represented
a time of recovery from recession. The economy then moved from recovery to
growth, and the decade ended in a tech-stock boom followed by a bursting of the
tech-stock price bubble and an attendant recession. The new decade brought
traditional financial policies intended to stimulate the economy which, in hindsight,
probably helped to lay the foundation for the housing price bubble which precip-
itated the 2007–2009 financial crisis. The time period from 1993 to 2008 seems to
be a very appropriate period in which to analyze bank capital ratios.
Previous empirical studies have used market measures of risk and various risk
measures derived from a market model based on return data. However, this
approach necessarily limits the sample to publically owned banks or bank holding
companies. In this study, we wish to determine the impact of capital on various
measures of risk and acknowledge the importance of small banks and bank holding
companies, as well as the largest bank holding companies. This concern is signif-
icant since public policy related to the banking industry must consider the broadest
sample and not only the largest organizations. As a result, we utilize a large panel
data set and turn to the typical accounting measures of a bank’s risk.
358
We see the significant events and the economic activity listed above in the statistics
in Table 13.1. First, the size of BHCs measured by either their asset values or equity
has increased, while the number of banks has decreased. This trend is still evident
after adjusting for changes in the reporting criteria for the FR Y-9C report. The
government deregulation noted above has resulted in increased concentration in the
banking industry. We note also the significant cross-sectional variation in scale of
BHCs that suggests the utilization and operation of their resources vary
considerably.
When we look in Table 13.1 at the basic leverage ratio of equity to assets (E/A),
we see a generally rising ratio. In 1993, the ratio was 8.5 %, while in 2008 it was
9.2 %. These ratios appear to be in line with mandatory capital requirements. We
also see variation in this trend consistent with prevailing economic activity. For
example, the decline from 9.4 % in 1998 to 8.9 % in 1999 reflects the tech-stock
problems of that time period. In Table 13.1, we also see a rising trend in RA, the
ratio of risk-based assets to total assets. Here, however, the trend is far more
pronounced, rising from 43.80 % in 1993 to 76.00 % in 2008. Confirmation of
these two trends comes from the trend in CAP, the ratio of tier 1 plus tier 2 capital to
risk-based assets. This ratio declines from 16.10 % in 1993 to 14.50 % in 2008.
While these ratios are substantially above the Basle Capital Accord standards, the
trend is clearly down.
Another dramatic trend over this time period is the increase in off-balance-sheet
activity. In Table 13.1, the off-balance-sheet activities-to-total assets ratio (OBS)
has increased from 12.00 % in 1993 to 31.50 % in 2008. While this trend is not
a surprise, we need to ask if there is capital to support this expansion and consider
the makeup of the components of off-balance-sheet activities. It is unclear whether
BHCs use off-balance-sheet activities to decrease or increase risk.
The time-varying overall performance measures of our sample of BHCs such as
pretax income (PTI), return on equity (ROE), nonperforming assets ratio (NPA),
and the interest-sensitive gap (Gap) are shaped by major economic occurrences and
policies. Return on equity has varied in a relatively narrow band over this time
period. With the exception of 2007 and 2008, the return on equity ranged from
12.20 % to 13.50 %. In line with the financial crisis that started in 2007, ROE
declined to 11.00 % in 2007 and to 8.40 % in 2008. It is also noteworthy that the
highest return on equity was in the first year of our sample period, 1993. Nonper-
forming assets appear to move in concert with economic activity. The recovery and
expansion period of 1993–1998 is marked by a steady decrease in the ratio of
nonperforming assets to equity. This is followed by an increase in this ratio during
the tech-stock bubble and recession after which we see another decline until the
crisis of 2007 and 2008.
Since the industrial structure of financial services changes intertemporally, we
analyze the risk and use of capital by BHCs year by year. The analysis suggests banks
progressively depend more on aggressive funding sources and new product lines over
our sample period. Given that financial leverage (e.g., equity/asset ratio) must remain
360 W.-J.P. Chiou and R.L. Porter
approximately stable due to regulatory requirements, bankers may try to improve their
ROE by (1) enhancing overhead efficiency (OHE), (2) engaging in more off-balance-
sheet activities (OBS), and (3) using interest-sensitive gap management in an attempt
to decrease their total risk-based capital ratio (Cap) while maintaining an attractive
ROE. The above developments in the banking industry generate potential improve-
ment in performance but also intensify uncertainties and complexities of bank man-
agement. Therefore, a study to investigate the impact of the use of capital on the
riskiness of banks is an indispensible element in bank management.
The statistical summary of our instrumental variable, d, for each year is shown in
Table 13.2. Consistent with the findings documented by Hughes et al. (2001), John
et al. (2000), Keeley (1990), and Kim and Santomero (1988), the use of equity
capital by banks, on average, triggers a loss in efficiency. The dispersion of d is
substantial both cross-sectionally and intertemporally. For our sample, the distri-
bution of d in the same year tends to be skewed to the left-hand side and leptokurtic
(i.e., has positive excess kurtosis). Therefore, we look at nonparametric statistics
and use a normality-free regression model in our analysis to avoid the possible
errors in estimation.
We investigate the risks faced by banks from various aspects. Table 13.3 displays
the measures of risk used in this study: credit risk, liquidity risk, interest rate risk,
off-balance-sheet (OBS) risk, market risk, and finally leverage risk. Credit risk is
concerned with the quality of a bank’s assets. Historically this has focused on
a bank’s loan portfolio, but recent events have shown the importance of looking at
all bank assets in light of potential default risk. Liquidity risk measures the ability of
a bank to meet all cash needs at a reasonable cost whenever they arise. Interest rate
risk is the extent to which banks have protected themselves from market-driven
changes in the level of interest rates. Banks have the opportunity to use asset/
liability management tools to mitigate the impact of changes in interest rates on
both bank earnings and bank equity. We also collect data on off-balance-sheet
activities and investigate their relationship with bank capital. Market risk is the risk
of changes in asset prices that are beyond the control of bank management. Finally,
leverage risk is the risk arising from the capital structure decisions of the BHC. The
first five measures of risk relate to the various elements of business risk confronting
bank management. Leverage risk, on the other hand, relates directly to the financial
decisions taken in terms of the amount of capital employed. From another perspec-
tive, it can be said that minimum capital requirements (i.e., maximum leverage
standards) are mandated by regulators to mitigate the various elements of business
risk that the BHC accepts.
Table 13.4 displays the Spearman correlation coefficients between bank size
and our instrumental variable over the sample period. We look at this nonpara-
metric test due to the non-normal distribution of the instrument and variables. We
believe that the generally insignificant correlation between our instrument and the
book value of assets in combination with the generally significant correlation of
our instrument and the book value of equity justifies the use of delta as an
instrument for capital. In addition, Table 13.4 shows that, measured by book
value of equity and pretax income, large BHCs tend to suffer a greater loss in
efficiency than their smaller counterparts at a statistically significant level. On the
other hand, the value of assets does not necessarily demonstrate a negative relation
with bank efficiency. These findings suggest that the inefficiency of BHCs comes
from the use of equity capital but is not directly led by the expansion of business
scale and/or scope. Therefore, a careful investigation of the impact of capital on
banking risks is appropriate.
We look at seven different measures of risk: credit risk, liquidity risk, interest rate
risk, off-balance-sheet risk, market risk, composite risk, and leverage risk. While
many of the results support the proposition that increased capital requirements
reduce risk in BHCs, there are some very significant results that suggest the
opposite – that BHCs increase risk as their capital ratios increase.
362 W.-J.P. Chiou and R.L. Porter
Our results are displayed in Tables 13.5, 13.6, 13.7, 13.8, 13.9, and 13.10 and
provide a number of interesting insights. To enhance robustness, we present both
Spearman’s rank correlation coefficient between the tested variable and the instru-
mental variable for capital, d, and the coefficient of d in GMM regressions. To
control for the size of the bank holding companies, each coefficient of d is generated
by GMM regression with a constant and the natural logarithm of the book value of
assets. While the coefficient of the control variable and the constant term are omitted
from the tables, they are available upon request. In Table 13.5, we find a positive
relationship between ratio of total equity to total assets and our instrument for capital
(see Eq/A). The coefficient on our instrument is strictly positive and statistically
significant. This is clearly what we would expect. As leverage decreases, so does
risk; therefore, higher capital should be associated with higher levels of this risk
measure. In other words, it should be a positive relationship, and it is. However, for
the ratio of risky assets to total assets, risk increases as the ratio increases. Therefore,
higher capital should be associated with lower levels of this risk measure (a negative
relationship), and again that is what we find. When we look at just tier 1 capital to
total risky assets, we find the expected positive relationship, and when we look at the
ratio of tier 2 to tier 1 capital, we find the expected negative relationship.
In Table 13.6, we look at some traditional measures of credit risk. As the ratio of
nonperforming loans to total loans increases, so does risk. Therefore, the coefficient
on capital should be negative, and they are with several exceptions over the years. Our
second measure of credit risk is the ratio of nonperforming loans to total equity. Here
again higher levels of the ratio imply higher risk, so we expect to find a negative
relationship and we do, and this time without exception and at high levels of
significance. When we look at the ratio of loan charge-offs to loans outstanding, we
have more exceptions, but in general we find an expected negative relationship.
364
Table 13.4 Explanatory variables and instrument for capital of BHCs: Spearman’s rank correlation coefficient
2008 2007 2006 2005 2004 2003 2002 2001 2000 1999 1998 1997 1996 1995 1994 1993
r(BVA, d) 0.27 0.02 0.01 0.01 0.01 0.07 0.01 0.03 0.03 0.01 0.03 0.25 0.07 0.07 0.00 0.01
p-value 0.001 0.409 0.422 0.429 0.412 0.101 0.413 0.323 0.327 0.450 0.337 0.000 0.145 0.136 0.495 0.467
r(BE, d) 0.51 0.27 0.23 0.32 0.34 0.26 0.32 0.31 0.30 0.33 0.34 0.51 0.34 0.33 0.26 0.20
p-value 0.000 0.000 0.001 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
r(PTI, d) 0.38 0.09 0.04 0.13 0.13 0.04 0.13 0.16 0.20 0.20 0.17 0.35 0.23 0.23 0.16 0.17
p-value 0.000 0.123 0.304 0.005 0.006 0.247 0.007 0.002 0.000 0.000 0.003 0.000 0.000 0.000 0.008 0.002
Spearman’s rank correlation coefficients between book value of assets (BVA), book value of equity (BE), and pretax income (PTI) with the instrumental
variable for capital (d) in each year are reported
W.-J.P. Chiou and R.L. Porter
13
p-value 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
Tier 0.38 0.45 0.50 0.44 0.42 0.46 0.44 0.44 0.45 0.44 0.44 0.44 0.44 0.44 0.44 0.44
1/RA
p-value 0.000 0.000 0.000 0.000 0.009 0.003 0.004 0.005 0.004 0.004 0.005 0.004 0.004 0.004 0.004 0.004
Tier 0.23 0.37 0.38 0.33 0.36 0.35 0.35 0.35 0.35 0.35 0.35 0.35 0.35 0.35 0.35 0.35
2/tier 1
p-value 0.006 0.000 0.000 0.002 0.001 0.001 0.006 0.002 0.003 0.004 0.003 0.003 0.003 0.003 0.003 0.003
Panel B: GMM regressions
2008 2007 2006 2005 2004 2003 2002 2001 2000 1999 1998 1997 1996 1995 1994 1993
Eq/A 0.312 0.102 0.144 0.206 0.202 0.161 0.161 0.165 0.174 0.165 0.222 0.098 0.221 0.215 0.202 0.204
p-value 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
RA/A 0.035 0.105 0.045 0.199 0.242 0.152 0.155 0.152 0.154 0.114 0.165 0.049 0.066 0.021 0.070 0.118
p-value 0.375 0.000 0.150 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.106 0.116 0.372 0.172 0.225
(continued)
365
366
p-value 0.442 0.052 0.010 0.083 0.052 0.052 0.065 0.005 0.010 0.005 0.006 0.160 0.008 0.004 0.022 0.142
Provision/E 0.24 0.33 0.39 0.41 0.38 0.35 0.37 0.41 0.40 0.41 0.38 0.24 0.34 0.33 0.27 0.26
p-value 0.004 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
Allowance/L 0.18 0.08 0.08 0.13 0.12 0.13 0.17 0.13 0.15 0.14 0.14 0.17 0.11 0.06 0.05 0.02
p-value 0.021 0.144 0.135 0.045 0.069 0.068 0.001 0.012 0.005 0.009 0.011 0.003 0.038 0.168 0.204 0.350
Allowance/E 0.40 0.54 0.57 0.60 0.57 0.50 0.56 0.57 0.55 0.52 0.50 0.33 0.43 0.44 0.43 0.49
p-value 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
Panel B: GMM regressions
2008 2007 2006 2005 2004 2003 2002 2001 2000 1999 1998 1997 1996 1995 1994 1993
NPL/LL 0.074 0.006 0.002 0.002 0.004 0.007 0.001 0.001 0.001 0.001 0.000 0.004 0.004 0.009 0.001 0.003
p-value 0.022 0.017 0.219 0.134 0.007 0.000 0.258 0.402 0.373 0.246 0.415 0.012 0.028 0.003 0.437 0.204
NPL/E 9.510 0.159 0.103 0.088 0.087 0.004 0.122 0.118 0.087 0.088 0.090 0.064 0.116 0.173 0.191 0.195
p-value 0.096 0.000 0.000 0.000 0.000 0.463 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
Charge-offs/L 0.007 0.001 0.000 0.001 0.002 0.005 0.000 0.003 0.000 0.000 0.001 0.003 0.002 0.003 0.002 0.004
p-value 0.074 0.138 0.431 0.134 0.037 0.000 0.358 0.142 0.435 0.313 0.199 0.010 0.045 0.044 0.113 0.021
(continued)
367
368
FFS/A 0.16 0.04 0.01 0.05 0.05 0.17 0.06 0.08 0.08 0.08 0.03 0.10 0.01 0.04 0.09 0.00
p-value 0.037 0.294 0.471 0.177 0.148 0.000 0.127 0.081 0.070 0.077 0.287 0.048 0.461 0.279 0.091 0.485
FFP/A 0.04 0.04 0.15 0.32 0.33 0.43 0.39 0.35 0.32 0.29 0.34 0.12 0.65 1.00 0.99 1.00
p-value 0.346 0.289 0.024 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.023 0.000 0.000 0.000 0.000
Cash/STPF 0.04 0.03 0.04 0.10 0.10 0.10 0.13 0.11 0.10 0.09 0.13 0.08 0.20 0.31 0.24 0.31
p-value 0.431 0.434 0.374 0.266 0.241 0.196 0.315 0.240 0.275 0.261 0.288 0.250 0.277 0.266 0.259 0.295
Gap 0.07 0.07 0.08 0.05 0.08 0.03 0.11 0.05 0.06 0.02 0.07 0.03 0.05 0.06 0.09 0.10
p-value 0.207 0.019 0.015 0.014 0.016 0.255 0.021 0.163 0.141 0.352 0.117 0.326 0.203 0.185 0.092 0.054
Panel B: GMM regressions
2008 2007 2006 2005 2004 2003 2002 2001 2000 1999 1998 1997 1996 1995 1994 1993
STPF/A 0.091 0.015 0.038 0.045 0.044 0.024 0.028 0.031 0.040 0.046 0.029 0.019 0.026 0.011 0.021 0.011
p-value 0.004 0.016 0.018 0.000 0.000 0.000 0.000 0.000 0.001 0.000 0.000 0.002 0.003 0.060 0.016 0.073
Cash/A 0.001 0.001 0.012 0.004 0.005 0.007 0.005 0.006 0.002 0.004 0.005 0.018 0.023 0.031 0.004 0.002
p-value 0.487 0.397 0.046 0.409 0.224 0.131 0.272 0.121 0.405 0.218 0.252 0.015 0.001 0.000 0.300 0.441
(continued)
369
370
p-value 0.319 0.227 0.195 0.140 0.129 0.225 0.168 0.201 0.208 0.306 0.202 0.288 0.240 0.225 0.176 0.175
IR Der 0.14 0.16 0.20 0.25 0.27 0.27 0.28 0.27 0.26 0.26 0.27 0.25 0.25 0.27 na na
p-value 0.256 0.276 0.179 0.119 0.087 0.103 0.115 0.111 0.121 0.149 0.152 0.149 0.158 0.143 na na
FX Der 0.16 0.48 0.56 0.59 0.60 0.60 0.61 0.62 0.60 0.59 0.58 0.59 0.62 0.63 na na
p-value 0.085 0.103 0.093 0.065 0.045 0.039 0.043 0.045 0.046 0.053 0.059 0.060 0.061 0.060 na na
Eq Der 0.71 0.35 0.38 0.42 0.43 0.44 0.44 0.44 0.43 0.42 0.43 0.43 na na na na
p-value 0.000 0.122 0.123 0.103 0.080 0.086 0.094 0.092 0.099 0.119 0.123 0.122 na na na na
Cmd Der 0.89 0.45 0.31 0.31 0.34 0.35 0.36 0.36 0.35 0.35 0.35 0.35 na na na na
p-value 0.000 0.109 0.135 0.110 0.083 0.073 0.079 0.082 0.084 0.095 0.106 0.109 na na na na
OBS/E 0.26 0.43 0.49 0.49 0.48 0.44 0.47 0.48 0.47 0.40 0.42 0.21 0.33 0.30 0.31 0.33
p-value 0.019 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000 0.000
Der/A 0.39 0.54 0.44 0.44 0.45 0.45 0.45 0.46 0.46 0.44 0.42 0.38 0.26 0.12 0.07 0.06
p-value 0.132 0.167 0.145 0.107 0.085 0.105 0.100 0.106 0.111 0.144 0.129 0.146 0.123 0.126 0.130 0.257
(continued)
371
372
p-value
Investment M/B 0.20 0.21 0.16 0.15 0.16 0.14 0.14 0.17 0.13 0.16 0.14 0.13 0.16 0.13 0.13 n.a
p-value 0.059 0.070 0.118 0.070 0.059 0.052 0.050 0.050 0.051 0.041 0.057 0.063 0.041 0.038 0.059 n.a
Panel B: GMM regressions
2008 2007 2006 2005 2004 2003 2002 2001 2000 1999 1998 1997 1996 1995 1994 1993
Trading assets 0.028 0.001 0.023 0.007 0.007 0.002 0.005 0.003 0.009 0.006 0.009 0.004 0.014 0.010 0.009 0.003
p-value 0.062 0.399 0.049 0.003 0.020 0.173 0.056 0.104 0.040 0.075 0.067 0.047 0.077 0.077 0.074 0.252
Trading A/L 16.484 0.227 0.970 1.365 1.467 3.105 0.992 0.925 0.752 1.869 0.169 2.241 1.526 1.025 0.366 na
p-value 0.000 0.497 0.388 0.183 0.188 0.148 0.230 0.146 0.285 0.167 0.324 0.147 0.128 0.057 0.391 na
Investment M/B 0.002 0.001 0.006 0.002 0.009 0.007 0.010 0.012 0.026 0.019 0.009 0.001 0.001 0.006 0.010 na
p-value 0.066 0.030 0.106 0.046 0.014 0.000 0.001 0.000 0.035 0.001 0.083 0.042 0.098 0.055 0.009 na
Please see the description of Table 13.5
373
Table 13.10 Performance
374
The ratio of the provision for loan loss to total loans is an ambiguous measure.
A high provision could indicate bank management is expecting high loan losses. On
the other hand, a high ratio could indicate conservative bank management that is taking
no chances on an underfunded allowance for loan loss. When we look at the provision
as a percentage of total equity, we again have the same ambiguous results possible. In
general, we find that both of these risk measures produce a negative coefficient on our
measure of capital. The ambiguity seems to be resolved in that as the provision for loan
losses increases, so does risk. The alternative explanation is that risk should decrease
with this ratio, but higher capital levels produce counterintuitive results. The ratio of
the allowance for loan loss to total loans moves inversely with capital while the ratio
of allowance for loan loss to total equity moves in the opposite direction.
We turn now to the allowance for loan losses. Like the provision for loan loss
ratios, we have the same ambiguous expectations, but now we find conflicting
results. There is no clear expectation for the impact of this ratio on risk. In other
words, both the ratio of the allowance for loan losses to total loans and the ratio of
the allowance for loan losses to total equity can be reflecting either high risk or low
risk. What we find is that the allowance for loan loss as a percentage of loans
produces a positive sign for the coefficient on capital, while the allowance for loan
loss as a percentage of equity produces a negative sign on the coefficient.
We find even more consistent counterintuitive results when we look at the
relationship between capital and liquidity risk in Table 13.7. Lower capital ratios are
generally related to higher levels of short-term purchased liabilities (see STPF/A
and FFP/A). Since short-term purchased money is more volatile than core deposits,
for example, we would expect high levels of purchased money to be associated with
high levels of capital, yet this is not what we find. On the other hand, we do find that
higher capital ratios are related to more liquid assets (HLA/A) and better coverage
of short-term liabilities (FFS/A). Since both of these ratios imply higher levels of
liquidity, we expected them to be related to lower levels of capital. Apparently
liquidity risk is not reflected in a BHC’s capital level.
When we look at a BHC’s exposure to interest rate risk, we again find counter-
intuitive results. As noted above, Table 13.7 shows that low capital ratios are related
to high levels of short-term purchased funds. This can result in a fundamental
liquidity problem if some markets for short-term borrowing completely dry up as
we have observed in the recent financial crisis. Our more direct measure of interest
rate risk is the interest-sensitive gap (gap) which we define as interest-sensitive
assets minus interest-sensitive liabilities divided by total assets. Here we find
ambiguous results. It is obvious that wider gaps expose banks to more risk if
interest rates move against the bank. However, wide gaps can be held in both
a positive and negative direction. A high positive gap indicates a BHC has a large
amount of interest-sensitive assets in relation to interest-sensitive liabilities and will
be hurt by falling interest rates. A high negative gap indicates a BHC has a large
amount of interest-sensitive liabilities in relation to interest-sensitive assets and will
be hurt if interest rates rise. Our results in Table 13.7 indicate that wider gaps are
associated with lower levels of capital, but this is only a true measure if BHCs
typically held a positive gap.
376 W.-J.P. Chiou and R.L. Porter
13.6 Conclusions
In this study, we thoroughly analyze a large cross section of bank holding company
data from 1993 to 2008 to determine the relationship between capital and bank risk-
taking. Our sample includes a minimum of almost 700 BHCs in 2008 and
a maximum of about 1,500 BHCs in 1993. This produces nearly 25,000
company-year observations of BHCs starting with the year that risk-based capital
requirements were first in place. Our data cover a period containing significant
changes in the banking industry and varying levels of economic activity. The
Riegle-Neal and Gramm-Leach-Bliley acts were passed during this time period,
and the tech-stock and housing bubbles both burst with attendant recessions. By
including a larger size range of BHCs in our analysis over a long sample period, our
results are applicable to relatively small BHCs as well as to the largest 200 or so
BHCs traditionally included in empirical studies.
13 Does Banking Capital Reduce Risk? 377
where PTIi is the pretax income of the bank i, i ¼ 1,..N, TEi denotes the technical
efficiency defined as the ratio of observed output to maximum feasible output, xi is
a vector of J inputs used by the bank i, f(xi, b) is the frontier, and b is a vector of
technology parameters to be estimated. Since the frontier provides an estimate of
the maximum feasible output, we then can measure the shortfall of the observed
output from the maximum feasible output. Considering a stochastic component that
describes random shocks affecting the production process in this model, the sto-
chastic frontier becomes
The shock, ℯvi , is not directly attributable to the bank or the technology but may
come from random white noises in the economy, which is considered as a two-sided
Gaussian distributed variable.
We further describe TEi as a stochastic variable with a specific distribution
function. Specifically,
X
H
PTI i ¼ a þ bh lnxi, h þ vi ui : (13.9)
h¼1
Because both vi and ui constitute a compound error term with a specific distri-
bution to be determined, hence the SFA is often referred as composed error model.
In our study, we use the above stochastic frontier with different inputs to
generate the net effect of bank capital without mixing the impact of risk. The
unrestricted model (without including bank equity) is
where BVA is the natural logarithm of book value of assets, x is statistical noise, B is
systematic shortfall (under management control), and B 0. Our restricted model is
as follows:
13 Does Banking Capital Reduce Risk? 379
where BVC is the natural logarithm of book value of capital, v is statistical noise,
and u denotes the inefficiency of a bank considering its use of both assets and
capital. The difference in the inefficiency between the restricted and unrestricted
model,
d ¼ u B, (13.12)
is our instrumental variable. The instrumental variable for capital can be used in
regressions of various measures of risk, as the dependent variable, on our instru-
ment for capital, as the independent variable, while controlling for BHC size.
The notation denotes the Kronecker product of the two vectors. Therefore,
ft (y) is a vector containing the cross product of each instrument in h with each
element of «. The expected value of this cross product is a vector with NeNΗ
elements of zeros at the parameter vector:
E½f t ðy0 Þ
¼ 0: (13.14)
Since we do not observe the true expected values of f, thus we must work instead
with the sample mean of f,
X
T X
T
0
gt ðyÞ T 1 f t ðyÞ ¼ T 1 ht «t ðyÞ ¼ T 1 H «t ðyÞ: (13.15)
t¼1 t¼1
380 W.-J.P. Chiou and R.L. Porter
0
QT ðyÞ gT ðyÞ WT gT ðyÞ, (13.16)
0
DT y^T WT gT y^T ¼ 0, (13.17)
0
DT ðyT Þ ¼ ∂gT ðyT Þ=∂y :
Note the above problem is nonlinear; thus, the optimization must be solved
numerically.
Applying the asymptotic distribution theory, the coefficient estimate y^T is
pffiffiffi d
T y^T y0 ! N ð0; OÞ, (13.18)
References
Berger, A. N., Herring, R. J., & Szego, G. P. (1995). The role of capital in financial institutions.
Journal of Banking and Finance, 24, 1383–1398.
Berger, A. N., DeYoung, R., Flannery, M. J., Lee, D., & Oztekin, O. (2008). How do large
banking organizations manage their capital ratios? Journal of Financial Services Research,
343, 123–149.
13 Does Banking Capital Reduce Risk? 381
Buser, S. A., Chen, A. H., & Kane, E. J. (1981). Federal deposit insurance, regulatory policy, and
optimal bank capital. Journal of Finance, 36, 51–60.
Calomiris, C. W. (1999). Building an incentive-compatible safety net. Journal of Banking and
Finance, 23, 1499–1519.
Calomiris, C. W., & Kahn, C. M. (1991). The role of demandable debt in structuring optimal
banking arrangements. American Economic Review, 81, 497–513.
Campbell, J., Lo, A., & MacKinlay, A. C. (1997). The econometrics of financial markets.
Princeton: Princeton University Press.
Diamond, D. W., & Dybvig, P. H. (1983). Bank runs, deposit insurance, and liquidity. Journal of
Political Economy, 91, 401–419.
Diamond, D. W., & Rajan, R. (2000). A theory of bank capital. Journal of Finance, 55,
2431–2465.
Duan, J., Moreau, A. F., & Sealey, C. W. (1992). Fixed rate deposit insurance and risk-shifting
behavior at commercial banks. Journal of Banking and Finance, 16, 715–742.
Esty, B. C. (1998). The impact of contingent liability on commercial bank risk taking. Journal of
Financial Economics, 47, 189–218.
Flannery, M. J., & Rangan, K. P. (2008). What caused the bank capital build-up of the 1990s?
Review of Finance, 12, 391–430.
Gatev, E., Schuermann, T., & Strahan, P. E. (2009). Managing bank liquidity risk: How
deposit-loan synergies vary with market conditions. Review of Financial Studies, 22,
995–1020.
Gorton, G., & Pennacchi, G. (1992). Financial innovation and the provision of liquidity services.
In J. R. Barth & R. Dan Brumbaugh (Eds.), Reform of federal deposit insurance. New York:
Harper Collins.
Hamilton, J. D. (1994). Time series analysis. Princeton: Princeton University Press.
Hansen, L. P. (1982). Large sample properties of Generalized Method of Moments estimators,
Econometrica, 50, 1029–1054.
Hovakimian, A., & Kane, E. J. (2000). Effectiveness of capital regulation at U.S. commercial
banks: 1985 to 1994. Journal of Finance, 55, 451–468.
Hughes, J. P., Mester, L. J., & Moon, C. (2001). Are scale economies in banking elusive or
illusive? Incorporating capital structure and risk into models of bank production. Journal of
Banking and Finance, 25, 2169–2208.
Hughes, J. P., Lang, W. W., Mester, L. J., Moon, C., & Pagano, M. (2003). Do bankers sacrifice
value to build empires? Managerial incentives, industry consolidation, and financial perfor-
mance. Journal of Banking and Finance, 27, 417–447.
John, K., Saunders, A., & Senbet, L. (2000). A theory of bank regulation and management
compensation. Review of Financial Studies, 13, 95–125.
Jondrow, J., Lovell, C. A., Materov, I. S., & Schmidt, P. (1982). On the estimation of technical
inefficiency in the stochastic frontier production function model. Journal of Econometrics, 19,
233–238.
Keeley, M. C. (1990). Deposit insurance, risk, and market power in banking. American Economic
Review, 80, 1183–1200.
Keeley, M. C. (1992). Bank capital regulation in the 1980s: Effective or ineffective? In
S. Anthony, G. F. Udell, & L. J. White (Eds.), Bank management and regulation: A book of
readings. Mountain View: Mayfield Publishing Company.
Kim, D., & Santomero, A. M. (1988). Risk in banking and capital regulation. Journal of Finance,
42, 1219–1233.
Laeven, L. A., & Levine, R. (2009). Bank governance, regulation and risk taking. Journal of
Financial Economics, 93, 259–275.
Marcus, A. J. (1984). Deregulation and bank financial policy. Journal of Banking and Finance, 8,
557–565.
382 W.-J.P. Chiou and R.L. Porter
Marcus, A. J., & Shaked, I. (1984). The valuation of FDIC deposit insurance using option-pricing
estimates. Journal of Money, Credit and Banking, 16, 446–460.
Merton, R. C. (1977). An analytic derivation of the cost of deposit insurance and loan guarantees.
Journal of Banking and Finance, 1, 3–11.
Prescott, E. S. (1997). The pre-commitment approach in a model of regulatory banking capital.
FRB of Richmond Economic Quarterly, 83, 23–50.
Santos, J. A. (2001). Bank capital regulation in contemporary banking theory: A review of the
literature. Financial Markets, Institutions and Instruments, 10, 41–84.
Shrieves, E., & Dahl, D. (1992). The relationship between risk and capital in commercial banks.
Journal of Banking and Finance, 16, 439–457.
VanHoose, D. (2007). Theories of bank behavior under capital regulation. Journal of Banking and
Finance, 31, 3680–3697.
Evaluating Long-Horizon Event Study
Methodology 14
James S. Ang and Shaojun Zhang
Contents
14.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 384
14.2 Fundamental Issues in Long-Horizon Event Studies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 386
14.2.1 The Buy-and-Hold Benchmark Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 386
14.2.2 The Calendar-Time Portfolio Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 389
14.3 A Review of Simulation Studies on Long-Horizon Event Study Methodology . . . . . . . 390
14.4 A Simulation Study of Large-Size Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 394
14.4.1 Research Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 394
14.4.2 Simulation Results for the Buy-and-Hold Benchmark Approach . . . . . . . . . . . . 396
14.4.3 Simulation Results for the Calendar-Time Portfolio Approach . . . . . . . . . . . . . . 402
14.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 404
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 405
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 409
Abstract
We describe the fundamental issues that long-horizon event studies face in
choosing the proper research methodology and summarize findings from
existing simulation studies about the performance of commonly used methods.
We document in details how to implement a simulation study and report our own
findings on large-size samples. The findings have important implications for
future research.
We examine the performance of more than 20 different testing procedures
that fall into two categories. First, the buy-and-hold benchmark approach uses
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 383
DOI 10.1007/978-1-4614-7750-1_14,
# Springer Science+Business Media New York 2015
384 J.S. Ang and S. Zhang
a benchmark to measure the abnormal buy-and-hold return for every event firm
and tests the null hypothesis that the average abnormal return is zero. Second, the
calendar-time portfolio approach forms a portfolio in each calendar month
consisting of firms that have had an event within a certain time period prior to
the month and tests the null hypothesis that the intercept is zero in the regression
of monthly portfolio returns against the factors in an asset-pricing model. We
find that using the sign test and the single most correlated firm being the
benchmark provides the best overall performance for various sample sizes and
long horizons. In addition, the Fama-French three-factor model performs better
in our simulation study than the four-factor model, as the latter leads to serious
over-rejection of the null hypothesis.
We evaluate the performance of bootstrapped Johnson’s skewness-adjusted
t-test. This computation-intensive procedure is considered because the distribution
of long-horizon abnormal returns tends to be highly skewed to the right. The
bootstrapping method uses repeated random sampling to measure the significance
of relevant test statistics. Due to the nature of random sampling, the resultant
measurement of significance varies each time such a procedure is used. We also
evaluate simple nonparametric tests, such as the Wilcoxon signed-rank test or the
Fisher’s sign test, which are free from random sampling variation.
Keywords
Long-horizon event study • Johnson’s skewness-adjusted t-test • Weighted
least squares regression • Bootstrap test • Calendar-time portfolio approach •
Fama-French three-factor model • Johnson’s skewness-adjusted t-statistic •
Large-scale simulations
14.1 Introduction
A large number of papers in finance literature have documented evidence that firms
earn abnormal returns over a long time period (ranging from 1 to 5 years) after
certain corporate events. Kothari and Warner (2007) report that a total of 565 papers
reporting event study results were published between 1974 and 2000 in five leading
journals: the Journal of Business (JB), Journal of Finance (JF), Journal of Finan-
cial Economics (JFE), Journal of Financial and Quantitative Analysis (JFQA), and
the Review of Financial Studies (RFS). Approximately 200 of the 565 event studies
use a maximum window length of 12 months or more.
The evidence of long-horizon abnormal returns contradicts the efficient market
hypothesis that stock prices adjust to information fully within a narrow time
window (a few days). To reconcile the contradiction, Fama (1998) argues that
“Most important, consistent with the market efficiency prediction that apparent
anomalies can be due to methodology, most long-term return anomalies tend to
disappear with reasonable changes in technique.” Several simulation studies such as
Kothari and Warner (1997) and Barber and Lyon (1997) document evidence that
statistical inference in long-horizon event studies is sensitive to the choice of
14 Evaluating Long-Horizon Event Study Methodology 385
Buy-and-hold
Calendar-Time
Benchmark
Portfolio Approach
Approach
Model Estimation
Specification, e.g., Technique, e.g., Parametric Nonparametric Parametric Nonparametric
Fama and French’s Tests, e.g., Tests, e.g., Tests, e.g., Tests, e.g.,
Three-Factor Model, OLS, t-test, Bootstrapped test, t-test, Bootstrapped test
Carhart’s Four- WLS Johnson’s test Sign test Johnson’s test Sign test
Factor Model
Fig 14.1 Overview of the two approaches to choose a methodology for long-horizon event study
where Ri is the long-term buy-and-hold return of firm i and BRi is the long-term
return on a particular benchmark of firm i. The buy-and-hold return of firm i over t
months is obtained by compounding monthly returns, that is,
Yt
Ri ¼ t¼1
ð1 þ rit Þ 1, (14.2)
where rit is firm i’s return in month t. Calculation of the benchmark return BRi is
given below. The benchmark return, BRi, estimates the return that an event firm
would have had if the event had not happened.
Several articles clearly show that long-term abnormal returns are very sensitive to
choice of benchmarks; see, e.g., Ikenberry et al. (1995), Kothari and Warner (1997),
Barber and Lyon (1997), and Lyon et al. (1999). If wrong benchmarks were used in
measuring long-term abnormal returns, inference on the significance of a certain event
14 Evaluating Long-Horizon Event Study Methodology 387
would be erroneous. Most existing studies use either a single matched firm or a matched
reference portfolio as the benchmark. Barber and Lyon (1997) point out that the control
firm approach eliminates the new listing bias, the rebalancing bias, and the skewness
problem. It also yields well-specified test statistics in virtually all the situations they
consider. Further, Lyon et al. (1999) advocate a reference portfolio of firms that match
on size and BE/ME. The issue on choice of the benchmark is practically unresolved.
Ang and Zhang (2004) additionally argue that the control firm method overcomes
another important problem that is associated with the event firm not being representative
in important aspects of the respective matched portfolio in the reference portfolio
approach. This leads to the matched portfolio return generating a biased estimate of
expected firm return. This problem is particularly severe with small firms.
A common practice in computing an event firm’s long-term abnormal return is to
utilize a benchmark that matches the event firm on size and BE/ME. The practice is
often justified by quoting the findings in Fama and French (1992) that size and
BE/ME combine to capture the cross-sectional variation in average monthly stock
returns and that market beta has no additional power in explaining cross-sectional
return differences. However, in a separate paper, Fama and French (1993)
demonstrate that expected monthly stock returns are related to three factors:
a market factor, a size-related factor, and a book-to-market equity ratio
(BE/ME)-related factor. To resolve this issue, Ang and Zhang (2004) show that
matching based on beta in addition to size and BE/ME does not improve the
performance of the approach.
A recent trend is to use computation-intensive bootstrapping-based tests, such as
the bootstrapped Johnson’s skewness-adjusted t-statistic (e.g., Sutton 1993 and Lyon
et al. 1999) and the simulated empirical p-values (e.g., Brock et al. 1992 and
Ikenberry et al. 1995). These procedures rely on repeated random sampling to
measure the significance of relevant test statistics. Due to the nature of random
sampling, the resultant measurement of significance varies every time such
a procedure is used. As a consequence, different researchers could reach contradic-
tory conclusions using the same procedure on the same sample of event firms. In
contrast, simple nonparametric tests, such as the Wilcoxon signed-rank test or the
Fisher’s sign test, are free from random sampling variation. Barber and Lyon (1997)
examined the performance of the Wilcoxon signed-rank test in a large-scale simula-
tion study. They show that the performance depends on choice of the benchmark. The
signed-rank test is well specified when the benchmark is a single size and BE/ME
matched firm and misspecified when the benchmark is a size and BE/ME matched
reference portfolio. However, Barber and Lyon (1997) present only simulation
results for 1-year horizon. No simulation study in the finance literature has examined
the performance of these simple nonparametric tests for 3- or 5-year horizons, which
are the common holding periods in long-horizon event studies.1
1
The sign test has an advantage over the signed-rank test in that it does not require a symmetric
underlying distribution, while the signed-rank test does.
388 J.S. Ang and S. Zhang
2
Variations of this approach have been used extensively; see, e.g., Ritter (1991); Ikenberry
et al. (1995); Ikenberry et al. (1996); and Desai and Jain (1997), among many others.
14 Evaluating Long-Horizon Event Study Methodology 389
event firms’ returns. The bad model problem arises because no benchmark gives
perfect estimate of the counterfactual (i.e., what if there was no event) return of an
event firm and benchmark errors are multiplied in computing long-term buy-and-
hold returns. Therefore, Fama (1998) advocates a calendar-time portfolio approach.3
In the calendar-time portfolio approach, for each calendar month, an event portfolio
is formed, consisting of all firms that have experienced the same event within the t
months prior to the given month. Monthly return of the event portfolio is computed
as the equally weighted average of monthly returns of all firms in the portfolio.
Excess returns of the event portfolio are regressed on the Fama-French three factors
as in the following model:
Rpt Rft ¼ a þ b Rmt Rft þ sSMBt þ hHMLt þ et , (14.3)
where Rpt is the event portfolio’s return in month t; Rft is the 1-month Treasury bill rate,
observed at the beginning of the month; Rmt is the monthly market return; SMBt is the
monthly return on the zero investment portfolio for the common size factor in stock
returns; and HMLt is the monthly return on the zero investment portfolio for the common
book-to-market equity factor in stock returns.4 Under the assumption that the Fama-
French three-factor model provides a complete description of expected stock returns, the
intercept, a, measures the average monthly abnormal return on the portfolio of event
firms and should be equal to zero under the null hypothesis of no abnormal performance.
A later modification that has gained popularity is the four-factor model that
added a momentum-related factor to the Fama-French three factors:
Rpt Rft ¼ a þ b Rmt Rft þ sSMBt þ hHMLt þ pPR12t þ et , (14.4)
3
Loughran and Ritter (1995), Brav and Gompers (1996), and Brav et al. (2000), among others,
have used the calendar-time portfolio approach.
4
See Fama and French (1993) for details on construction of the mimicking portfolios for the
common size and book-to-market equity factors. We thank Eugene Fama for providing us with
returns on Rft, Rmt, SMBt, and HMLt.
390 J.S. Ang and S. Zhang
time series conforms to the asset-pricing model, the event is said to have had no
significant long-term effect; otherwise, the event has produced significant
long-term abnormal returns. Lyon et al. (1999) report that the calendar-time
portfolio approach together with the Fama-French three-factor model, which shall
be referred to as the Fama-French calendar-time approach later, is well specified for
random samples in their simulation study.
However, we do not know how much power the Fama-French calendar-time
approach has. Loughran and Ritter (1999) criticize the approach as having very low
power. They argue that reduction in power is caused by using returns on contam-
inated portfolios as factors in the regression, by weighting each month equally and
by using value-weighted returns of the calendar-time portfolios. However, their
empirical evidence is based only on one carefully constructed sample of firms and is
hardly conclusive. No large-scale simulation study has been done to examine power
of the Fama-French calendar-time approach, which we will remedy in this paper.
The Fama-French calendar-time approach, estimated with the ordinary least
squares (OLS) technique, could suffer from a potential heteroskedasticity problem
due to unequal and changing number of firms in the calendar-time portfolios. The
weighted least squares (WLS) technique, which is helpful in addressing the
heteroskedasticity problem, has been suggested as a way to deal with the changing
size of calendar-time portfolios. When applying WLS, we use the monthly number
of firms in the event portfolio as weights.
(2000)
Cowan and Size, BE/ME t-test, two-group test with 1,000 simulated samples 1, 3, and 3 years
Sergeant winsorized data with sample size of 5 years
(2001) 50, 200, and 1,000
Ang and Three-factor OLS, Size, BE/ME, t-test, 250 simulated samples 1, 3, and 1, 3, and 5 years
Zhang (2004) model, four- WLS beta, and Johnson’s test, with sample size of 5 years
factor model correlation bootstrapped test, sign 200 and 1,000
coefficient test
Jegadeesh and Size, BE/ME t-test, t-test adjusted for 1,000 simulated samples 1, 3, and 1, 3, and 5 years
Karceski heteroskedasticity and of 200 firms each 5 years
(2009) serial correlation
391
392 J.S. Ang and S. Zhang
In their simulation study, Kothari and Warner (1997) measure the long-term
(up to 3 years) impact of an event by cumulative monthly abnormal returns, where
monthly abnormal returns are computed against four common models: the market-
adjusted model, the market model, the capital asset-pricing model, and the Fama-
French three-factor model. They find that tests for cumulative abnormal returns are
severely misspecified. They identify sample selection, survival bias, and bias in
variance estimation as potential sources of the misspecification and suggest that
nonparametric and bootstrap tests are likely to reduce misspecification.
Barber and Lyon (1997) address two main issues in their simulation study. First,
they argue that buy-and-hold return is a better measure of investors’ actual experience
over a long horizon and should be used in long-term event study (up to 5 years). They
show simulation evidence that approaches using cumulative abnormal returns cause
severe misspecification, which is consistent with the observation in Kothari and
Warner (1997). Second, they use simulations to measure both size and power of
testing procedures that follow the buy-and-hold benchmark approach. An important
finding is that using a single control firm as benchmark yields well-specified tests,
whereas using reference portfolio causes substantial over-rejection.
In a later paper, Lyon et al. (1999) report another simulation study (for up to the
5-year horizon) that investigates the performance of both buy-and-hold benchmark
approach and calendar-time portfolio approach. They find that using the
Fama-French three-factor model yields a well-specified test. However, they advo-
cate a test that uses carefully constructed reference portfolio as benchmark and the
bootstrapped Johnson’s statistic for testing abnormal returns. They present
evidence that this test is well specified and has high power at the 1-year horizon.
Two questions remain unanswered in Lyon et al. (1999). First, how much power
does the bootstrap test have for event horizons longer than 1 year (e.g., 3 or 5 years
that is common in long-horizon studies)? It is known in statistics literature that
a bootstrap test is not as powerful as simple nonparametric tests in many occasions
(see Efron and Tibshirani 1993, Chap. 16 and Davison and Hinckley 1997,
Chap. 4). It is necessary to know the actual power of such test for event horizons
beyond 1 year. Second, is the calendar-time portfolio approach as powerful as the
buy-and-hold benchmark approach? Loughran and Ritter (2000) argue that the
calendar-time portfolio approach has low power, using simulations and empirical
evidence from a sample of new equity issuers. However, they do not measure how
much power the approach actually has, which makes it impossible to compare the
two approaches directly in more general settings.
Mitchell and Stafford (2000) is the only study that empirically measures power
of the calendar-time portfolio approach using simulations. Their main focus is to
assess performance of several testing procedures in three large samples of major
managerial decisions, i.e., mergers, seasoned equity offerings, and share
repurchases (up to 3 years). They find that different procedures lead to contradicting
conclusions and argue that the calendar-time portfolio approach is preferred. To
resolve Loughran and Ritter’s (2000) critique that the calendar-time portfolio
approach has low power, they conduct simulations to measure the empirical
power and find that the power is actually very high with an empirical rejection
14 Evaluating Long-Horizon Event Study Methodology 393
rate of 99 % for induced abnormal returns of 15 % over a 3-year horizon. Since they
have a large sample size, this finding is actually consistent with what we document in
Table 14.5. However, their simulations focus on only samples of 2,000 firms. Many
event studies have much smaller sample sizes, especially after researchers slice and
dice a whole sample into subsamples. More evidence is needed in order to have great
confidence in applying the calendar-time portfolio approach in such studies.
Cowan and Sergeant (2001) focus on the buy-and-hold benchmark approach in their
simulations. They find that using the reference portfolio approach cannot overcome the
skewness bias discussed in Barber and Lyon (1997) and that the larger the sample size,
the smaller the magnitude of the skewness bias. They also argue that cross-sectional
dependence among event firms’ abnormal returns increases in event horizon due to
partially contemporaneous holding periods, which may cause the overlapping horizon
bias. They propose a two-group test using abnormal returns winsorized at three
standard deviations to deal with these two biases and report evidence that this test
yields correct specifications and considerable power in many situations.
All previous simulation studies use only size and BE/ME to construct
benchmarks, which is often justified by the findings in Fama and French (1992)
that size and BE/ME together adequately capture the cross-sectional variations in
average monthly stock returns. Ang and Zhang (2004) use two other matching
criteria to explore whether better benchmarks could be used for future studies. The
two criteria are market beta and pre-event correlation coefficient. Using market beta
is motivated by the fact that Fama and French’s (1993) three-factor model has
a market factor, a size-related factor, and a BE/ME-related factor. Matching on the
basis of size and BE/ME does not account for the influence of the market factor. The
rationale for using pre-event correlation coefficient is that matching on size and
BE/ME may fail to control for other factors that could influence stock returns, such as
industry factor, seasonal factor, momentum factor, and other factors shared by only
firms of the same characteristics, such as geographical location, ownership, and
governance structures. Matching on the basis of pre-event correlation coefficient
helps remove the effect of these factors on the event firm’s long-term return.
The main findings in Ang and Zhang (2004) include the following. First, the four-
factor model is inferior to the well-specified three-factor model in the calendar-time
portfolio approach in that the former causes too many rejections of the null hypothesis
relative to the specified significance level. Second, WLS improves the performance of
the calendar-time portfolio approach over OLS, especially for long event horizons.
Third, the Fama-French three-factor model has relatively high power in detecting
abnormal returns, although power decreases sharply as event horizon increases. Fourth,
the simple sign test is well specified when it is applied with a single firm benchmark,
but misspecified when used with reference portfolio benchmarks. More importantly,
the combination of the sign test and the benchmark with the single most correlated firm
consistently has much higher power than any other test in our simulations and is the
only testing procedure that performs well in samples of small firms.
Jegadeesh and Karceski (2009) propose a new test of long-run performance that
allows for heteroskedasticity and autocorrelation. Previous tests used in Lyon
et al. (1999) implicitly assume that the observations are cross-sectionally uncorrelated.
394 J.S. Ang and S. Zhang
In this simulation study, we construct 250 samples each consisting of 1,000 event
firms. To produce one sample, we randomly select, with replacement, 1,000 event
months between January 1980 and December 1992, inclusively.5,6 This allows us to
calculate 5-year abnormal returns until December 1997. For each selected event
month, we randomly select, without replacement, one firm from a list of qualified
firms. The qualified firms satisfy the following requirements: (i) They are publicly
traded firms, incorporated in the USA, and have ordinary common shares with
Center for Research in Security Prices (CRSP) share codes 10 and 11; (ii) they have
return data found in the CRSP monthly returns database for the 24-month period
5
We use a pseudorandom number generator developed by Matsumoto and Nishimura (1998) to
ensure high quality of random sampling.
6
Kothari and Warner (1997) use 250 samples, each of 200 event months between January 1980 and
December 1989 inclusively. Barber and Lyon (1997) use 1,000 samples, each of 200 event months
in a much longer period from July 1963 to December 1994. The period under our study, between
January 1980 and December 1992, is of similar length to Kothari and Warner’s.
14 Evaluating Long-Horizon Event Study Methodology 395
prior to the event month; (iii) they have nonnegative book values on COMPUSTAT
prior to the event month so that we can calculate their book-to-market equity ratios.
The 250 samples, each of 1,000 randomly selected firms, comprise the
simulation setting for comparing the performance of different testing procedures.7
We apply all testing procedures under our study to the same samples. Such
controlled comparison is more informative because it eliminates difference in
performance due to variation in the samples.
For the buy-and-hold approach, we compute the long-term buy-and-hold abnor-
mal return of firm i as the difference between the long-term buy-and-hold return of
firm i and the long-term return of a benchmark. The buy-and-hold return of firm
i over t months is obtained by compounding monthly returns. In case that firm
i does not have return data for all t months, we replace missing returns by the
same-month returns of a size and BE/ME matched reference portfolio.8 We evaluate
a total of five benchmarks and four test statistics in this study. We briefly describe
them in the following and give the details in the Appendix.
Three of the benchmarks are reference portfolios. The first reference portfolio
consists of firms that are similar to the event firm in both size and BE/ME. We
follow the same procedure as in Lyon et al. (1999) to construct the two-factor
reference portfolio. We use the label “SZBM” for this benchmark. The second
reference portfolio consists of firms that are similar to the event firm not only in size
and BE/ME but also in market beta. We use the label “SZBMBT” for this bench-
mark. The third reference portfolio consists of ten firms that are most correlated
with the event firm prior to the event. We use the label “MC10” for this benchmark.
The other two of the five benchmarks consist of a single firm. The first single
firm benchmark is the firm that matched the event firm in both size and BE/ME. To
find the two-factor single firm benchmark, we first identify all firms whose market
value is within 70–130 % of the event firm’s market value and then choose the firm
that has the BE/ME ratio closest to that of the event firm. We use the label
“SZBM1” for this benchmark. The second single firm benchmark is the firm that
has the highest correlation coefficient with the event firm prior to the event. We use
the label “MC1” for this benchmark.
We apply four test statistics to test the null hypothesis that the mean long-term
abnormal return is zero. They include Student’s t-test, Fisher’s sign test, Johnson’s
skewness-adjusted t-test, and the bootstrapped Johnson’s t-test. Fisher’s sign test is
a nonparametric test and is described in details in Hollander and Wolfe (1999, Chap. 3).
Johnson’s skewness-adjusted t-statistic was developed by Johnson (1978) to deal with
the skewness-related misspecification error in Student’s t-test. Sutton (1992) proposes
to apply Johnson’s t-test with a computationally intensive bootstrap resampling
7
Ang and Zhang (2004) examine two other simulation settings. Under one setting, they have another
250 samples of 200 event firms, a smaller sample size than the setting in this chapter. Under the other
setting, they have the sample size of 200 with the requirement that event firms belong to the smallest
quintile sorted by NYSE firm size. The second setting is used to examine the effect of small firms.
8
Filling in missing returns is a common practice in calculating long-term buy-and-hold returns;
e.g., see Barber and Lyon (1997), Lyon et al. (1999), and Mitchell and Stafford (2000).
396 J.S. Ang and S. Zhang
technique when the population skewness is severe and the sample size is small. Lyon
et al. (1999) advocate use of the bootstrapped Johnson’s t-test because long-term
buy-and-hold abnormal returns are highly skewed when buy-and-hold reference port-
folios are used as benchmarks. We follow Lyon et al. (1999) and set the resampling
size in the bootstrapped Johnson’s t-test to be one quarter of the sample size.
For the Fama-French calendar-time approach, we use both the Fama-French
three-factor model and the four-factor model. We apply both ordinary least
squares (OLS) and weighted least squares (WLS) techniques to estimate parameters
in the pricing model. The WLS is used to correct the heteroskedasticity problem due
to the monthly variation in the number of firms in the calendar-time portfolio. When
applying WLS, we use the number of event firms in the portfolio as weights.
because reference portfolios have lower volatility due to averaging. As for kurtosis, all
five benchmarks produce highly leptokurtic abnormal returns, with kurtosis coeffi-
cients ranging from 41.4 to 67.5, which are far greater than three, the kurtosis
coefficient of any normal distribution. At last, skewness coefficients for the two single
firm benchmarks are close to zero regardless of event horizons, while skewness
coefficients for the three portfolio benchmarks are excessively positive.
To sum up, probability distributions of long-term abnormal returns exhibit
different properties, depending on whether the benchmark is a reference portfolio
or a single firm. Under a reference portfolio benchmark, the distribution is highly
leptokurtic and positively skewed, with a close-to-zero mean but a highly negative
median. Under a single firm benchmark, the distribution is highly leptokurtic but
symmetric, with both mean and median close to zero. Statistical properties of
long-term abnormal returns have important bearings on performance of test
398 J.S. Ang and S. Zhang
statistics. Overall, it seems single firm benchmarks have more desirable properties.
Between the two single firm benchmarks, MC1 shows better performance than
SZBM1, because the abnormal returns based on MC1 have both mean and median
being closer to zero and smaller standard deviation.
A superior test should control for the probability of committing two errors. First,
it is important to control for the probability of misidentifying an insignificant event
as having statistical significance; in other words, the empirical size of the test,
which is computed from simulations, is close to the prespecified significance level
at which the test is conducted. When this happens, the test is well specified. Second,
power of the test should be large, that is, the probability of finding a statistically
significant event if one did exist.
Table 14.3 reports empirical size of all 20 tests for three holding periods.
Empirical size is calculated as the proportion of 250 samples that rejects the null
hypothesis at the 5 % nominal significance level. With only a few exceptions,
Student’s t-test is well specified against the two-sided alternative hypothesis.
Despite excessively high skewness in abnormal returns from reference portfolio
benchmarks, Student’s t-test is well specified against two-sided alternative
hypothesis because the effect of skewness at both tails cancels out (see, e.g.,
Pearson and Please 1975). When testing against the two-sided alternative
hypothesis, Johnson’s skewness-adjusted t-test is in general misspecified, but its
bootstrapped version is well specified in most situations. The sign test is
misspecified when applied to abnormal returns from reference portfolio
benchmarks, and the extent of misspecification is quite serious and increases in
the length of holding period. This is not surprising because abnormal returns from
reference portfolio benchmarks have highly negative medians.
Table 14.4 reports empirical power of testing the null hypothesis of zero abnormal
return against the two-sided alternative hypothesis. We follow Brown and Warner
(1980, 1985) to measure empirical power by intentionally forcing the mean
abnormal return away from zero with induced abnormal returns. We induce nine levels
of abnormal returns ranging from 20 % to 20 % at an increment of 5 %. To induce an
abnormal return of 20 %, for example, we add 20 % to the observed holding period
return of an event firm. Empirical power is calculated as the proportion of 250 samples
that rejects the null hypothesis at 5 % significance level.
With a large sample size of 1,000, the power of these tests remains reasonably
high at the longer holding period. Ang and Zhang (2004) report that, with the sample
size of 200, the power of all tests deteriorates sharply as holding period lengthens
from 1 to 3 and to 5 years and is alarmingly low at the 5-year horizon. For example,
when the induced abnormal return is 20 % over a 5-year horizon, the highest power
of the bootstrapped Johnson’s t-test is 13.6 % for a sample of 200 firms, whereas the
highest power is 62.8 % for a sample of 1,000 firms.
We compare the power of the three test statistics: Student’s t-test, the
bootstrapped Johnson’s skewness-adjusted t-test, and the sign test. All three
test statistics are applied together with the most correlated single firm benchmark.
The evidence shows that all three tests are well specified. However, the sign
test clearly has much higher power than the other two tests.
14
SZBM 11.2* 14.4* 12.8* 99.6* 1.2* 1.2* 0.8* 100.0* 17.6* 22.8* 21.6* 0.0*
SZBMBT 5.2 5.2 5.6 100.0* 7.6 7.6 5.6 100.0* 2.8 3.2 3.6 0.0*
MC10 4.8 6.8 5.6 100.0* 6.8 5.6 4.4 100.0* 3.2 6.0 5.6 0.0*
SZBM1 6.0 7.6 5.2 9.2* 2.0* 2.8 1.6* 0.4* 11.2* 13.6* 10.0* 15.6*
MC1 6.8 8.4* 6.4 6.4 2.8 2.8 2.8 7.6 7.6 8.0* 6.8 1.2*
(continued)
399
400
Table 14.5 reports the rejection frequency of the calendar-time portfolio approach
in testing the null hypothesis that the intercept is zero in the regression of monthly
calendar-time portfolio returns, against the two-sided alternative hypothesis. Rejec-
tion frequency is measured as the proportion of the total 250 samples that reject the
null hypothesis. We compute rejection frequencies at nine nominal levels of
induced abnormal returns, ranging from 20 % to 20 % at an increment of 5 %.
Since monthly returns of the calendar-time portfolio are used in fitting the model, to
examine the power of testing the intercept, we need to induce abnormal returns by
14 Evaluating Long-Horizon Event Study Methodology 403
Table 14.5 Rejection frequency of calendar-time portfolio approach in samples of 1,000 firms
Panel A: 1-year holding period
Average effective induced holding period return (%)
20.4 15.7 10.7 5.5 0 5.7 11.7 17.9 24.4
Three factors OLS 100.0 100.0 99.2 53.2 2.4 78.8 100.0 100.0 100.0
WLS 100.0 100.0 99.6 74.4 2.0* 82.8 100.0 100.0 100.0
Four factors OLS 100.0 99.2 90.8 18.0 28.0* 97.6 100.0 100.0 100.0
WLS 100.0 99.6 93.2 20.8 25.2* 98.8 100.0 100.0 100.0
Panel B: 3-year holding period
Average effective induced holding period return (%)
25.2 19.3 13.2 6.8 0 7.1 14.5 22.3 30.4
Three factors OLS 98.0 86.8 38.0 3.6 2.4 32.0 84.8 99.6 99.6
WLS 100.0 97.2 65.2 10.0 1.2* 36.0 91.6 100.0 100.0
Four factors OLS 69.2 22.0 1.6 6.4 55.2* 94.0 99.6 100.0 100.0
WLS 92.0 38.0 4.0 10.4 75.6* 99.6 100.0 100.0 100.0
Panel C: 5-year holding period
Average effective induced holding period return (%)
31.1 23.9 16.3 8.3 0 8.7 17.9 27.4 37.5
Three factors OLS 64.8 31.2 10.0 0.8 4.0 27.6 62.4 90.8 99.6
WLS 94.4 58.4 14.8 0.4 4.0 36.0 81.2 99.2 100.0
Four factors OLS 12.4 1.6 5.2 32.8 70.8* 89.2 98.8 100.0 100.0
WLS 14.0 1.2 14.8 62.4 94.0* 100.0 100.0 100.0 100.0
This table reports rejection frequency in testing the null hypothesis that the intercept in the
regression of monthly calendar-time portfolio returns is zero, in samples of 1,000 firms. Both
the Fama-French three-factor model and the four-factor model are used in the regression. Model
parameters are estimated with both OLS and WLS estimation techniques. Rejection frequency is
equal to the proportion of 250 samples that reject the null hypothesis at 5 % significance level. We
measure rejection frequency at nine levels of induced abnormal returns. We induce abnormal
returns by adding an extra amount to monthly returns of every event firm before forming the
calendar-time portfolios. The effective induced holding period return of an event firm is equal to
the difference in the firm’s holding period return between before and after adding the monthly
extra amount. The average effective induced holding period return is computed over all event
firms in the 250 samples
adding an extra amount to actual monthly returns of every event firm before
forming the calendar-time portfolios. For example, in order to induce the 20 %
nominal level of abnormal holding period return, we add the extra amount
of 1.67 % (¼20 %/12) to an event firm’s 12 monthly returns for a 1-year
horizon, or add the abnormal amount of 0.56 % (¼ 20 %/36) to the firm’s
24 monthly returns for a 3-year horizon, or the abnormal amount of 0.33 %
(¼20 %/60) to the firm’s 60 monthly returns for a 5-year horizon.
Note that the nominal induced holding period return is different from the
effective induced abnormal holding period return, because adding the abnormal
amount each month does not guarantee that an event firm’s holding period return
will be increased or decreased by the exact nominal level. We measure the effective
induced holding period return of an event firm as the difference in the firm’s holding
404 J.S. Ang and S. Zhang
period return between before and after adding the monthly abnormal amount. The
average effective induced holding period return is computed over all event firms in
the 250 samples. The average induced holding period return allows us to compare
power of the buy-and-hold benchmark approach with that of the calendar-time
portfolio approach at the scale of holding period return.
We first examine empirical size of the calendar-time portfolio approach, which is
equal to the rejection frequency when no abnormal return is induced. In Table 14.5,
the empirical size is in the column with zero induced return. It is very surprising that
when the four-factor model is used, the test has excessively high rejection frequency
at 3-year and 5-year horizons. The rejection frequency, for example, is 94.0 % at the
5-year horizon with the WLS estimation! In contrast, when the Fama-French three-
factor model is used, the empirical sizes are not significantly different from the 5 %
significance level. The evidence strongly suggests that the three-factor model is
preferred for the calendar-time portfolio approach, whereas the four-factor model
suffers from overfitting and should not be used.
Table 14.5 shows that, for a sample of 1,000 firms, the power of this approach
remains high as event horizon increases. WLS estimation does improve the power
of the procedure over the OLS, and the extent of improvement becomes greater as
holding period gets longer. By comparing Tables 14.4 and 14.5, we find that the
power of the Fama-French calendar-time approach implemented with WLS tech-
nique (i.e., FF, WLS) has almost the same power as the buy-and-hold benchmark
approach implemented with the most correlated single firm and the sign test (i.e.,
MC1, sign), at the 1-year horizon, but slightly less at the 3- and 5-year horizons.
14.5 Conclusion
Comparing the simulation results in Sect. 14.4 with those in Ang and Zhang (2004),
we find that sample size has a significant impact on the performance of tests in long-
horizon event studies. With a sample size of 1,000, a few tests perform reasonably
well, including the Fama-French calendar-time approach implemented with WLS
technique and the buy-and-hold benchmark approach implemented with the most
correlated single firm (MC1) and the sign test. In particular, they have reasonably
high power even for the long 5-year holding period. On the contrary, with a sample
size of 200, Ang and Zhang (2004) find that the power of most well-specified tests is
very low for the 5-year horizon, only in the range of 10–20 % against a high level of
induced abnormal returns, while the combination of the most correlated single firm
and the sign test stands out with a power of 41.2 %. Thus, the most correlated single
firm benchmark dominates for most practical sample sizes, and in addition, the
simplicity of the sign test is appealing.
The findings have important implications for future research. For long-horizon
event studies with a large sample, it is likely to be more fruitful to spend efforts on
understanding the characteristics of the sample firms, than on implementing
various sophisticated testing procedures. The simulation results here show that the
commonly used tests following both the Fama-French calendar-time approach and
14 Evaluating Long-Horizon Event Study Methodology 405
Appendix
This appendix includes the details on the benchmarks and the test statistics that are
used in our simulation studies. We use five benchmarks. The first benchmark is
a reference portfolio constructed on the basis of firm size and BE/ME. We follow
Lyon et al. (1999) to form 70 reference portfolios at the end of June in each
year from 1979 to 1997. At the end of June of year t, we calculate the size of
every qualified firm as price per share multiplied by shares outstanding. We sort all
NYSE firms by firm size into ten portfolios, each having the same number of firms,
and then place all AMEX/NASDAQ firms into the ten portfolios based on firm size.
Since a majority of NASDAQ firms are small, approximately 50 % of all firms fall
in the smallest size decile. To obtain portfolios with the same number of firms, we
further partition the smallest size decile into five subportfolios by firm size
without regard to listing exchange. We now have 14 size portfolios. Next, we
calculate each qualified firm’s BE/ME as the ratio of the book equity value
(COMPUSTAT data item 60) of the firm’s fiscal year ending in year t 1 to its
market equity value at the end of December of year t 1. We then divide each of
the 14 portfolios into five subportfolios by BE/ME and conclude the procedure
with 70 reference portfolios on the basis of size and BE/ME.
406 J.S. Ang and S. Zhang
The size and BE/ME matched reference portfolio of an event firm is taken to
be the one of the 70 reference portfolios constructed at the month of June prior to the
event month that matches the event firm in size and BE/ME. The return on a size and
BE/ME matched reference portfolio over t months is calculated as
2 Xnt 3
Y
t1 r
j¼1 jt
BRSZBM ¼ 41 þ 5 1, (14.5)
i
t¼0
nt
where month t ¼ 0 is the event month, nt is the number of firms in month t, and rjt is
the monthly return of firm j in month t. We use the label “SZBM” for the benchmark
that is based on firm size and BE/ME.
The second benchmark is a reference portfolio constructed on the basis of firm size,
BE/ME, and market beta. The Fama-French three-factor model suggests that expected
stock returns are related to three factors: a market factor, a size-related factor, and a BE/
ME-related factor. Reference portfolios constructed on the basis of size and BE/ME
account for the systematic portion of expected stock returns due to the size and BE/ME
factors, but not the portion due to the market factor. Our second benchmark is based on
firm size, BE/ME, and market beta to take all three factors into account.
To build a three-factor reference portfolio for a given event firm, we first
construct the 70 size and BE/ME reference portfolios as above and identify the
one that matches the event firm. Next, we pick firms within the matched portfolio
that have returns in CRSP monthly returns database for all 24 months prior to the
event month and compute their market beta by regressing the 24 monthly returns on
the value-weighted CRSP return index. Lastly, we divide these firms that have
market beta into three portfolios by their rankings in beta and pick the one that
matches the event firm in beta as the three-factor reference portfolio. The return on
a three-factor portfolio over t months is calculated as
2 Xlt 3
Y
t1 r
j¼1 jt
BRSZBMBT ¼ 41 þ 5 1, (14.6)
i
t¼0
nt
where month t ¼ 0 is the event month, nt is the number of firms in month t, and rjt is
the monthly return of firm j in month t. We use the label “SZBMBT” to indicate that
the benchmark is based on firm size, BE/ME, and market beta.
The third benchmark is a reference portfolio constructed on the basis of firm size,
BE/ME, and pre-event correlation coefficient. The rational for using pre-event
correlation coefficient as an additional dimension is that returns of highly correlated
firms are likely to move in tandem in response to not only changes in “global” risk
factors, such as the market factor, the size factor, and the BE/ME factor in the Fama-
French model, but also changes in other “local” factors, such as the industry factor,
the seasonal factor, liquidity factor, and the momentum factor. Over a long time
period following an event, both global and local factors experience changes that
affect stock returns. It is reasonable to expect more correlated stocks would be
14 Evaluating Long-Horizon Event Study Methodology 407
affected by these factors similarly and should have resulting stock return patterns
that are closer to each other. Therefore, returns of a reference portfolio on the basis
of pre-event size, BE/ME, and pre-event correlation coefficient are likely to
be better estimate of the status quo (i.e., what if there was no event) return of
an event firm.
To build a reference portfolio on the basis of size, BE/ME, and pre-event
correlation coefficient, we first construct the same 70 size and BE/ME reference
portfolios as above and identify the combination that matches the event firm. Next,
we pick firms within the matched size and BE/ME reference portfolio that have
returns in CRSP monthly returns database for all 24 months prior to the event month
and compute their correlation coefficients with the event firm over the pre-event
24 months. Lastly, we choose the ten firms that have the highest pre-event
correlation coefficient with the event firm to form the reference portfolio. Return
of the portfolio over t months is calculated as
Yt1
X
10 1 þ rjt 1
BRMC10
i ¼ t¼0
, (14.7)
j¼1
10
where month t ¼ 0 is the event month and rjt is the monthly return of firm j in
month t. We use the label “MC10” to indicate that the benchmark consists of the
most correlated ten firms. The benchmark return is the return of investing equally in
the ten most correlated firms over the t months beginning with the event month.
The benchmark is to be considered as a hybrid between the reference portfolio
discussed above and the matching firm approach shown below.
The fourth benchmark is a single firm matched to the event firm in size and
BE/ME. Barber and Lyon (1997) report that using a size and BE/ME matched firm
as benchmark gives measurements of long-term abnormal return that is free of the
new listing bias, the rebalancing bias, and the skewness bias documented in Kothari
and Warner (1997) and Barber and Lyon (1997). To select the size and BE/ME
matched firm, we first identify all firms that have a market equity value between
70 % and 130 % of that of the event firm and then choose the firm with BE/ME
closest to that of the event firm. The buy-and-hold return of the matched firm
is computed as in Eq. 14.2. We use the label “SZBM1” to represent the single
size and BE/ME matched firm.
The fifth and last benchmark is a single firm that has the highest pre-event correlation
coefficient with the event firm. Specifically, to select the firm, we first construct the
70 size and BE/ME reference portfolios and identify the one that matches the event firm.
Next, we pick firms within the matched size and BE/ME reference portfolio that
have returns in CRSP monthly returns database for all 24 months prior to the event
month and compute their correlation coefficients with the event firm over the pre-event
24 months. We choose the firm with the highest pre-event correlation coefficient with the
event firm as the benchmark. The buy-and-hold return of the most correlated firm
is computed as in Eq. 14.2. We use the label “MC1” to represent the most
correlated single firm.
408 J.S. Ang and S. Zhang
We apply four test statistics to test the null hypothesis of no abnormal returns:
(a) Student’s t-test, (b) Fisher’s sign test, (c) Johnson’s skewness-adjusted t-test, and
(d) bootstrapped Johnson’s t-test.
(a) Student’s t-test
Given the long-term buy-and-hold abnormal returns for a sample of n event
firms, we compute Student’s t-statistic as follows:
AR
t¼ pffiffiffi , (14.8)
sðARÞ= n
where AR is the sample mean and s(AR) the sample standard deviation of the
given sample of abnormal returns. The Student’s t-statistic tests the null
hypothesis that the population mean of long-term buy-and-hold abnormal
returns is equal to zero. The usual assumption for applying the Student’s
t-statistic is that abnormal returns are mutually independent and follow the
same normal distribution.
(b) Fisher’s sign test
To test the null hypothesis that the population median of long-term buy-and-
hold abnormal returns is zero, we compute Fisher’s sign test statistic as follows:
X
n
B¼ IðARi > 0Þ, (14.9)
i¼1
where I(ARi > 0) equals 1 if the abnormal return on the ith firm is greater than
zero and 0 otherwise. At the chosen significance level of a, the null hypothesis
is rejected in favor of the alternative of nonzero median if B b(a/2, n, 0.5) or
B < [n b(a/2, n, 0.5)], or in favor of positive median if B b(a, n, 0.5), or in
favor of negative median if B < [n b(a, n, 0.5)]. The constant b(a, n, 0.5) is
the upper a percentile point of the binomial distribution with sample size n and
success probability of 0.5. The usual assumption for applying the sign test is
that abnormal returns are mutually independent and follow the same continuous
distribution. Note that application of the sign test does not require the popula-
tion distribution to be symmetric. When the population distribution is symmet-
ric, the population mean equals the population median, and the sign test then
indicates the significance of the population mean (see Hollander and Wolfe
2000, Chap. 3).
(c) Johnson’s skewness-adjusted t-test
Johnson (1978) developed the following skewness-adjusted t-test to correct the
misspecification of Student’s t-test caused by the skewness of the population
distribution. Johnson’s test statistic is computed as follows:
1 1
J ¼ t þ pffiffiffi t2 g þ pffiffiffi g, (14.10)
3 n 6 n
where t is Student’s t-statistic given in Eq. 14.8 and g is an estimate of the
Xn 3
coefficient of skewness given by g ¼ ARi AR =sðARÞ3 n . Johnson’s
i¼1
14 Evaluating Long-Horizon Event Study Methodology 409
t-test is applied to test the null hypothesis of zero mean under the assumption
that abnormal returns are mutually independent and follow the same continuous
distribution. At the chosen significance level of a, the null hypothesis is rejected
in favor of the alternative of nonzero mean if J > t(a/2, u) or J < t(a/2, u),
or in favor of positive mean if J > t(a, u), or in favor of negative mean if
J < t(a, u). The constant t(a, u) is the upper a percentile point of the Student’s
t distribution with the degrees of freedom u ¼ n 1.
(d) Bootstrapped Johnson’s skewness-adjusted t-test
Sutton (1992) proposes to apply Johnson’s t-test with a computer-intensive
bootstrap resampling technique when the population skewness is severe and the
sample size is small. He demonstrates it by an extensive Monte Carlo study that
the bootstrapped Johnson’s t-test reduces both type I and type II errors com-
pared to Johnson’s t-test. Lyon et al. (1999) advocate the bootstrapped
Johnson’s t-test in that long-term buy-and-hold abnormal returns are highly
skewed when buy-and-hold reference portfolios are used as benchmarks. They
report that the bootstrapped Johnson’s t-test is well specified and has consider-
able power in testing abnormal returns at the 1-year horizon. In this paper, we
document its power at 3- and 5-year horizons.
We apply the bootstrapped Johnson’s t-test as follows. From the given
sample of n event firms, we draw m firms randomly with replacement counted
as one resample until we have 250 resamples. We calculate Johnson’s test
statistic as in Eq. 14.10 for each resample and end up with 250 J values, labeled
as J1, , J250. Let J0 denotes the J value of the original sample. To test the null
hypothesis of zero mean at the significance level of a, we first determine two
critical values, C1 and C2, such that the percentage of J values less than C1
equals a/2 and the percentage of J values greater than c2 equals a/2, and then
reject the null hypothesis if J0 < C1 or J0 > C2. We follow Lyon et al. (1999) to
apply the bootstrapped Johnson’s t-test with m ¼ 50.9
References
Ang, J. S., & Zhang, S. (2004). An evaluation of testing procedures for long horizon event studies.
Review of Quantitative Finance and Accounting, 23, 251–274.
Barber, B. M., & Lyon, J. D. (1997). Detecting long-run abnormal stock returns: The empirical
power and specification of test statistics. Journal of Financial Economics, 43, 341–372.
Brav, A., & Gompers, P. A. (1997). Myth or reality? The long-run underperformance of initial
public offerings: Evidence from venture and nonventure capital-backed companies. Journal of
Finance, 52, 1791–1821.
Brav, A., Geczy, C., & Gompers, P. A. (2000). Is the abnormal return following equity issuances
anomalous? Journal of Financial Economics, 56, 209–249.
9
Noreen (1989, ▶ Chap. 4) cautions that bootstrap hypothesis tests can be unreliable and that
extensive research is necessary to determine which one of many possible specifications can be
trusted in a particular hypothesis testing situation. We also apply the bootstrapped Johnson’s t-test
with m ¼ 100, 200. We find no significant difference in the test’s performance.
410 J.S. Ang and S. Zhang
Brown, S. J., & Warner, J. B. (1980). Measuring security price performance. Journal of Financial
Economics, 8, 205–258.
Brown, S. J., & Warner, J. B. (1985). Using daily stock returns: The case of event studies. Journal
of Financial Economics, 14, 3–31.
Butler, A. W., & Wan, H. (2010). Stock market liquidity and the long-run stock performance of
debt issuers. Review of Financial Studies, 23, 3966–3995.
Carhart, M. M. (1997). On persistence in mutual fund performance. Journal of Finance, 52, 57–82.
Cowan, A. R., & Sergeant, A. M. A. (2001). Interacting biases, non-normal return
distributions and the performance of tests for long-horizon event studies. Journal of Banking and
Finance, 25, 741–765.
Dahlquist, M., & de Jong, F. (2008). Pseudo market timing: A reappraisal. Journal of Financial
and Quantitative Analysis, 43, 547–580.
Davison, A. C., & Hinkley, D. V. (1997). Bootstrap methods and their application. Cambridge,
UK: Cambridge University Press.
Eckbo, B. E., Masulis, R. W., & Norli, O. (2000). Seasoned public offerings: Resolutions of the
‘new issues puzzle’. Journal of Financial Economics, 56, 251–291.
Eckbo, B. E., & Norli, O. (2005). Liquidity risk, leverage, and long-run IPO returns. Journal of
Corporate Finance, 11, 1–35.
Efron, B., & Tibshirani, R. J. (1993). An introduction to the bootstrap. New York: Chapman & Hall.
Fama, E. F. (1998). Market efficiency, long-term returns and behavioral finance. Journal of
Financial Economics, 49, 283–306.
Fama, E. F., & French, K. R. (1992). The cross-section of expected stock returns. Journal of
Finance, 47, 427–465.
Fama, E. F., & French, K. R. (1993). Common risk factors in the returns on stocks and bonds.
Journal of Financial Economics, 33, 3–56.
Hollander, M., & Wolfe, D. A. (1999). Nonparametric statistical methods. New York: Wiley.
Jegadeesh, N. (2000). Long-term performance of seasoned equity offerings: Benchmark errors
and biases in expectations. Financial Management, 29, 5–30.
Jegadeesh, N., & Karceski, J. (2009). Long-run performance evaluation: Correlation and
heteroskedasticity-consistent tests. Journal of Empirical Finance, 16, 101–111.
Jegadeesh, N., & Titman, S. (1993). Returns to buying winners and selling losers: Implications for
stock market efficiency. Journal of Finance, 48, 65–91.
Jegadeesh, N., & Titman, S. (2001). Profitability of momentum strategies: An evaluation of
alternative explanations. Journal of Finance, 56, 699–720.
Johnson, N. J. (1978). Modified t tests and confidence intervals for asymmetrical populations.
Journal of the American Statistical Association, 73, 536–544.
Kolari, J. W., & Pynnonen, S. (2010). Event study testing with cross-sectional correlation of
abnormal returns. Review of Financial Studies, 23, 3996–4025.
Kothari, S. P., & Warner, J. B. (1997). Measuring long-horizon security price performance.
Journal of Financial Economics, 43, 301–340.
Kothari, S. P., & Warner, J. B. (2007). Econometrics of event studies. In B. E. Eckbo (Ed.),
Handbooks of corporate finance: Empirical corporate finance. Amsterdam:
Elsevier/North-Holland.
Loughran, T., & Ritter, J. R. (1995). The new issues puzzle. Journal of Finance, 50, 23–52.
Loughran, T., & Ritter, J. R. (2000). Uniformly least powerful tests of the market efficiency.
Journal of Financial Economics, 55, 361–389.
Lyon, J. D., Barber, B. M., & Tsai, C.-L. (1999). Improved methods for tests of long-run abnormal
stock returns. Journal of Finance, 54, 165–201.
Matsumoto, M., & Nishimura, T. (1998). Mersenne twister: A 623-dimensionally equidistributed
uniform pseudorandom number generator. ACM Transactions on Modeling and Computer
Simulation, 8, 3–30.
Mitchell, M. L., & Stafford, E. (2000). Managerial decisions and long-term stock price
performance. Journal of Business, 73, 287–329.
14 Evaluating Long-Horizon Event Study Methodology 411
Contents
15.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 414
15.2 Theoretical Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 417
15.3 Empirical Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 418
15.3.1 Asymmetric Nonlinear Smooth Transition GARCH Model . . . . . . . . . . . . . . . . . . 418
15.3.2 Empirical Models for the Intertemporal Relation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 419
15.4 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 420
15.4.1 The Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 420
15.4.2 Estimation Results, Interpretations, and Diagnostics . . . . . . . . . . . . . . . . . . . . . . . . . 420
15.4.3 GJR Sample Period . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 424
15.4.4 The Link Between Asymmetric Mean Reversion and
Intertemporal Relation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 428
15.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 433
Appendix 1: Method to Derive the Variance of the Parameters with Restrictions . . . . . . . . . . . 433
Appendix 2: News Impact Curve . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 434
Appendix 3: Sign Bias Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 435
Abstract
We suggest that an unexpected volatility shock is an important risk factor to
induce the intertemporal relation, and the conflicting findings on the relation
could be attributable to an omitting variable bias resulting from ignoring the
effect of an unexpected volatility shock on the relation. With the effect of an
unexpected volatility shock incorporated in estimation, we find a strong positive
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 413
DOI 10.1007/978-1-4614-7750-1_15,
# Springer Science+Business Media New York 2015
414 K. Nam et al.
Keywords
Intertemporal risk-return relation • Unexpected volatility shocks • Time-varying
rational expectation hypothesis • Stock market overreaction • Expected market
risk premium • Volatility feedback effect • Asymmetric mean reversion •
Asymmetric volatility response • Time-varying volatility • Volatility regime
switching • ANST-GARCH model
15.1 Introduction
The trade-off between risk and return is a core tenet in financial economics.
In particular, the intertemporal risk-return relation is a key element to explain the
predictable variation of expected asset returns.1 Despite its importance in asset
pricing, there has been a long-standing debate on the empirical sign of the
intertemporal relation, with findings that are mixed and inconclusive.
Criticisms of the mixed results often refer to a lack of conditional information. If
the predetermined conditional information set does not contain an important vari-
able that affects the risk-return trade-off, the econometric modeling of market
expectations suffers from the model misspecification problem and leads to
a wrong conclusion on the empirical nature of the intertemporal risk-return relation.
1
Fama and French (1989) argue that systematic patterns in the predictable variations of expected
returns are consistent with the intertemporal asset pricing model by Lucas (1978) and Breeden
(1979) and the consumption smoothing idea by Modigliani and Brumberg (1955) and Friedman
(1957). Ferson and Harvey (1991) and Evans (1994) also document the relative importance of the
time-varying risk premia to the conditional betas to explain predictable variations in expected
returns.
15 The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation 415
2
He suggested that a substantial portion of time variation in the expected risk premium is
associated with time-varying risk factors in investment opportunities.
3
They argue that investors may not require a large premium for bearing risk, but rather may reduce
the risk premium when they perceive exceptionally optimistic expectations on the future perfor-
mance of stock prices.
4
Brandt and Kang (2004) find that the conditional mean and volatility are negatively correlated
contemporaneously but positively correlated unconditionally due to the positive lead-lag relation
between the two moments of stock returns. Poterba and Summers (1986) suggested that, due to the
low level of volatility persistence, the volatility effect on the expected risk premium dissipates so
quickly that it cannot have a major effect on stock price movements. Note that there are some
studies that report weak evidence of the intertemporal relation. See Baillie and DeGennaro (1990),
Whitelaw (1994, 2000), Boudoukh et al. (1997), Yu and Yuan (2011), and M€ uller et al. (2011).
416 K. Nam et al.
5
To avoid the endogeneity problem, French et al. (1987) examined the volatility feedback effect
using ex post unexpected volatility changes. They found a strong negative relation between
unexpected returns and ex post unexpected volatility changes and interpreted it as evidence
supporting a positive intertemporal relation.
15 The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation 417
2
EðRmt rf t jOt1 Þ ¼ f 1 s ^ 2mt1 ,
^ mt þ f 2 e2mt1 s (15.2)
where e 2mt1 and s ^ 2mt1 are the actual realized volatility series and the predicted
volatility series, respectively, such that e2mt1 s ^ 2mt1 represents a prior unexpected
volatility shock. Many studies employ the GARCH models to conditionally
estimate the predictable volatility series s ^ 2mt . Equation 15.2 implies that
@EðRmt rf t jOt1 Þ @f 1
consists of two components: @^ and @ e2 @f^ 2
. While the first
@s2mt s 2mt ð mt1 s 2mt1 Þ
@f 1
term @^s 2 measures the effect of predictable market volatility on the relation, the
second term @ e2 @f^
mt
2
captures the effects of an unexpected volatility shock on
ð mt1 s 2mt1 Þ
the intertemporal relation. However, most of the previous empirical studies
on this topic completely ignore the second term and focus only on the first
term. Consequently, their empirical results show only a partial intertemporal
relation.
Thus, we examine the empirical nature of the full intertemporal relation by
considering not only the predictable conditional volatility but also the effect of an
unexpected volatility shock on the relation. Especially, we define an unexpected
volatility shock in two separate cases of a positive and negative unexpected
volatility shock to examine an asymmetrical effect of a positive and negative
unexpected shock on the relation, if any. A positive (negative) unexpected volatility
shock is denoted as e2mt1 > s ^ 2mt1 (e2mt1 < s
^ 2mt1 ), implying that the actual market
volatility is higher (lower) than the predicted conditional market volatility. We then
examine the sign of intertemporal relation under each case of e2mt1 > s ^ 2mt1 and
emt1 < s
2
^ mt1 , respectively.
2
418 K. Nam et al.
^ mt þ l2 s
Rmt rf t ¼ a þ l1 s ^ mt dt þ emt , (15.3)
where emt is a series of white noise innovations, and s ^mt is the conditional standard
deviation of market portfolio. Note that we use s ^mt instead of s^mt
2
as the conditional
forecasts of stock market volatility. The use of s ^mt is suggested as the slope of the
capital market line by Merton (1980). Also, in estimation, using s ^mt is expected to
yield an improvement in the statistical efficiency, mainly due to a reduction in the
mean square error of the regression. dt is the dummy variable for a positive or
negative unexpected volatility shock. It takes the value 1 with a prior unexpected
^ 2mt1 ) and 0 otherwise. The intertemporal
positive volatility shock (i.e., e2mt1 > s
relation is thus measured by l1 + l2 when e2mt1 > s ^ 2mt1 with dt ¼ 1 or by l1
otherwise with dt ¼ 0. The asymmetrical effect of a positive and negative unex-
pected volatility shock on the intertemporal relation, if any, is captured by l2.
Specifically, l2 > 0 (or l1 + l2 > l1) implies that a positive volatility shock has
a positive impact on the intertemporal relation.
ht ¼ a0 þ a1 e2t1 þ a2 ht1 þ b0 þ b1 e2t1 þ b2 ht1 Fðet1 Þ, (15.4)
negative volatility shock (^e 2t1 < ht1 ). Thus, t captures an asymmetrical effect of
a positive and negative volatility shock on the relation. We put an empirical focus
on t > 0 (or d + t > d), which indicates a positive impact of a positive unexpected
volatility shock on the intertemporal relation.6
We specify Model 3 to examine whether allowing the asymmetry for the
constant term affects the estimation of the intertemporal relation. A possibility is
that the asymmetrical effect of a positive and negative volatility shock on the
relation might disappear.
Model 3:
pffiffiffiffi
r t ¼ ðm1 þ m2 Mt Þ þ fr t1 þ ðd þ tMt Þ ht þ et
(15.7)
ht ¼ a0 þ a1 e2t1 þ a2 ht1 þ b0 þ b1 e2t1 þ b2 ht1 Fðet1 Þ,
where m2Mt captures the asymmetric effect of a prior positive and negative
volatility shock on the level of the conditional mean return. In Model 3, we focus
on whether the estimated value of t is still statistically significant even with
a presence of m2Mt.
We employ the excess market returns as the expected market risk premiums.
To generate the excess returns, we use the monthly nominal returns of the value-
and equal-weighted market portfolio index of the NYSE, AMEX, and NASDAQ
from the CRSP data files from 1926:01 to 2008:12. The monthly excess return series
is constructed by subtracting the 1-month US T-bill returns reported by Ibbotson
Associates from the monthly nominal index returns. The excess return series is
computed as percentage returns. We employ three sample periods: the full period
(1926:01–2008:12), the pre-87 Crash period (1926:01–1987:09), and the GJR period
(1951:04–1989:12). Table 15.1 reports the summary statistics for the data. The
descriptive statistics indicate that both the nominal and the excess returns series of
the two market indexes exhibit significant excess kurtosis and positive first-order
autocorrelation, characterizing the nonnormality of the short-horizon stock returns.
We employ the maximum likelihood method with the analytical derivatives of each
parameter provided in the Gauss code. All the statistical inferences are based on
the Bollerslev-Wooldrige (1992) robust standard errors. Estimation results of
Note that t ¼ 0 supports the volatility irrelevance argument by Poterba and Summers (1986).
6
15
Table 15.1 Summary statistics for monthly nominal and excess returns
Nominal value-weighted returns 1-monthT-bill rates Excess value-weighted returns
Statistics Full period Subperiod Full period Subperiod Full period Subperiod
Observations 996 741 996 741 996 741
Mean (100) 0.895 0.961 0.303 0.287 0.592 0.674
Std. dev. (100) 5.352 5.706 0.253 0.276 5.363 5.722
Skewness 0.18 0.368 1.104 1.261 0.227 0.406
Kurtosis 11.221 10.888 4.684 4.624 11.246 10.875
1st order Autocorrelation 0.112 (0.000) 0.107 (0.003) 0.966 (0.000) 0.967 (0.000) 0.115 (0.000) 0.111 (0.002)
The nominal return series are the monthly value-weighted market index returns for NYSE, AMEX, and NASDAQ stocks and were retrieved from the CRSP
tapes for the period from 1926:01 to 2008:12. The monthly nominal return series are the value-weighted market indexes retrieved from the CRSP tapes for the
same. The monthly excess return series is computed by subtracting 1-month treasury bill returns as reported by Ibbotson Associates from the nominal returns.
All returns are computed as percentage value. The analysis employs the full period (1926:01–2008:12) and the pre-87 Crash period (1926:01–1987:09). The
value in the parentheses is the p-value for the Ljung-Box Q test
The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation
421
422 K. Nam et al.
Model 1 for the full period (1926:01–2008:12) and the pre-87 Crash period
(1926:01–1987:09) are reported in Table 15.2. Model 1 examines the sign of the
simple linear relation between the expected risk premium and the predictable
market volatility. Estimation results for Model 1 show that the intertemporal
relation in a simple linear form is significantly positive for both the full sample
and subsample periods. The estimated value of d is 0.077 (0.080 for the pre-87
Crash period) and statistically significant at the 1 % level for both sample periods.
With regard to the conditional variance equation, the asymmetric
volatility response of a positive and negative return shock is well captured by
b1 + b2 < 0 with a statistical significance. The average estimated value of ht for
the full sample period under a lower volatility regime is 36.37, while it is 24.00
under a higher volatility regime. Also, the estimation results show a high estimated
value of the transition parameter g, which indicates that the transition between
volatility regimes occurs very quickly. This implies that the volatility regime is
divided into only two extreme regimes: the upper and lower volatility regimes. The
upper (lower) regime is induced by any negative (positive) return shock and
exhibits a high volatility persistence (a low volatility persistence) in the conditional
volatility process.
Estimation results of Model 2 are reported in Table 15.2. There are several
notable findings. First, a positive volatility shock has a positive impact on the
intertemporal relation. The results show that the estimated value of t is significantly
positive (0.052 for the full period and 0.041 the pre-87 Crash period) and statisti-
cally significant at the 1 % level for both sample periods. This implies that an
unexpected volatility shock is priced such that, for a positive volatility shock, the
expected risk premium for the full sample period increases p byffiffiffiffi5.2 % (4.1 % for the
pre-87 Crash period) of the predicted conditional volatility ht. Second, the results
show that the intertemporal coefficients are all positive for both sample periods.
The estimated value for the full period d + t ¼ 0.0878 is (0.093 for the pre-87 Crash
period) under prior positive volatility shock, while it is d ¼ 0.035 for the full period
and 0.052 for the pre-87 Crash period under prior negative volatility shock. The
results indicate that the sign of the intertemporal relation is indeed positive when
the effect of an unexpected volatility shock is incorporated in estimation. Third, the
estimation result of t > 0 indicates that the slope of the capital market line is
relatively steeper under a positive volatility shock than under a negative volatility
shock. This implies that a positive volatility shock increases the degree of risk
aversion.
Estimation results of Model 3 are also reported in Table 15.2. In Model 3, we
examine the possibility that the asymmetrical effect of a positive and
negative volatility shock may disappear under the presence of asymmetry in the
constant term allowed in the conditional mean equation. The estimation
results show a positive value of m2 for both periods, indicating that a prior
positive volatility shock raises the conditional mean returns. This asymmetrical
effect of a prior volatility shock on the constant term is more profound for the
pre-87 Crash period (m2 is 0.351 and statistically significant at the 1 % level).
With respect to the intertemporal relation, t is still positive and statistically
15
significant at the 1 % level for both periods (t1 is 0.054 for the full period and 0.048
for the pre-87 Crash period). This implies that, even with the asymmetry allowed in
the constant term, there still exists a significant asymmetrical effect of
a positive and negative volatility shock on the intertemporal relation. Also, d is
still significantly positive for both periods, confirming the positive full
intertemporal relation.
Table 15.3 reports the summary of diagnostics for the estimation results, such as
skewness, kurtosis, the Jarque-Bera normality test, and the Ljung-Box Q test on
the normalized and the squared normalized residuals. The Ljung-Box Q statistics
on the normalized residuals checks serial correlation in the residuals. Rejection
of the null of no autocorrelation up to a certain lag length indicates that either
the dynamics of the conditional mean or the lag structure of the conditional
variance process is not well specified or that both equations are not well specified
by the model. The Ljung-Box statistics on the squared normalized residuals
ascertains if the serial dependence in the conditional variance is well captured by
the equation.
We also perform the negative sign bias test (NSBT) suggested by Engle and Ng
(1993) to examine the ability of the model to capture the so-called leverage effect
of a negative return shock on the conditional variance process. The negative sign
bias test shows insignificant t-values for all estimations and indicates that the
asymmetric volatility response to a positive and negative return shock is well
captured by the ANST-GARCH model.7 The Ljung-Box Q(10) test indicates that
the serial dependence of the conditional mean and variance is well captured by
models 1–3.
7
The negative sign
bias
pffiffiffiffitest is performed with the regression equation v2t ¼ a + bS 0 *
t 1et 1 + p zt +
et, where v2t ¼ et = ht . S
2
t 1 ¼ 1 if et 1 < 0, and S
t 1 ¼ 0 otherwise. Also, z
¼ e
h ð C Þ=h t,
t
where eh ðCÞ ¼ @ht =@C evaluated at the values of maximum likelihood estimates of parameter C.
The test statistic of the NSBT is defined as the t-ratio of the coefficient b in the regression.
A statistically significant t-value implies the failure of the model to absorb the effect of sign bias
and indicates that the volatility model considered is misspecified.
15
where Gt is a dummy variable that takes a value 1 for the GJR sample period or
0 otherwise. The coefficient dG captures the differential effect of the GJR period, if
any, on the intertemporal relation, while d measures the relation for the full period.
The negative intertemporal relation reported by Glosten et al. (1993) can be
confirmed by d + dG < 0. One of the important features of the above modified
Model 1 is its capacity to distinguish the relation for the GJR sample period from
that for the entire sample period. We also estimate the same model with and without
the 1-month T-bill returns rft included in the conditional variance equation.8
Estimation results of Model 4 are reported in Table 15.4. A notable finding is
that the estimated value of dG is strongly negative (–0.103 with rft and –0.092
without rft) and highly significant with d + dG < 0. This result implies that,
comparing to the full period, the GJR sample period is especially characterized
by a strong negative intertemporal relation, and this result is consistent with that of
Glosten et al. (1993).
As mentioned earlier, however, this result does not consider the effect of an
unexpected volatility shock on the relation, reflecting only a simple partial
intertemporal relation. In order to examine the full relation for the GJR sample
period, we specify Model 5 as follows:
Model 5 (Modified Model 2 for GJR Period):
pffiffiffiffi
r t ¼ m þ fr t1 þ ðd þ tMt Þ þ dG þ tG Mt Gt ht þ et
(15.9)
ht ¼ a0 þ a1 e2t1 þ a2 ht1 þ yrf t þ b0 þ b1 e2t1 þ b2 ht1 Fðet1 Þ,
where Gt is a time dummy variable that takes the value 1 for the GJR period
or 0 otherwise. The full intertemporal relation for the GJR period is measured by
d + dG + t + tG under a positive volatility shock and by d + dG under a negative
volatility shock, such that the differential effect of the GJR period on the full
relation is captured by dG + tG. We also estimate the same model with and without
rft included in the conditional variance equation.
Estimation results of Model 5 for the GJR sample period are reported
in Table 15.4. It shows that the estimated value of all four important parameters
(d, t, dG, and tG) capturing the intertemporal relation are statistically significant
at the 1 % level. There are several notable findings. First, the estimation result of
8
Several studies show that the estimation results are sensitive to the inclusion of 1-month T-bill
return in the conditional variance equation. See Campbell (1987), Glosten et al. (1993), and
Scruggs (1998).
15 The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation 427
Table 15.4 Estimation results of models 4 and 5 for the GJR sample period
Model 4 for GJR sample period Model 5 for GJR sample period
Coeff. With rf Without rf With rf Without rf
m 0.546 (9.944) 0.443 (9.288) 0.531 (6.136) 0.543 (5.166)
f 0.056 (5.930) 0.072 (7.909) 0.051 (5.143) 0.051 (5.394)
d 0.059 (5.925) 0.080 (6.392) 0.103 (4.164) 0.100 (3.678)
t 0.046 (4.312) 0.046 (3.778)
dG 0.103 (12.369) 0.092 (8.882) 0.155 (12.540) 0.156 (11.720)
tG 0.067 (6.560) 0.091 (6.903)
a0 0.001 (0.766) 0.001 (0.581) 0.007 (0.238) 0.014 (0.202)
a1 0.110 (3.199) 0.110 (3.291) 0.104 (3.359) 0.102 (3.303)
a2 1.060 (25.142) 1.072 (28.727) 1.061 (26.070) 1.071 (29.469)
b0 2.482 (2.564) 3.007 (3.464) 2.270 (2.538) 2.689 (2.971)
b1 0.029 (0.457) 0.042 (0.795) 0.019 (0.337) 0.021 (0.399)
b2 0.412 (4.672) 0.428 (4.883) 0.404 (4.420) 0.416 (4.840)
y 0.885 (1.065) 0.901 (1.128)
g 121.735 (1.944) 217.354 (2.441) 287.546 (2.128) 254.235 (1.963)
The GJR sample period (1951:04–1989:12) is evaluated by incorporating the period as a dummy
variable in the mean equation, with the 1-month T-bill rates included or excluded in the condi-
tional variance equation. The modified models to estimate the GJR
samplep period
ffiffiffiffi are as follows
Model 4 (modified Model 1 for GJR period): r t ¼ m þ fr t1 þ d þ dG Gt ht þ et pffiffiffiffi
Model 5 (modified Model 2 for GJR period): r t ¼ m þ fr t1 þ ðd þ tMt Þ þ dG þ tG Mt Gt ht þ et
where the indicator function Gt is a dummy variable for the GJR sample period, and rft is
the yield on 1-month T-bill from Ibbotson Associates. The function Mt is specified to capture
the asymmetric effect of an unexpected volatility shock on the relation, such that Mt ¼ 1 if ^e 2t1
> ht1 ; otherwise, Mt ¼ 0. The conditional variance equation for the two estimation
models is specified as ht ¼ [a0 + a1e2t1 + a2ht1] + [b0 + b1e2t1 + b2ht1]F(et1), where
F(et1) ¼ {1 + exp[g(et1)]} 1. The values in parentheses are the Bollerslev-Wooldridge robust
t-statistics
Table 15.5 Diagnostics of the models 4 and 5 for the GJR sample period
Model 4 for GJR sample period Model 5 for GJR sample period
Without rf With rf Without rf Without rf
Skewness of vt 0.636 0.623 0.646 0.626
Kurtosis of vt 5.251 5.083 5.216 5.018
JB-Normality 246.664 (0.000) 217.536 (0.000) 242.89 (0.000) 208.29 (0.000)
Q(10) on vt 10.190 (0.424) 10.397 (0.406) 10.111 (0.431) 10.846 (0.370)
Q(10) on v2t 10.514 (0.397) 10.418 (0.405) 9.7167 (0.466) 10.122 (0.430)
NSBT on ht 1.198 (0.231) 1.162 (0.246) 1.205 (0.228) 1.105 (0.269)
This table presents a summary of diagnostics on the normalized residuals and the
squared normalized residuals from the estimations. The normalized residual series is defined as
pffiffiffiffi
vt ¼ et = ht. JB-Normality refers to the Jarque-Bera normality test statistic, which is distributed as
w with two degrees of freedom under the null hypothesis of normally distributed residuals.
2
Q(10) is the Ljung-Box Q(10) test statistic for checking serial dependence in the normalized
residuals and the squared normalized residuals from the estimations. NSBT refers to the negative
sign bias test suggested by Engle and Ng (1993). It is a diagnostic test that examines the ability of
the specified model to capture the so-called leverage effect of a negative return shock on the
conditional volatility process. The test is performed with the regression equation v2t ¼ a +
pffiffiffiffi2
bSt1 et1 + p0 z*t + et, where v2t ¼ et = ht . St1 ¼ 1 if et1 < 0, and St1 ¼ 0 otherwise. Also,
e e
zt ¼ h ðCÞ=ht , where h ðCÞ ¼ @ht =@C , is evaluated at the values of the maximum likelihood
estimates of parameter C. The test statistic of the NSBT is defined as the t-ratio of the coefficient
b in the regression. The value in the parentheses is the p-value of the individual test statistics
considered
It has been known that the expected market returns exhibit an asymmetric mean-
reverting pattern that negative returns are more likely to revert to positive returns
than positive returns reverting to negative returns. However, the quicker reversion
of negative returns is hardly justified under a positive intertemporal relation. Under
a positive intertemporal relation, a negative return shock raises the risk premium to
compensate for the excess volatility, and an increase in risk premium reduces the
current stock price, which in turn reduces the concurrent stock price. Thus, if
a positive intertemporal relation is correct, a negative return should be more likely
to be accompanied by another negative return for the subsequent periods.
Nam et al. (2001) suggest that the quicker reversion of a negative return is
attributed to a reduction in the expected risk. They show that the intertemporal
relation is significantly negative under a prior negative return shock. They argue
that a negative return shock generates an optimistic expectation by investors, of
the future performance of a stock experiencing a recent price drop, thereby reducing
the expected market risk premium. As a reduction in risk premium in turn raises the
current stock price, negative returns are more likely to revert to positive returns.
While successfully explaining the link between the asymmetric mean-reverting
property and the intertemporal relation, Nam et al. (2001) do not consider the effect
of an unexpected volatility shock on the link. In this section, we investigate the
15 The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation 429
impact of an unexpected volatility shock on the observed link. First, we specify the
following nonlinear autoregressive model to confirm the asymmetric mean rever-
sion of the expected market risk premium:
Model 6:
where the asymmetry is allowed in both the conditional mean and variance, such
that the asymmetry in both processes is controlled by a prior return shock.
The return serial correlation varies between f1 and f1 + f2 depending on the
value of the transition function F(et1). Under an extreme negative prior return
shock that causes F(et1) ¼ 0, serial correlation is measured by f1, while it is
measured by f1 + f2 for an extreme positive return shock yielding F(et1)¼ 1.9
A quicker reversion of a negative return is captured by f1 > 0 (or f1 + f2 > f1).
Note that the condition f1 < 0 (a negative serial correlation under et1 < 0)
indicates a stronger reverting tendency of a negative return.
Estimation results of Model 6 are reported in Table 15.6. It shows that the
estimated value of f2 is positive and highly significant for both the full
period and the pre-87 Crash period. The measured serial correlation is negative
(f1 ¼ 0.078 and –0.076, respectively, for the two periods) under a prior negative
return shock, while it is positive (f1 + f2 ¼ 0.056 and 0.094, respectively, for the
two periods) under a prior positive return shock. This result confirms the asymmet-
rical reverting pattern of the expected returns that a negative return reverts more
quickly, while a positive return tends to persist.
Secondly, we specify Model 7 to examine if there is a link between the
asymmetric mean reversion and the intertemporal relation.
Model 7:
pffiffiffiffi
r t ¼ ½m1 þ m2 Fðet1 Þ þ f1 r t1 þ ½d1 þ d2 Fðet1 Þ ht þ et
(15.11)
ht ¼ a0 þ a1 e2t1 þ a2 ht1 þ b0 þ b1 e2t1 þ b2 ht1 Fðet1 Þ:
9
Stationarity condition of rt is satisfied with |f1 + f2F(et 1)| < 1, i.e., |f1| < 1 for et 1 < 0 or
|f1 + f2| < 1 for et 1 > 0.
430
a strong asymmetry in the partial intertemporal relation under a prior positive and
negative return shock. Second, the partial intertemporal relation is indeed negative
under a prior negative return shock (d1 ¼ 0.155 for the full period and –0.182
for the pre-87 Crash period), while it is positive under a prior positive return shock
(d1 + d2 ¼ 0.103 for the full period and 0.152 for the pre-87 Crash period). This result
implies that while a positive return shock complies with the conventional
positive intertemporal relation, a negative return shock indeed induces a negative
intertemporal behavior of the expected market returns. More importantly, the results
of Models 6 and 7 confirm the highly significant asymmetric link between the mean
reversion and the intertemporal relation. The quicker reversion of negative returns is
attributed to the negative intertemporal relation under a prior negative return shock.
Model 7 does not incorporate the asymmetrical impact of a positive and negative
volatility shock in the estimation of the intertemporal relation. To get empirically
more reliable results on the link between the mean reversion and the full
intertemporal relation, we propose Model 8 in the following specification.
Model 8:
pffiffiffiffi
r t ¼ ½m1 þ m2 Fðet1 Þ þ f1 r t1 þ ½ðd1 þ t1 Mt Þ þ ðd2 þ t2 Mt ÞFðet1 Þ ht þ e t
ht ¼ a0 þ a1 e2t1 þ a2 ht1 þ b0 þ b1 e2t1 þ b2 ht1 Fðet1 Þ,
(15.12)
where the full intertemporal relation is measured separately under a positive and
negative return shock. When a negative return shock causing F(et1) ¼ 0 is realized,
the full intertemporal relation is measured by d1 + t1, for which t1 captures an
asymmetrical impact of a prior unexpected volatility shock on the relation. With
a prior positive return shock, the relation is measured by d1 + d2 + t1 + t2, for which
t1 + t2 measures an asymmetrical effect of a prior volatility shock on the relation, if any.
The estimation results of Model 8 are reported in Table 15.6. There are several
notable findings. First, the results show that three out of the four important parameters
(d1, t1, d2, and t2) to capture the full intertemporal relation are statistically significant
at the 5 % level. Second, the result of d1 + t1 ¼ 0.181 (–0.220 for the pre-87 Crash
period) and d1 + d2 + t1 + t2 ¼ 0.020 (0.057 for the pre-87 Crash period) implies that
the full intertemporal relation is still negative (positive) under a prior negative
(positive) return shock. Third, more importantly, the results of t1 ¼ 0.079 and t1 +
t2 ¼ 0.051 (0.024 and 0.076 for the pre-87 Crash period) indicate that a positive
volatility shock has a positive impact on the intertemporal relation, regardless of the
sign of a prior return shock. This implies that a positive unexpected volatility shock
consistently induces a positive impact on the intertemporal risk-return relation.
Table 15.7 reports the results of diagnostic tests. The Ljung-Box Q(10) test
results on vt and v2t indicate that serial dependence is well captured by the
specified conditional mean and variance processes. The negative sign bias test
shows insignificant t-values for all estimations, confirming the capability of the
ANST-GARCH model to capture the excess volatility response caused by
a negative return shock.
432
15.5 Conclusions
We suggest that the intertemporal risk-return relation is driven not only by the
underlying market volatility but also by an unexpected volatility shock. Most of the
previous literature on this topic ignores the effect of an unexpected volatility shock
on the relation. Thus, their results reflect only a partial intertemporal relation. With
the effect of an unexpected volatility shock incorporated in the estimation of the
relation, we find a strong positive intertemporal relation for the US monthly excess
returns for the period of 1926:01–2008:12.
We also reexamine the relation for the GJR sample period with the effect of
a volatility shock incorporated in estimation. The estimation results show that the
GJR sample period is indeed characterized by a strong positive (negative) relation
under a positive (negative) volatility shock. This implies that the negative relation
reported by Glosten et al. (1993) is attributed to ignoring the effect of an unexpected
volatility shock on the relation.
Lastly, we examine the observed link between the mean reversion property and
the intertemporal relation under a consideration of the impact of an unexpected
volatility shock on the link. We confirm that the quicker reversion of negative
returns is attributed to the negative intertemporal relation under a prior negative
return shock. We interpret this negative intertemporal relation as reflective of
strong optimistic expectations as perceived by investors of the future performance
of a stock experiencing a recent price drop.
where ^b is the actual coefficient estimated from computation, and b^ is the true
parameter estimate we want to transform from Eq. 15.13. Using Taylor expan-
sion, Eq. 15.13 can be expressed as
f ^b ¼ f ðbÞ þ f 0 ðbÞ ^b b , (15.14)
or
434 K. Nam et al.
f b^ f ðbÞ ¼ f 0 ðbÞ ^b b : (15.15)
where the true value of f 0 (b) is not known. We thus use the MLE of b, ^b, to get
the value of f 0 (b). Then the variance of b^ can be calculated as follows:
2
var b^ ¼ f 0 ^b var ^b , (15.18)
where f 0 ^
b ¼ f ^b 1 f ^b :
^
b^ ¼ f ^b ¼ eb : (15.19)
We derive the functions of the news impact curve (NIC) for the ANST-GARCH and
the GJR models as follows:
ANST-GARCH (1,1) Model:
Model : ht ¼ a0 þ a1 e2t1 þ g1 ht1 þ b0 þ b1 e2t1 þ b2 ht1 Fðet1 Þ (15.21)
and
15 The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation 435
1, if et1 > 0
where I ðet1 > 0Þ ¼ :
0, if et1 < 0
Based on the news impact curve, we perform three diagnostic tests for examining
the ability of a model to capture asymmetric effect of news on conditional variance:
the sign bias test (SBT), the negative sign bias test (NSBT), and the positive sign
bias test (PSBT). These tests are performed by the t-statistic on the coefficient
b under the following regression equations:
0
v2t ¼ a þ bS
t1 þ b zt þ et (15.25)
0
v2t ¼ a þ bS
t1 et1 þ b zt þ et (15.26)
0
v2t ¼ a þ bSþ
t1 et1 þ b zt þ et , (15.27)
pffiffiffiffi2
where v2t ¼ et = ht , S
t1 ¼ 1 if et1 < 0 and St1 ¼ 0 otherwise, and
e e
St1 ¼ 1 St1. zt ¼ h ðyÞ=ht , where h ðyÞ ¼ @ht =@y evaluated at the
+
References
Abel, A. B. (1988). Stock prices under time-varying dividend risk: An exact solution in an infinite-
horizon general equilibrium model. Journal of Monetary Economics, 22, 373–393.
Ang, A., Hodrick, R., Xing, Y., & Zhang, X. (2006). The cross-section of volatility and expected
returns. Journal of Finance, 61, 259–299.
Backus, D., & Gregory, A. W. (1993). Theoretical relations between risk premiums and condi-
tional variances. Journal of Business and Economic Statistics, 11, 177–185.
436 K. Nam et al.
Baillie, R., & DeGennaro, R. P. (1990). Stock returns and volatility. Journal of Financial and
Quantitative Analysis, 25, 203–214.
Bali, T. G. (2008). The intertemporal relation between expected returns and risk. Journal of
Financial Economics, 87, 101–131.
Ball, R., & Kothari, S. (1989). Nonstationary expected returns: Implications for tests of market
efficiency and serial correlation in returns. Journal of Financial Economics, 25, 51–74.
Barky, R. B. (1989). Why don’t the prices of stocks and bonds move together? American
Economic Review, 79, 1132–1145.
Bollerslev, T., & Wooldridge, J. M. (1992). Quasi-maximum likelihood estimation and inference
in dynamic models with time-varying covariances. Econometric Review, 11, 143–172.
Boudoukh, J., Richardson, M., & Whitelaw, R. F. (1997). Nonlinearities in the relation between
the equity risk premium and the term structure. Management Science, 43, 371–385.
Brandt, W., & Kang, Q. (2004). On the relationship between the conditional mean and volatility of
stock returns: A latent VAR approach. Journal of Financial Economics, 72, 217–257.
Breeden, D. T. (1979). An intertemporal asset pricing model with stochastic consumption and
investment opportunities. Journal of Financial Economics, 7, 265–296.
Breen, W., Glosten, L., & Jagannathan, R. (1989). Economic significance of predictable variations
in stock index returns. Journal of Finance, 44, 1177–1189.
Campbell, J. Y. (1987). Stock returns and the term structure. Journal Financial Economics, 18,
373–399.
Campbell, J. Y., & Hentschel, L. (1992). No news is good news: An asymmetric model of
changing volatility in stock returns. Journal Financial Economics, 31, 281–318.
Campbell, J. Y., Lo, A. W., & MacKinlay, A. C. (1997). The econometrics of financial markets.
Princeton: Princeton University Press.
Cecchetti, S. G., Lam, P., & Mark, N. C. (1990). Mean reversion in equilibrium asset prices.
American Economic Review, 80, 398–418.
Darrat, A. F., Gilley, O. W., Li, B., & Wu, Y. (2011). Revisiting the risk/return relations in the
Asian pacific markets: New evidence from alternative models. Journal of Business Research,
64, 199–206.
Engle, R. F., & Ng, V. K. (1993). Measuring and testing the impact of news on volatility. Journal
of Finance, 48, 1749–1778.
Evans, M. D. (1994). Expected returns, time-varying risk, and risk premia. Journal of Finance, 49,
655–679.
Fama, E. F., & French, K. R. (1988). Permanent and temporary components of stock prices.
Journal of Political Economy, 96, 246–273.
Fama, E. F., & French, K. R. (1989). Business conditions and expected returns on stocks and
bonds. Journal of Financial Economics, 25, 23–49.
Ferson, W., & Harvey, C. (1991). The variation of economic risk premiums. Journal of Political
Economy, 99, 385–415.
Fornari, F., & Mele, A. (1997). Sign- and volatility-switching ARCH models: Theory and
applications to international stock markets. Journal of Applied Econometrics, 12, 49–65.
French, K. R., Schwert, G. W., & Stambaugh, R. F. (1987). Expected stock returns and volatility.
Journal of Financial Economics, 19, 13–29.
Friedman, M. (1957). A theory of the consumption function. Princeton: Princeton University Press.
Ghysel, E., Santa-Clara, P., & Valkanov, R. (2005). There is a risk-return trade-off after all.
Journal of Financial Economics, 76, 509–548.
Glosten, L., Jagannathan, R., & Runkle, D. (1993). On the relation between the expected value
and the volatility of the nominal excess return on stocks. Journal of Finance, 48, 1779–1801.
Gonzalez-Rivera, G. (1998). Smooth transition GARCH models. Studies in Nonlinear Dynamics
and Econometrics, 3, 61–78.
Gourieroux, C., & Monfort, A. (1997). Time series and dynamic models. Cambridge, UK:
Cambridge University Press.
Hamilton, J. D. (1994). Time series analysis. Princeton: Princeton University Press.
15 The Effect of Unexpected Volatility Shocks on Intertemporal Risk-Return Relation 437
Hamilton, J. D., & Susmel, R. (1994). Autoregressive conditional heteroskedasticity and changes
in regime. Journal of Econometrics, 64, 307–333.
Harvey, C. R. (1989). Time-varying conditional covariances in tests of asset pricing models.
Journal of Financial Economics, 24, 289–317.
Harvey, A. C. (1993). The econometric analysis of time series. Cambridge, MA: The MIT Press.
Harvey, C. R. (2001). The specification of conditional expectations. Journal of Empirical Finance,
8, 573–638.
Haugen, R. A., Talmor, E., & Torous, W. N. (1991). The effect of volatility changes on the level of
stock prices and subsequent expected returns. Journal of Finance, 46, 985–1007.
Kim, C., Morley, J. C., & Nelson, C. R. (2001). Does an intertemporal tradeoff between risk and
return explain mean reversion in stock price? Journal of Empirical Finance, 8, 403–426.
Lucas, R. E. (1978). Asset prices in an exchange economy. Econometrica, 46, 1429–1446.
Ludvigson, S. C., & Ng, S. (2007). The empirical risk-return relation: A factor analysis approach.
Journal of Financial Economics, 83, 171–222.
utkepol, H. (1993). Introduction to multiple time series analysis. New York: Springer.
L€
Merton, R. C. (1973). An intertemporal asset pricing model. Econometrica, 41, 867–888.
Merton, R. C. (1980). On estimating the expected return on the market: An exploratory investi-
gation. Journal of Financial Economics, 8, 323–361.
Modigliani, F., & Brumberg, R. (1955). Utility analysis and the consumption function. In
K. Kurihara (Ed.), Post Keynesian economics. London: G. Allen.
M€uller, G., Durand, R. B., & Maller, R. A. (2011). The risk–return tradeoff: A COGARCH
analysis of Merton’s hypothesis. Journal of Empirical Finance, 18, 306–320.
Nam, K., Pyun, C., & Avard, S. (2001). Asymmetric reverting behavior of short-horizon stock
returns: An evidence of stock market overreaction. Journal of Banking and Finance, 25,
805–822.
Nelson, D. B. (1991). Conditional heteroskedasticity in asset returns: A new approach.
Econometrica, 59, 347–370.
Nyberga, H. (2012). Risk-return tradeoff in U.S. stock returns over the business cycle. Journal of
Financial and Quantitative Analysis, 47, 137–158.
Pagan, A. R., & Hong, Y. S. (1989). Nonparametric estimation and the risk premium. In W.A. Barnett,
J. Powell, & G. E. Tauchen (Eds.), Semiparametric and nonparametric methods in econometrics
and statistics (pp. 51–75). Cambridge: Cambridge University Press.
Pastor, L., Sinha, M., & Swaminthan, B. (2008). Estimating the intertemporal risk-return tradeoff
using the implied cost of capital. Journal of Finance, 63, 2859–2897.
Pindyck, R. (1984). Risk, inflation, and the stock market. American Economic Review, 74,
335–351.
Poterba, J. M., & Summers, L. H. (1986). The persistence of volatility and stock market fluctu-
ations. American Economic Review, 76, 1142–1151.
Rothman, P. (1999). Nonlinear time series analysis of economic and financial data. Boston:
Kluwer.
Scruggs, J. T. (1998). Resolving the puzzling intertemporal relation between the market risk
premium and conditional market variance: A two-factor approach. Journal of Finance, 52,
575–603.
Turner, C. M., Startz, R., & Nelson, C. R. (1989). A Markov model of heteroskedasticity, risk, and
learning in the stock market. Journal of Financial Economics, 25, 3–22.
Whitelaw, R. F. (1994). Time variations and covariations in the expectation and volatility of stock
market returns. Journal of Finance, 49, 515–541.
Whitelaw, R. F. (2000). Stock market risk and return: An equilibrium approach. Review of
Financial Studies, 13, 521–547.
Yu, J., & Yuan, Y. (2011). Investor sentiment and the mean–variance relation. Journal of
Financial Economics, 100, 367–381.
Combinatorial Methods for Constructing
Credit Risk Ratings 16
Alexander Kogan and Miguel A. Lejeune
Contents
16.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440
16.1.1 Importance of Credit Risk Ratings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 440
16.1.2 Contribution and Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 442
16.2 Logical Analysis of Data: An Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 444
16.3 Absolute Creditworthiness: Credit Risk Ratings of Financial Institutions . . . . . . . . . . . . 447
16.3.1 Problem Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 447
16.3.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 448
16.3.3 An LAD Model for Bank Ratings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 450
16.3.4 LAD Model Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 450
16.3.5 Remarks on Reverse Engineering Bank Ratings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 454
16.4 Relative Creditworthiness: Country Risk Ratings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 455
16.4.1 Problem Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 455
16.4.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 456
16.4.3 Rating Methodologies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 458
16.4.4 Evaluation of the Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 467
16.4.5 Importance of Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 474
16.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 474
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 476
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 481
A. Kogan (*)
Rutgers Business School, Rutgers, The State University of New Jersey, Newark–New Brunswick,
NJ, USA
Rutgers Center for Operations Research (RUTCOR), Piscataway, NJ, USA
e-mail: kogan@rutgers.edu
M.A. Lejeune
George Washington University, Washington, DC, USA
e-mail: mlejeune@gwu.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 439
DOI 10.1007/978-1-4614-7750-1_16,
# Springer Science+Business Media New York 2015
440 A. Kogan and M.A. Lejeune
Abstract
This study uses a novel method, the Logical Analysis of Data (LAD), to reverse
engineer and construct credit risk ratings which represent the creditworthiness of
financial institutions and countries. LAD is a data mining method based on
combinatorics, optimization, and Boolean logic that utilizes combinatorial
search techniques to discover various combinations of attribute values that are
characteristic of the positive or negative character of observations. The proposed
methodology is applicable in the general case of inferring an objective rating
system from archival data, given that the rated objects are characterized by
vectors of attributes taking numerical or ordinal values. The proposed
approaches are shown to generate transparent, consistent, self-contained, and
predictive credit risk rating models, closely approximating the risk ratings
provided by some of the major rating agencies. The scope of applicability of
the proposed method extends beyond the rating problems discussed in this study
and can be used in many other contexts where ratings are relevant.
We use multiple linear regression to derive the logical rating scores.
Keywords
Credit risk rating • Reverse engineering • Logical Analysis of Data • Combina-
torial optimization • Data mining • Creditworthiness • Financial strength •
Internal rating • Preorder • Logical rating score
16.1 Introduction
Credit ratings published by such agencies as Moody’s, Standard & Poor’s, and Fitch
are considered important indicators for financial markets, providing critical infor-
mation about the likelihood of future default. The importance of credit ratings is
recognized by international bodies, as manifested by the Basel Capital Accord
(2001, 2006). The progressively increasing importance of credit risk ratings is
driven by the dramatic expansion of investment opportunities associated with the
globalization of the world economies. Since these opportunities are often risky, the
internationalized financial markets have to rely on agencies’ ratings for the assess-
ment of credit risk.
The importance of credit risk rating systems is manifold, as shown below:
• Credit approval: the credit risk rating plays a major role in the credit
preapproval decisions. Indeed, a binary decision model in which the credit risk
rating of the obligor is an explanatory variable is typically used to preapprove the
decision to grant or not the credit.
• Pricing: in case of loan preapproval, the credit risk rating impacts the conditions
(interest rate, covenants, collaterals, etc.) under which the final credit is granted
(Treacy and Carey 2000). The credit risk rating is also utilized in the operations
subsequent to the preapproval stage.
16 Combinatorial Methods for Constructing Credit Risk Ratings 441
• Provisioning: the credit risk rating is a variable used for calculating the expected
losses, which, in turn, are used to determine the amount of economic capital
that a bank must keep to hedge against possible defaults of its borrowers.
The expected loss of a credit facility is a function of the probability of default
of the borrower to which the credit line is granted, as well as the exposure at
default and the loss given default associated with the credit facility. Since there
is usually a mapping between credit risk rating and the borrower’s probability of
default, and provided that the rating of a borrower is a key predictor to assess the
recovery rate associated with a credit facility granted to this borrower, the
importance of the credit risk rating in calculating the amounts of economic
capital is evident.
• Moral hazard: the reliance upon an objective and accurate credit risk rating
system is a valuable tool against some moral hazard situations. Some financial
institutions do not use a credit risk rating model but instead let lending officers
assign credit ratings based on their judgment. The lending officers are in charge
of the marketing of banking services, and their performance and therefore
compensation are determined with respect to the “profitability” of the relation-
ships between the bank and its customers. Clearly, the credit risk ratings
assigned by the lending officers will affect the volume of the approved loans
and the compensation of the officer who, as a result, could have an incentive to
assign ratings in a way that is not consistent with the employer’s interests. Thus,
the use of a reliable credit risk rating system could lead to avoiding such perverse
incentive situations.
• Basel compliance: the New Basel Capital Accord (Basel II) requires banks to
implement a robust framework for the evaluation of credit risk exposures of
financial institutions and the capital requirements they must bear. This involves
the construction and the cross-validation of accurate and predictive credit risk
rating systems.
Financial institutions, while taking into account external ratings (e.g., those
provided by Fitch, Moody’s, S&P’s), have increasingly been developing efficient
and refined internal rating systems over the past years. The reasons for this trend are
multiple.
First, the work of rating agencies, providing external, public ratings, has recently
come under intense scrutiny and criticism, justified partly by the fact that some of
the largest financial collapses of the decade (Enron Corp, etc.) were not anticipated
by the ratings. At his March 2006 testimony before the Senate Banking Committee,
the president and CEO of the CFA Institute1 highlighted the conflicts of interest in
credit ratings agencies (Wall Street Letter 2006). He regretted that credit rating
agencies “have been reluctant to embrace any type of regulation over the services
they provide” and reported that credit rating agencies “should be held to the highest
standards of transparency, disclosure and professional conduct. Instead, there are no
standards.” The problem is reinforced by the fact that rating agencies, charging fees
1
Wall Street Letter. 2006. CFA To Senate: Follow Our Lead On Credit Rating.
442 A. Kogan and M.A. Lejeune
The above discussion outlines the impact of the credit risk ratings, the possible
problems of objectivity and transparency of external (i.e., provided by rating
agencies) credit risk ratings, and the importance and need for financial institutions
to develop their own, internal credit risk rating systems.
2
The Economist, July 15, 1995, 62.
16 Combinatorial Methods for Constructing Credit Risk Ratings 443
In this chapter, we use the novel combinatorial pattern extraction method called
the logical analysis of data (LAD) to learn a given credit risk rating system and to
develop on its basis a rating system having the following characteristics:
• Self-containment: the rating system does not use as predictor variables any other
credit risk rating information (non-recursiveness). Clearly, this requirement pre-
cludes the use of lagged ratings as independent variables. It is important to note that
this approach is in marked contrast with that of the current literature (see Hammer
et al. 2006, 2011 for a discussion). The significant advantage of the non-recursive
nature of the rating system is its applicability to not-yet-rated obligors.
• Objectivity: the rating system only relies on measurable characteristics of the
rated entities.
• Transparency: the rating system has formal explicit specification.
• Accuracy: it is in close agreement with the learned, opaque rating system.
• Consistency: the discrepancies between the learned rating system and the
constructed one are resolved by subsequent changes in the learned rating system.
• Generalizability: applicability of the rating system to evaluate the creditwor-
thiness of obligors at subsequent years or of obligors that were not previously
rated.
• Basel compliance: it satisfies the Basel II Accord requirements (cross-
validation, etc.).
In this study, we derive a rating system for two types of obligors:
• Countries: Eliasson (2002) defines country risk as the “risk of national gov-
ernments defaulting on their obligations,” while Afonso et al. (2007) state that
“sovereign credit ratings are a condensed assessment of a government’s ability
and willingness to repay its public debt both in principal and in interests on
time.” Haque et al. (1996) define country credit risk ratings compiled by
commercial sources as an attempt “to estimate country-specific risks, particu-
larly the probability that a country will default on its debt-servicing obligations.”
• Financial institutions: bank financial strength ratings represent the “bank’s
intrinsic safety and soundness” (Moody’s 2006).
Two different approaches developed in this chapter for reverse engineering and
constructing credit risk ratings will be based on the following:
• Absolute creditworthiness, which evaluates the riskiness of individual obligors.
• Relative creditworthiness, which first evaluates the comparative riskiness of
pairs of obligors; the absolute riskiness of entities is then derived from their
relative riskiness using the combinatorial techniques of partially ordered sets.
As was noted in the literature (de Servigny and Renault 2004), banks and other
financial service organizations are not fully utilizing the opportunities provided by
the significant increase in the availability of financial data. More specifically, the
area of credit rating and scoring is lacking in up-to-date methodological advances
(Galindo and Tamayo 2000; de Servigny and Renault 2004; Huang et al. 2004).
This provides a tremendous opportunity for the application of modern data mining
and machine learning techniques, based on statistics (Jain et al. 2000) and combi-
natorial pattern extraction (Hammer 1986). The above described approaches are
implemented using the LAD combinatorial pattern extraction method, and it turns
444 A. Kogan and M.A. Lejeune
out to be a very conclusive case for the application and the contribution of data
mining to the credit risk industry.
This chapter is structured as follows. Section 16.2 provides an overview of LAD
(Hammer 1986), which is used to develop a new methodology for reverse engineer-
ing and rating banks and countries with respect to their creditworthiness. Section 16.3
details the absolute creditworthiness model constructed for rating financial institu-
tions and describes the obtained results. Section 16.4 is devoted to the description of
two relative creditworthiness models for rating countries and contains an extensive
description of the results. Section 16.5 provides concluding remarks.
The logical analysis of data (LAD) is a modern data mining methodology based on
combinatorics, optimization, and Boolean logic3. LAD can be applied for the analysis
and classification of archives containing both binary (Hammer 1986; Crama
et al. 1988) and numerical (Boros et al. 1997) data. The novelty of LAD consists in
utilizing combinatorial search techniques to discover various combinations of attri-
bute values that are characteristic of the positive or negative character of observations
(such as whether a bank is solvent or not or whether a patient is healthy or sick). Then
LAD selects (usually small) subsets of such combinations (usually optimizing
a certain quality objective) to construct what is called a model (Boros et al. 2000).
We briefly describe below the basic concepts of LAD, referring the reader for a more
detailed description to Boros et al. (2000) and Alexe et al. (2007).
Observations in archives analyzed by LAD are represented by n-dimensional
real-valued vectors which are called positive or negative based on the value of the
additional binary (0,1) attribute called the outcome or the class of the observation.
Consider a dataset as a collection of M points (a j, z j) where the outcome z j of
observation j has value 1 for a positive outcome and 0 for a negative one and a j is an
n-dimensional vector. Figure 16.1 illustrates a dataset containing five observations
described by three variables. Each component a j[i] of the [5 3]-matrix in
3
The presentation in this section is partially based on Hammer et al. (2006).
16 Combinatorial Methods for Constructing Credit Risk Ratings 445
Observations
Variables a[1] a[2] a[3]
Outcome
c1,1 c1,2 c1,3 c2,1 c2,2 c3,1 c3,2 c3,3
Cutpoints
3 2.4 1.5 3 2 4 3 2
j
yi,k zj
1 1 1 1 1 1 0 0 1 1
Binary 2 0 1 1 0 0 1 1 1 1
Variables
3 0 0 0 0 1 0 1 1 1
4 1 1 1 0 0 0 1 1 0
5 0 0 1 0 1 0 0 0 0
Fig. 16.1 gives the value taken by variable i in observation j. We will use a[i] to
denote the variables corresponding to the components of this dataset. The rightmost
column provides the outcome of the observation.
LAD discriminates positive and negative observations by constructing a binary-
valued function f depending on the n input variables, in such a way that it closely
approximates the unknown actual discriminator. LAD constructs this function f as
a weighed sum of combinatorial patterns.
In order to specify how such a function f is found, we first transform the original
dataset into a binarized dataset in which the variables can only take the values 0 and
1. We shall achieve this goal by using indicator variables which show whether the
values the variables take in a particular observation are “large” or “small”; more
precisely, each indicator variable shows whether the value of a numerical variable
does or does not exceed a specified level. This is achieved by defining, for each
variable a[i], a set of K(i) values {ci,k j k ¼ 1,..., K(i)}, called cut points to which
binary variables {yi,k j k ¼ 1,. . .,K(i)} are associated. The values of these binary
variables for each observation (z j, a j) are then defined as:
1 if a j ½i ci, k
yij, k ¼
0 otherwise
Figure 16.2 provides the binarized dataset corresponding to the data displayed in
Fig. 16.1 and shows the values ci,k of the cut points k for each variable a[i] and those
j
of the binary variables yi,k associated with any cut point k of variable i in observa-
tion j. For example, y1,1 ¼ 1 since a1[1 ¼ 3.5 is greater than c1,1 ¼ 3.
1
and the corresponding classification is determined by the sign of this expression. LAD
leaves unclassified any observation for which D(o) ¼ 0, since in this case either the
model does not provide sufficient evidence or the evidence it provides is contradictory.
Computational experience with real-life problems has shown that the number of
unclassified observations is usually small. The results of classifying the observations
in a dataset can be represented in the form of a classification matrix (Table 16.1).
Here, the percentage of positive (negative) observations that are correctly
classified is represented by a (respectively d). The percentage of positive
(negative) observations that are misclassified is represented by c (respectively b).
The percentage of positive (negative) observations that remain unclassified is
represented by e (respectively f). Clearly, a + c + e ¼ 100 % and b + d + f ¼
100 %. The quality of the classification is defined by
1 1
Q ¼ ða þ dÞ þ ðe þ f Þ (16.2)
2 4
The capability of evaluating the credit quality of banks has become extremely
important in the last 30 years given the increase in the number of bank failures:
during the period from 1950 to 1980, bank failures averaged less than seven per
year, whereas during the period from 1986 to 1991, they averaged 175 per year
(Barr and Siems 1994).4 Curry and Shibut (2000) report that the so-called savings
and loan crisis cost around $123.8 billion. Central banks are afraid of widespread
bank failures since they could exacerbate cyclical recessions and result in more
severe financial crises (Basel Committee on Banking Supervision 2004). More
accurate credit risk models for banks could enable the identification of problematic
banks early, which is seen as a necessary condition by the Bank for International
Settlements (2004) to avoid failure, and could serve the regulators in their efforts to
minimize bailout costs.
4
The presentation in this section is based on Hammer et al. (2012).
448 A. Kogan and M.A. Lejeune
The evaluation and rating of the creditworthiness of banks and other financial
organizations is particularly challenging, since banks and insurance companies
appear to be more opaque than firms operating in other industrial sectors. Morgan
(2002) attributes this to the fact that banks hold certain assets (loans, trading assets,
etc.), the risks of which change fast and are very difficult to assess, and it is further
compounded by banks’ high leverage. Therefore, it is not surprising that the main
rating agencies (Moody’s and S&P’s) disagree much more often about the ratings
given to banks than about those given to obligors in other sectors. The difficulty of
accurately rating those organizations is also due the fact that the rating migration
volatility of banks is historically significantly higher than it is for corporations and
countries and that banks tend to have higher default rates than corporations
(de Servigny and Renault 2004). Another distinguishing characteristic of the
banking sector is the external support (i.e., from governments) that banks receive
and other corporate sectors do not (Fitch Rating 2006). A thorough review of the
literature pertaining to the rating and evaluation of credit risk of financial institu-
tions can be found in Hammer et al. (2012).
In the next sections, we shall:
• Identify a set of variables that provides sufficient information to accurately
replicate the Fitch bank ratings.
• Construct LAD patterns to discriminate between banks with high and low ratings.
• Construct an optimized model utilizing (some of) the LAD patterns which is
capable of distinguishing between banks with high and low ratings.
• Define an accurate bank rating system on the basis of the discriminant values
provided by the constructed model.
• Cross-validate the proposed rating system.
16.3.2 Data
Our dataset consists of 800 banks rated by Fitch and operating in 70 different
countries (247 in Western Europe; 51 in Eastern Europe; 198 in Canada and the
USA; 45 in developing Latin American countries; 47 in the Middle East; 6 in
Oceania; 6 in Africa; 145 in developing Asian countries; 55 in Hong Kong, Japan,
and Singapore).
Our design of an objective and transparent bank rating system on the basis of the LAD
methodology is guided by the properties of LAD as a classification system. Therefore,
we define a classification problem associated to the bank rating problem, construct
an LAD model for it, and then define a bank rating system rooted in this LAD model.
We define as positive observations the banks which have been rated by Fitch
as A, A/B, or B and as negative observations those whose Fitch rating is D, D/E, or E.
In the binarization process, cut points were introduced for the 19 of the 24 numer-
ical variables shown in Table 16.3. Actually, the other five numerical variables
were also binarized, but since it turned out that they were redundant, only the
variables shown in Table 16.3 were retained for constructing the model. Table 16.3
provides all the cut points used in pattern and model construction. For example, two
cut points (24.8 and 111.97) are used to binarize the numerical variable “profit
before tax” (PbT), that is, two binary indicator variables replace PbT, one telling
whether PbT exceeds 24.8 and the other telling whether PbT exceeds 111.97.
The first step of applying the LAD technique to the problem binarized with the
help of these variable cut points was the identification of a collection of powerful
patterns. One example of such a powerful negative pattern is (i) the country risk
rating is strictly lower than A and (ii) the profits before tax are at most equal to
€111.96 millions. One can see that these conditions describe a negative pattern,
since none of the positive observations (i.e., banks rated A, A/B, or B) satisfy both
of them, while no less than 69.11 % of the negative observations (i.e., those banks
rated D, D/E, or E) do satisfy both conditions. This pattern has degree 2, prevalence
69.11 %, and homogeneity 100 %.
The model we have developed for bank ratings is very parsimonious, consisting
of only 11 positive and 11 negative patterns, and is built on a support set of only
19 out of the 24 original variables. All the patterns in the model are of degree at
most 3, have perfect homogeneity (100 %), and very substantial prevalence
(averaging 50.9 % for the positive and 37.7 % for the negative patterns).
observations to a training and a testing sets of equal size. We use the training set to
build the model, and we apply it to classify the observations in the testing set. In the
second half of the experiment, we reverse the role of the two sets. The average
accuracy and the standard deviation of the accuracy are, respectively, equal to
95.12 % and 0.03. These results highlight the predictive capability and the robust-
ness of the derived LAD model.
We have also computed the correlation between the discriminant values of the
LAD model and the bank credit risk ratings (represented on their numerical scale).
Despite the fact that the correlation measure takes into account all the banks in the
dataset (i.e., not only those which were used in creating the LAD model but also
those rated B/C, C, or C/D, which were not used in the learning process), the
correlation is very high, equal to 80.70 %, attesting to the strong predictive power of
the LAD model. The ten two-folding experiments described above were then used
to verify the stability of the correlation between the LAD discriminant values and
the bank ratings. The average correlation is equal to 80.04 %, with a standard
deviation of 0.04, another testimony of the stability of the close positive association
between the LAD discriminant values and the bank ratings.
subject to di xjðiÞþ1 , i 2 N
(16.3)
xjðiÞ < di , i 2 N
1 ¼ x0 x1 x2 , . . . , xj , . . . , x8 x9 ¼ 1
1 di 1, i 2 N
on the NEOS server (Czyzyk et al. 1998) with the solver Lancelot to determine the cut
points xj and the adjusted discriminant values di that will be used for rating the banks
not only in the training sample but also those which are not. In this case, the bank rating
is defined by the particular subinterval containing the LAD discriminant value.
16 Combinatorial Methods for Constructing Credit Risk Ratings 453
discrepancy counts (Table 16.5) of the bank ratings predicted for the testing sets.
The fact that on the average the difference between the Fitch and the LAD ratings is
only 0.98 is a very strong indicator of the LAD model’s stability and the absence of
overfitting.
The results presented above demonstrate that the LAD approach can be used to
derive rating models with varying granularity levels:
• A binary classification model to be used for the preapproval operations
• A model with the same granularity as the benchmarked rating model
• A model with higher discrimination power, that is, with higher granularity than
that of the benchmarked rating model, to allow the bank to refine its pricing
policies and the allocation of regulatory capital
We show that the LAD model cross-validates extremely well and therefore is
highly generalizable and could therefore be used by financial institutions to develop
internal, Basel-compliant rating models.
The availability and processing of data have been major obstacles in the way of
using credit risk rating models. Until recently, many banks did not maintain such
16 Combinatorial Methods for Constructing Credit Risk Ratings 455
datasets and were heavily dependent on qualitative judgments. It is only after the
currency crises of the 1990s and the requirements imposed by the Basel Accord that
financial institutions have seen an incentive in collecting the necessary data and
maintaining the databases.
The move towards a heavier reliance on rating models is based on the assump-
tion that models produce more consistent ratings and that, over the long haul,
operating costs will diminish since less labor will be required to produce ratings.
The model proposed in this chapter will reinforce the incentives to develop and rely
upon credit risk models in bank operations due to the following:
• The accuracy and predictive ability of the proposed model will guarantee
dependable ratings.
• Its parsimony will alleviate the costs of extracting and maintaining large
datasets.
• It will result in leaner loan approval operations and faster decisions and could
thus reduce the overall operating costs.
Country risk ratings have critical importance in the international financial markets
since they are the primary determinants of the interest rates at which countries can
obtain credit.5 There are numerous examples of countries having to pay higher rates
on their borrowing following their rating downgrade, an often cited example being
Japan. As mentioned above, another critical aspect of country risk ratings concerns
their influence on the ratings of national banks and companies that would make
them more or less attractive to foreign investors. That is why the extant literature
calls country risk ratings the “pivot of all other country’s ratings” (Ferri et al. 1999)
and considers them the credit risk ceiling for all obligors located in a country
(Eliasson 2002; Mora 2006). Historical record shows the reluctance of raters to give
a company a higher credit rating than that of the sovereign where the company
operates. Contractual provisions sometimes prohibit institutional investors from
investing in debt rated below a prescribed level. It has been demonstrated (Ferri
et al. 1999) that variations in sovereign ratings drastically affect the ratings of banks
operating in low-income countries, while the ratings of banks operating in high-
income countries (Kaminsky and Schmukler 2002; Larrain et al. 1997) do not
depend that much on country ratings. Banks, insurance companies, and public
institutions frequently assess the amount of exposure they have in each country
and establish lending limits taking into account the estimated level of country risk.
Credit managers in multinational corporations have to assess evolving conditions in
foreign countries in order to decide whether to request letters of credit for particular
5
The presentation in this section is based on Hammer et al. (2006, 2012).
456 A. Kogan and M.A. Lejeune
transactions. Country risk estimates and their updates are utilized on a real-time
basis by multinational corporations facing constant fluctuation in international
currency values and difficulties associated with moving capital and profits across
national boundaries. Financial institutions have to rely on accurate assessments of
credit risk to comply with the requirements of the Basel Bank for International
Settlements. Feeling the pressure to get higher ratings could lead to fraud attempts.
A notorious case was Ukraine’s attempt to obtain IMF credits through misleading
reporting of its reserve data on foreign exchanges.
The existing literature on country risk, defined by Bourke and Shanmugam
(1990) as “the risk that a country will be unable to service its external debt due to
an inability to generate sufficient foreign exchange,” recognizes both financial/
economic and political components of country risk. There are two basic approaches
to the interpretation of the reasons for defaulting. The first one is the debt-service
capacity approach which considers the deterioration of solvency of a country as
preventing it from fulfilling its commitments. This approach views country risk as
a function of various financial and economic country parameters. The second one is
the cost-benefit approach which considers a default on commitments or
a rescheduling of debt as a deliberate choice of the country. In this approach the
country accepts possible long-term negative effects (e.g., the country’s exclusion
from certain capital markets (Reinhart 2002) as preferable to repayment). Being
politically driven, this approach includes political country parameters in addition to
the financial and economic ones in country risk modeling (Brewer and Rivoli 1990,
1997; Citron and Neckelburg 1987).
16.4.2 Data
16.4.2.1 Ratings
We analyze in this chapter Standard & Poor’s foreign currency country ratings, as
opposed to the ratings for local currency debt. The former is the more important
problem, since the sovereign government has usually a lower capacity to repay
external (as opposed to domestic) debt, and as an implication, the international
bond market views foreign currency ratings as the decisive factor (Cantor and
Packer 1996). This is manifested by much higher likelihood for international
investors to acquire foreign currency obligations rather than domestic ones. In
evaluating foreign currency ratings, one has to take into account not only the
economic factors but also the country intervention risk, that is, the risk that
a country imposes, for example, exchange controls or a debt moratorium. The
evaluation of local currency ratings need not take into account this country
intervention risk.
Table 16.16 lists the different country risk levels used by S&P’s and also pro-
vides descriptions associated with these labels. A rating inferior to BB+ indicates
that a country is non-investment grade (speculative). A rating of CCC+ or lower
indicates that a country presents serious default risks. BB indicates the least degree
of speculation and CC the highest. The addition of a plus or minus sign modifies the
16 Combinatorial Methods for Constructing Credit Risk Ratings 457
rating (between AA and CCC) to indicate relative standing within the major rating
category. These subcategories are treated as separate ratings in our analysis.
In the next subsections, we derive two new combinatorial models for country risk
rating by reverse engineering and “learning” from past S&P’s ratings. The models
are developed using the novel combinatorial-logical technique of logical analysis of
data which derives a new rating system only from the qualitative information
representing pairwise comparisons of country riskiness. The approach is based on
a relative creditworthiness concept, which posits that the knowledge of (pre)order
of obligors with respect to their creditworthiness should be the sole source for
creating a credit risk rating system. Stated differently, only the order relation
between countries should determine the ratings. Thus, the inference of a model
for the order relation between countries is the objective of this study. This is in
perfect accordance with the general view of the credit risk rating industry. Altman
and Rijken (2004) state that the objective of rating agencies is “to provide an
accurate relative (i.e., ordinal) ranking of credit risk,” which is confirmed by
Fitch ratings (2006) saying that “Credit ratings express risk in relative rank order,
which is to say they are ordinal measures of credit.” Bhatia (2002) adds that:
“Although ratings are measures of absolute creditworthiness, in practice, the ratings
exercise is highly comparative in nature. . . On one level, the ratings task is one of
continuously sorting the universe of rated sovereigns – assessed under one uniform
set of criteria – to ensure that the resulting list of sovereigns presents a meaningful
global order of credit standing. On another level, the sorting task is constrained by
a parallel need to respect each sovereign’s progression over time, such that shifting
peer comparisons is a necessary condition – but not a sufficient one – for upward or
downward ratings action.”
A self-contained model of country risk ratings can hardly be developed by
standard econometric methods since the dataset contains information only
about 69 countries, each described by 12 explanatory variables. An alternative at
hand possible with combinatorial techniques is to examine the relative riskiness of
one country compared to another one, rather than modeling the riskiness of each
individual country. This approach has the advantage of allowing the modeling to be
based on a much richer dataset (2,346 pairs of countries), which consists of the
comparative descriptions of all pairs of countries in the current dataset.
The models utilize the values of nine economic and three political variables
associated to a country, but do not use directly or indirectly previous years’ ratings.
This is a very important feature, since the inclusion of information from past ratings
(lagged ratings, rating history) does not allow the construction of a self-contained
rating system and does not make possible to rate the creditworthiness of not-yet-
rated countries. We refer to Hammer et al. (2006, 2011) for a more detailed
discussion of the advantages of building a non-recursive country risk rating system.
Moreover, the proposed LAD models completely eliminate the need to view the
ratings as numbers. Section 16.4.3.1 presents the common features of the two
developed LAD models, while Sects. 16.4.3.2 and 16.4.3.3 discuss the specifics
of the LAD-based Condorcet ratings and the logical rating scores, respectively.
16 Combinatorial Methods for Constructing Credit Risk Ratings 459
16.4.3.1 Commonalities
Pairwise Comparison of Countries: Pseudo-observations
Every country i 2 I ¼ {1, . . ., 69} in this study is described by the 13-dimensional
vector Ci, whose first component is the country risk rating given by Standard &
Poor’s, while the remaining 12 components specify the values of the nine eco-
nomic/financial and of the three political variables. A pseudo-observation Pij,
associated to every pair of countries i, j 2 I, provides in a way specified below
a comparative description of the two countries.
Every pseudo-observations is also described by a 13-dimensional vector. The
first component is an indicator which takes the value 1 if the country i in the pseudo-
observation Pij has a higher rating than the country j, –1 if the country j has a higher
rating than the country i, and 0 if the two countries have the same rating. The other
components k, k ¼ 2, . . ., 13 of the pseudo-observation Pij[k] are derived by taking
the differences of the corresponding components of Ci and Cj:
to show that the achieved high classification accuracy is not due to overfitting. This
technique removes from the dataset one observation at a time, learns a model from
all the remaining observations, and evaluates the resulting model on the removed
observation; then it repeats all these steps for each observation in the dataset. If on
the average the predicted evaluations are “close to” the actual ones, then the model
is not affected by overfitting.
In the case at hand, it would not be statistically sound to implement the jackknife
technique in a straightforward way because of the dependencies among pseudo-
observations (Hammer et al. 2006). Indeed, even after explicitly eliminating a single
pseudo-observation Pij from the dataset, it would still remain in the dataset implicitly,
since Pih + Phj + Pij. This problem can be resolved by modifying the above described
procedure so that at each step, instead of just removing a single pseudo-observation Pij,
all the pseudo-observations which involve a particular country i are removed. Then, the
LAD discriminant is derived on the basis of the remaining pseudo-observations and
used to evaluate the relative preferences for every removed pseudo-observation,
resulting in a row of relative preferences of all pseudo-observations Pij which involve
the country i. This modified procedure is repeated for every country in the dataset, and
the obtained rows are combined into a matrix of relative preferences denoted by DJK.
The absence of overfitting is indicated by a very high correlation level of 96.48 %
between the matrix DJK and the original relative preference matrix D.
To further test for overfitting, we use the obtained matrix of relative preferences
DJK to classify the dataset of 4,360 pseudo-observations. The results of this classifi-
cation are presented in Table 16.8, and the overall classification quality of the LAD
model (according to formula (16.2)) is 95.10 %. The results in Table 16.8 are virtually
identical to those shown in Table 16.7, thus proving the absence of overfitting.
Since the derived LAD model does not suffer from overfitting, it is tempting to
interpret the signs of relative preferences as indicative of rating superiority and
infer that a positive value for D(Pi,j) indicates that country i is more creditworthy
than country j, while a negative one for D(Pi,j) justifies the opposite conclusion.
However, this naı̈ve approach towards relative rating superiority ignores the poten-
tial noise in the data and in relative preferences, which would make it difficult to
transform classifications of pseudo-observations to a consistent ordering of coun-
tries by their creditworthiness. It is shown in Hammer et al. (2006) that the
relationship based on the naı̈ve interpretation of the relative preferences can violate
the transitivity requirement of an order relation and therefore does not provide
a consistent partially ordered set of countries.
The following section is devoted to overcoming this issue by relaxing the overly
constrained search for (possibly nonexistent) country ratings whose pairwise
16 Combinatorial Methods for Constructing Credit Risk Ratings 463
orderings are in precise agreement with the signs of relative preferences. Instead,
we utilize a more flexible search for a partial order on the set of countries, which
satisfies the transitivity requirements and approximates well the set of relative
preferences.
define the external preference of country i over country j with respect to country k, let
X
Sij ðkÞ
j2I
Sij ¼ (16.6)
jI j
• In all the other cases, countries i and j are said to be not comparable; this can be
due to the lack of evidence, or to conflicting evidence about the dominance of
i over j.
Note that the larger the value of is, the stronger the conditions are, and the
fewer pairs of countries are comparable. If is sufficiently large, then the domi-
nance relationship is transitive and is a partial order, since it becomes progressively
sparser until being reduced to empty. On the other hand, if is small (and in
particular for ¼ 0), then the dominance relationship becomes denser, but is not
necessarily transitive.
We are looking for a “rich” dominance relationship (applying to as many
country pairs as possible) which is transitive. Formally, our objective is to maxi-
mize the number of comparable country pairs subject to preserving the transitivity
of the dominance relationship.
The richest dominance relationship defined by the two conditions (16.8) and
(16.9) in the case of ¼ 0 will be called the base dominance relationship. If i and
j are any two countries comparable in the base dominance relationship, then it is
possible to calculate the smallest value of the parameter ij ¼ j Sij/sij j such that the
i and j are not comparable in any dominance relationship defined by a parameter
Table 16.9 Correlation levels between LAD and canonical relative preferences
D dS&P dM dII dLRS
D 100 % 93.21 % 92.89 % 91.82 % 97.57 %
dS&P 93.21 % 100 % 98.01 % 96.18 % 95.54 %
dM 92.89 % 98.01 % 100 % 96.31 % 95.20 %
dII 91.82 % 96.18 % 96.31 % 100 % 94.11 %
dLRS 97.57 % 95.54 % 95.20 % 94.11 % 100 %
16 Combinatorial Methods for Constructing Credit Risk Ratings 465
value exceeding or equal to ij. This calculation is based on the fact that given
a value of , it is possible to check in polynomial time (Tarjan 1972) whether the
corresponding dominance relationship is transitively closed. Then the algorithm
that determines in polynomial time the minimum value * for which the
corresponding dominance relationship is still transitive, sorts at most j I j2 numbers
ij in ascending order, and then checks one by one the transitivity of the
corresponding dominance relationships. When the dominance relationship becomes
transitive for the first time, the algorithm stops and outputs * equal to the
corresponding value of the parameter ij. This study utilizes this value * and the
corresponding dominance relationship between countries, called here the logical
dominance relationship and denoted by the subscript of LAD (e.g., ).
LAD
The definition of dominance relationship between countries on the basis of
average external preferences bears some similarities to the so-called “column
sum methods” (Choo and Wendley 2004) utilized in reconciling inconsistencies
resulting from the application of pairwise comparison matrix methods.
remaining set of countries are assigned the lowest remaining level until every
country is assigned a level denoted by PEi.
The method utilized above to construct OE and PE is known as the Condorcet
method. It represents a specific type of voting system (Gehrlein and Lepelley
1998), and it is often used to determine the winner of an election. The Condorcet
winner(s) of an election is generally defined as the candidate(s) who, when
compared in turn with every other candidate, is preferred over each of them.
Given an election with preferential votes, one can determine weak Condorcet
winners (Ng et al. 1996) by constructing the Schwartz set as the union of all
possible candidates such that (i) every candidate inside the set is pairwise
unbeatable by any other candidate outside the set (ties are allowed) and (ii) no
proper subset of the set satisfies the first property.
The Schwartz set consists exactly of all weak Condorcet winners. The weak
Condorcet losers are the reverse of the weak Condorcet winners, that is, those
losing pairwise to every other candidate. OE assigns the highest level to those
countries that are the weak Condorcet winners, that is, better than or incomparable
with every other country. PE assigns the lowest level to those countries that are
the weak Condorcet losers, that is, worse than or are incomparable with every
other country. Note that both OE and PE have the minimum possible number of
levels. Indeed, in a directed graph whose vertices are the countries and whose arcs
represent comparable countries in the dominance relationship, the length of the
longest directed path bounds from below the number of levels in any linear
preorder extending the dominance relationship. This length equals the number
of levels in OE and PE.
The values of the b’s providing the best L2 approximation of the D’s can be
found by solving the following multiple linear regression problem:
X
DðpÞ ¼ bk xk ðpÞ þ eðpÞ, (16.12)
k2I
8
< 1, for k ¼ i
where p ¼ {(i, j)ji, j 2 I, i 6¼ j} and xk ði; jÞ ¼ 1, for k ¼ j .
:
0, otherwise
The logical rating scores bk obtained by fitting the regression model are given in
Column 8 of Table 16.16.
In this section, we analyze the results obtained with the proposed rating systems.
The evaluation of the results involves the analysis of the following:
• The relative preferences
• The partially ordered set (i.e., the logical dominance relationship)
• The LRS scores and the Condorcet ratings (extremal linear extensions)
with respect to the rating system of S&P’s as well as that of Moody’s and The
Institutional Investor.
partially ordered sets associated with The Institutional Investor’s scores and the
logical rating scores. The partially ordered sets corresponding to The Institutional
Investor’s scores and the logical rating scores are obtained as follows:
i j if si sj > y
i≺j if si sj < y
ijjj otherwise
where y is the positive number chosen so as to obtain a partially ordered set of the
same density as the dominance relationship and si represents the numerical score
given to country i by the respective rating system. The incomparability between two
countries means, for S&P’s and Moody’s, that the two countries are equally
creditworthy, while for the logical dominance relationship, it means that the
evidence about the relative creditworthiness of the two countries is either missing
or conflicting. The concept of density of a partially ordered set is defined in
Hammer et al. (2006) and represents the extent to which a partial order on a set
of countries differentiates them by their creditworthiness.
To assess the extent to which the preference orders agree with each other, the
following concepts are introduced (Hammer et al. 2006). Given a pair of countries
(i, j), two partially ordered sets are:
• In concordance if one of the following relations i j, i ≺ j, or i || j holds for both
partially ordered sets
• In discordance if i j for one of the partially ordered sets and i ≺ j for the other one
• Incomparable otherwise, that is, if i || j for one of the partially ordered sets, and
either i j or i ≺ j for the other partially ordered set.
We measure the levels of concordance, discordance, or incomparability between
two partially ordered sets by the fractions of pairs of countries for which the two
partially ordered sets are, respectively, in concordance, discordance, or incompa-
rable. Hammer et al. (2006) have shown that there is a very high level of agreement
between the following:
• The logical dominance relationship and the preference orders associated with
S&P’s and Moody’s ratings and The Institutional Investor scores
• The logical dominance relationship and the logical rating scores
subject to
S Si M ai , for all i 2 I (16.13)
i
where ai takes the value 1 if the S&P’s rating of country i must be modified and the
value 0 if otherwise, Si is the original S&P’s rating of country i, S*i is the adjusted
S&P’s rating of country i, and M is a sufficiently large positive number (e.g., M ¼ 22).
The optimal solution of Eq. 16.13 shows that the 0 % discordance level can be
achieved by adjusting the S&P’s ratings of nine countries: France, India, Japan,
Colombia, Latvia, Lithuania, Croatia, Iceland, and Romania. To check the rele-
vance of the proposed rating adjustments, we examine the S&P’s ratings posterior
to December 1998. We observe that Romania, Japan, and Columbia’s S&P’s
ratings have been modified in the direction suggested by our model. More precisely,
Columbia was downgraded by S&P’s twice, moving from BBB in December
1998 to BB+ in September 1999 and then to BB in March 2000. Japan was
downgraded to AA+ in February 2001 and AA in November 2001. Romania was
upgraded to B in June 2001. The S&P’s rating of the other countries (Iceland,
France, India, Croatia, Latvia, and Lithuania) has remained unchanged.
Relative Preferences
Table 16.11 shows that the LAD relative preferences are highly correlated with
those of the S&P’s rating system, as well as with the logical rating scores. The LAD
relative preferences and the LRS were obtained by applying to the 1999 data the
models derived from the 1998 data.
The high levels of pairwise correlations between the S&P’s 1999 ratings, the
relative preferences given by the LAD discriminant, and the canonical relative
preferences corresponding to LRS show that the LRS model has a very strong
temporal stability and indicate its high predictive power.
Preorders
The logical dominance relationship is compared to the 1999 preference order of the
S&P’s and the partially ordered set associated with the logical rating scores.
Table 16.12 displays the concordance, discordance, and incomparability levels
16 Combinatorial Methods for Constructing Credit Risk Ratings 471
between the logical dominance relationship, the preference order of S&P’s, and the
partial order associated with the logical rating scores and underlines their very
strong level of agreement.
Hammer et al. (2007b) say that there is a discrepancy between S&P’s rating RSP j
and the logical
rating score if
• Jamaica’s first S&P’s rating (1999) was B. Its OE/PE levels (OE ¼ 3, PE ¼ 2)
are identical to these of Paraguay, Brazil, the Dominican Republic, and Bolivia,
which had 1999 S&P’s ratings of B, B+, B+, and BB-, respectively.
• The first S&P’s rating for Papua New Guinea was B+. Its OE/PE levels (OE ¼ 3,
PE ¼ 3) are the same as those of Peru and Mexico, which both had 1999 S&P’s
ratings of BB.
• Ecuador’s OE/PE levels (OE ¼ 3, PE ¼ 2) are the same as those of Paraguay, Brazil,
the Dominican Republic, and Bolivia, which had 1999 S&P’s ratings of B, B+, B+,
and BB-, respectively. Interestingly, while the initial S&P’s rating of Ecuador was
SD (in July 2000), it was upgraded in August 2000 (1 month later) to B.
The striking similarity between the initial S&P’s rating of each of the four
countries discussed above and the S&P’s ratings of those countries which have
the same OE/PE levels validates the proposed model, indicating its power to predict
the creditworthiness of previously unrated countries.
LRS Approach
The LAD discriminant, which does not involve in any way the previous years’
S&P’s ratings, allows the rating of previously unrated countries in the following
way. First, we construct all the pseudo-observations involving the new countries to
be evaluated. Second, we calculate the relative preferences for these pseudo-
observations, and we add the resulting columns and rows to the matrix of relative
preferences. Third, we determine the new LRS for all the countries (new and old) by
running the multiple linear regression model (16.12). Fourth, we apply the linear
transformation defined above to the LRS so that the transformed LRS and the
S&P’s ratings are on the same scale.
The evaluation of the ability of LRS to accurately predict S&P’s ratings is
carried out by comparing the predicted LRS (obtained as described above) and
the S&P’s ratings (when they first become available). We compute the confidence
intervals (16.14) for the transformed LRS of four countries never rated by S&P’s by
December 1998. The predictions for Guatemala, Jamaica, and Papua New Guinea
correspond perfectly to the first time (subsequent) S&P’s ratings. The comparison
between the LRS and the first (July 2000) S&P’s rating (SD) to Ecuador shows that
S&P’s rated it too harshly, since 1 month later S&P’s raised its rating to B-,
justifying the LRS prediction.
474 A. Kogan and M.A. Lejeune
16.5 Conclusions
performing cross-validation and applying the LAD model derived by using the
banks in the training set to those in the testing one.
This study also shows that the LAD-based approach to reverse engineering bank
ratings provides a model that is parsimonious and robust. This approach allows to
derive rating models with varying levels of granularity that can be used at different
stages in the credit granting decision process and can be employed to develop
internal rating systems that are Basel 2 compliant. Besides impacting the credit risk
of a financial institution, the use of the generalizable and accurate credit risk rating
system proposed here will also be critical in mitigating the financial institution’s
operational risk due to breakdowns in established processes and risk-management
operations or to inadequate process mapping within business lines. In particular, the
reliance on such risk rating system will reduce the losses due to mistakes made in
executing transactions, such as settlement failures, failures to meet capital regula-
tory requirements, or untimely debt collections, or the losses due to the offering of
inappropriate financial products or credit conditions, or giving incorrect advice to
counterparty.
The evaluation of the creditworthiness of countries is also of utmost importance,
since the country’s risk rating is generally viewed as the upper bound on the rating
that entities within a given country can be assigned. This study proposes an LAD
methodology for inducing a credit risk system from a set of country risk rating
evaluations. It uses nine economic and three political variables to construct the
relative preferences of countries on the basis of their creditworthiness. Two
methods are then developed to construct countries’ credit rating systems on the
basis of their relative creditworthiness. The first one is based on extending the
preorder of countries using the Condorcet voting technique and provides two rating
systems (weak Condorcet winners and losers), while the second one uses linear
regression to determine the logical rating scores.
The proposed rating systems correlate highly with those of the utilized rating
system (S&P’s) and those of other rating agencies (Moody’s and The Institutional
Investor) and are shown to be stable, having an excellent classification accuracy
when applied to the following years’ data or to the ratings of previously unrated
countries. Rating changes implemented by the S&P’s in subsequent years have
resolved most of the (few) discrepancies between the constructed partially ordered
set and S&P’s initial ratings. This study provides new insights on the importance of
variables by supporting the necessity of including in the analysis, in addition to
economic variables, also political variables (“political stability”), and by identify-
ing “financial depth and efficiency” as a new critical factor in assessing
country risk.
The rating systems proposed here for banks as well as countries are as follows:
• Avoid overfitting as attested by the back-testing analysis (i.e., extremely high
concordance between in- and out-of-sample rating predictions calculated using
the k-folding and jackknife cross-validation methods).
• Distinguish themselves from the rating models in the existing literature by their
self-contained nature, that is, by their non-reliance on any information derived
from lagged ratings. Therefore, the high level of correlation between predicted
476 A. Kogan and M.A. Lejeune
and actual ratings cannot be attributed to the reliance on lagged ratings and is
a reflection of the predictive power of the independent variables included in
these models. An important advantage of the non-recursive nature of the pro-
posed models is their applicability to not-yet-rated obligors.
The scope of the proposed methodology extends beyond the rating problems
discussed in this study and can be used in many other contexts where ratings are
relevant. The proposed methodology is applicable in the general case of inferring an
objective rating system from archival data, given that the rated objects are charac-
terized by vectors of attributes taking numerical or ordinal values.
Appendix
Republic
Denmark AA+ 20 20 84.7 15 15 0.048 AA+ 20 15 15 0.0492
Dominican B+ 8 10 28.1 3 2 0.3568 B+ 8 3 2 0.3431
Rep
Egypt BBB 12 11 44.4 6 5 0.2915 BBB 12 6 6 0.3067
El Salvador BB 10 12 31.2 4 4 0.3379 BB+ 11 5 4 0.3301
Estonia BBB+ 14 14 42.8 9 8 0.2518 BBB+ 14 8 8 0.245
Finland AA 19 21 82.2 14 14 0.064 AA+ 20 14 13 0.0458
477
(continued)
478
References
Afonso, A. (2003). Understanding the determinants of sovereign debt ratings: Evidence for the two
leading agencies. Journal of Economics and Finance, 27(1), 56–74.
Afonso, A., Gomes, P., & Rother, P. (2007). What “hides” behind sovereign debt ratings?
(European Central Bank Working Paper Series 711). https://www.repository.utl.pt/bitstream/
10400.5/2134/1/ecbwp711.pdf
Alexe, S. (2002). Datascope – A new tool for logical analysis of data. DIMACS Mixer Series.
DIMACS, Rutgers University.
Alexe, G., Alexe, S., Bonates, T. O., & Kogan, A. (2007). Logical analysis of data – The vision of
Peter L. Hammer. Annals of Mathematics and Artificial Intelligence, 49, 265–312.
Altman, E. I., & Rijken, H. A. (2004). How rating agencies achieve rating stability. Journal of
Banking and Finance, 28, 2679–2714.
Barr, R. S., & Siems, T. F. (1994). Predicting bank failure using DEA to quantify management
quality. Federal Reserve Bank of Dallas Financial Industry Studies, 1, 1–31.
Basel Committee on Banking Supervision. (2001). The internal ratings based approach. http://
www.bis.org/publ/bcbsca05.pdf
Basel Committee on Banking Supervision. (2004). Bank failures in mature economies (Working
Paper 13).
Basel Committee on Banking Supervision. (2006). International convergence of capital measure-
ment and capital standards: A revised framework (Basel II).
Bhatia, A.V. (2002). Sovereign credit risk ratings: An evaluation (IMF Working Paper WP/03/
170).
Boros, E., Hammer, P. L., Ibaraki, T., & Kogan, A. (1997). Logical analysis of numerical data.
Mathematical Programming, 79, 163–190.
Boros, E., Hammer, P. L., Ibaraki, T., Kogan, A., Mayoraz, E., & Muchnik, I. (2000). An
implementation of logical analysis of data. IEEE Transactions on Knowledge and Data
Engineering, 12(2), 292–306.
Bouchet, M. H., Clark, E., & Groslambert, B. (2003). Country risk assessment: A guide to global
investment strategy. Chichester: Wiley.
Bourke, P., & Shanmugam, B. (1990). An introduction to bank lending (Addison-Wesley Business
Series). Sydney: Addison-Wesley.
Brewer, T. L., & Rivoli, P. (1990). Politics and perceived country creditworthiness in international
banking. Journal of Money, Credit and Banking, 22, 357–369.
Brewer, T. L., & Rivoli, P. (1997). Political instability and country risk. Global Finance Journal,
8(2), 309–321.
Cantor, R., & Packer, F. (1996). Determinants and impact of sovereign credit ratings. FRBNY
Economic Policy Review, 2, 37–53.
Choo E.U., & Wedley W.C., (2004). A Common Framework for Deriving Preference Values from
Pairwise Comparison Matrices. Computers and Operational Research 31, 893–908.
Citron, J. T., & Neckelburg, G. (1987). Country risk and political instability. Journal of Develop-
ment Economics, 25, 385–395.
Crama, Y., Hammer, P. L., & Ibaraki, T. (1988). Cause-effect relationships and partially defined
boolean functions. Annals of Operations Research, 16, 299–326.
Curry, T., & Shibut, L. (2000). The Cost of the Savings and Loan Crisis: Truth and Consequences.
FDIC Banking Review, 2(2), 26–35.
Czyzyk, J., Mesnier, M. P., & Moré, J. J. (1998). The NEOS server. IEEE Computer Science
Engineering, 5(3), 68–75.
de Servigny, A., & Renault, O. (2004). Measuring and managing credit risk. New York:
McGraw-Hill.
Eliasson, A. (2002). Sovereign credit ratings (Working Papers 02-1). Deutsche Bank.
Ferri, G., Liu, L.-G., & Stiglitz, J. (1999). The procyclical role of rating agencies: Evidence from
the east asian crisis. Economic Notes, 3, 335–355.
482 A. Kogan and M.A. Lejeune
Fitch Ratings. (2001). Fitch simplifies bank rating scales. Technical Report.
Fitch Ratings. (2006). The role of support and joint probability analysis in bank ratings. Fitch
Special Report.
Galindo, J., & Tamayo, P. (2000). Credit risk assessment using statistical and machine
learning/basic methodology and risk modeling applications. Computational Economics,
15, 107–143.
Gehrlein, W. V., & Lepelley, D. (1998). The condorcet efficiency of approval voting and the
probability of electing the condorcet loser. Journal of Mathematical Economics, 29, 271–283.
Hammer, P. L. (1986). Partially defined boolean functions and cause-effect relationships. Inter-
national conference on multi-attribute decision making via OR-based expert systems. Passau:
University of Passau.
Hammer, P. L., Kogan, A., & Lejeune, M. A. (2006). Modeling country risk ratings using partial
orders. European Journal of Operational Research, 175(2), 836–859.
Hammer, P. L., Kogan, A., & Lejeune, M. A. (2011). Reverse-engineering country risk ratings:
Combinatorial non-recursive model. Annals of Operations Research, 188, 185–213.
Hammer, P. L., Kogan, A., & Lejeune, M. A. (2012). A logical analysis of banks’ financial
strength ratings. Expert Systems with Applications, 39(9), 7808–7821.
Haque, N. U., Kumar, M. S., Mark, N., & Mathieson, D. (1996). The economic content of
indicators of developing country creditworthiness. International Monetary Fund Working
Paper, 43(4), 688–724.
Haque, N. U., Kumar, M. S., Mark, N., & Mathieson, D. (1998). The relative importance of
political and economic variables in creditworthiness ratings. International Monetary Fund
Working Paper, 46, 1–13.
Hu, Y.-T., Kiesel, R., & Perraudin, W. (2002). The estimation of transition matrices for sovereign
credit ratings. Journal of Banking and Finance, 26(7), 1383–1406.
Huang, Z., Chen, H., Hsu, C.-J., Chen, W.-H., & Wu, S. (2004). Credit rating analysis with support
vector machines and neural networks: A market comparative study. Decision Support Systems,
37, 543–558.
Jain, K., Duin, R., & Mayo, J. (2000). Statistical pattern recognition: A review. IEEE Transactions
on Pattern Analysis and Machine Intelligence, 22, 4–37.
Kaminsky, G., & Schmukler, S. L. (2002). Emerging market instability: Do sovereign ratings
affect country risk and stock returns? World Bank Economic Review, 16, 171–195.
Kaufmann, D., Kraay A., & Zoido-Lobaton, P. (1999a). Aggregating governance indicators.
(World Bank Policy Research Department Working Paper, 2195).
Kaufmann, D., Kraay, A., & Zoido-Lobaton, P. (1999b). Governance matters (World Bank Policy
Research Department Working Paper, 2196).
Kunczik, M. (2001). Globalization: News media, images of nations and the flow of international
capital with special reference to the role of rating agencies. Paper presented at the IAMCR
Conference, Singapore, pp. 1–49.
Larrain, G., Reisen, H., & von Maltzan, J. (1997). Emerging market risk and sovereign credit
ratings (pp. 1–30). OECD Development Centre Working Papers 124, OECD Publishing.
Moody’s. (2006). Bank financial strength ratings: Revised methodology. Moody’s Global Credit
Research Report.
Mora, N. (2006). Sovereign credit ratings: Guilty beyond reasonable doubt? Journal of Banking
and Finance, 30(7), 2041–2062.
Morgan, D. P. (2002). Rating banks: Risk and uncertainty in an opaque industry. The American
Economic Review, 92(4), 874–888.
Ng, W.-Y., Choi, K.-W., & Shum, K.-H. (1996). Arbitrated matching: Formulation and protocol.
European Journal of Operational Research, 88(2), 348–357.
Poon, W. P. H., Firth, M., & Fung, H.-G. (1999). A multivariate analysis of the determinants of
Moody’s bank financial strength ratings. Journal of International Financial Markets, Institu-
tions & Money, 9(3), 267–283.
16 Combinatorial Methods for Constructing Credit Risk Ratings 483
Quenouille, M. (1949). Approximate tests of correlation in time series. Journal of the Royal
Statistical Society, Series B, 11, 18–84.
Reinhart, C. M. (2002). Default, currency crises, and sovereign credit ratings. World Bank
Economic Review, 16, 151–170.
Sarkar, S., & Sriram, R. S. (2001). Bayesian models for early warning of bank failures. Manage-
ment Science, 47(11), 1457–1475.
Sobehart, J. R., Keenan, S. C., & Stein, R. M. (2000). Benchmarking quantitative default risk
models: A validation methodology. New York: Moody’s Investors Service.
Stein, R. M. (2002). Benchmarking default prediction models: Pitfalls and remedies in model
validation. New York: Moody’s KMV.
Sy, A. N. R. (2004). Rating the rating agencies: Anticipating currency crises or debt crises?
Journal of Banking and Finance, 28(11), 2845–2867.
Tarjan, R. E. (1972). Depth-first search and linear graph algorithms. SIAM Journal on Computing,
1, 146–160.
Treacy, W. F., & Carey, M. S. (2000). Credit risk rating systems at large US banks. Journal of
Banking & Finance, 24(1–2), 167–201.
Dynamic Interactions Between
Institutional Investors and the Taiwan 17
Stock Returns: One-Regime and Threshold
VAR Models
Contents
17.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 486
17.2 Sample Data and Basic Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 491
17.3 Lead-Lag Relation Among the Three Groups of Institutional Investors
in the TSE . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 496
17.3.1 The Unrestricted VAR Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 496
17.3.2 The Structural VAR Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 500
17.3.3 The Threshold VAR Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 502
17.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 511
Appendix 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 513
The Unrestricted VAR Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 513
The Structural VAR Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 515
The Threshold VAR Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 516
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 518
B.-N. Huang
National Chung-Cheng University, Minxueng Township, Chiayi County, Taiwan
e-mail: ecdbnh@ccu.edu.tw
K. Hung (*)
Division of International Banking & Finance Studies, Texas A&M International University,
Laredo, TX, USA
e-mail: ken.hung@tamiu.edu
C.-H. Lee
National Kaohsiung University of Applied Sciences, Kaohsiung, Taiwan
e-mail: chlee@cc.kuas.edu.tw
C.W. Yang
Clarion University of Pennsylvania, Clarion, PA, USA
e-mail: yang@mail.clarion.edu; ecdycw@ccu.edu.tw
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 485
DOI 10.1007/978-1-4614-7750-1_17,
# Springer Science+Business Media New York 2015
486 B.-N. Huang et al.
Abstract
This paper constructs a six-variable VAR model (including NASDAQ returns, TSE
returns, NT/USD returns, net foreign purchases, net domestic investment compa-
nies (dic) purchases, and net registered trading firms (rtf) purchases) to examine:
(i) the interaction among three types of institutional investors, particularly to test
whether net foreign purchases lead net domestic purchases by dic and rtf (the
so-called demonstration effect); (ii) whether net institutional purchases lead market
returns or vice versa; and (iii) whether the corresponding lead-lag relationship is
positive or negative? The results of unrestricted VAR, structural VAR, and multi-
variate threshold autoregression models show that net foreign purchases lead net
purchases by domestic institutions and the relation between them is not always
unidirectional. In certain regimes, depending on whether previous day’s TSE
returns are negative or previous day’s NASDAQ returns are positive, we find
ample evidence of a feedback relation between net foreign purchases and net
domestic institutional purchases. The evidence also supports a strong positive-
feedback trading by institutional investors in the TSE. In addition, it is found that
net dic purchases negatively lead market returns in Period 4. The MVTAR results
indicate that net foreign purchases lead market returns when previous day’s
NASDAQ returns are positive and have a positive influence on returns.
Readers are well advised to refer to chapter appendix for detailed discussion
of the unrestricted VAR model, the structural VAR model, and the threshold
VAR analysis.
Keywords
Demonstration effect • Multivariate threshold autoregression model • Foreign
investment • Lead-lag relationship • Structural VAR • Block Granger causality •
Institutional investors • Domestic investment companies • Registered trading
firms • Qualified foreign institutional investors
17.1 Introduction
Table 17.1 Inflows and outflows of foreign capital investment in the TSE Unit: millions of
US dollars
Securities
investment Foreign institutional
companies investors Natural person Foreign total
Period Inflow Outflow Inflow Outflow Inflow Outflow Net inflow
1991 263 53 448 0 0 0 658
1992 57 61 447 17 0 0 426
1993 653 93 1,859 97 0 0 2,322
1994 451 207 2,279 634 0 0 1,889
1995 664 457 3,509 1,506 0 0 2,210
1996 565 477 6,213 3,881 334 8 2,747
1997/1 4 12 644 4,442 2,529 261 78 1,465
Cumulative amount 2,664 1,992 19,198 8,664 595 85 11,716
Source of data: Securities and Futures Commission, Ministry of Finance, Taiwan
process. First, it directed a number of local trust companies to issue investment funds
abroad for the local stock market. Second, qualified foreign investment companies
were allowed to invest in Taiwan’s stock market. Third, foreign individual as well as
institutional investors were permitted to directly participate in trading in the Taiwan
Stock Exchange Corporation (hereafter TSE). The first stage took effect from
September 1983 until December 29, 1990, when the second stage replaced the first
stage with a maximum investment limit of $2.5 billion. In response to the strong
demand to invest in the TSE, the ceiling was raised to $5 billion and to $7.5 billion in
August 1993 and March 1994, respectively. Foreign investments later increased
substantially (see Table 17.1), and as a result the ceiling was lifted entirely in 1996
except for a maximum investment limit to each individual stock.
The relatively slow pace in allowing foreign investment in Taiwan was the result
of ongoing debates between the Central Bank of China and the Securities and
Futures Commission of the Ministry of Finance regarding the stability of foreign
investment. The focus of the discussion was on the following three questions. First,
are there differences in the trading behaviors of different type of institutions? One
objective of opening up the domestic market to foreign investments is to utilize its
advantages in information acquisition, information processing, and trade execution
to improve the overall performance of local institutional investors.1 Second, will
1
This is because foreign institutions may well have better research teams and buy stocks according
to the fundamentals such as the firm’s future profitability. In contrast, local institutions and
individual investors usually choose stocks based on insider information or what the newspapers
write about. However, if stocks bought by foreign investors have a better performance than that of
local institutions and individual investors, the latter tend to buy the stocks bought by successful
foreign investors the previous day. Hence this gives rise to the so-called “demonstration effect.”
Such a concept is similar to the “herding” which Lakonishok et al. (1992) referred to as correlated
trading across institutional investors. Nonetheless, their definition is close to the contemporaneous
correlation rather than the lead-lag relation under study, and as such this leads to the demonstration
effect which we documented.
488 B.-N. Huang et al.
the trades of the three groups of classified institutional investors affect stock returns
of the TSE? Third, will the lifting of restrictions on foreign capital contribute to the
volatility of both the foreign exchange and stock markets in Taiwan?
The purpose of this paper is to provide answers to two of the aforementioned
issues: (1) Are there differences in the trading behaviors of different types of
institutions? Typically, we examine interaction among the three types of institu-
tions: qualified foreign institution investors (qfii), domestic investment companies
(dic), and registered trading firms (rtf). (2) Does institutional trading “cause” stock
returns or do institution investments follow movements in stock prices? The
distinction between this paper and previous literature lies in following aspects:
(i) Most previous studies uses low-frequency – yearly, quarterly, or at best weekly –
data (Nofsinger and Sias 1999; Cai and Zheng 2002; Karolyi 2002). In this paper we
employ daily data to help explore these issues in detail and to provide new evidence
on the short-term dynamics among institutional investors and stock returns.2
(ii) Unlike previous studies that used a bivariate VAR model (Froot et al. 2001),
we use a six-variable VAR model, which includes three types of institutional trades,
stock returns in the TSE, NT/USD exchange rate changes, and NASDAQ index
returns to test related relationships. (iii) Improving on the conventional linear VAR
analysis in the previous studies, we employ the threshold concept and split data into
two regimes based on whether previous trading day’s market returns are positive or
negative.
A number of studies have examined the relationship between investment flows
and stock returns. Brennan and Cao (1997) develop a theoretical model of interna-
tional equity flows that relies on the informational difference between foreign and
domestic investors. The model predicts that if foreign and domestic investors are
differently informed, portfolio flows between two countries will be a linear function
of the contemporaneous return on all national market indices. Moreover, if domes-
tic investors have a cumulative information advantage over foreign investors about
domestic securities, the coefficient of the host market return is expected to be
positive.
Nofsinger and Sias (1999) use US annual data to investigate the relationship
between stock returns and institutional and individual investors. They identify
a strong positive correlation between changes in institutional ownership and stock
returns measured over the same period. Their results suggest that (i) institutional
investors practiced positive-feedback trade more than individual investors and
(ii) institutional herding impacted price more than that by individual investors.
Nofsinger and Sias show that institutional herding is positively correlated with lag
returns and appears to be related to stock return momentum.
Choe et al. (1999) use order and trade data from November 30, 1996 to the end of
1997 to examine the impact of foreign investors on stock returns. They found strong
2
Froot et al. (2001) also use daily data, but they examine the behavior of capital flows across
countries. In addition, our models and approaches used in the estimation differ drastically from
theirs.
17 Dynamic Interactions 489
3
Lakonishok et al. (1992) refer to the positive-feedback trading or trend chasing as buying winners
and selling losers and the negative feedback trading or contrarian as buying losers and selling
winners. Cai and Zheng (2002) point out that feedback trading occurs when lagged returns act as
the common signal that the investors follow.
4
Karolyi (2002) reaches such a conclusion because there is little evidence of any impact of foreign
net purchases on future Nikkei returns or currency returns.
490 B.-N. Huang et al.
buy-sell imbalances are positively related to previous day’s returns and the institu-
tional buy-sell imbalances (previous day) are not associated with current return.
The results are consistent with the finding by Sias and Starks (1997) using US data.
Griffin et al. (2003) estimate a structural VAR with the contemporaneous returns in
the institutional imbalance equation and discover a strong contemporaneous rela-
tionship between daily returns and institutional buy-sell imbalances.
Kamesaka et al. (2003) use Japanese weekly investment flow data over 18 years
to investigate the investment patterns and performance of foreign investors, indi-
vidual investors, and five types of institutional investors. Not surprisingly, they find
individual investors perform poorly, while securities firms, banks, and foreign
investors perform admirably over the sample period.
Several related studies focus mainly on Taiwan’s stock market. Huang and Hsu
(1999) detect decreased volatility in the weighted TSE using Levene’s F-statistic
following market liberalization. Lee and Oh (1995), implementing a vector
autoregression (VAR) model, find a reduction in the explanatory power of macro-
economics variables. Wang and Shen (1999) indicate that foreign investments exert
a positive impact on the exchange rate with only a limited effect on the TSE. In
addition, by using the turnover rate as a proxy for non-fundamental factors and
earnings per share for fundamental factors within the framework of a panel data
model, Wang and Shen are able to identify that (i) the non-fundamental factors
impacted the returns of the TSE before market liberalization and (ii) both the
fundamental and non-fundamental factors exerted an impact following market
liberalization.
Lee et al. (1999) investigate interdependence and purchasing patterns among
institutional investors, large, and small individual investors. Their results, based on
15-min intraday transaction data (3 months for 30 companies), highlight the
important role played by large individual investors, whose trading affects not
only stock returns but also small individual investors. However, net buys (i.e., the
difference between total buy and total sell) by institutional investors have no effect
on the TSE returns, and vice versa.
The previous literature is predominantly focused on the relationship between
institutional trading and stock returns, rarely on the interaction among institutional
investors. For example, the majority of prior studies find evidence of positive-
feedback trading by institutions, with the exception of Froot et al. (2001), who
discover that in Latin America and emerging East Asian markets, the trading by
institutions positively predicts future returns. Karolyi (2002) also detects that
foreign investors in Japan are positive-feedback traders, while Japanese financial
institution and companies are contrarian investors.
Most of the studies to date on these issues have been on the USA and Japanese
markets despite that some of the literature gives scant attention to Taiwan’s stock
market. When investigating the related issues in large countries such as the USA
and Japan, the influence of the foreign sector on the domestic market could be
neglected; however, for a small country such as Taiwan, it should not be ignored.
This is because the electronics industry in Taiwan is closely connected to the US
companies listed on the NASDAQ. Ultimately, after examining the interaction
17 Dynamic Interactions 491
among institutional investors and the dynamic relationship between stock returns
and institutional investors, the conclusion may also be affected by whether returns
of domestic market are positive or negative.
To circumvent the above problems, this paper employs a six-variable VAR
model, which takes into account trades of three types of institutional investors
(qfii, dic, and rtf), foreign returns, domestic returns, and changes in the NT/USD
exchange rate to jointly test hypotheses under different market conditions. Using
daily data, we find that net foreign purchases lead net domestic purchases. How-
ever, such a relation is not unidirectional. Under certain conditions (either when
previous day’s TSE returns are negative or previous day’s NASDAQ returns are
positive), we identify a feedback relation between net foreign purchases and net
domestic purchases. It highlights the well-known argument in Taiwan regarding
foreign investors: The demonstration effect on domestic institutional investors is
not entirely correct. As for the lead-lag relation between market returns and
institutional trading, we find that in most cases market returns at least lead both
net foreign and dic purchases; however, market returns also lead net rtf purchases if
the relationship between contemporaneous returns and institutional trading is
considered. On the other hand, our results also indicate that net dic purchases
lead market returns and are negatively associated with market returns in the fourth
period. The MVTAR analysis shows that when previous day’s NASDAQ returns
are positive, net foreign purchases positively lead stock returns.
The remainder of this paper is organized as follows. Section 17.2 describes the
sample data and the basic statistics. Section 17.3 investigates the lead-lag relation
for three groups of institutional investors in order to explore the issue of whether
foreign investments give rise to demonstration effects in Taiwan’s stock market and
examines the relationship between institutional trading activity and stock returns of
the TSE. To further explore the interaction among three types of market partici-
pants, the sample is divided into two regimes based on either the sign of the market
returns or that of the NASDAQ index returns of previous trading day, respectively.
The last section provides a conclusion.
This paper employs daily data from December 13, 1995 to May 13, 2004 for a large
sample analysis.5 The variables considered include purchases (qfiibuy) and sales
(qfiisell) by qfii, purchases (dicbuy) and sales (dicsell) by dic, purchases (rtfbuy) and
sales (rtfsell) by rtf, TSE daily weighted stock index (pt), NASDAQ stock index
(naspt), and the NT/USD exchange rate (et). The data are from the Taiwan Eco-
nomic Journal (TEJ). The changes in the exchange rate and the logarithmic returns
on TSE and NASDAQ indices are defined as
5
The data began on December 13, 1995, since the inception of the TEJ.
492 B.-N. Huang et al.
Table 17.2 Unit root tests for the six time series
Test nasrt rt Det qfiibst dicbst rtfbst
PP 44.04* 42.76* 43.32* 28.47* 31.12* 29.90*
The sample period starts from December 13, 1995, to May 13, 2004, a total of 1989 observations.
qfiibuyt and qfiisellt ¼ purchases and sales by qualified foreign institutional investors, and
qfiibst ¼ qfiibuyt-qfiisellt. dicbuyt and dicsellt ¼ purchases and sales by domestic investment
companies, and dicbst ¼ dicbuyt-dicsellt. rtfbuyt and rtfsellt ¼ purchases and sales by registered
trading firms, and rtfbst ¼ rtfbuyt-rtfsellt. nasrt are the NASDAQ index returns. rt is the TSE index
return. Det is changes in the NT/USD exchange rate. qfiibst is net purchases by qfii. dicbst is net
purchases by dic, and rtfbst is net purchases by rtf. y denotes the level of the variable. Dy denotes
the first difference of the variable. * denotes statistical significance at 1 % level
Table 17.3 Summary statistics for the net institutional purchases, stock returns and NT/USD
currency returns
Mean Median Maximum Minimum Std. dev.
Det 0.0105 0.0000 3.4014 2.9609 0.3331
rt 0.0083 0.0438 8.5198 12.6043 1.7839
nasrt 0.0302 0.1300 13.2546 10.4078 2.0210
qfiibuyt 5,386.30 4,130 31,415 151 4,513.78
qfiisellt 4,675.17 3,601 44,000 74 4,027.41
qfiibst 711.11 372 19,408 23772 3,171.87
dicbuyt 3,191.66 2,924 14,980 192 1,643.79
dicsellt 3,306.13 3,096 11,854 141 1,556.99
dicbst 114.49 120 10,070 8,876 1,277.43
rtfbuyt 1,814.00 1,418 10,972 49 1,433.95
rtfsellt 1,837.15 1,504 18,024 32 1,384.79
rtfbst 23.15 36 6,379 11,177 934.55
For variable definitions, see Table 17.2
Net foreign purchases (qfiibst) are computed as the daily purchases (qfiibuyt) less
sales (qfiisellt) of Taiwan stocks by foreigners. Similarly, net dic purchases (dicbst)
are computed as the daily purchases (dicbuyt) less sales (dicsellt) of Taiwan stocks
by dic, and net rtf purchases (rtfbst) are computed as the daily purchases (rtfbuyt)
less sales (rtfsellt) of Taiwan stocks by rtf. The VAR analysis used here depends on
whether the time series are stationary; hence, a unit root test is to be performed in
advance to avoid spurious regression. The Phillips and Perron test is applied and the
results are illustrated in Table 17.2.
The Phillips and Perron test results indicate that all time series are statistically
significant at 1 % level. There is no further differencing needed before applying VAR.
Table 17.3 presents the summary statistics for the time series used in this paper.
17 Dynamic Interactions 493
The average percentage change in the exchange rate on the daily basis is
0.011 %; the average daily TSE return is 0.0083 %; and the average daily NASDAQ
return equals 0.0302 %. Overall, qfii are net purchasers on average and two other
domestic institutional investors are net sellers of equity over the sample period,
reflecting different trading strategies adopted by foreign and domestic institutional
investors. Such distinct trading activities among institutional investors can also be
seen in Fig. 17.1.
Figure 17.1 presents the cumulative net purchases and daily net purchases by
qfii, dic, and rtf and how they are associated with the TSE returns, NASDAQ
returns, and NT/USD exchange rate. Over the entire period, the cumulative net
purchases by qfii suggest an upward trend in general, while those of dic and rtf
tend to present a downward trend. Overall, the NASDAQ index is more volatile
than the TSE index (1995/12/13 ¼ 100), and it seems that there exists some
correlation between the two indices. During the Asia financial crisis in 1997, the
NT/USD exchange rate suffered a great upward swing (depreciation of New
Taiwan Dollar) followed by a slight downward slide in 1999 and then rose
again from 2002 onwards. Over the sample period, the volatility of net purchases
by foreigners seemed to have increased since 2002. As for the relationship
between net purchases by institutions and stock returns, no clear correlation
could be detected as shown in Fig. 17.1. To grasp a better understanding on
their linkages, the contemporaneous correlation of net purchases by the three
types of institutional investors, stock returns, and currency returns are displayed
in Table 17.4.
An inspection of Table 17.4 points out that returns on the NT/USD
exchange rate (currency returns) are negatively correlated with both the
TSE returns and net purchases by the three types of institutional investors,
especially by foreign investors. Such relations are very much in line with the
expectation. When stock prices rise following the influx of foreign capital,
the local currency is expected to appreciate to a degree and as such negative
correlations among them is expected. In addition, we find that returns on the
TSE and NASDAQ are positively correlated. It is noteworthy that there exists
a positive contemporaneous correlation between net purchases by the three
types of institutional investors and the TSE returns with the correlation
coefficients ranging from about 0.3 (qfiibst and rt) to 0.4419 (rtfbst and rt).
In short, this finding largely echoes the previous results (e.g., Froot
et al. 2001; Karolyi 2002).
The greatest correlation between institutional trading and NASDAQ
returns is that of rtfbst and nasrt at 0.0938 followed by dicbst and qfiibst,
respectively. Owing to the time difference, the TSE returns may be
influenced by NASDAQ index returns. If nasrt1 is used instead, a higher
correlations between nasrt1 and net purchases by institutions are found:
0.3441 for qfiibst, 0.2369 for dicbst, and 0.1428 for rtfbst, respectively. It
implies that previous day’s NASDAQ returns exert a greater impact on net
purchases by each institutional investor in the TSE than do the current
NASDAQ returns.
494 B.-N. Huang et al.
Index
1000 (95/12/13=100)
1) Cumulative Net qfii Purchase
NT Dollars 500
a 97/07/02 Nasdaq
98/06/30 400
Asia Flu
300
200
1600000 100
Taiwan Stock Index
1200000 0
800000
400000 Cumulative Net QFII Purchase
0
−400000
97/01/07 98/02/05 99/03/03 00/03/24 01/04/02 02/05/09 03/05/20
1000
NT Dollars NT/USD
2) Net qfii Purchase
40000 36
30000 a 35
20000 34
10000 33
0 32
−10000 31
−20000 Net QFII Purchase 30
−30000 29
NT/USD
−40000 28
−50000 27
97/01/07 98/02/05 99/03/03 00/03/24 01/04/02 02/05/09 03/05/20
Index
1000 1) Cumulative Net dic Purchase (95/12/13=100)
NT Dollars 500
b 97/07/02 Nas daq
98/06/30 400
Asia Flu
300
200
50000
100
0 Taiwan Stock Index
0
−50000
−100000
Cumulative Net DIC Purchase
−150000
−200000
−250000
97/01/07 98/02/05 99/03/03 00/03/24 01/04/02 02/05/09 03/05/20
Index
1000
NT Dollars 1) Cumulative Net rtf Purchase (95/12/13=100)
500
c 97/07/02 Nasdaq
98/06/30 400
Asia Flu
300
200
20000
Taiwan Stock Index 100
0
0
−20000
Cumulative Net RTF Purchase
−40000
−60000
−80000
97/01/07 98/02/05 99/03/03 00/03/24 01/04/02 02/05/09 03/05/20
Fig. 17.1 Trends of cumulative net purchase, net purchases by the three types of institutions,
Taiwan stock prices, NASDAQ index, and the NT/USD exchange rate. a-1 Cumulative net qfii
purchase, a-2 Net qfii purchase, b-1 Cumulative net dic purchase, b-2 Net dic purchase, c-1
Cumulative net rtf purchase, c-2 Net rtf purchase. Notes: qfii qualified foreign institutional
investors, dic domestic investment companies, rtf registered trading firms
496 B.-N. Huang et al.
Table 17.4 Correlation matrix of net purchases by institutions, stock returns, and exchange rate
changes
Det rt nasrt qfiibst dicbst rtfbst
Det 1.0000
rt 0.1338 1.0000
nasrt 0.0124 0.1322 1.0000
qfiibst 0.1270 0.2976 0.0482 1.0000
dicbst 0.0991 0.3750 0.0586 0.2400 1.0000
rtfbst 0.0767 0.4419 0.0938 0.3559 0.3779 1.0000
See also Table 17.3
Note that prices of many Taiwanese electronics securities are affected by the
NASDAQ returns and hence foreign portfolio inflows may induce fluctuations of
exchange rate. To investigate interactions emanated from the three types of insti-
tutional investors and the relationship between institutional trading activity and
stock returns in Taiwan, we employ a six-variable VAR model using the NASDAQ
returns (nasrt), currency returns (Det), TSE returns (rt), net foreign purchases
(qfiibst), net dic purchases (dicbst), and net rtf purchases (rtfbst) as the underlying
variables. We attempt to answer the issues pertaining to (i) the interaction among
trading activities of the three types of institutions and (ii) the relationship between
stock returns and institutional trading. First, we propose a six-variable unrestricted
VAR model shown below:
2 3 2 3
nasr t f11 ðLÞ f12 ðLÞ f13 ðLÞ f14 ðLÞ f15 ðLÞ f16 ðLÞ
6 De 7 6 7
6 t 7 6 f21 ðLÞ f22 ðLÞ f23 ðLÞ f24 ðLÞ f25 ðLÞ f26 ðLÞ 7
6 7 6 7
6 rt 7 6 f31 ðLÞ f32 ðLÞ f33 ðLÞ f34 ðLÞ f35 ðLÞ f36 ðLÞ 7
6 7¼6 7
6 qfiibs 7 6 f ðLÞ f ðLÞ f ðLÞ f ðLÞ f ðLÞ f ðLÞ 7
6 t7 6 41 42 43 44 45 46 7
6 7 6 7
4 dicbst 5 4 f51 ðLÞ f52 ðLÞ f53 ðLÞ f54 ðLÞ f55 ðLÞ f56 ðLÞ 5
rtfbst f61 ðLÞ f62 ðLÞ f63 ðLÞ f64 ðLÞ f65 ðLÞ f66 ðLÞ
2 3 2 3
nasr t1 e1t
6 De 7 6e 7
6 t1 7 6 2t 7
6 7 6 7
6 r t1 7 6 e3t 7
66 7þ6 7 (1)
7 6 7
6 qfiibst1 7 6 e4t 7
6 7 6 7
4 dicbst1 5 4 e5t 5
rtfbst1 e6t
17 Dynamic Interactions 497
where fij(L) is the polynomial lag of the jth variable in the ith equation.
To investigate the lead-lag relation among three types of institutional investors,
we
2 need to test the hypothesis3 that each off-diagonal element in the sub-matrix
f44 ðLÞ f45 ðLÞ f46 ðLÞ
4 f54 ðLÞ f55 ðLÞ f56 ðLÞ 5 is zero.
f64 ðLÞ f65 ðLÞ f66 ðLÞ
On the other hand, to determine whether the TSE returns of the previous day lead
net purchases by the three types of institutional investors,0 we test the hypothesis that
each polynomial lag in the vector ½ f43 ðLÞf53 ðLÞf63 ðLÞ is zero. Conversely, if we
want to determine whether previous day’s net purchases by institutional investors
lead current market returns, we test the hypothesis that each element in the vector
[f34(L)f35(L)f36(L)] is zero. Before applying the VAR model, an appropriate lag
structure needs to be specified. A 3-day lag is selected based on the Akaike
information criterion (AIC). Table 17.5 presents the lead-lag relation among the
six time2series using block exogeneity
3 tests.
f44 ðLÞ f45 ðLÞ f46 ðLÞ
The 4 f54 ðLÞ f55 ðLÞ f56 ðLÞ 5 block represents the potential interaction among
f64 ðLÞ f65 ðLÞ f66 ðLÞ
three types of institutional investors. The results indicate that net foreign purchases lead
net dic purchases and the dynamic relationship between these two variables can be
provided by the impulse response function (IRF) in Fig. 17.2a. Clearly, a one-unit
standard error shock to net foreign purchases leads to an increase in net dic purchases,
but this effect dissipates quickly by Period 2. Figure 17.2b, c indicate a feedback
relation between net purchases by qfii and rtf. A one-unit standard error shock to net
foreign purchases results in a positive response to net rtf purchases over the next two
periods, which become negative in Period 3, followed by a positive response again after
Period 4. Furthermore, a one-unit standard error shock to net rtf purchases also gives rise
to an increase in net foreign purchases, which decays slowly over ten-period horizon.
Figure 17.2d shows that net purchases by dic lead net rtf purchases. A one-unit
standard error shock to net dic purchases produces an increase in net rtf purchases in the
first three periods and then declines thereafter. Overall, these impulse responses suggest
that previous day’s net foreign purchases exert a noticeable impact on net rtf purchases,
while previous day’s net rtf purchases also has an impact on net foreign purchases. It
implies that not only do foreign capital flows affect the trading activity of domestic
institutional investors but also the relation is not unidirectional. To be specific, there is
a feedback relation between net rtf purchases and net foreign purchases.
As for the effect of the three types of institutional trading activity on stock
returns in the TSE, Table 17.5 reveals that net dic purchases on previous day lead
the TSE returns. We can also see in Fig. 17.2e that after Period 3, net dic purchases
exert a negative (and thus destabilizing) effect on market returns, while the other
two institutional investors do not have such an effect over the sample period.
Examining the relationship between market returns and trading activity of the
three types of institutional investors, we find that either the net foreign purchases
or net dic purchases on previous trading day are affected by the previous day’s TSE
498
Table 17.5 Results of Granger causality tests using the unrestricted VAR models
All nasrt Det rt qfiibst dicbst rtfbst
X3 5.23 (0.16) 103.73* (0.00) 329.58* (0.00) 103.46* (0.00) 40.92* (0.00)
nasr ti
i¼1
3
X 2.77 (0.43) 2.66 (0.45) 2.01 (0.57) 5.98 (0.11) 4.66 (0.20)
Deti
i¼1
3
X 0.04 (1.00) 17.43* (0.00) 14.22* (0.00) 97.30* (0.00) 0.27 (0.97)
r ti
i¼1
3
X 12.67* (0.01) 5.81 (0.12) 4.38 (0.22) 9.55** (0.02) 33.95* (0.00)
qfiibsti
i¼1
3
X 9.38** (0.02) 0.85 (0.84) 15.97* (0.00) 5.83 (0.12) 12.70* (0.01)
dicbsti
i¼1
3
X 0.66 (0.88) 0.81 (0.85) 2.33 (0.51) 52.84* (0.00) 0.85 (0.84)
rtfbsti
i¼1
*, **, and *** denote statistical significance at 1 %, 5 %, and 10 % levels, respectively. Values in parentheses are p values. The optimal lag length of three is
selected based on the Akaike information criterion
B.-N. Huang et al.
17 Dynamic Interactions 499
Fig. 17.2 Impulse responses to innovations in the unrestricted VAR models (up to ten periods).
Notes: Solid lines represent response path and dotted lines are bands for the 95 % confidence
interval around response coefficients. r, TSE returns; nasr, NASDAQ returns; de, exchange rate
changes; qfiibs, net qfii purchases; dicbs, net dic purchases; rtfbs, net rtf purchases
500 B.-N. Huang et al.
returns. Moreover, the IRFs in Fig. 17.2f, g also reveal a significant positive relation
between TSE returns and net foreign purchases up to four periods and a significant
positive relation between TSE returns and net dic purchases for two periods, which
becomes negative after Period 3. In other words, foreign investors in the TSE
engage in positive-feedback trading, while those of the dic tend to change their
strategy and adopt negative-feedback trading after Period 3.
On the other hand, Table 17.5 indicates that previous day’s NASDAQ returns
significantly affect both current TSE returns and net purchases by the three types of
institutional investors. The impulse response in Fig. 17.2h–k also confirms that
previous day’s NASDAQ returns are positively related to both current returns and
net purchases by these institutional investors, with the exception that a negative
relation between net dic purchases and previous day’s NASDAQ returns is found
after Period 4. Such results are much in sync with the expectation since the largest
sector that comprises the TSE-weighted stock index is the electronics industry to
which many listed companies on NASDAQ have a strong connection. In addition,
although previous day’s net qfii and dic purchases also lead the NASDAQ returns,
we find that no significant relation exists except for Period 4 with a significantly
negative relation between them.
The liberalization of Taiwan’s stock market has ushered in significant amount of
short-term inflows and outflows of foreign capital, which have induced fluctuations in
the exchange rate. As is seen from Table 17.5, the TSE returns lead currency returns
and it appears that the initial significant effect of stock returns on currency returns is
negative for the first three periods and then turns to be significantly positive thereafter
(Fig. 17.2n). Given that foreign investors are positive-feedback traders, the capital
inflows is expected to grow in order to increase their stakes in TSE securities when
stock prices rise. Consequently, NT/USD is expected to appreciate.
The unrestricted VAR model does not consider the effect of current returns on net
purchases by institutions. The prior study by Griffin et al. (2003) includes current
returns in the institutional imbalance equation and finds a strong contemporaneous
positive relation between institutional trading and stock returns.6 Therefore, to
further examine the relationship between institutional trading and stock returns,
we introduce the current TSE returns (rt) in the net purchases equations of the three
types of institutional investors and reestimate the VAR model before conducting
the corresponding block exogeneity tests. Table 17.6 presents the estimation results.
As indicated in the last row of Table 17.6, we find evidence of a strong
contemporaneous correlation between current returns and net institutional
6
The main purpose of this paper is to improve our understanding on the interaction among
institutional investors and the relationship between institutional trading and stock returns. The
following discussion will therefore focus on these two issues.
17
Table 17.6 Results of Granger causality tests using the structural VAR models
SVAR nasrt Det rt qfiibst dicbst rtfbst
X3 5.23 (0.16) 103.73* (0.00) 265.94* (0.00) 54.67* (0.00) 6.75*** (0.08)
Dynamic Interactions
nasr ti
i¼1
3
X 2.77 (0.43) 2.66 (0.45) 1.85 (0.60) 7.06*** (0.07) 4.45 (0.22)
Deti
i¼1
3
X 1.26 (0.74) 20.63* (0.00) 98.84* (0.00) 264.71* (0.00) 75.05* (0.00)
r ti
i¼1
3
X 12.67* (0.01) 5.81 (0.12) 4.38 (0.22) 14.05* (0.00) 36.64* (0.00)
qfiibsti
i¼1
3
X 9.38** (0.02) 0.85 (0.84) 15.97* (0.00) 5.09 (0.17) 11.09* (0.01)
dicbsti
i¼1
3
X 0.66 (0.88) 0.81 (0.85) 2.33 (0.51) 50.56* (0.00) 0.83 (0.84)
rtfbsti
i¼1
rt 332.55* [10.91] 247.47* [18.10] 213.66* [21.43]
The structural VAR model includes current TSE returns in the institutional trading equations to take the contemporaneous correlations into consideration.
See also Table 17.5 for definitions
501
502 B.-N. Huang et al.
Fig. 17.3 Selected impulse responses to innovations up to ten periods in the structural VAR
models. Notes: See also Fig. 17.2
We pool all the data together when estimating either the unrestricted or restricted
VAR model; however, the trading activity of institutional investors may depend on
whether stock prices rise or fall.8 A small economy like Taiwan also depends to
a large degree on the sign of NASDAQ index returns. Consequently, to investigate
institutional trading under distinct regimes based on market returns, we use the
multivariate threshold autoregression (MVTAR) model proposed by Tsay (1998)
7
Figure 17.2f, g show significantly positive responses of qfiibst and dicbst to rt in Period 1 if the
impact of current returns is not considered.
8
Recall that both the positive-feedback and negative-feedback trading are associated with the sign
of market returns on the previous trading day.
17 Dynamic Interactions 503
to test the relevant hypotheses. Let yt ¼ [nasrt, Det, rt, qfiibst, dicbst, rtfbst]0 be
a 6 1 vector and the MVTAR model can be described as
!
X
p
yt ¼ f0;1 þ fi, 1 yti ½1 I ðztd > cÞ
i¼1
! (2)
X
p
þ f0;2 þ fi, 2 yti I ðztd > cÞ þ et
i¼1
where E(e) ¼ 0, E(ee0 ) ¼ S, and I() is an index function, which equals 1 if the
relation in the bracket holds. It equals zero otherwise. ztd is the threshold variable
with a delay (lag) d.
In order to explore whether institutional trading activity would change during
different domestic and foreign market return scenarios, the potential threshold
variables used are rt1 and nasrt1.9 Before estimating Eq. 2, we need to test for
possible potential nonlinearity (threshold effect) in this equation. Tsay (1998)
suggests using the arranged regression concept to construct the C(d) statistic to
test the hypothesis Ho:fi,1 ¼ fi,2, i ¼ 0, . . . p. If H0 can be rejected, it implies that
there exists the nonlinearity in data with ztd as the threshold variable. Tsay (1998)
proves that C(d) is asymptotically a chi-square random variable with k(pk + 1)
degrees of freedom, where p is the lag length of the VAR model and k is the number
of endogenous variables yt.10 Table 17.7 presents the estimation results of the C(d)
statistic.
As shown in Table 17.7, the null hypothesis H0 is rejected using either past
returns on the TSE or NASDAQ, suggesting that our data exhibit nonlinear
threshold effect. Theoretically, one needs to rearrange the regression based on the
size of the threshold variable ztd before applying a grid search method to find the
optimal threshold value c*. Nonetheless, our goal is to know whether the institu-
tional trading behavior depends on the sign of market returns, as such the threshold
is set to zero in a rather arbitrary way.11 Table 17.8 lists the results of block
9
Here, we assume that net purchases by institutions are only affected by market returns on the
previous trading day.
10
For more details see Tsay (1998).
11
A previous study by Sadorsky (1999) also splits data into two regimes based on the sign of the
variable to discuss whether variables used would change their behaviors under different regimes.
504
Table 17.8 Results of Granger causality tests using the MVTAR models with threshold variable rt1
rt1 < 0 nasrt Det rt qfiibst dicbst rtfbst
X 3 3.07 (0.38) 84.21* (0.00) 167.80* (0.00) 53.71* (0.00) 30.12* (0.00)
nasr ti
i¼1
3
X 0.51 (0.92) 2.01 (0.57) 1.00 (0.80) 3.11 (0.37) 0.65 (0.89)
Deti
i¼1
3
X 2.28 (0.52) 12.41* (0.01) 8.57** (0.04) 22.40* (0.00) 3.14 (0.37)
r ti
i¼1
3
X 7.75** (0.05) 3.04 (0.39) 4.33 (0.23) 9.79** (0.02) 32.73* (0.00)
qfiibsti
i¼1
3
X 3.83 (0.28) 0.46 (0.93) 4.92 (0.18) 2.38 (0.50) 4.62 (0.20)
dicbsti
i¼1
3
X 1.23 (0.75) 1.62 (0.65) 1.62 (0.66) 19.99* (0.00) 2.83 (0.42)
rtfbsti
i¼1
3
X 3.72 (0.29) 27.57* (0.00) 168.61* (0.00) 49.94* (0.00) 12.25* (0.01)
nasr ti
i¼1
B.-N. Huang et al.
17
3
X 5.97 (0.11) 1.78 (0.62) 2.73 (0.44) 5.16 (0.16) 7.01*** (0.07)
Deti
i¼1
3
X 0.88 (0.83) 13.00* (0.00) 3.02 (0.39) 43.44* (0.00) 1.68 (0.64)
r ti
i¼1
3
X 4.83 (0.18) 5.87 (0.12) 1.85 (0.60) 5.72 (0.13) 5.16 (0.16)
qfiibsti
i¼1
Dynamic Interactions
3
X 8.99** (0.03) 2.04 (0.56) 13.17* (0.00) 3.90 (0.27) 12.54* (0.01)
dicbsti
i¼1
3
X 0.22 (0.97) 0.58 (0.90) 2.53 (0.47) 39.46* (0.00) 2.15 (0.54)
rtfbsti
i¼1
The results are premised on the condition that when previous day’s TSE returns are positive. See also Table 17.5
505
506 B.-N. Huang et al.
exogeneity tests for the lead-lag relation in the rt1 < 0 and rt1 0 regimes,
respectively.
The interaction among institutional investors is depicted in Table 17.8: Current
net purchases by foreign investors affect that by domestic institutions when previ-
ous day’s TSE returns are negative. Note that no such relation is evidenced when
previous day’s TSE returns are positive. A feedback relation between rtf and qfii is
observed when rt1 is positive or negative. However, dic is found to lead rtf only
when rt1 is positive. Such results reveal different institutional trading strategies
under distinct return regimes. The demonstration effect – previous day’s net foreign
purchases have on domestic institutions using the unrestricted VAR model – seems
to surface only when previous day’s market returns are negative. Therefore, it may
produce misleading results if we fail to consider the sign of previous returns.
The impulse responses in Fig. 17.4 illustrate that the responses of dic and rtf
from the qfii shock are quite similar to the ones in Fig. 17.2f, g.12 As for the impact
of previous day’s market returns on current net purchases by institutions, it can be
shown via the MVTAR model that market returns lead net purchases by qfii and dic
when previous day’s market returns are negative, which is consistent with the
finding using the one-regime VAR model. When previous day’s market returns
are positive, market returns lead net purchases by the dic only. Obviously, returns
have more influence on net institutional purchases when previous day’s
returns were negative. In addition, we find that net dic purchases on the previous
day may affect current returns when the one-regime VAR model is used. Actually,
the MVTAR analysis reveals that such a relation emerges only when previous day’s
market returns are positive. The impulse responses depicted in Fig. 17.4 (Panel B)
demonstrate that a one-unit standard error shock to net dic purchases produces an
increase in market returns in Period 2, and then they turns to be negative after
Period 4, a result similar to those using the one-regime VAR model.
Among the listed companies on the TSE, the electronics sector has the largest market
share, which accounts for more than 60 % of all trades. This being the case, Taiwan’s
stock market is closely related to the NASDAQ index as is evidenced using the
conventional VAR model. To further investigate whether the interaction among insti-
tutions and the relationship between institutional trading and stock returns are affected
by the sign of previous day’s NASDAQ index return, nasrt1 is used as the threshold
variable. That is, block Granger causality tests are performed by splitting our data into
two regimes based on the sign of the variable nasrt1. Table 17.9 reports the results.
The results indicate that net qfii purchases lead to that of two domestic institu-
tional investors regardless of the sign of previous day’s NASDAQ returns. More-
over, when nasrt1 < 0, net rtf purchases lead net qfii purchases, which is in line
with that using the one-regime VAR model. However, when nasrt1 0, the net
purchases by either dic or rtf lead net qfii purchases, and net qfii purchases lead net
purchases by either dic or rtf. In other words, we find strong evidence of a feedback
12
To economize space, only relevant impulse responses are presented here; the remaining are
available upon request.
17 Dynamic Interactions 507
rt-1< 0
a Response to Cholesky One S.D. Innovations ± 2 S.E.
a) Response of dicbs to qfiibs b) Response of rtfbs to qfiibs c) Response of qfiibs to rtfbs
1200 800 2400
600 2000
800
1600
400 1200
400
200 800
0 400
0
0
−400 −200 −400
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
d) Response of qfiibs to r e) Response of dicbs to r
2400 1200
2000
1600 800
1200
400
800
400 0
0
−400 −400
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
rt-1 ≥ 0
b Response to Cholesky One S.D. Innovations ± 2 S.E.
a) Response of qfiibs to rtfbs b) Response of rtfbs to dicbs
3000 800
2500
600
2000
1500 400
1000 200
500
0 0
−500 −200
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
c) Response of dicbs to r d) Response of r to dicbs
1200 2.0
1.6
800
1.2
400 0.8
0.4
0
0.0
−400 −0.4
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
Fig. 17.4 Selected impulse responses to innovations up to ten periods in the MVTAR models.
(a) rt1 < 0, (b) rt1 0. Notes: See also Fig. 17.2
relation between net foreign purchases and two domestic net purchases when
previous day’s NASDAQ returns are positive. The results pertaining to the impact
of previous day’s returns on institutional trading parallel those using the
unrestricted VAR model: Previous day’s returns have an impact on the net pur-
chases by qfii and dic, but not on net purchases by rtf regardless of the sign of
previous day’s NASDAQ returns. As for the impact of net institutional purchases
508
Table 17.9 Results of Granger causality tests using the MVTAR models with threshold variable nasrt1
nasrt 1 0 nasrt Det rt qfiibst dicbst rtfbst
X3 0.92 (0.82) 36.02* (0.00) 50.15* (0.00) 15.42* (0.00) 12.53* (0.01)
nasr ti
i¼1
3
X 0.50 (0.92) 4.95 (0.18) 0.29 (0.96) 3.81 (0.28) 3.08 (0.38)
Deti
i¼1
3
X 2.13 (0.55) 8.09** (0.04) 8.54** (0.04) 48.61* (0.00) 1.80 (0.62)
r ti
i¼1
3
X 4.62 (0.20) 9.55** (0.02) 3.25 (0.35) 7.43*** (0.06) 14.74* (0.00)
qfiibsti
i¼1
3
X 6.94*** (0.07) 1.05 (0.79) 7.97** (0.05) 2.77 (0.43) 9.64* (0.02)
dicbsti
i¼1
3
X 0.12 (0.99) 1.48 (0.69) 1.33 (0.72) 28.94* (0.00) 1.93 (0.59)
rtfbsti
i¼1
3
X 2.64 (0.45) 29.58* (0.00) 115.60* (0.00) 59.06* (0.00) 16.76* (0.00)
nasr ti
i¼1
B.-N. Huang et al.
17
3
X 6.72*** (0.08) 5.21 (0.16) 3.85 (0.28) 5.63 (0.13) 5.73 (0.13)
Det-i
i¼1
3
X 1.49 (0.68) 27.44* (0.00) 14.75* (0.00) 55.66* (0.00) 0.27 (0.96)
r ti
i¼1
3
X 11.39* (0.01) 0.60 (0.90) 7.76** (0.05) 7.00*** (0.07) 33.48* (0.00)
qfiibsti
i¼1
Dynamic Interactions
3
X 3.48 (0.32) 3.25 (0.35) 11.08* (0.01) 6.74*** (0.08) 3.47 (0.32)
dicbsti
i¼1
3
X 1.82 (0.61) 2.59 (0.46) 4.12 (0.25) 25.09* (0.00) 4.65 (0.20)
rtfbsti
i¼1
The results are premised on the condition that when previous day’s NASDAQ returns are positive. See also Table 17.5
509
510 B.-N. Huang et al.
a nasrt−1 < 0
Response to Cholesky One S.D.Innovations ± 2 S.E.
a) Response of qfiibs b) Response of dicbs c) Response of rtfbs
to rtfbs to qfiibs to qfiibs
3000 1200 800
2500 600
2000 800
1500 400
400
1000 200
500 0 0
0
−500 −400 −200
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
d) Response of rtfbs e) Response of qfiibs f) Response of dicbs
to dicbs to r to r
800 3000 1200
600 2500
2000 800
400 1500
400
200 1000
500 0
0
0
−200 −500 −400
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
g) Response of r
to dicbs
2.0
1.6
1.2
0.8
0.4
0.0
−0.4
1 2 3 4 5 6 7 8 9 10
b nasrt-1 ≥ 0
Response to Cholesky One S.D. Innovations ± 2 S.E.
a) Response of qfiibs b) Response of qfiibs c) Response of dicbs
to dicbs to rtfbs to qfiibs
2500 2500 1200
2000 2000
800
1500 1500
1000 1000 400
500 500
0
0 0
−500 −500 −400
1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10 1 2 3 4 5 6 7 8 9 10
Fig. 17.5 Impulse responses to innovations up to ten periods in the MVTAR models (portion).
(a) nasrt1 < 0, (b) nasrt1 0. Notes: See also Fig. 17.2
on previous day’s stock returns, only net dic purchases still lead stock returns when
nasrt1 < 0, as was the case in the one-regime model. However, we find that net qfii
purchases also lead stock returns when nasrt1 0.
The results from Panel B of Fig. 17.5g indicate that a one-unit standard error shock
from net qfii purchases made on previous days produces a positive response to stock
returns in Period 2, but no significantly negative responses are found during other
periods. The results of the MVTAR model also capture the phenomenon in the
one-regime model: Net dic purchases exert a negative impact on market returns.
However, such an effect is witnessed when previous day’s NASDAQ returns are
negative.
17.4 Conclusion
In this paper we investigate whether the trading behavior of foreign investors leads
that of Taiwanese institutional investors (i.e., the demonstration effect) and whether
institutional trading has a destabilizing effect on the stock market. The reason we
select Taiwan in our study is due to her unique role of being gradually opened up to
foreign investment and her high stock returns volatility. To provide more informa-
tion to these issues, this paper has constructed a six-variable VAR model including
trading activities of three types of institutional investors, the TSE returns,
NASDAQ returns, and currency returns so as to examine the interaction and the
dynamic relationship between institutional trading and stock returns using daily
data from December 13, 1995 to May 13, 2004.
512 B.-N. Huang et al.
The results from the conventional unrestricted VAR model indicate that net
purchases by foreign investors lead those by domestic institutions (and thus the
demonstration effect), while net purchases by rtf also lead those by foreign inves-
tors. That is, there exists a feedback relation between them. As for the relationship
between institutional trading and stock returns, we find that except for rtf, both
foreign and dic net purchases are positively affected by previous day’s TSE returns.
That is, both the qfii and dic engage in positive-feedback trading. We also find that
previous day’s net dic purchases first produce a positive and then a negative impact
on stock returns. Furthermore, we employ a structural VAR model with the
contemporaneous returns included in the three net institutional purchase equations.
A comparison of the structural and unrestricted VAR models suggests that the TSE
returns positively lead net rtf purchases using the structural VAR model, which
cannot be observed when the unrestricted VAR model is used. In other words, if the
contemporaneous relation between returns and net institutional purchases is taken
into account, we find that rtf are also positive-feedback traders.
On the other hand, the sign of market returns does affect trading activities of the
institutions. As a result, this paper makes use of the MVTAR model introduced by
Tsay (1998). By splitting data into two regimes based on the sign of both TSE and
NASDAQ returns on the previous trading day, we find that the demonstration effect
that foreign investors have on domestic institutions arises only when previous day’s
TSE returns are negative. In addition, when previous day’s TSE returns are nega-
tive, stock returns lead both net purchases by qfii and dic. However, stock returns
lead only net dic purchases when previous day’s TSE returns are positive. Finally,
we find the relation that net dic purchases lead market returns using the unrestricted
VAR model tends to emerge only when previous day’s TSE returns are positive.
As for the effect of NASDAQ returns on institutional trading, the results from
this paper suggest that when previous day’s NASDAQ returns are positive,
a feedback relation between net foreign purchases and net domestic purchases
prevails. Moreover, it is found that the net dic purchases lead the TSE returns, as
do the net foreign purchases. The latter, however, exert a positive influence on the
TSE returns.
In summary, our results suggest that net foreign purchases do lead net domestic
purchases, but more details manifest when the threshold model is applied. When
previous day’s TSE returns are negative (or previous day’s NASDAQ returns are
positive), a feedback relation between net foreign purchases and net domestic
purchases is observed. It implies that the widespread argument that foreign inves-
tors have a demonstration effect on domestic institutions in Taiwan is not entirely
correct. In examining the relation between market returns and institutional trading,
we find that market returns at least lead net purchases by both qfii and dic in most
cases. Market returns also lead net rtf purchases if the relationship of contempora-
neous returns and institutional trading is considered. Our analysis also indicates that
net dic purchases lead market returns and are negatively associated with market
returns in Period 4. The results of the MVTAR model suggest that when previous
day’s NASDAQ returns are positive, net foreign purchases positively lead stock
returns and thus will not exert a destabilizing influence on the market.
17 Dynamic Interactions 513
Appendix 1
The prices of many Taiwanese electronics securities are affected by the NASDAQ
returns and hence foreign portfolio inflows may induce fluctuations of exchange
rate. To investigate interactions emanated from the three types of institutional
investors and the relationship between institutional trading activity and stock
returns in Taiwan, we employ a six-variable VAR model using the NASDAQ
returns (nasrt), currency returns (Det), TSE returns (rt), net foreign purchases
(qfiibst), net dic purchases (dicbst), and net rtf purchases (rtfbst) as the underlying
variables. We attempt to answer the issues pertaining to (i) the interaction among
trading activities of the three types of institutions and (ii) the relationship between
stock returns and institutional trading. First, we propose a six-variable unrestricted
VAR model shown below:
2 3 2 3
nasr t f11 ðLÞ f12 ðLÞ f13 ðLÞ f14 ðLÞ f15 ðLÞ f16 ðLÞ
6 De 7 6 f22 ðLÞ f23 ðLÞ f24 ðLÞ f25 ðLÞ f26 ðLÞ 7
6 t 7 6 f21 ðLÞ 7
6 7 6 7
6 rt 7 6 f31 ðLÞ f32 ðLÞ f33 ðLÞ f34 ðLÞ f35 ðLÞ f36 ðLÞ 7
6 7 6 7
6 qfiibs 7 ¼ 6 f ðLÞ f42 ðLÞ f43 ðLÞ f44 ðLÞ f45 ðLÞ f46 ðLÞ 7
6 t7 6 41 7
6 7 6 7
4 dicbst 5 4 f51 ðLÞ f52 ðLÞ f53 ðLÞ f54 ðLÞ f55 ðLÞ f56 ðLÞ 5
rtfbst f61 ðLÞ f62 ðLÞ f63 ðLÞ f64 ðLÞ f65 ðLÞ f66 ðLÞ
2 3 2 3
nasr t1 e1t
6 De 7 6e 7
6 t1 7 6 2t 7
6 7 6 7
6 r t1 7 6 e3t 7
66 7þ6 7
7 6 7
6 qfiibst1 7 6 e4t 7
6 7 6 7
4 dicbst1 5 4 e5t 5
rtfbst1 e6t
Where fij(L) is the polynomial lag of the jth variable in the ith equation. To
investigate the lead-lag relation among three types of institutional investors, we
need
2 to test the hypothesis
3 that each off-diagonal element in the sub-matrix
f44 ðLÞ f45 ðLÞ f46 ðLÞ
4 f54 ðLÞ f55 ðLÞ f56 ðLÞ 5 is zero.
f64 ðLÞ f65 ðLÞ f66 ðLÞ
On the other hand, to determine whether the TSE returns of the previous day lead
net purchases by the three types of institutional investors,0 we test the hypothesis that
each polynomial lag in the vector ½ f43 ðLÞf53 ðLÞf63 ðLÞ is zero. Conversely, if we
want to determine whether previous day’s net purchases by institutional investors
lead current market returns, we test the hypothesis that each element in the vector
[f34(L)f35(L)f36(L)] is zero. Before applying the VAR model, an appropriate lag
structure needs to be specified. A 3-day lag is selected based on the Akaike
514 B.-N. Huang et al.
information criterion (AIC). Table 17.5 presents the lead-lag relation among the six
time series
2 using block exogeneity 3 tests.
f44 ðLÞ f45 ðLÞ f46 ðLÞ
The 4 f54 ðLÞ f55 ðLÞ f56 ðLÞ 5 block represents the potential interaction among
f64 ðLÞ f65 ðLÞ f66 ðLÞ
three types of institutional investors. The results indicate that net foreign purchases
lead net dic purchases and the dynamic relationship between these two variables
can be provided by the impulse response function (IRF) in Fig. 17.2a. Clearly,
a one-unit standard error shock to net foreign purchases leads to an increase in net
dic purchases, but this effect dissipates quickly by Period 2. Figure 17.2b, c indicate
a feedback relation between net purchases by qfii and rtf. A one-unit standard error
shock to net foreign purchases results in a positive response to net rtf purchases over
the next two periods, which become negative in Period 3, followed by a positive
response again after Period 4. Furthermore, a one-unit standard error shock to net rtf
purchases also gives rise to an increase in net foreign purchases, which decays
slowly over ten-period horizon.
Figure 17.2d shows that net purchases by dic lead net rtf purchases. A one-unit
standard error shock to net dic purchases produces an increase in net rtf purchases in the
first three periods and then declines thereafter. Overall, these impulse responses suggest
that previous day’s net foreign purchases exert a noticeable impact on net rtf purchases,
while previous day’s net rtf purchases also has an impact on net foreign purchases. It
implies that not only do foreign capital flows affect the trading activity of domestic
institutional investors but also the relation is not unidirectional. To be specific, there is
a feedback relation between net rtf purchases and net foreign purchases.
As for the effect of the three types of institutional trading activity on stock
returns in the TSE, Table 17.5 reveals that net dic purchases on previous day lead
the TSE returns. We can also see in Fig. 17.2e that after Period 3, net dic purchases
exert a negative (and thus destabilizing) effect on market returns, while the other
two institutional investors do not have such an effect over the sample period.
Examining the relationship between market returns and trading activity of the
three types of institutional investors, we find that either the net foreign purchases
or the net dic purchases on previous trading day are affected by the previous day’s
TSE returns. Moreover, the IRFs in Fig. 17.2f, g also reveal a significant positive
relation between TSE returns and net foreign purchases up to four periods and
a significant positive relation between TSE returns and net dic purchases for two
periods, which becomes negative after Period 3. In other words, foreign investors in
the TSE engage in positive-feedback trading, while those of the dic tend to change
their strategy and adopt negative-feedback trading after Period 3.
Table 17.5 indicates that previous day’s NASDAQ returns significantly affect
both current TSE returns and net purchases by the three types of institutional
investors. The impulse response in Fig. 17.2h–k also confirms that previous day’s
NASDAQ returns are positively related to both current returns and net purchases by
these institutional investors, with the exception that a negative relation between
net dic purchases and previous day’s NASDAQ returns is found after Period 4.
17 Dynamic Interactions 515
Such results are much in sync with the expectation since the largest sector that
comprises the TSE-weighted stock index is the electronics industry to which many
listed companies on NASDAQ have a strong connection. In addition, although
previous day’s net qfii and dic purchases also lead the NASDAQ returns, we find
that no significant relation exists except for Period 4 with a significantly negative
relation between them (Figs. 17.1 and 17.2).
The liberalization of Taiwan’s stock market has ushered in significant amount of
short-term inflows and outflows of foreign capital, which have induced fluctuations in
the exchange rate. As is seen from Table 17.5, the TSE returns lead currency returns
and it appears that the initial significant effect of stock returns on currency returns is
negative for the first three periods and then turns to be significantly positive thereafter
(Fig. 17.2n). Given that foreign investors are positive-feedback traders, the capital
inflows is expected to grow in order to increase their stakes in TSE securities when
stock prices rise. Consequently, NT/USD is expected to appreciate.
The unrestricted VAR model does not consider the effect of current returns on net
purchases by institutions. The prior study by Griffin et al. (2003) includes current
returns in the institutional imbalance equation and finds a strong contemporaneous
positive relation between institutional trading and stock returns. Therefore, to
further examine the relationship between institutional trading and stock returns,
we introduce the current TSE returns (rt) in the net purchases equations of the three
types of institutional investors and reestimate the VAR model before conducting
the corresponding block exogeneity tests. Table 17.6 presents the estimation results.
As indicated in the last row of Table 17.6, we find evidence of a strong contem-
poraneous correlation between current returns and net institutional purchases, which
confirms the finding by the previous research. As shown in Table 17.5, we find no
evidence that past returns lead net rtf purchases when the unrestricted VAR model is
used. When the contemporaneous impact of stock returns on net institutional pur-
chases is considered, we find that past returns also lead net rtf purchases as well as net
purchases by qfii and dic when the structural VAR model is used (Table 17.6). In other
words, net purchases by the three types of institutions are affected by past stock returns
as was evidenced by previous studies. The corresponding impulse response relations
are presented in Fig. 17.3.
Comparing the impulse response relations in Figs. 17.2 and 17.3, it is clear that
when the impact of current returns on net institutional purchases is considered,
a one-unit standard error shock from rt does not produce a positive impulse
response in institutional trading until Period 2. The responses of foreign investors
are rather distinct from those of domestic institutional investors after Period 3. In
general, a sustained positive response from foreign investors is observed, while
a negative response is witnessed for dic and sometimes, an insignificant response
for rtf manifests itself after Period 3.
516 B.-N. Huang et al.
We pool all the data together when estimating either the unrestricted or restricted
VAR model; however, the trading activity of institutional investors may depend on
whether stock prices rise or fall. A small economy like Taiwan also depends to
a large degree on the sign of NASDAQ index returns. Consequently, to investigate
institutional trading under distinct regimes based on market returns, we use the
multivariate threshold autoregression (MVTAR) model proposed by Tsay (1998) to
test the relevant hypotheses. Let yt ¼ [nasrt, Det, rt, qfiibst, dicbst, rtfbst]0 be a 6 1
vector and the MVTAR model can be described as
!
X
p
yt ¼ f0;1 þ fi, 1 yti ½1 I ðztd > cÞ
i¼1
!
X
p
þ f0;2 þ fi, 2 yti I ðztd > cÞ þ et
i¼1
where E(e) ¼ 0, E(ee0 ) ¼ S, and I() is an index function, which equals 1 if the
relation in the bracket holds. It equals zero otherwise. ztd is the threshold variable
with a delay (lag) d.
In order to explore whether institutional trading activity would change during
different domestic and foreign market return scenarios, the potential threshold
variables used are rt1 and nasrt1. Before estimating Eq. 2, we need to test for
possible potential nonlinearity (threshold effect) in this equation. Tsay (1998)
suggests using the arranged regression concept to construct the C(d) statistic to
test the hypothesis Ho:fi,1 ¼ fi,2, i ¼ 0, . . .p. If H0 can be rejected, it implies that
there exists the nonlinearity in data with ztd as the threshold variable. Tsay (1998)
proves that C(d) is asymptotically a chi-square random variable with k(pk + 1)
degrees of freedom, where p is the lag length of the VAR model and k is the number
of endogenous variables yt. Table 17.7 presents the estimation results of the C(d)
statistic.
As shown in Table 17.7, the null hypothesis H0 is rejected using either past
returns on the TSE or NASDAQ, suggesting that our data exhibit nonlinear
threshold effect. Theoretically, one needs to rearrange the regression based on the
size of the threshold variable ztd before applying a grid search method to find the
optimal threshold value c*. Nonetheless, our goal is to know whether the institu-
tional trading behavior depends on the sign of market returns, as such the threshold
is set to zero in a rather arbitrary way. Table 17.8 lists the results of block
exogeneity tests for the lead-lag relation in the rt1 < 0 and rt1 0 regimes,
respectively.
The interaction among institutional investors is depicted in Table 17.8: Current
net purchases by foreign investors affect that by domestic institutions when previ-
ous day’s TSE returns are negative. Note that no such relation is evidenced when
17 Dynamic Interactions 517
previous day’s TSE returns are positive. A feedback relation between rtf and
qfii is observed when rt1 is positive or negative. However, dic is found to lead
rtf only when rt1 is positive. Such results reveal different institutional trading
strategies under distinct return regimes. The demonstration effect – previous day’s
net foreign purchases have on domestic institutions using the unrestricted VAR
model – seems to surface only when previous day’s market returns are negative.
Therefore, it may produce misleading results if we fail to consider the sign of
previous returns.
The impulse responses in Fig. 17.4 illustrate that the responses of dic and rtf
from the qfii shock are quite similar to the ones in Fig. 17.2f, g. As for the impact of
previous day’s market returns on current net purchases by institutions, it can be
shown via the MVTAR model that market returns lead net purchases by qfii and dic
when previous day’s market returns are negative, which is consistent with the
finding using the one-regime VAR model. When previous day’s market returns
are positive, market returns lead net purchases by the dic only. Obviously, returns
have more influence on net institutional purchases when previous day’s
returns were negative. In addition, we find that net dic purchases on the previous
day may affect current returns when the one-regime VAR model is used. Actually,
the MVTAR analysis reveals that such a relation emerges only when previous day’s
market returns are positive. The impulse responses depicted in Fig. 17.4 (Panel B)
demonstrate that a one-unit standard error shock to net dic purchases produces an
increase in market returns in Period 2, and then they turns to be negative after
Period 4, a result similar to those using the one-regime VAR model.
To further investigate whether the interaction among institutions and the
relationship between institutional trading and stock returns are affected by the
sign of previous day’s NASDAQ index return, nasrt1 is used as the threshold
variable. That is, block Granger causality tests are performed by splitting our
data into two regimes based on the sign of the variable nasrt1. Table 17.9 reports
the results.
The results indicate that net qfii purchases lead that of two domestic institutional
investors regardless of the sign of previous day’s NASDAQ returns. Moreover,
when nasrt1 < 0, net rtf purchases lead net qfii purchases, which is in line with that
using the one-regime VAR model. However, when nasrt1 0, the net purchases
by either dic or rtf lead net qfii purchases, and net qfii purchases lead net purchases
by either dic or rtf. In other words, we find strong evidence of a feedback relation
between net foreign purchases and two domestic net purchases when previous day’s
NASDAQ returns are positive. The results pertaining to the impact of previous
day’s returns on institutional trading parallel those using the unrestricted VAR
model: Previous day’s returns have an impact on the net purchases by qfii and dic,
but not on net purchases by rtf regardless of the sign of previous day’s NASDAQ
returns. As for the impact of net institutional purchases on previous day’s stock
returns, only net dic purchases still lead stock returns when nasrt1 < 0, as was the
case in the one-regime model. However, we find that net qfii purchases also lead
stock returns when nasrt1 0.
518 B.-N. Huang et al.
References
Brennan, M., & Cao, H. (1997). International portfolio investment flows. Journal of Finance, 52,
1851–1880.
Cai, F., & Zheng, L. (2002). Institutional trading and stock returns (Working paper). Ann Arbor:
University of Michigan Business School.
Choe, H., Kho, B. C., & Stulz, R. (1999). Do foreign investors destabilize stock markets? The
Korean experience in 1997. Journal of Financial Economics, 54, 227–264.
Froot, K. A., O’Connell, P. G. J., & Seasholes, M. S. (2001). The portfolio flows of international
investors. Journal of Financial Economics, 59, 151–193.
Griffin, J. M., Harris, J. H., & Topaloglu, S. (2003). The dynamics of institutional and individual
trading. Journal of Finance, 58, 2285–2320.
Grinblatt, M., & Keloharju, M. (2000). The investment behavior and performance of various
investor-types: A study of Finland’s unique data set. Journal of Financial Economics, 55,
43–67.
Hamao, Y., & Mei, J. (2001). Living with the “enemy”: An analysis of foreign investment in the
Japanese equity market. Journal of International Money and Finance, 20, 715–735.
Huang, C.-H., & Hsu, Y. Y. (1999). The impact of foreign investors on the Taiwan stock exchange.
Taipei Bank Monthly, 29(4), 58–77.
Kamesaka, A., Nofsinger, J. R., & Kawakita, H. (2003). Investment patterns and performance of
investor groups in Japan. Pacific-Basin Finance Journal, 11, 1–22.
Karolyi, G. A. (2002). Did the Asian financial crisis scare foreign investors out of Japan? Pacific-
Basin Finance Journal, 10, 411–442.
Lakonishok, J., Shleifer, A., & Vishny, R. (1992). The impact of institutional trading on stock
price. Journal of Financial Economics, 32, 23–44.
Lee, T. S., & Oh, Y. L. (1995). Foreign investment, stock market volatility and macro variables.
Fundamental Financial Institutions, 31, 45–47 (in Chinese).
Lee, Y.-T., Lin, J. C., & Liu, Y. J. (1999). Trading patterns of big versus small players in an
emerging market: An empirical analysis. Journal of Banking and Finance, 23, 701–725.
Nofsinger, J. R., & Sias, W. (1999). Herding and feedback trading by institutional and individual
investors. Journal of Finance, 54, 2263–2295.
Sadorsky, P. (1999). Oil price shocks and stock market activity. Energy Economics, 21, 449–469.
Sias, R. W., & Starks, L. T. (1997). Return autocorrelation and institutional investors. Journal of
Financial Economics, 46, 103–131.
Tsay, R. S. (1998). Testing and modeling multivariate threshold models. Journal of American
Statistical Association, 93(443), 1188–1202.
Wang, L.-R., & Shen, C. H. (1999). Do foreign investments affect foreign exchange and stock
markets? – The case of Taiwan. Applied Economics, 31(11), 1303–1314.
Methods of Denoising Financial Data
18
Thomas Meinl and Edward W. Sun
Contents
18.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 520
18.2 Denoising (Trend Extraction) of Financial Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 520
18.3 Linear Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 524
18.3.1 General Formulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 524
18.3.2 Transfer Functions: Time Versus Frequency Domain . . . . . . . . . . . . . . . . . . . . . . . . 525
18.4 Nonlinear Filters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 527
18.4.1 General Perception . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 527
18.5 Specific Jump Detection Models and Related Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 531
18.5.1 Algorithms for Jump Detection and Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 532
18.5.2 Further Related Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 533
18.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 536
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 536
Abstract
Denoising analysis imposes new challenges for financial data mining due to the
irregularities and roughness observed in financial data, particularly, for instan-
taneously collected massive amounts of tick-by-tick data from financial markets
for information analysis and knowledge extraction. Inefficient decomposition of
the systematic pattern (the trend) and noises of financial data will lead to
erroneous conclusions since irregularities and roughness of the financial data
make the application of traditional methods difficult.
In this chapter, we provide a review to discuss some methods applied for
denoising analysis of financial data.
T. Meinl (*)
Karlsruhe Institute of Technology (KIT), Karlsruhe, Germany
e-mail: thomas.meinl@kit.edu
E.W. Sun
KEDGE Business School and BEM Management School, Bordeaux, France
e-mail: edward.sun@bem.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 519
DOI 10.1007/978-1-4614-7750-1_18,
# Springer Science+Business Media New York 2015
520 T. Meinl and E.W. Sun
Keywords
Jump detection • Linear filters • Nonlinear filters • Time series analysis • Trend
extraction • Wavelet
18.1 Introduction
X t ¼ # t þ st þ Y t , (18.1)
However, as these extreme events can have an enormous impact on the stochastic
variance analysis and its succeeding usage, and furthermore could lead to mislead-
ing results in the regions of the signal without any jumps, it is generally
preferred to include them into the trend component # rather than the stochastic
component Y.
In, for example, Wang (1995), a definition of an a-cusp in a continuous function
f at x0 is given, that is, if there exists an e > 0 so that
holds for all h 2 [x0 e, x0 + e]. For a ¼ 0, f is said to have a jump at x0. It is
commonly agreed that the jump should significantly differ from the other fluctua-
tions (i.e., the noise) in the signal. As said above, jumps are just one particular
pattern of extreme events we are interested in. Others are steep slopes, roofs, and
valleys, which in Joo and Qiu (2009) are defined by at this point having a jump in
the first-order derivative of the regression curve.
Other extreme events frequently occurring in many practical applications are spikes
and outliers. However, these are usually undesirable features that should not be
included in the trend of affect it by any means. This is due to the following reasons.
First, in many cases, these outliers or spikes consist only of one or very few points often
caused by measurement errors, and it is obvious that they were caused by some factor
that plays no vital role in the ongoing time series analysis (unless the focus is on what
caused these outliers). Second, while jumps imply a permanent change in the
whole time series, outliers do not contribute to this. While we are aware that
the distinction between a few (adjacent) outliers and roofs/valleys is not precise,
from the context of the time series in most cases, it is evident whether an occurrence
should be considered as an outlier that is to be neglected or a significant feature to be
included in the trend.
In this work we only consider homogeneous financial time series data. However,
in many applications and particularly for high-frequency financial time series data,
this data is initially irregularly spaced. Homogeneity in this context means that for
a given series Xt, t 2 N, holds t + 1 t ¼ c, with a constant c > 0, that is, all time
steps are equally spaced. As we will see, this is not always the case for empirical
time series, especially in the area of financial high-frequency data. In this case, it is
necessary to preprocess the inhomogeneous (i.e., irregularly spaced) time series by
interpolation methods in order to regularize the raw data. Though there exist models
which can handle inhomogeneous time series directly (see Dacorogna (2001), but
they also remark that most today’s models are suited for regularly spaced time
series only), regarding the methods we discuss in the following sections, we restrict
ourselves to homogeneous ones.
In this chapter we focus on nonparametric methods for trend extraction. This is
due to the reason that in most time series we analyze in this work, we cannot
reasonably assume any model for the underlying trend. Yet, as noted in the
framework above, in case such assumptions hold, we can expect those models to
18 Methods of Denoising Financial Data 523
perform better than nonparametric models, since they exploit information that
nonparametric approaches cannot. Furthermore, the commitment to certain para-
metric time series or trend models can be seen as a restriction when considering the
general case, which may lead even to misleading results in case the trend does not
match model, as certain patterns might not be captured or considered by the
model itself. This can easily be seen at a most basic example, in case a linear trend
is expected, which in most cases will only be a poor estimator for any nonlinear
trend curve. In this case nonparametric approaches are less restrictive and can
more generally be applied, while of course not delivering the same accuracy
as parametric models which exactly match the underlying trend, with only
their parameters to be calibrated. If, e.g., the trend follows sinusoidal curve, a
sinusoidal curve with its parameters being estimated by, e.g., the least-squares
method will almost surely provide a better accuracy than any other nonparametric
approach. On the other hand, if the underlying trend is linear or even contains
only little deviations from a perfect sinusoidal curve, a sinusoidal fit to this trend,
it has been shown at simple examples that the parametric sinusoidal approach can
lead to confusing results and conclusions.
We further require that the method used for trend extraction be robust, that is,
the results are reliable and the error can be estimated or is at least bounded in some
way. In many cases (see, e.g., Donoho and Johnstone (1994)), the robustness
of a method is shown by proving its asymptotic consistency, that is, its conver-
gence towards a certain value for certain parameters tending towards infinity.
It should be remarked that the robustness should be independent of the
time series itself and/or any specific algorithm parameter sets, in order to be
applicable in practice. Of course this does not exclude specific assumptions on the
time series that must be met or parameter ranges for which the algorithm is
defined.
Therefore, as we cannot reasonably assume any model for the any time series
in general, in this work, we focus on nonparametric approaches only. Within these
approaches we focus on two main branches for trend extraction: linear
and nonlinear filters. This is for the reason because linear filters are known and
have proven to deliver a very smooth trend (given the filtering window size is
large enough), while nonlinear filters excel at preserving characteristic patterns
in a time series, i.e., especially jumps. Both methods in general require only
very few (or none at all) information about the underlying data, be-
sides their configuration of weights and calibration parameters, and are thus
applicable to a wide range of time series, independent of the field the data was
measured in.
Although there exists a variety of other nonparametric methods, most of these
already rely on specific assumptions or choices of parameters which in general
cannot easily be derived for any time series data or different analysis goals.
Nevertheless, for the sake of completeness, later in this chapter we list some
alternative methods, also including parametric approaches, which have been
applied in economic, financial, or related time series data.
524 T. Meinl and E.W. Sun
Linear filters are probably the most common and well-known filters used for trend
extraction and additive noise removal. We first provide the most general notion of
this filter class and then depict the two viewpoints of linear filters and how they can
be characterized. While this characterization on the one hand is one of the most
distinguishable advantages of linear filters, on the other hand at the same time, it
leads to the exact problem we are facing in this work, that is, the representation of
sharp edges in otherwise smooth trends.
The filtered output depends linearly from the time series input. Using the notation of
Fan and Yao (2003), a linear filter of length 2h + 1 can be defined as
X
h
^t ¼
X wi Xtþi , (18.3)
i¼h
with X ^ t the filtered output and wi the filter weights. These kinds of filters are
also known as (discrete) convolution filters, as the outcome is the convolution of
the input signal with a discrete weight function, where often the following notation
is used:
X
h Xh
w Xt :¼ wi Xti ¼ wti Xi : (18.4)
i¼h i¼h
Thus, for every data point Xt, the filtered output X ^ t is the result of weighted
summation of data points around t. Applied for the whole time series, this results in
weighted average window of size L ¼ 2h + 1 which is moved throughout the whole
series. The size of this window is also called the bandwidth of the filter. It is also
common notation to set h 1 even if the window filter size should be finite, with
wi ¼ 0 for jij > h.
The probably best known linear filter is the mean filter, with wi ¼ 2h + 1, that is,
all filter weights are uniformly distributed. A more general viewpoint is given by
the notion of kernel filters. Given a kernel function w.l.o.g. with support [1, 1],
this function assigns the weights according to
K ði=hÞ
wi ¼ Xh : (18.5)
j¼h
K ðj=h Þ
3
K E ð uÞ ¼ 1 u2 þ , (18.6)
4
18 Methods of Denoising Financial Data 525
1 g
K Bg ðuÞ ¼ 1 u2 I juj1 : (18.8)
Bð1=2, g þ 1Þ
For the values g 2 {0, 1, 2, 3}, the kernel KBg (u) corresponds to the uniform,
Epanechnikov, biweight, and triweight kernel functions, respectively.
The previous section depicts the linear filtering method in the time domain, i.e., we
look at the time series Xt and its respective filtered output X ^ t and how they evolve
over time t. Another perception can be given by taking the frequency domain into
account. For all linear filters, we cannot only give its definition as depicted earlier,
but also in respect to the frequencies, the filters let pass. This notion can be derived
as follows.
While the sequence of filter weights wi, also called impulse response sequence,
determines the filtered output in the time domain (or equivalent are the linear filter’s
time domain representation), via the discrete Fourier transform (DFT), we can
derive the transfer function
X
1
Wð f Þ ¼ wj ei2pf j , (18.9)
j¼1
its counterpart in the frequency domain, also called frequency response function.
Alternatively, if this formulation is given in the beginning, we can also derive the
weights via the inverse transform:
Z 1=2
wj ¼ W ð f Þei2pf j df : (18.10)
1=2
Obviously, these two formulations are equivalent, as one can be derived from the
other, and vice versa. By considering the transfer function’s polar representation,
W ð f Þ ¼ jW ð f Þjeiyðf Þ , (18.11)
526 T. Meinl and E.W. Sun
with jW ð f Þj the gain function. The magnitude in gain jW ð f Þj describes the linear
filters behavior in the frequency domain, that is, what kind of frequencies and their
respective proportions will let passed or be blocked. For our needs, it
satisfies to distinguish between high – and low-pass filters, that is, filters that
let pass either the high frequencies and block the lower ones, or vice versa. Of
course, other filter types exist; for example, by combining high – and low-pass
filters (e.g., in a filter cascade), one can derive band-pass and stop filters,
so that the frequency domain output will be located only in a certain frequency
range. In this work we are specifically interested in low-pass filters, as they
block the high-frequency noise and the output consists of the generally
low-frequency trend.
As we assume that the weights wj are real valued, one can show
(see, e.g., Percival and Walden (2006)) that W ðf Þ ¼ W ð f Þ, and, with jW ðf Þj ¼
jW ð f Þj , it follows that jW ðf Þj ¼ jW ð f Þj . Therefore, the transfer functions
are symmetric around zero. Because of its periodicity, we need to consider W ð f Þ
only an interval of unit length. For convenience, this interval is often taken
to be [1/2, 1/2], i.e., j fj 1/2. Therefore, with above-depicted
symmetry, it suffices to consider f 2 [0, 1/2] in order to fully specify the transfer
function.
While saying that certain frequencies are blocked and others are passed, this
holds only approximately true, since the design of such exact frequency filters is not
possible, but always a transition between the blocked and passing frequencies. The
goal of many linear filters is either to minimize these transitions (i.e., the range of
by this affected frequencies), which, on the other hand, inevitably causes ripples in
the other frequencies, that is, they are not any longer blocked, or to let them pass
completely (see Bianchi and Sorrentino (2007) and the references therein for
further details about this topic).
As we have seen at above examples, linear filters can be designed either from
a time or from a frequency perspective. The time domain usually focuses on putting
more weights on the surrounding events (i.e., events that recently before or
occurred shortly after) and, thus, gives an (economic) interpretation similar to
the, for example, ARMA and GARCH models. On the contrary, the frequency
domain is based on the point of view that certain disturbances (almost) exclusively
are located in a certain frequency range and are thus isolated from the rest of the
signal. Also, the slowly evolving trend can be seen to occupy only the lower
frequency ranges. Thus, the (economic) meaning lies here in the frequency of
events; see, for example, Dacorogna (2001).
Although linear filters can be designed to block or let pass certain frequencies
nearly optimal, at the same time this poses a serious problem when facing trends
that exhibit jumps or slopes. As these events are also located in the same (or in case
of slopes the adjacent) frequency range as the high-frequency noise, this has the
effect that jumps and edged frontiers are blurred out, while steep slopes mostly are
captured with poor precision only. Hence, from a frequency perspective, a smooth
trend and edge preservation are two conflicting goals. This is as the linear filters are
not capable to distinguish between the persistent noise and a single events that,
18 Methods of Denoising Financial Data 527
while located in the same frequency range, usually have a higher energy and, thus,
significantly larger amplitude. Thus, the same filtering rule is applied throughout
the whole signal, without any adaption. Note, however, that linear filters still give
some weight to undesirable events like outliers, due to their moving average
nature. Thus, while some significant features like jumps are not displayed in
enough detail, other unwanted patterns, like spikes, still partially carry over to
the filtered output. To overcome all these drawbacks for that kind of trends or
signals, besides the class of linear filters, the class of nonlinear filters has been
developed.
As we have seen, linear filters tend to blur out edges and other details though they
form an elementary part of the time series’ trend. In order to avoid this, a wide range
of nonlinear filters has been developed which on one hand preserve those details
while on the other try to smooth out as much of the noise as possible. We do not find
nonlinear filters only in time series, but they were in many cases developed
specifically for two-dimensional signals, specifically images, where the original
image, probably corrupted by noise during data transmission, consists mainly of
edges, which form the image.
While linear filters generally provide a very smooth trend achieved through aver-
aging, two characteristics pose a problem for this class of filters:
• Outliers and spikes
Single, extreme outliers and spikes can cause the whole long-term trend to
deviate in the same direction, though they obviously do not play a part in it.
• Jumps, slopes, and regime changes
Whenever there occurs a sudden external event in the underlying main
driving force, it causes the trend to jump, that is, contrary to spikes it
changes permanently onto another plane. While slopes are not that extreme,
they also show a similar behavior as they decay or rise with the trend’s unusual
degree.
The reasons for the deviation sensitivity to these events are given by one of the
most favorable linear filters’ characteristics themselves: It follows directly from
them being characterizable in terms of frequency passbands we explained earlier
that all frequencies are treated the same (i.e., filtered according to the same rule)
throughout the whole signal. This means that no distinction is made (and even
cannot be made) between noise and those patterns, as they are located in approx-
imately the same frequency range. Technically, as long as an outlier or a jump is
contained in the weighted moving average filtering window, also a weight is
assigned to these outlier data points or the points on the other plane, i.e., before
528 T. Meinl and E.W. Sun
and after the jump. Nonlinear filtering procedures try to avoid this by using
a different approach, for example, by considering a single value only (instead of
multiple weighted ones) that was selected from an ordered or ranked permutation of
the original values in the filtering window.
Although we cannot characterize nonlinear filters in the same way as we do with
linear filters (i.e., by transfer functions) besides their characteristics not being
classifiable in that way, according to Peltonen et al. (2001), we can divide the
whole class of these filters into several subclasses that share the same or similar
approaches. Among them, there are stack filters, weighted median filters, polyno-
mial filters, and order statistic filters. Astola and Kuosmanen (1997) provide two
different taxonomies for further classification, though they remark that these divi-
sions are not unique. In their work, they extensively show how those different filters
behave (i.e., their characteristics) when applied onto different benchmark signals in
respect to the mean absolute error (MAE) and mean squared error (MSE) measures.
The behavior of nonlinear filters is generally characterized by their impulse and
step response, i.e., the filtered output when the input consists of a single impulse or
step only. These impulses generally are given by the sequence [. . . , 0, 0, 0, a, 0, 0,
0, . . .] and [. . . , 0, 0, 0, a, a, a, . . .], respectively, with a 6¼ 0. Though in most cases
no analytical result can be given, these characteristics help us to understand how the
nonlinear filter behaves in respect to those patterns, for which the linear filters
generally fail to deliver adequate results.
Despite their undoubtedly good ability to preserve outstanding features in a time
series while extracting the trend, nonlinear filters also suffer some drawbacks:
1. Insufficient smoothness
Though most nonlinear filters try to deliver a smooth trend and a good resolution
of edged frontiers, beyond the jumps’ surrounding regions, they usually fail to
deliver a smooth trend as accurate as even simple linear filter (e.g., the mean
filter) provides. Yet, by applying further smoothing procedures (e.g., by recur-
sive filters or some kind of linear filtering on the nonlinear output, with a smaller
bandwidth, as most of the noise is already smoothed out) comes at the price that
the prior preserved details of jumps or slopes tend to get lost.
This effect is even aggravated when the high-frequency noise present in the
signal is extremely high, i.e., the trend which evolves, besides the jumps, quite
slowly is dominated by the noise with extremely high amplitudes. Some filters
try to counter this effect, but they either:
• Provide only a tradeoff between an overall smooth signal and poor jump
resolution, or a trend still exhibiting ripples but preserved edges, or
• Rely on further information about the time series itself, that is, the noise
component and its structure, the jumps or the trend itself
This is due to the problem that most filtering rules are applied throughout the
whole signal, that is, they do not adapt themselves sufficiently when the filtering
window approaches a jump. An overview of these and other nonlinear filters’
performance is given by Astola and Kuosmanen (1997), proving again the
well-known insight that there cannot exist one solution that performs optimal
for all cases. Though the authors also report some approaches that try to
18 Methods of Denoising Financial Data 529
incorporate that behavior, these filters provide only dissatisfactory results (see
the examples below).
2. Lack of frequency control
Another feature nonlinear filters are lacking is the ability to regulate the filtered
output in terms of frequency passbands, as linear filters do. Since Pollock (2001)
defines the trend in terms of frequency bands and Ramsey (1999, 2002) points
out that frequency analysis is an important aspect in financial time series, so is
frequency control. Though not necessary for all applications, the ability to
a priori control and regulate the filters output (in contrast to an only a posteriori
frequency analysis of the filtered result) may come handy when one wants to ensure
that certain frequencies are not contained in the output, that is, when certain
information about the noise frequencies is at hand, the analyst can decide before
the actual filtering process (and thus, without any try and error procedures) what
frequency parts should be filtered out. Although a nonlinear filter can also provide
the same or a similar result, no theoretical results or statements are available before
the filtering procedure has been carried out completely. This incapacity of the
nonlinear filter follows directly from the fact that nonlinear filters do not rely on
frequency passbands, as they must be able to filter events, even though they occupy
nearly the same frequency range (i.e., noise and jumps), due to different rules.
As Astola and Kuosmanen (1997) classify linear filters to be a subclass of the
class of nonlinear filters, above statement does not exactly hold true, i.e., it would
mean that a subclass of nonlinear filters can be characterized by their transfer
function and frequency output. We, however, separate the classes of linear and
nonlinear filters by whether or not a filter can be described by a transfer function or
not, which marks a strict division of these two classes.
18.4.1.1 Examples
To illustrate the different courses of action of nonlinear filters and give the reader an
idea of their general procedure, in this section we outline several examples of above
named subclasses. As this list cannot be exhaustive by any means, of course, we
note that we do not take filters into account that already rely on specific assumptions
of the systems beneath the time series themselves.
1 X
N s
X ð iÞ : (18.12)
N r s i¼rþ1
!
1 X
N s
r Xðrþ1Þ XðiÞ þ s XðNsÞ : (18.13)
N i¼rþ1
In these methods, the (r, s) tuple is dependent of the data itself. Other filters
consider to make these values independent from the data or dependent from the
central sample itself, i.e., nearest neighbor techniques. All those filters have in
common that they discard all samples from the ordered series being too far away
(respectively) according to some measure.
X
N
wi XðiÞ , (18.14)
i¼1
with wi the filter weights. A similar notion is given by weighted median filters, where
the weights are assigned to the time ordered sample Xt and where the weights denote
a duplication operation, i.e., wi ∘Xt = Xt, 1 , . . . , Xt, wi. The output is then given by
Hybrid Filters
Another approach is the design of nonlinear filters consisting of filter cascades,
that is, the repeated application of different filters on the respective
outputs. A general formulation is, for example, given by rth-order statistic
F1(X1, . . . , Xn), . . . , FM(X1, . . . , Xn), where F1, . . . , FM can denote any other
filtering procedure. A concrete example is the median hybrid filter that combines
prior linear filtering procedure with a succeeding median ordering operation, i.e.,
( )
Xk X N
median ð1=kÞ Xi , Xkþ1 , ð1=kÞ Xi : (18.17)
i¼1 i¼kþ2
18 Methods of Denoising Financial Data 531
Selective Filters
An interesting approach is given by the principle of switching between different
output rules depending on some selection rule, for example, based on the fact that
the mean filter delivers a larger (smaller) output than the median filter when the
filtering window approaches an upward (downward) jump. Thus, a selection rule
would be given by
mean X1 , ..., XN median X1 , ..., XN : (18.18)
A certain drawback of this selection rule is that it is one-sided, that is, it considers
only the first half of the region around the jump. This is due to the fact that the mean
for the second half, after the jump has occurred, is generally smaller (larger) than the
median. Other rules can include thresholds and aim at deciding whether there has
actually happened a jump or there was an impulse in the signal, which is not caused
by the noise distribution, but due to some other explanatory effect.
Above-depicted examples of nonlinear filters should give the reader an overview
over the most common methods applied in practice. For detailed information about
each filter’s characteristics, advantages, and drawbacks, we refer to Astola and
Kuosmanen (1997), where there are also the references to the original works to be
found. Yet, we see that basically most of these filters rely on some ordered statistics,
with their input or output modified prior or afterwards, respectively. Since this basic
principle applies to most filters not directly dependent on some specific character-
istic or assuming a certain structure of the original series to be filtered, the different
methods pointed out above can be combined in numerous ways. In many cases,
however, even though we portrait only the very basic methods, we see that almost
all of them already incorporate an implicit or explicit choice of additional param-
eters besides the filter bandwidth, either by weights, rules, or thresholds. These
choices introduce further biases into the filtering process. Though some of these
parameters can be chosen to be optimal in some sense (i.e., minimize a certain
distance measure, e.g., the MSE for the L-filters), they lack the concrete meaning of
weights we have for linear filters.
In this section we list further related methods that are concerned with the estimation
of long-term trends exhibiting edged frontiers, i.e., jumps and/or steep slopes. We
first review methods being explicitly developed either only for the detection of
jumps in a signal corrupted by noise or approaches that also include capturing (i.e.,
modeling) those very jumps. We show the advantages and limits of applications of
these methods, highlighting in which aspects further research is still necessary. We
conclude this chapter by listing some of the most in practice well-known methods.
532 T. Meinl and E.W. Sun
The issue of detecting and modeling jumps in time series has been recognized as an
essential task in time series analysis and therefore has already been considered
extensively in the literature. Though we introduce wavelet methods in the next
chapter only, we list the works based on these methods here as well without going
into details. We only note that wavelets, based on their characteristics, make
excellent tools for jump and spike detection, as it is this what they were developed
for in the first place (see Morlet (1983)).
Generally, the most appreciated procedures in the recent literature can be seen as
two different general approaches. One is via wavelets, and the other uses local
(linear) estimators and derivatives.
One of the first approaches using wavelets for jump detection in time series,
besides the classical wavelet literature, for example, Mallat and Hwang (1992), was
given by Wang (1995, 1998). He uses wavelets together with certain data-
dependent thresholds in order to determine where in the signal jumps have hap-
pened and whether they are significantly different from short-varying fluctuations,
and provides several benchmark signals. Assumptions about the noise
structure were made according to Donoho and Johnstone (1994), that is, the
approach is applicable for white (i.e., uncorrelated) Gaussian noise only. This
work was extended by Raimondo (1998) to include even more general cusp
definitions. More recent contributions extend these works to stationary noise
(Yuan and Zhongjie (2000)) and other distributions (Raimondo and Tajvidi
(2004)) and also provide theoretical results about asymptotic consistency.
A further application specifically on high-frequency data is given by Fan and
Wang (2007).
The other line is given by Qiu and Yandell (1998) who estimate jumps using
local polynomial estimators. This work is continued in Qiu (2003) where jumps are
not only detected but also represented in the extracted time series. Gijbels
et al. (2007) further refine the results by establishing a compromise between
a smooth estimation for the continuous parts of a curve and a good resolution of
jumps. Again, this work is limited to pure jumps only and, since it uses local linear
estimators as the main method, has no frequency interpretation available. Sun and
Qiu (2007) and Joo and Qiu (2009) use derivatives to test the signal for jumps and to
represent them.
Finally, we note a completely different approach that is presented in Kim and
Marron (2006): a purely graphical tool for the recognition of probable jumps (and
also areas where almost certainly no jumps occurred). Yet, the authors confess that
their work is only to be seen as a complementary approach, and refer to Carlstein
et al. (1994) for further thorough investigation.
We note that there exists already a good body of work about how to detect (i.e.,
estimate their location), and even how to model jumps (i.e., estimate their height),
though in most works bounded to strict requirements on the noise or the trend
model. However, although most models will also automatically include the
detection of steep slopes, they fail at modeling the slope itself. While a jump
18 Methods of Denoising Financial Data 533
can easily be presented (either by indicator functions or any other methods used in
the cited works above), matters are different with slopes: Since there can be given
no general formulation or model of the exact shape of a slope, any parametric
approach will fail or deliver only a poor approximation if the model does not fit
the occurred slope. Examples of such different kinds of slopes are innumerous:
Sine, exponential, and logarithmic decays are only the most basic forms to
approximate such events, which in practical examples rarely follow such ideal-
ized curves. Naturally, only nonparametric approaches will adapt themselves to
the true shape of the underlying trend but generally suffer the same drawbacks as
all linear and nonlinear methods pointed out above, i.e., they always have to
bargain a tradeoff between bias and signal fidelity.
We conclude this section by outlining the most popular and established filtering
methods.
component. Thus, indicator functions excel at modeling jumps but perform poorly
with other types of sudden changes. Second, for this approach, it is extremely
important to determine the location of the jump as exact as possible, as otherwise
the estimated trend in this area may be highly inaccurate.
It is therefore dubious, whether such parametric approaches are appropriate to
deal with economic and financial time series, though they will perform very well, if
their requirements are met.
X
N ð
ðXt mðtÞÞ2 þ o ðm00 ðxÞÞ dx:
2
min (18.19)
m
t¼1
It follows directly from this definition that for o ¼ 0 this yields us an interpo-
lation while for o ! 1 m will approximate a linear regression.
In practice, there emerges another difficulty: In many cases, the (optimal) choice
of o remains unclear. Although there exist several works that have established some
data-dependent rules for this, in many cases, when the assumptions about the noise do
not hold or the time series incorporates additional deterministic (e.g., cycles) or
stochastic components (e.g., outliers that are part of the system and not due to
measurement or other errors), the choice of the penalizing smoothing parameter is
a difficult task that has been and is still undergoing extensive research (see Morton
et al. (2009), Lee (2003), Hurvich et al. (1998), Cantoni and Ronchetti (2001),
Irizarry (2004)). We only mention particularly the cross-validation method which is
used to determine the optimal smoothing parameters (see Green and Silverman
(1994)). Furthermore, though o is eventually responsible for the degree of smooth-
ness (i.e., on which scale or level the trend shall be estimated), one can hardly neither
impose nor derive any additional meaning on or from this parameter.
X
N
min ðXt tt Þ2 þ oððttþ1 tt Þ ðtt tt1 ÞÞ2 : (18.20)
t
t¼1
We note that (besides from not using a spline basis for approximation)
this approach can be seen as a discretized formulation of the smoothing spline.
18 Methods of Denoising Financial Data 535
The smoothing parameter o plays the same role, while the penalized smoothness
measure is the discretized version of the second derivative. Thus, though several
authors propose explicit rules (of thumb) for choosing o (see, e.g., Ravn and Uhlig
(2002), Dermoune et al. (2008), Schlicht (2005)), some researchers like Harvey and
Trimbur (2008) also recognize that in some way this choice still remains kind of
arbitrary or problematic for many time series which cannot be associated in the same
time terms.
with A the state transition model that relates in conjunction with the (optional)
control input B, the respective input ut, and the noise component
wt the previous state to the next. In the above equation, A and B are ass-
umed to be constant but may also change over time. Of this model (i.e., the true
state xt), only
zk ¼ Hxk þ vk (18.22)
With A, B, Q, and R assumed to be known, the filter predicts the next state xt
based on xt1 and also provides an estimate of the accuracy of the actual prediction.
Since its first development, the Kalman filter has become popular in many areas;
see, for example, Grewal et al. (2001). However, a serious drawback of this Kalman
procedure is that many real-world models do not fit the assumptions of the model, for
example, above requirement of a linear underlying system is often not met. Though
there exist extensions for nonlinear systems (see e.g., Julier and Uhlmann (1997)),
there still exists the problem that one or more of the required parameters are unknown.
While for many technical systems (e.g., car or missile tracking systems) based on
physical laws the state transition model A is exactly known, this becomes a difficult
issue in many other application areas, including finance. Additionally, as, e.g., Mohr
(2005) notes, the performance of the Kalman filter can be very sensitive to initial
conditions of the unobserved components and their variances, while at the same time it
536 T. Meinl and E.W. Sun
18.6 Summary
In this chapter we provided an overview of the different tasks of time series analysis
in general and the specific challenges of time series trend extraction. We pointed
out several methods and outlined their advantages and disadvantages, together with
their requirements. The reader should keep in mind the following key points:
• Linear filters provide a very smooth trend but fail to capture sudden changes.
• Nonlinear filters capture jumps extremely well but are very sensitive to higher
levels of noise.
• Advanced methods may provide good results depending on the scenario but
often rely on very specific assumptions that can yield misleading results, in case
they are not completely met.
• The specific task of jump detection and modeling is well understood but lacks
appropriate methods for more general changes like steep slopes.
• And the most important: No methods universally perform best for all tasks!
References
Astola, J., & Kuosmanen, P. (1997). Fundamentals of nonlinear filtering. Boca Raton: CRC Press.
Au, S., Duan, R., Hesar, S., & Jiang, W. (2010). A framework of irregularity enlightenment for
data pre-processing in data mining. Annals of Operations Research, 174(1), 47–66.
Bianchi, G., & Sorrentino, R. (2007). Electronic filter simulation & design. New York: McGraw-
Hill. ISBN 0071494677.
Cantoni, E., & Ronchetti, E. (2001). Resistant selection of the smoothing parameter for smoothing
splines. Statistics and Computing, 11(2), 141–146.
Carlstein, E., Mueller, H. G., & Siegmund, D. (1994). Change-point problems. Hayward,
CA: Institute of Mathematical Statistics.
Dacorogna, M. M. (2001). An introduction to high-frequency finance. San Diego, CA: Academic Press.
Dermoune, A., Djehiche, B., & Rahmania, N. (2008). Consistent estimators of the smoothing
parameter in the hodrick-prescott filter. Journal of the Japan Statistical Society, 38, 225–241.
Donoho, D. L., & Johnstone, I. M. (1994). Ideal spatial adaptation by wavelet shrinkage.
Biometrika, 81(3), 425–455.
Fan, J., & Wang, Y. (2007). Multi-scale jump and volatility analysis for high-frequency financial
data. Journal of the American Statistical Association, 102(480), 1349–1362.
Fan, J., & Yao, Q. (2003). Nonlinear time series: Nonparametric and parametric methods.
New York: Springer.
Gijbels, I., Lambert, A., & Qiu, P. (2007). Jump-preserving regression and smoothing using local
linear fitting: A compromise. Annals of the Institute of Statistical Mathematics, 59(2), 235–272.
18 Methods of Denoising Financial Data 537
Green, P. J., & Silverman, B. W. (1994). Nonparametric regression and generalized linear models:
A roughness penalty approach. New York: Chapman & Hall/CRC. ISBN 0412300400.
Grewal, M. S., Andrews, A. P., & Ebooks Corporation. (2001). Kalman filtering theory and
practice using MATLAB. New York: Wiley Online Library. ISBN 0471392545.
Harvey, A. C. (1989). Structural time series models and the Kalman filter forecasting. Cambridge:
Cambridge University Press.
Harvey, A., & Trimbur, T. (2008). Trend estimation and the Hodrick-Prescott filter. Journal of the
Japan Statistical Society, 38(1), 41–49.
Hodrick, R. J., & Prescott, E. C. (1997). Postwar US business cycles: An empirical investigation.
Journal of Money, Credit and Banking, 29(1), 1.
Hurvich, C. M., Simonoff, J. S., & Tsai, C. L. (1998). Smoothing parameter selection in nonpara-
metric regression using an improved Akaike information criterion. Journal of the Royal
Statistical Society, Series B: Statistical Methodology, 60(2), 271–293.
Irizarry, R. A. (2004). Choosing smoothness parameters for smoothing splines by minimizing an
estimate of risk (p. 30). Johns Hopkins University, Department of Biostatistics Working Papers.
Joo, J. H., & Qiu, P. (2009). Jump detection in a regression curve and its derivative.
Technometrics, 51(3), 289–305.
Julier, S. J., & Uhlmann, J. K. (1997). A new extension of the Kalman filter to nonlinear systems.
In International symposium on aerospace/defense sensing, simulation and controls
(Vol. 3, p. 26). Citeseer.
Kalman, R. E. (1960). A new approach to linear filtering and prediction problems. Journal of Basic
Engineering, 82(1), 35–45.
Kim, C. S., & Marron, J. S. (2006). Sizer for jump detection. Journal of Nonparametric Statistics,
18(1), 13–20.
Lee, T. (2003). Smoothing parameter selection for smoothing splines: A simulation study.
Computational Statistics & Data Analysis, 42(1–2), 139–148.
Leser, C. E. V. (1961). A simple method of trend construction. Journal of the Royal Statistical
Society: Series B: Methodological, 23(1), 91–107.
Lucia, J. J., & Schwartz, E. S. (2002). Electricity prices and power derivatives: Evidence from the
nordic power exchange. Review of Derivatives Research, 5(1), 5–50.
Mallat, S. G., & Hwang, W. L. (1992). Singularity detection and processing with wavelets. IEEE
Transactions on Information Theory, 38(2), 617–643.
Meinl, T., & Sun, E. W. (2012). A nonlinear filtering algorithm based on wavelet transforms for
high-frequency financial data analysis. Studies in Nonlinear Dynamics and Econometrics, 16(3),
1–24.
Mohr, M. F. (2005). A trend-cycle (-season) filter. Working Paper Series European Central Bank, (499).
Morlet, J. (1983). Sampling theory and wave propagation. Issues in Acoustic Signal/Image
Processing and Recognition, 1, 233–261.
Morton, R., Kang, E. L., & Henderson, B. L. (2009). Smoothing splines for trend estimation and
prediction in time series. Environmetrics, 20(3), 249–259.
Peltonen, S., Gabbouj, M., & Astola, J. (2001). Nonlinear filter design: Methodologies and
challenges. In Proceedings of the 2nd International Symposium on Image and Signal
Processing and Analysis, 2001. ISPA 2001, (pp. 102–107).
Percival, D. B., & Walden, A. T. (2006). Wavelet methods for time series analysis. Cambridge:
Cambridge University Press.
Pollock, D. S. G. (2001). Methodology for trend estimation. Economic Modelling, 18(1), 75–96.
Qiu, P. (2003). A jump-preserving curve fitting procedure based on local piecewise-linear kernel
estimation. Journal of Nonparametric Statistics, 15(4), 437–453.
Qiu, P., & Yandell, B. (1998). A local polynomial jump-detection algorithm in nonparametric
regression. Technometrics, 40(2), 141–152.
Rachev, S. T. (2003). Handbook of heavy tailed distributions in finance. North-Holland.
Raimondo, M. (1998). Minimax estimation of sharp change points. Annals of Statistics, 26(4),
1379–1397.
538 T. Meinl and E.W. Sun
Raimondo, M., & Tajvidi, N. (2004). A peaks over threshold model for change-point detection by
wavelets. Statistica Sinica, 14(2), 395–412.
Ramsey, J. B. (1999). The contribution of wavelets to the analysis of economic and financial data.
Philosophical Transactions: Mathematical, Physical and Engineering Sciences, 357(1760),
2593–2606.
Ramsey, J. B. (2002). Wavelets in economics and finance: Past and future. Studies in Nonlinear
Dynamics & Econometrics, 6(3), 1–27.
Ravn, M. O., & Uhlig, H. (2002). On adjusting the Hodrick-Prescott filter for the frequency of
observations. Review of Economics and Statistics, 84(2), 371–376.
Schlicht, E. (2005). Estimating the smoothing parameter in the so-called Hodrick-Prescott filter.
Journal of the Japan Statistical Society, 35(1), 99–119.
Seifert, J., & Uhrig-Homburg, M. (2007). Modelling jumps in electricity prices: Theory and
empirical evidence. Review of Derivatives Research, 10(1), 59–85.
Sun, E. W., & Meinl, T. (2012). A new wavelet-based denoising algorithm for high-frequency
financial data mining. European Journal of Operational Research, 217(3), 589–599.
Sun, J., & Qiu, P. (2007). Jump detection in regression surfaces using both first-order and second-
order derivatives. Journal of Computational and Graphical Statistics, 16(2), 289–311.
Sun, W., Rachev, S. T., Fabozzi, F., & Stoyan S. (2008). Multivariate skewed student’s t copula in
analysis of nonlinear and asymmetric dependence in German equity market. Studies in
Nonlinear Dynamics and Econometrics, 12(2).
Tsay, R. S. (2005). Analysis of financial time series. Hoboken: Wiley.
Wang, Y. (1995). Jump and sharp cusp detection by wavelets. Biometrika, 82(2), 385–397.
Wang, Y. (1998). Change curve estimation via wavelets. Journal of the American Statistical
Association, 93(441), 163.
Welch, G., & Bishop, G. (1995). An introduction to the Kalman filter. Chapel Hill: University of
North Carolina at Chapel Hill.
Yuan, L., & Zhongjie, X. (2000). The wavelet detection of the jump and cusp points of a regression
function. Acta Mathematicae Applicatae Sinica (English Series), 16(3), 283–291.
Analysis of Financial Time Series Using
Wavelet Methods 19
Philippe Masset
Contents
19.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 540
19.2 Frequency-Domain Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 541
19.2.1 Spectral Analysis: Some Basics and an Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 541
19.2.2 Filtering Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 546
19.3 Scale-by-Scale Decomposition with Wavelets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 549
19.3.1 Theoretical Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 552
19.3.2 Implementation and Practical Issues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 560
19.3.3 Illustration: Home Price Indices for Different US Cities . . . . . . . . . . . . . . . . . . . . . 563
19.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 571
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 571
Abstract
This chapter presents a set of tools, which allow gathering information about the
frequency components of a time series. In a first step, we discuss spectral
analysis and filtering methods. Spectral analysis can be used to identify and to
quantify the different frequency components of a data series. Filters permit to
capture specific components (e.g., trends, cycles, seasonalities) of the original
time series. Both spectral analysis and standard filtering methods have two main
drawbacks: (i) they impose strong restrictions regarding the possible processes
underlying the dynamics of the series (e.g., stationarity) and (ii) they lead to
a pure frequency-domain representation of the data, i.e., all information from the
time-domain representation is lost in the operation.
In a second step, we introduce wavelets, which are relatively new tools in
economics and finance. They take their roots from filtering methods and Fourier
analysis, but overcome most of the limitations of these two methods. Their principal
P. Masset
Ecole Hôtelière de Lausanne, Le-Chalet-à-Gobet, Lausanne 25, Switzerland
e-mail: philippe.masset@ehl.ch
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 539
DOI 10.1007/978-1-4614-7750-1_19,
# Springer Science+Business Media New York 2015
540 P. Masset
advantages derive from (i) combined information from both time domain and
frequency domain and (ii) their flexibility as they do not make strong assumptions
concerning the data-generating process for the series under investigation.
Keywords
Filtering methods • Spectral analysis • Fourier transform • Wavelet filter •
Continuous wavelet transform • Discrete wavelet transform • Multiresolution
analysis • Scale-by-scale decomposition • Analysis of variance • Case-Shiller
home price indices
19.1 Introduction
The purpose of this chapter is to present a set of methods and tools belonging to the
so-called frequency-domain analysis and to explain why and how they can be used
to enhance the more conventional time-domain analysis. In essence, time-domain
analysis studies the evolution of an economic variable with respect to time, whereas
frequency-domain analysis shows at which frequencies the variable is active. We
focus on concepts rather than technicalities and illustrate each method with exam-
ples and applications using real datasets.
The usual time-domain approach aims at studying the temporal properties
of a financial or economic variable, whose realizations are recorded at
a predetermined frequency. This approach does not convey any information
regarding the frequency components of a variable. Thus, it makes the implicit
assumption that the relevant frequency to study the behavior of the variable
matches with its sampling frequency. An issue arises, however, if the variable
realizations depend (in a possibly complicate manner) on several frequency
components rather than just one. In such a case, the time-domain approach
will not be able to efficiently process the information contained in the original
data series.
In this chapter, we start by discussing methods belonging to the frequency-
domain analysis. These tools are very appealing to study economic variables that
exhibit a cyclical behavior and/or are affected by seasonal effects (e.g., GDP,
unemployment). Spectral analysis and Fourier transforms can be used to quantify
the importance of the various frequency components of the variable under investi-
gation. In particular, they allow inferring information about the length of a cycle
(e.g., business cycle) or a phase (e.g., expansion or recession). The presence of such
patterns also imposes the use of appropriate methods when it comes to model the
dynamics of the variable. Filtering methods have proven useful in this context.
Notably, filters may serve to remove specific frequency components from the
original data series.
In a second step, we introduce wavelets. During the last two decades, wavelets have
become increasingly popular in scientific applications such as signal processing and
functional analysis. More recently, these methods have also started to be applied to
financial datasets. They are very attractive as they possess the unique ability to provide
19 Analysis of Financial Time Series Using Wavelet Methods 541
a complete representation of a data series from both the time and frequency perspec-
tives simultaneously. Hence, they allow breaking down the activity on the market into
different frequency components and to study the dynamics of each of these compo-
nents separately. They do not suffer from some of the limitations of standard
frequency-domain methods and can be employed to study a financial variable, whose
evolution through time is dictated by the interaction of a variety of different frequency
components. These components may also behave according to nontrivial (noncyclical)
dynamics – e.g., regime shifts, jumps, and long-term trends.
For instance, the presence of heterogeneous agents with different trading
horizons may generate very complex patterns in the time series of stock prices
(see M€ uller et al. 1995; Lynch and Zumbach 2003). This heterogeneity may in
particular induce long memory in stock return volatility. In such a case, studying the
properties of a time series and trying to model it from the perspective of a single
frequency can be misleading. Much information will be lost because of the naive
and implicit aggregation of the different frequency components into a single com-
ponent. Furthermore, as these components may interact in a complicated manner
and may be time varying or even nonstationary, standard methods like Fourier
analysis are not appropriate. Therefore, one has to resort to more flexible filtering
methods like wavelets.
The remaining of this chapter is structured as follows. In Sect. 19.2, we
discuss spectral analysis and filtering methods. Section 19.3 is devoted to the
presentation of wavelets, Sect. 19.3.1 explains the relevant theoretical back-
ground, Sect. 19.3.2 discusses the implementation of these methods, and
Sect. 19.3.3 presents a complete case study. Section 19.4 offers a short
conclusion.
component f in the dynamics of the original series. This is achieved using the
discrete version of the Fourier transform1
X1
X ðf Þ ¼ t¼1
xðtÞei2pft , (19.1)
where f denotes the frequency at which X(f) is evaluated. In order to gain a deeper
insight into this decomposition, one may think about the De Moivre’s (Euler’s)
theorem, which allows to write e–i2pft as
ei2pft ¼ cosð2pf tÞ i sinð2pftÞ: (19.2)
Equation 19.3 shows that X(f) determines how much of each frequency compo-
nent is needed to synthesize the original signal x(t).
1 X1
sx ðoÞ ¼ g eioj ,
j¼1 j
(19.4)
2p
where gj is the jth autocovariance of x(t)2; o ¼ 2pf is a real scalar, which is related to the
frequency f ¼ 1/t at which the spectrum is evaluated; and t is the period length of one
cycle at frequency f.3 One may notice that the right part of Eq. 19.4 is indeed the
discrete-time Fourier transform of the autocovariance series. There is also a close link
between this expression and the autocovariance generating function, which is defined by
X1
gx ð z Þ ¼ j¼1 j
g zj , (19.5)
where z denotes a complex scalar. This implies that one can easily recover the
autocovariance generating function from the spectrum. In the same spirit, an
1
The discrete version of the Fourier transform is used because the time series is recorded at
discrete-time intervals.
2
The jth autocovariance of x(t) is given by gj ¼ E[(x(t) m)(x(t j) m)], where m denotes the
expected value of x(t).
3
As an example, let us consider an economic variable, whose evolution is fully determined by the
state of the economy. A complete business cycle lasts on average 36 months and therefore
f ¼ 1/36 months.
19 Analysis of Financial Time Series Using Wavelet Methods 543
1 1
^s x ðoÞ ¼ j X ðf Þj 2 , (19.6)
2p T
where T is the length of the time series x(t). │X(f)│2 is also known as the power
spectrum of x(t). This approach is usually called the “periodogram.” As noted in
Hamilton (1994), its accuracy seems questionable as the confidence interval for the
estimated spectrum is typically very broad. Furthermore, the variance of the
periodogram does not tend to zero as the length of the data series tends to infinity.
This implies that the periodogram is not a consistent estimator of the population
spectrum. Therefore, modified versions of the periodogram have been put forward.
For instance, smoothed periodogram estimates have been suggested as a way to
reduce the noise of the original estimator and to improve its accuracy. The idea
underlying this approach is that sx(o) will be close to sx(l) when o is close to l.
This suggests that sx(o) might be estimated with a weighted average of the values of
sx(o) for values of l in a neighborhood around o, where the weights depend on the
distance between o and l (Hamilton 1994). The weights are typically determined
by a kernel weighting function. Welch’s method (Welch 1967) and Childers (1978)
constitute another alternative based on a simple idea: instead of estimating a single
periodogram for the complete sample, one divides the original sample into sub-
samples, estimates the periodogram for each subsample, and computes the average
periodogram over all subsamples.
The second approach is based on some parameterization of the data-
generating process of x(t). Methods belonging to this category are close in spirit
to the population spectrum, i.e., to a direct application of Eq. 19.4. Typically
some specification based on an ARMA (autoregressive moving average) repre-
sentation is chosen to represent the temporal dynamics of the variable. The
model is then calibrated, i.e., the ARMA coefficients are estimated from the
realizations of the process x(t). These estimated coefficients are employed to
calculate the spectrum. As long as the autocovariances are reasonably well
estimated, the results would also be reasonably close to the true values.
A detailed discussion of the various parametric methods (e.g., the covariance,
Yule-Walker and Burg methods) is beyond the scope of this introduction, but
parametric methods are particularly effective when the length of the observed
sample is short. This is due to their ability to distinguish the noise from the
information contained in the data.
544 P. Masset
19.2.1.3 Example
We now turn to the discussion of a simple example. We consider a time series,
which has the following dynamic:
2pt 2pt
xðtÞ ¼ a cos þ b sin þ eðtÞ,
21 63
where e(t) is a random term that follows a normal distribution with mean zero and
unit variance. One may observe that the process is driven by two cyclical compo-
nents, which repeat themselves, respectively, each 21 and 63 units of time.
The full line in Fig. 19.1 shows the first 100 (simulated) realizations of x(t); the
dotted lines are for the cos and sin functions. At first glance, it seems difficult to
distinguish the realizations of x(t) from a purely random process. Figure 19.2
reports the autocorrelations (left panel) and partial autocorrelations (right panel)
of x(t) (upper panel) and of the cos and sin components (bottom panel). Again, it
remains difficult, when looking at this figure, to gather conclusive evidence
concerning the appropriate model specification for x(t).
On the other hand, results from the Fourier analysis, reported in Fig. 19.3, clearly
show that two cyclical components drive the evolution of x(t) and repeat themselves
around each 21 and 63 units of time. This demonstrates the effectiveness of Fourier
methods for the study of processes featuring cyclical components.
Evolution of xt
4
−1
−2
−3 xt cos sin
−4
0 10 20 30 40 50 60 70 80 90 100
ACF: xt PACF: xt
0.4 0.4
0.2 0.2
0 0
−0.2 −0.2
0 5 10 15 20 25 0 5 10 15 20 25
0 0 0 0
Fig. 19.2 Autocorrelations and partial autocorrelations of x(t). The upper panel reports the
autocorrelations (left panel) and partial autocorrelations (right panel) of x(t). The lower panel
shows similar statistics for the cosinus and sinus functions
(e.g., periodogram estimates). On the other hand, the difference between the
Yule-Walker and the Burg methods is minimal. Nevertheless, the key message
remains remarkably similar: strong seasonalities affect home prices with
a frequency of recurrence of 12 months.
546 P. Masset
Power Spectrum of xt
700 Periodogram
Population Spectrum
600
500
400
300
200
100
0
0 10 20 30 40 50 60 70 80 90 100
Periods
Fig. 19.3 Spectral analysis. The periodogram has been estimated directly from the realizations of
x(t). The population spectrum has been estimated on the basis of the theoretical autocovariances of x(t)
A filter is a mathematical operator that serves to convert an original time series x(t)
into another time series y(t). The filter is applied by convolution of the original
series x(t) with a coefficient vector w:
X1
yðtÞ ¼ ðw xÞðtÞ ¼ k¼1
wðkÞxðt kÞ, (19.7)
Index level
200
150
100
Returns [in %]
2
−2
1990 1995 2000 2005 2010
Fig. 19.4 Case-Shiller home price index for the city of New York. Price levels and returns are
reported in the upper and lower panel respectively.
ACF PACF
1 1
0.5 0.5
0 0
−0.5 −0.5
0 10 20 30 40 0 10 20 30 40
Fig. 19.5 Autocorrelations and partial autocorrelation of the returns on the New York home
price index
The frequency response, H(f), can be further split into two parts:
where G(f) is the gain function, eiy(f) the phase function, and y the phase angle
(or equivalently the argument of H( f )). The gain function is the magnitude of the
frequency response, i.e., G(f) ¼ jH( f )j. If the application of the filter on x(t) results in
a phase shift, i.e., if the peaks and lows of x(t) and y(t) have a different timing, the
phase angle y will be different from zero. The use of uncentered moving average
548 P. Masset
Power Spectrum
3
Periodogram
2.5
Welch
Yule-Walker
2 Burg
1.5
0.5
0
0 5 10 15 20
Fig. 19.6 Spectral analysis of the return series. Comparison between parametric (Burg and
Yule-Walker) and non-parametric (Periodogram and Welch) methods
filters leads to this (often) undesirable feature because turning points will be recorded
earlier in the original series than in the filtered series. On the other hand, centered
(symmetric) moving averages have y( f ) ¼ 0; hence, there is no phase shift for this
class of filters. For instance, the frequency response of a two-period uncentered
moving average filter with coefficients w(k) ¼ 0.5 for k ¼ 0, 1 is4
X1
H ðf Þ ¼ k¼0
0:5ei2pfk
¼ 0:5 þ 0:5ei2pf
¼ 0:5 eipf þ eipf eipf
¼ cosðpf Þeipf
This result shows the existence of a phase shift as the phase angle is y( f ) ¼ pf.5
Hence, the turning points from the original series will be shifted to the right in the
filtered series. On the other hand, a filter with coefficients w(k) ¼ 1/3 for k ¼ 1, 0, 1
has a zero phase angle.
Based on their gain functions, filters can be categorized as follows:
• High-pass filters should be able to capture the high-frequency components of
a signal, i.e., the value of their gain function G(f) should equal one for frequen-
cies f close or equal to 1/2.
4
See also Gençay et al. (2002) who uses a similar example.
Im½H ðf Þ
In full generality, the phase angle can be computed as yðf Þ = arctan
5
Re½Hðf Þ , where Im[H(f)] and
Re[H(f)] are, respectively, the imaginary part and the real part of H(F).
19 Analysis of Financial Time Series Using Wavelet Methods 549
19.2.2.2 Example
A basic example of a high-pass filter is a filter that takes the difference between two
adjacent values from the original series; its coefficients are whi ¼ [0.5, 0.5]. Similarly,
the most simple low-pass filter is p a ffiffiffi 2-period pmoving
ffiffiffi average; in this case
wlo ¼ [0.5, 0.5]. In wavelet theory, whi = 2 and wlo = 2 form the Haar wavelet family.
In this case, the low-pass filter wlo is basically an averaging filter, while the high-pass
filter whi is a differencing filter. The gain functions for these two filters are reported in
Fig. 19.7.
19.2.2.3 Illustration
The full line in the left panels of Fig. 19.8 shows the monthly (unadjusted) returns on the
Case-Shiller New York home price index from 1987 to 2011. We add to the top panel
the output series resulting from the application of both a centered and an uncentered
(causal) 3-period moving average on the original data. The three-filter coefficients have
a value of 1/3. This implies that the output series of the centered moving average at time
t is basically the average return from months t – 1 to t + 1. Similarly, the uncentered
moving returns the average return from t – 2 to t. It is apparent from the figure that the
uncentered moving average leads to a phase shift of 1 month. The bottom panel shows
the outputs of a 7-period moving average. The filter coefficients are equal to 1/7. Again,
we consider both centered and uncentered filters. The use of an uncentered moving
average leads to a phase shift of 3 months as compared to the centered moving average.
The right part of the figure reports the gain functions for the 3-period and the
7-period moving averages. The gain function is similar for both the uncentered and
the centered moving average. The only element that distinguishes these two filters
is indeed their phase function. One may notice that (i) both filters are low-pass
filters, and (ii) the longer filter captures more efficiently the low-frequency
components of the original signal than the short filter.
To a large extent, wavelets can be seen as a natural extension to spectral and Fourier
analysis as (i) wavelets do not suffer the weaknesses of Fourier analysis, and
(ii) wavelets provide a more complete decomposition of the original time series
than Fourier analysis does.
550 P. Masset
Haar filter
1
whi
0.8 wlo
0.6
Gain
0.4
0.2
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Frequency
Fig. 19.7 Gain functions of the high-pass (differencing) and low-pass (averaging) Haar filters
0
0.4
−1
0.2
−2
0
2000 0 0.1 0.2 0.3 0.4 0.5
Time Frequency
0
0.4
−1
0.2
−2
0
2000 0 0.1 0.2 0.3 0.4 0.5
Time Frequency
Ch. CPI Centered Uncentered
Fig. 19.8 Two moving averages and their frequency responses. Left panels show the outputs from
a 3-period (upper panel) and a 7-period (lower panel) moving average filters applied on the returns on
the Case-Shiller New York home price index. Right panels report the corresponding gain functions
19 Analysis of Financial Time Series Using Wavelet Methods 551
There are some problems with spectral methods and Fourier transforms. Notably,
these methods require the data under investigation to be stationary. This is often not the
case in economics and finance. In particular, volatility is known to exhibit complicated
patterns like jumps, clustering, and long memory. Furthermore, the frequency decom-
position delivered by Fourier analysis only makes sense if the importance of the various
frequency components remains stable over the sample period. Ex ante, there is good
reason to expect this assumption not to hold for a variety of economic and financial
variables. For instance, volatility changes are likely to exhibit a different frequency
spectrum when trading activity is intense than when the market is quiet. The short-time
Fourier transform (which is also known as the Gabor or windowed Fourier transform)
has been suggested to overcome these difficulties. The idea is to split the sample into
subsamples and to compute the Fourier transform on these subsamples. Hence, this
extension achieves a better trade-off between the time and the frequency representation
of the original data. Nevertheless, this provides at best a partial solution to the
aforementioned issues as the strong restrictions regarding the possible data-generating
process over each subsample are maintained.
Wavelets do not make any of these assumptions. Furthermore, wavelets pro-
vide a complete decomposition of the original series, which is located both in time
and in frequency. From a mathematical viewpoint, a wavelet is a function, which
enables to split a given signal into several components, each reflecting the
evolution trough time of the signal at a particular frequency. Wavelet analysis
has originally been used in signal processing (e.g., image processing and data
compression). Its applications to economics and finance are relatively recent.
Nevertheless, the range of application of wavelets is potentially wide: denoising
and seasonality filtering, decorrelation and estimation of fractionally integrated
models, identification of regime shift and jumps, robust estimation of the covari-
ance and correlation between two variables at different time scales, etc.
From a physicist perspective, but with application to time series analysis,
Percival and Walden (2000) and Struzik (2001) provide a mathematically rigorous
and exhaustive introduction to wavelets. Struzik (2001) particularly emphasizes the
unique ability of nonparametric methods (like wavelets) to let the data speak by
themselves. Thus, such methods avoid making misleading interpretations of the
coefficients obtained from the calibration of misspecified models. Gençay
et al. (2002) discuss the use of wavelets for specific purposes in economics and
finance and adopt a more intuitive approach (with many illustrations and examples),
while Ramsey (2002) surveys the most important properties of wavelets and
discusses their fields of application in both economics and finance. Crowley
(2007) proposes a genuine guide to wavelets for economists. His article can be
considered as a complete and easily understandable toolkit, precisely explaining in
which circumstances to use wavelets and how to proceed. Schleicher (2002) is
a complementary reference to those already named; Schleicher focuses on some
mathematical concepts underlying the use of wavelets and discusses them in details
using examples.
This short literature review focuses only on textbook-style references. There are
however quite a large amount of economic and financial articles that have employed
552 P. Masset
wavelets for empirical purposes. Ramsey6 and Gençay7 and their respective coau-
thors can be considered as the pioneers of the use of wavelets in economics and
finance. Other recent contributions include Nielsen and Frederiksen (2005),
Vuorenmaa (2005), Oswiecimka et al. (2006), Elder and Jin (2007), Fan
et al. (2007), Fernandez and Lucey (2007), Rua and Nunes (2009), Manchaldore
et al. (2010), Curci and Corsi (2012), and Hafner (2012).
Taken together, these conditions imply (i) that at least some coefficients of the
wavelet function must be different from zero and (ii) that these departures from zero
must cancel out. Clearly the sin function does not meet these two requirements.
A vast variety of functions meets conditions (19.10) and (19.11). Nevertheless, these
conditions are very general and not sufficient for many practical purposes. Therefore,
one has to impose additional conditions in order to run a specific analysis with
wavelets. One of these conditions is the so-called admissibility condition, which
states that a wavelet function is admissible if its Fourier transform
Z 1
Cðf Þ ¼ cðtÞei2pft dt, (19.12)
1
is such that
Z 1
jCðf Þj2
CC ¼ df satisfies 0 < CC < 1: (19.13)
0 f
These conditions allow reconstructing a function from its continuous wavelet
transform (see Percival and Walden (2000) for more details).
6
See Ramsey et al. (1995), Ramsey and Zhang (1997), and Ramsey (1999).
7
See Gençay et al. (2003), Gençay et al. (2010), Gençay and Fan (2010), and Gençay and
Gradojevic (2011).
8
The Morlet wavelet is actually similar to a sin curve modulated by a Gaussian envelope.
19 Analysis of Financial Time Series Using Wavelet Methods 553
Wave or wavelet?
1
Morlet Wavelet
0.8 sin(x)
0.6
0.4
0.2
−0.2
−0.4
−0.6
−0.8
−1
−20 −15 −10 −5 0 5 10 15 20
1 t u
cu,s ðtÞ ¼ pffiffi c , (19.14)
s s
where u and s are the location and scale parameters. The term p1ffis ensures that the
norm of cu,s(t) is equal to one. The CWT, W(u, s), which is a function of the two
parameters u and s, is then obtained by projecting the original function x(t) onto the
mother wavelet cu,s(t):
Z 1
W ðu; sÞ¼ xðtÞcu,s ðtÞdt: (19.15)
1
To assess the variations of the function on a large scale (i.e., at a low frequency),
a large value for s will be chosen and vice versa. By applying the CWT for
a continuum of location and scale parameters to a function, one is able to
554 P. Masset
Where j and k are integers representing the set of discrete translations and discrete
dilatations. Gençay et al. (2002) refer to this procedure as the critical sampling of
the CWT. This implies that the wavelet transform of the original function or
signal is calculated only at dyadic scales, i.e., at scales 2j. A further implication is
that for a time series with T observations, the largest number of scales for the
DWT is equal to the integer J such that J ¼ blog2(T)c ¼ blog(T)/log(2)c. It is not
possible to directly apply the DWT if the length of the original series is not dyadic
(i.e., if J < log2(T) < J + 1). In such case, one has either to remove some
observations or to “complete” the original series in order to have a series of
dyadic length. Several methods exist to deal with this kind of boundary problems
(see Sect. 19.3.2).
The DWT is based on two discrete wavelet filters, which are called the mother
wavelet hl ¼ (hl, . . .,hL–1) and the father wavelet g1 ¼ (g1,. . .,gL–1). The mother
wavelet is characterized by three basic properties:
These three properties ensure that (i) the mother wavelet is associated with
a difference operator, (ii) the wavelet transform preserve the variance of the original
data, and (iii) a multiresolution analysis can be performed on a finite variance data
series. The first property implies that the mother wavelet (also called “differencing
19 Analysis of Financial Time Series Using Wavelet Methods 555
XL1
l¼0
gl ¼ 1: (19.18)
The application of both the mother and the father wavelets allows separating the
low-frequency components of a time series from its high-frequency components.
Furthermore, a band-pass filter can be constructed by recursively applying
a succession of low-pass and high-pass filters.
Let’s assume that we have observed a sample of size T of some random variable
x(t), {x(1), x(2), . . ., x(T)}. The wavelet and scaling coefficients at the first level of
decomposition are obtained by convolution of the data series with the mother and
the father wavelets:
XL1 XL1
w1 ðtÞ ¼ l¼0
hl xðt0 Þ and v1 ðtÞ ¼ l¼0
gl xðt0 Þ (19.19)
9
The low-pass filter can be directly obtained from the high-pass filter using the quadrature mirror
relationship; see Percival and Walden (2000, p. 75).
10
For two integer a and b, a modulus b is basically the remainder after dividing a by b, i.e., a mod
b ¼ a–c b with c ¼ ba/bc.
556 P. Masset
XL1
XL1
0
w 1 ðt Þ ¼ l¼0
h l x ð t Þ and v 1 ð tÞ ¼ g xðt0 Þ,
l¼0 l
(19.21)
where t ¼ 0,1,. . .T and t0 ¼ t – l mod T. As for the DWT, the MODWT coefficients
for scales j > 1 can be obtained using the pyramid algorithm. For instance, w j and v j
are calculated as
11
See Crowley (2007) for more details about the properties of MODWT.
19 Analysis of Financial Time Series Using Wavelet Methods 557
XL1
XL1
w j ðt Þ ¼ h v ðt0 Þ and v j ðtÞ ¼
l¼0 l j1
g v ðt0 Þ,
l¼0 l j1
(19.22)
where t0 ¼ 2j–1 l mod T.
Using matrix notation, we can conveniently calculate
the wavelet and scaling
coefficients up to scale J. We first define a matrix W that is composed of J + 1
sub-matrices, each of them T x T:
2 3
W1
6 7
6 W 7
6 27
W¼ 6 ⋮ 7 (19.23)
6 7
4 WJ 5
VJ
Each W j has the following structure:
2 3
h 0 =2j=2 0 0 0 h L1 =2j=2 h 1 =2j=2
6 j=2 7
6 h 1 =2 h 0 =2j=2 0 0 0 ⋮ 7
6 7
6 ⋮ h 1 =2j=2 h 0 =2j=2 h L1 =2 7j=2 7
6 0 0
6 7
6 h L1 =2j=2 ⋮ h 1 =2j=2 0 0 0 7
6 7
6 h L1 =2j=2 ⋮ 7
6 0 0 0 0 7
Wj ¼ 6 7
6 0 0 h L1 =2j=2 0 0 0 7
6 7
6 ⋮ ⋮ ⋮ ⋱ ⋮ ⋮ ⋮ 7
6 7
6 0 0 0 h 0 =2j=2 0 0 7
6 7
6 h 1 =2j=2 h 0 =2j=2 7
6 0 0 0 0 7
4 0 0 0 ⋮ ⋮ 0 5
0 0 0 h L1 =2j=2 h L2 =2j=2 h 0 =2 j=2
(19.24)
V J has a similar structure as W J but it contains the coefficients associated to the
father wavelet instead of the mother wavelet.
We can now directly calculate all wavelet and scaling coefficients via
w ¼ W x, (19.25)
where w is a vector made up of J + 1 length T vectors of wavelet and scaling
T
coefficients, w, . . . , w j and vj ; i.e., w = w 1 w 2 w J v J .
12
Our presentation of the multiresolution analysis is restricted to the case of the MODWT.
Nevertheless, a very similar procedure exists for the DWT; see Percival and Walden (2000).
558 P. Masset
W is an orthonormal matrix, as such W TW = 1. Hence, if we multiply both side
of Eq. 19.25 by W T , we get
W T w ¼ W T Wx ¼ x: (19.26)
h i
T
As W T = W 1 W 2 W J V J and w = w 1 w2 w J v J , we can further
rearrange Eq. 19.26 and show that
XJ T T
x¼ j¼1
W j wj þ V J vJ : (19.27)
T T
Setting Dj = W j w j and Sj = V j v j , we can reconstruct the original time
series as
x ¼ D 1 þ . . . þ D J þ SJ : (19.28)
xS ¼ DS þ . . . þ DJ þ SJ : (19.29)
Equation 19.29 can be used, for instance, to filter out noise or seasonalities from
a time series. In image processing, Eq. 19.29 serves for data compression.
Equation 19.27 has been specifically derived for the MODWT but similar results
are available for the DWT (see Percival and Walden 2000).
Using the wavelet and scaling coefficients of the discrete wavelet transform, it is
possible to decompose the energy of the original series on a scale-by-scale basis:
XT1 XJ XT=2J 1
kx k2 ¼ t¼0
xðtÞ2 ¼ j¼1 t¼0
wj ðtÞ2 þ vJ ðtÞ2 , (19.30)
where kxk2 denotes the energy of x. The wavelet coefficients capture the deviations
of x from its long-run mean at the different frequency resolutions. Therefore, at
scale j ¼ J ¼ log2(T), the last remaining scaling coefficient is equal to the sample
mean of x,
Eð x Þ ¼ v J : (19.31)
The variance of the scaling coefficients has to be taken into account because vJp
incorporates deviations of x from its mean at scales Jp < j < J.13
An alternative way to decompose the energy of x is based on the smooth and
detail coefficients of the MRA. As above, kxk2 can be computed as the sum of the
energy of the smooth and detail coefficients. This approach is, however, valid only
for the DWT (See Gençay et al. 2002).
It is important to note that some of the wavelet coefficients involved in Eq. 19.32
are affected by boundary conditions. One should remove the corresponding wavelet
coefficients in order to get an unbiased estimator of the wavelet variance:
" #
XJ 1 X 2Tj 1
^ ðxÞ ¼
V 0 wj ðtÞ
2
, (19.34)
j¼1 ^j
2lj T t¼Lj
where lj is the scale that is associated to the frequency interval [1/2j+1 1/2j].
0
Lj ¼ d(L 2) (1 2 j)e is the number of DWT coefficients computed using the
boundaries. Hence, T^ j = T=2j L0 is the number of coefficients unaffected by the
j
boundary.
13
One may notice that the variance of the scaling coefficient at scale J is 0 as vJ is a scalar (the
sample mean of x).
560 P. Masset
meet this requirement, that is, their scaling and wavelet coefficients are orthog-
onal by construction.
• Smoothness: The degree of smoothness is measured by the number of continu-
ous derivatives of the basis function. As such, the Haar wavelet is the least
smooth wavelet. The choice of a more or less smooth filter depends essentially
on the data series to be represented. If the original time series is very smooth,
then one will opt for a smooth wavelet. For instance, the Haar wavelet is
appropriate for the analysis of a pure jump process.
• Number of vanishing moments: The number of vanishing moments of the
wavelet function has a direct implication on the ability of the wavelet
to account for the behavior of the signal. That is, if a signal has
a polynomial structure or if it can be approximated by a polynomial of
order q, then the wavelet transform will be able to properly capture this
polynomial structure only if it has q vanishing moments. For instance,
Daubechies wavelets have a number of vanishing moments which is half
the length of the filter. Thus, the Haar and D(8) have, respectively, 1 and
4 vanishing moments.
The last two properties depend not only on the wavelet filter but also on its
length. In fact, the most crucial point is probably not to choose the “right” filter but
to choose a filter with an appropriate length. Increasing filter length allows better
fitting the data. Unfortunately, this also renders the influence of boundary condi-
tions more severe. Hence, a trade-off has to be found.
Problems due to boundary conditions arise in two situations. The first case
concerns the DWT. To use the DWT, one requires a time series with a dyadic
length. If the series does not meet this requirement, i.e., if its length N is such that
2j < N < 2j+1, one has the choice between removing observations until N ¼ 2j and
completing the series such that N ¼ 2 j+1. Removing data might be the best
alternative but it leads to a loss of information.
The second case concerns both the DWT and the MODWT. The wavelet filter
has to be applied on all observations, including observations recorded at the
beginning (t ¼ 1). A problem arises because the convolution operator requires
that L – 1 observations are available before t. In this case, removing data is
useless. Therefore, one has to complete the data series. One solution is to pad
each end of the series with zeros; this technique is known as “zero padding.” An
alternative is to use the fit from a polynomial model to replace nonexisting data at
each end of the series (“polynomial approximation”). One may also complete
each end of the series either by mirroring the last observations (“mirror” or
“reflection”) or by taking the values observed at the beginning of the other end of
the series (“circular”). The choice depends on the data considered. For instance,
if working on stock returns, the use of a “mirror” seems to be the most suitable
approach as it accounts for the presence of volatility clustering. Moreover, after
the multiresolution decomposition of the original signal, one may obviously
discard the coefficients that are affected due to their proximity to the boundaries
of the series.
562 P. Masset
Instead of selecting a priori specific wavelet function, one may also use the
so-called optimal transforms. The idea is to choose the wavelet function that
minimizes a loss function (e.g., the entropy cost function; see Crowley 2007).
19.3.2.3 Example
Let’s consider a variable x, whose dynamics is primarily driven by an AR(1) process
and three cyclical components:
X5
2pt
yðtÞ ¼ 0:90yðt 1Þ þ s¼3
5 cos þ eðtÞ, (19.36)
s
e(t) is an i.i.d. Gaussian process with mean zero and unit variance and
t ¼ 1, . . . ,10,000. In the absence of seasonalities and noise, the autocorrelation
function of y should take value 0.90k at lag k. Wavelets can be used to remove the
impact of both noise and seasonalities. From Eq. 19.36, one may notice that the
cyclical components have a period length of 3–5 periods. Hence, they have an
impact on frequencies between 1/5 and 1/3.
At scale j, the wavelet detail Dj captures frequencies 1/2 j+1 f 1/2j and
the wavelet smooth Sj captures frequencies f < 1/2 j+1. If we use a level
2 multiresolution analysis, the wavelet smooth S2 will thus capture the compo-
nents of the time series, which have a frequency f < 1/8. This means that S2 will
take into account changes in y that are associated with a period length of at least
8 units of time. Therefore, S2 should keep the AR(1) dynamics of y, while
removing its cyclical behavior and noise.
Figure 19.14 reports the autocorrelation coefficients for the original time series y,
for the theoretical AR(1) process, and for the wavelet smooth S2. In order to assess the
impact of choosing a different wavelet filter, we use both the Haar and the LA
(8) wavelets. The results are very similar even if for short lags the LA(8) seems to
provide some improvements over the Haar. All in one, this example demonstrates the
ability of a wavelet filter to deal with a complex cyclical structure and with noise.
19 Analysis of Financial Time Series Using Wavelet Methods 563
Haar D4
0.1
0.05
0.05
0
0
−0.05 −0.05
D8 LA8
0.1
0.1
0.05 0.05
0 0
−0.05
−0.05
Fig. 19.11 Mother wavelet filters. The Haar, D(4), D(8), and LA(8) filters at scale j ¼ 4
Table 19.1 reports some descriptive statistics for the returns on each of the
12 home price indices. There are huge disparities in performance. The largest price
increases are to be found in Portland (+216 % since 1987 but 31 % since the
14
Los Angeles (LA), San Francisco (SF), Denver (De), Washington (Wa), Miami (Mi), Chicago
(Chi), Boston (Bos), Las Vegas (LV), New York (NY), Portland (Po), Charlotte (Cha), and
Cleveland (Cl).
564 P. Masset
Fig. 19.12 Gain function for the Haar, D(4), D(8), and LA(8) wavelets
19.3.3.2 Autocorrelations
Figure 19.15 shows the autocorrelations (up to 24 lags, i.e., 2 years) of the index
returns. The full lines are for the original series, while the dotted lines show the
autocorrelations computed from the smooth coefficients obtained using the MRA
from a partial MODWT. We employ a LA(8) filter with Jp ¼ 3. Hence, the wavelet
smooth (S3) should capture the frequency components that are associated with
a period length of at least 16 months and should therefore be free of seasonal
effects. In order to deal with the boundary conditions, we use the reflection method.
Some indices are very affected by seasonal effects (Portland, Boston, Denver,
Charlotte, and Cleveland) while other indices are much less affected
19 Analysis of Financial Time Series Using Wavelet Methods 565
G5(f)
1
Haar
0.9 D(4)
D(8)
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0 0.05 0.1 0.15 0.2 0.25 0.3 0.35 0.4 0.45 0.5
Fig. 19.13 Gain function for the Haar, D(4), and D(8) at scale j ¼ 5
1
0.8 AR(1) y S2 Haar S2 LA(8)
0.6
0.4
0.2
0
−0.2
−0.4
−0.6
−0.8
0 5 10 15 20 25 30 35 40 45 50
Lags
Fig. 19.14 Autocorrelations estimated before and after having removed some specific frequency
components of the original time series
566
LA SF Denver Washington
1 1 1 1
0 0 0 0
−1 −1 −1 −1
0 10 20 0 10 20 0 10 20 0 10 20
Miami Chicago Boston LV
1 1 1 1
0 0 0 0
−1 −1 −1 −1
0 10 20 0 10 20 0 10 20 0 10 20
NY Portland Charlotte Cleveland
1 1 1 1
0 0 0 0
−1 −1 −1 −1
0 10 20 0 10 20 0 10 20 0 10 20
Fig. 19.15 Autocorrelations of rj(t). This figure shows the autocorrelations for both the original
returns series and the wavelet smooth series for each home price index
(Miami, Las Vegas, and Los Angeles). In general the indices that display the least
significant seasonal patterns are also those that have been the most affected by the
recent crisis. This observation may suggest that the (quasi) absence of these patterns
is the result of the predominant role of speculation on price changes in these cities.
One may again observe the ability of wavelets to remove seasonal patterns. The
autocorrelations estimated from the wavelet smooth show the long-run temporal
dynamics of home prices. In contrast to financial markets, whose evolution is
almost unpredictable, home prices are strongly autocorrelated. The autocorrelation
remains positive even after 2 years.
0.6
Percentage of variance at scale j
0.5
0.4
0.3
0.2
0.1
0
1 2 3 4 5 6
Scale j
Fig. 19.16 Distribution of the wavelet variance across scales. Each bar corresponds to the
variance recorded at a specific scale j and for a particular city. Cities are ordered as follows
(from left to right): Miami, Las Vegas, Los Angeles, Washington, New York, San Francisco,
Chicago, Portland, Boston, Denver, Charlotte, Cleveland
cycle components (scales j > 5). Thus, it comes as no surprise that they do not have
much information content.
Scale 3 corresponds to periods of 8–16 months, and as such the variance observed
at this scale reflects the importance of seasonal patterns in the dynamics of the various
home price indices. As before, we observe that seasonal effects have a very limited
impact on prices in Miami; they are also much less important in cities like Las Vegas,
Los Angeles, Washington, and New York than in Boston, Denver, Charlotte, and
Cleveland. Interestingly the ordering is reversed when considering the variance at
scale 6. That is, the index returns series on which seasonalities have a weak impact
exhibit the largest percentage of long-term volatility.
Figure 19.17 reports the correlation between the various city indices at different
scales. The largest correlations are observed at scales larger than 2. In particular at
scale 3, the correlations are very significant. This is because seasonalities affect
most indices simultaneously. One may notice that indices 5 and 8 (Miami and Las
Vegas) show less correlation with the other indices. At scales 4 and 5, the correla-
tions become highly significant. This demonstrates that the series tend to behave
very similarly in the long run.
An extension to this correlation analysis is to study how many (economic)
factors are important to explain the structure of the correlation matrix at each
19 Analysis of Financial Time Series Using Wavelet Methods 569
W1 W2 W3
1
2 2 2
4 4 4
0.9
6 6 6
8 8 8
0.8
10 10 10
12 12 12
0.7
2 4 6 8 10 12 2 4 6 8 10 12 2 4 6 8 10 12
W4 W5 0.6
V6
2 2
2 0.5
4 4 4
6 6 6
0.4
8 8 8
10 10 10
0.3
12 12 12
2 4 6 8 10 12 2 4 6 8 10 12 2 4 6 8 10 12
Fig. 19.17 Correlations between home price indices at different frequencies. Panels denoted by
W1 to W5 show the correlations between the wavelet coefficients at scales j ¼ 1, . . ., 5, while panel
denoted V5 show the correlations between the scaling coefficients at scale j ¼ 5. The 12 cities are
order as follows: Los Angeles, San Francisco, Denver, Washington, Miami, Chicago, Boston,
Las Vegas, New York, Portland, Charlotte, Cleveland
scale. In order to address this question, one may resort to random matrix theory
(RMT). A good introduction is provided by Bouchaud and Potters (2004) (See also
Sharifi et al. (2004) for a literature review). Here, we concentrate on the main
premises of RMT. That is, this approach should help to (i) assess if the correlation
coefficients have a genuine information content or if they are merely due to the
noise inherent in the data and (ii) estimate the number of factors that are necessary
to “explain” the correlation matrix. This is done by comparing the eigenvalues of
the empirical correlation matrix with those from a theoretical distribution.
Let’s consider a dataset with N time series of length T. We assume that the
theoretical random matrix belongs to the ensemble of Wishart matrices. On this
basis, we can derive the theoretical distribution of the eigenvalues of the correlation
matrix. Under the null of pure randomness, the eigenvalues must be confined within
the bounds15:
sffiffiffi! sffiffiffi!
1 1 1 1
lmin ¼ 1þ 2 and lmax ¼ 1þ þ2 , (19.37)
q q q q
where q = NT . If the correlation matrix is random, then the probability that any of
its eigenvalues lies outside the bounds defined by [lmin, lmax] is zero. Hence, the
15
See Sharifi et al. (2004) and Kwapien et al. (2007).
570 P. Masset
W1 W2
0.02 0.02
0.01 0.01
0 0
0 0.5 1 1.5 2 0 0.5 1 1.5 2 2.5
-λ -
W3 W4
0.03
0.02
0.02
0.01
0.01
0 0
0 0.5 1 1.5 2 2.5 3 0 1 2 3 4
W5 V5
0.08
0.2
0.06
0.04 0.1
0.02
0 0
0 1 2 3 4 5 6 0 1 2 3 4 5 6
presence of eigenvalues larger than the upper bound can be taken as an evidence
that the structure of the correlation matrix is not due to chance, i.e., that there are
deviations from RMT. Furthermore, the number of such eigenvalues can be
interpreted as corresponding to the number of factors underlying the evolution of
the N time series. A potential problem with the RMT approach is that it
requires N ! 1 and T ! 1. In our case, these conditions are evidently not fulfilled
as T ¼ 302 and N ¼ 12. To account for this, we also estimate lmax from the empirical
distribution of the correlation matrix eigenvalues. To this aim, we resample (without
replacement) the original time series of returns and then apply the MODWT on the
resampled data and calculate the correlation matrix and its eigenvalues at each scale
of the MODWT. This procedure is done 100,000 times. In Fig. 19.18, we report lmax
as computed on the basis of Eq. 19.37 and the empirical values of lmax obtained from
our simulations. In the latter case, we consider two different values for lmax, which
correspond to the 5 %- and, respectively, 1 %-percentile of the empirical cumulative
distribution of l. The theoretical (RMT) maximum eigenvalue is always much lower
than its empirical counterparts. This is probably due to the small sample size.
In Table 19.2, we report the eigenvalues of the correlation matrix for each
scale of the MODWT. Comparing the eigenvalues lk, k ¼ 1,. . .,12 with ^l max
demonstrates that there is a single factor underlying the evolution of home
19 Analysis of Financial Time Series Using Wavelet Methods 571
Table 19.2 Eigenvalues of the correlation matrix at each scale of the MODWT
w1 w2 w3 w4 w5 v5
l1 0.54 0.29 0.04 0.04 0.01 0.00
l2 0.60 0.30 0.07 0.07 0.02 0.00
l3 0.69 0.33 0.08 0.13 0.08 0.00
l4 0.78 0.46 0.12 0.15 0.13 0.01
l5 0.84 0.48 0.14 0.20 0.16 0.01
l6 0.86 0.61 0.18 0.23 0.18 0.05
l7 0.99 0.70 0.22 0.25 0.34 0.15
l8 1.02 0.77 0.29 0.44 0.50 0.16
l9 1.16 0.92 0.38 0.62 0.66 0.37
l10 1.38 1.06 0.77 0.82 0.69 1.10
l11 1.49 1.37 1.11 1.33 1.38 1.18
l12 1.64a 4.70a 8.60a 7.72a 7.84a 8.98a
lmax 1.30 1.31 1.31 1.32 1.34 1.34
^l max ð95 %Þ 1.43 1.53 1.77 2.16 2.79 3.25
^
l max ð99 %Þ 1.53 1.66 1.96 2.47 3.25 3.89
We report the 12 eigenvalues of the correlation matrix at each scale j as well as the theoretical
(RMT) maximum eigenvalue (lmax in the table) and the empirical 95 %- and 99 %-percentiles
of the empirical (bootstrapped) cumulative distribution of eigenvalues at scale j (^
l max ð95%Þ and
^ max ð99%Þ). a denotes significance at the 95 %-level
l
prices. At each scale, the eigenvalue that is attached to this factor is significant at
the 99 % level. This unique factor can be interpreted as a sort of national-wide
home price index.
19.4 Conclusion
This chapter discusses spectral and wavelet methods. It aims at being an easy-to-
follow introduction and it is structured around conceptual and practical explana-
tions. It also offers many supporting examples and illustrations. In particular, the
last section provides a detailed case study, which analyzes the evolution of home
prices in the USA over the last 20 years using wavelet methodology.
References
Bouchaud, J.-P., & Potters, M. (2004). Theory of financial risk and derivative pricing: From
statistical physics to risk management. Cambridge: Cambridge University Press.
Childers, D. (1978). Modern spectrum analysis. New York: IEEE Press.
Crowley, P. (2007). A guide to wavelets for economists. Journal of Economics Surveys,
21, 207–264.
Curci, G., & Corsi, F. (2012). Discrete sine transform for multi-scale realized volatility measures.
Quantitative Finance, 12, 263–279.
572 P. Masset
Elder, J., & Jin, H. (2007). Long memory in commodity futures volatility: A wavelet perspective.
The Journal of Futures Markets, 27, 411–437.
Fan, J., Fan, Y., & Yazhen, W. (2007). Multiscale jump and volatility analysis for high-frequency
financial data. Journal of American Statistical Association, 102, 618–631.
Fernandez, V., & Lucey, B. (2007). Portfolio management under sudden changes in volatility and
heterogeneous investment horizons. Physica A: Statistical Mechanics and its Applications,
375, 612–624.
Gençay, R., & Fan, Y. (2010). Unit root tests with wavelets. Econometric Theory, 26, 1305–1331.
Gençay, R., & Gradojevic, N. (2011). Errors-in-variables estimation with wavelets. Journal of
Statistical Computation and Simulation, 81, 1545–1564.
Gençay, R., Selçuk, F., & Whitcher, B. (2002). An introduction to wavelets and other filtering
methods in finance and economics. San Diego: Academic Press.
Gençay, R., Selçuk, F., & Whitcher, B. (2003). Systematic risk and timescales. Quantitative
Finance, 3, 108–116.
Gençay, R., Gradojevic, N., Selçuk, F., & Whitcher, B. (2010). Asymmetry of information flow
between volatilities across time scales. Quantitative Finance, 10, 895–915.
Gu, K. (2002). The predictability of house prices. Journal of Real Estate Research, 24, 213–234.
Hafner, C. (2012). Cross-correlating wavelet coefficients with applications to high-frequency
financial time series. Journal of Applied Statistics, 39, 1363–1379.
Hamilton, J. (1994). Time series analysis. Princeton.
Hodrick, R., & Prescott, E. (1997). Postwar U.S. business cycles: An empirical investigation.
Journal of Money, Credit, and Banking, 29, 1–16.
Iacobucci, A. (2003). Spectral analysis for economic time series. Documents de Travail de
l’OFCE.
Kuo, C. (1996). Autocorrelation and seasonality in the real estate market. Journal of Real Estate
Finance and Economics, 12, 139–162.
Kwapien, J., Oswiecimka, P., & Drozdz, S. (2007). Empirics versus RMT in financial cross-
correlations. Acta Physica Polonica B, 38, 4027–4039.
Lynch, P., & Zumbach, G. (2003). Market heterogeneities and the causal structure of volatility.
Quantitative Finance, 3, 320–331.
Manchaldore, J., Palit, I., & Soloviev, O. (2010). Wavelet decomposition for intra-day volume
dynamics. Quantitative Finance, 10, 917–930.
M€uller, U., Dacorogna, M., Davé, R., Pictet, O., Olsen, R., & Ward, J. (1995). Fractals and
intrinsic time – A challenge to econometricians (Working Papers). Olsen and Associates.
Z€urich
Nielsen, M., & Frederiksen, P. (2005). Finite sample comparison of parametric, semiparametric,
and wavelet estimators of fractional integration. Econometric Reviews, 24, 405–443.
Oswiecimka, P., Kwapien, J., & Drozdz, S. (2006). Wavelet versus detrended fluctuation analysis
of multifractal structures. Physical Review E, 74, 016103.
Percival, D., & Walden, A. (2000). Wavelet methods for time series analysis. Cambridge:
Cambridge University Press.
Ramsey, J. B. (1999). The contribution of wavelets to the analysis of economic and financial data.
Philosophical Transactions of the Royal Society of London A, 357, 2593–2606.
Ramsey, J. B. (2002). Wavelets in economics and finance: Past and future. Studies in Nonlinear
Dynamics & Econometrics, 3, 1–29.
Ramsey, J. B., & Zhang, Z. (1997). The analysis of foreign exchange data using waveform
dictionaries. Journal of Empirical Finance, 4, 341–372.
Ramsey, J. B., Zaslavsky, G., & Usikov, D. (1995). An analysis of U. S. stock price behavior using
wavelets. Fractals, 3, 377–389.
Rua, A., & Nunes, L. (2009). International comovement of stock market returns: A wavelet
analysis. Journal of Empirical Finance, 16, 632–639.
Schleicher, C. (2002). An introduction to wavelets for economists (Working Paper). Bank of
Canada.
19 Analysis of Financial Time Series Using Wavelet Methods 573
Sharifi, S., Crane, M., Shamaie, A., & Ruskin, H. (2004). Random matrix theory for portfolio
optimization: A stability approach. Physica A, 335, 629–643.
Struzik, Z. (2001). Wavelet methods in financial time-series processing. Physica A, 296, 307–319.
Subbotin, A. (2008). A multi-horizon scale for volatility. University of Paris-1: CES Working
Paper 2008.20.
Vuorenmaa, T. (2005). A wavelet analysis of scaling laws and long-memory in stock market
volatility, Bank of Finland Research Discussion Paper.
Welch, P. (1967). The use of fast Fourier transforms for the estimation of power spectra: A method
based on time averaging over short modified periodograms. IEEE Transactions on Audio and
Electroacoustics, 15, 70–73.
Composite Goodness-of-Fit Tests for
Left-Truncated Loss Samples 20
Anna Chernobai, Svetlozar T. Rachev, and Frank J. Fabozzi
Contents
20.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 576
20.2 Problem Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 577
20.3 EDF Statistics for Left-Truncated Loss Samples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 580
20.3.1 Supremum Class Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 581
20.3.2 Quadratic Class Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 582
20.3.3 Cramér-von Mises Statistic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 583
20.4 “Upper Tail” Anderson-Darling Statistic . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 584
20.4.1 Supremum Class “Upper Tail” Anderson-Darling Statistic . . . . . . . . . . . . . . . . . . 584
20.4.2 Quadratic Class “Upper Tail” Anderson-Darling Statistic . . . . . . . . . . . . . . . . . . . 585
20.5 Application to Loss Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 588
20.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 590
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 591
Derivation of AD2* Computing Formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 591
Derivation of W2* Computing Formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 592
Derivation of AD2up Computing Formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 594
Derivation of AD2* up Computing Formula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 595
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 596
A. Chernobai (*)
Department of Finance, M.J. Whitman School of Management, Syracuse University, Syracuse,
NY, USA
e-mail: annac@syr.edu
S.T. Rachev
Department of Applied Mathematics and Statistics, College of Business, Stony Brook University,
SUNY, Stony Brook, NY, USA
FinAnalytica, Inc, New York, NY, USA
e-mail: rachev@pstat.ucsb.edu
F.J. Fabozzi
EDHEC Business School, EDHEC Risk Institute, Nice, France
e-mail: fabozzi321@aol.com; frank.fabozzi@edhec.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 575
DOI 10.1007/978-1-4614-7750-1_20,
# Springer Science+Business Media New York 2015
576 A. Chernobai et al.
Abstract
In many financial models, such as those addressing value at risk and ruin
probabilities, the accuracy of the fitted loss distribution in the upper tail of the
loss data is crucial. In such situations, it is important to test the fitted loss
distribution for the goodness of fit in the upper quantiles, while giving lesser
importance to the fit in the low quantiles and the center of the distribution of the
data. Additionally, in many loss models the recorded data are left truncated with
the number of missing data unknown. We address this gap in literature by
proposing appropriate goodness-of-fit tests.
We derive the exact formulae for several goodness-of-fit statistics that
should be applied to loss models with left-truncated data where the fit of
a distribution in the right tail of the distribution is of central importance.
We apply the proposed tests to real financial losses, using a variety of
distributions fitted to operational loss and the natural catastrophe insurance
claims data, which are subject to the recording thresholds of $1 and $25 million,
respectively.
Keywords
Goodness-of-fit tests • Left-truncated data • Minimum recording threshold •
Loss distribution • Heavy-tailed data • Operational risk • Insurance • Ruin
probability • Value at risk • Anderson-Darling statistic • Kolmogorov-Smirnov
statistic
20.1 Introduction
In most loss models, the central attention is devoted to studying the distribu-
tional properties of the loss data. The shape of the dispersion of the data
determines the vital statistics such as the expected loss, variance, and ruin
probability, value at risk, or conditional value at risk where the shape in the
right tail is crucial. Parametric procedures for testing the goodness of fit
(GOF) include the likelihood ratio test and chi-squared test. A standard
semi-parametric procedure to test how well a hypothesized distribution fits
the data involves applying the in-sample GOF tests that provide a comparison
of the fitted distribution to the empirical distribution. These tests, referred to
as empirical distribution function (EDF) tests, include the Kolmogorov-
Smirnov test, Anderson-Darling test, and the Cramér-von Mises tests. Related
works on the discussion of these widely used tests include Anderson and
Darling (1952, 1954), D’Agostino and Stephens (1986), and Shorack and
Wellner (1986).
In many applications, the data set analyzed is incomplete, in the sense that the
observations are present in the loss database only if they exceed a predetermined
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 577
threshold level. This problem is usually absent in risk models involving market
risk and credit risk. However, it is a common problem in operational risk or
insurance claims models. In operational risk, banks’ internal databases are subject
to a minimum recording threshold of roughly $6,000–$10,000, and external
databases usually collect operational losses starting from $1 million, BCBS
(2003). Similarly, in non-life insurance models, the thresholds are set at $5 million,
$25 million, or other levels. Consequently, in the analysis of operational losses,
recorded loss data are left truncated, and, as a result, it is inappropriate to employ
standard GOF tests.
GOF tests for truncated and censored data have been studied by Dufour and
Maag (1978), Gastaldi (1993), and Guilbaud (1998), among others. In this paper,
we derive the exact formulae for several GOF test statistics that should be applied
where there exist incomplete samples with an unknown number of missing data in
low quantiles and propose two new statistics to determine the goodness of fit in the
upper tail that can be used for loss models where the accuracy of the upper tail
estimate is of central concern.
The paper is organized as follows. In Sect. 20.2 we describe the problem of left-
truncated samples and explain the necessary adjustments that are required to the
GOF tests to make them applicable for the truncated samples. In Sect. 20.3 we
review the widely used test statistics for complete samples and derive the exact
formulae for the statistics to be used for left-truncated samples. We propose in
Sect. 20.4 two new EDF statistics to be used for the situations when the fit in the
upper tail is of the central concern. Application of the modified EDF tests to
operational loss data, obtained from Zurich IC2 FIRST Database, and the USA
natural catastrophe insurance claims data, obtained from Insurance Services Office
Inc. Property Claim Services, is presented in Sect. 20.5, with final remarks in
Sect. 20.6. Necessary derivations are provided in the Appendix.
8
>
> 0 x < xð 1Þ
# observations xðjÞ <
j
Fn ðxÞ :¼ ¼ xðjÞ x < xðjþ1Þ , j ¼ 1, 2, . . . , n 1 :
n >
> n
:
1 x xn ,
(20.1)
578 A. Chernobai et al.
Assuming that F is the true distribution, the estimated number of missing points
^ and the number of observed points n are related as1
m
^
m zH
¼ (20.3)
n ð1 zH Þ
resulting in
zH n
n^c ¼ n þn¼ (20.4)
ð1 zH Þ ð1 zH Þ
where zH :¼ F^y ðHÞ is the estimated distribution evaluated at the truncation point.
Then, the estimated empirical distribution function F^nc ðxÞ of complete data
sample is
where the numerator refers to the total number of observations of the complete data
sample, not exceeding in magnitude the jth-order statistic of the incomplete
(observed) data sample such that X(1) X(2) . . . X(n). By Eqs. 20.3 and 20.4,
Eq. 20.5 becomes
m^ þj n zH þ jð1 zH Þ j
F^nc ðxÞ ¼ ¼ ¼ zH þ ð1 zH Þ, j ¼ 1, 2, . . . , n:
n=ð1 zH Þ n n
l m
1
More accurately, m ^ should be estimated as m
^ ¼ n ð1z
zH
H Þ , but this detail can be ignored for all
practical purposes.
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 579
Rearranging terms leads to the fitted distribution function of the observed sample
of the following form:
8
< F^y ðxÞ F^y ðH Þ
F^y ðxÞ ¼ xH
^ (20.6)
: 1 F y ðH Þ
0 x < H,
so that Fy(X) U[Fy(H), 1] and Fy ðXÞ U ½0; 1 under the null hypothesis that the
fitted distribution function is true. Therefore, the estimated empirical distribution
function of the observed part of the data, using Eq. 20.1, is represented by
8
> ^ y ðH Þ
F x < xð1 Þ
>
<
^ y ðH Þ þ F
F n ðx Þ 1 F ^ y ðH Þ ¼ j
1F ^ y ðH Þ xðjÞ x < xðjþ1Þ , j ¼ 1,2, . .. ,n 1 :
^ y ðH Þ þ F
>
>
:n
1 x xn ,
(20.7)
Figure 20.1 gives a visual illustration of the idea we just described. With
these modifications, the in-sample GOF tests can be applied to the left-truncated
samples.
In this paper we consider tests of a composite hypothesis that the
empirical distribution function of an observed incomplete left-truncated loss data
sample belongs to a family of hypothesized distributions (with parameters not
specified), i.e.,
H0 : Fn ðxÞ 2 Fy ðxÞ vs: = Fy ðxÞ:
H A : Fn ð x Þ 2 (20.8)
Under the null Eq. 20.8, Fy ðXÞ
U ½0; 1, and the null is rejected if the p-value is
lower than the level a, such as a from 5 % to 10 %. Letting D be the
observed value of a GOF statistic (such as Kolmogorov-Smirnov or Anderson-
Darling) and d the critical value for a given level a, the p-value is computed as
p-value ¼ P(D d). Since the distribution of the statistic is not parameter-free, one
way to compute the p-values and the critical values is by means of Monte Carlo
simulation, for each hypothesized fitted distribution (Ross 2001). Under the
procedure, the observed value D is computed. Then, for a given level a, the
following algorithm is applied:
1. Generate a large number of samples I (such as I ¼ 1,000) from the fitted
truncated distribution of size n (the number of observed data points), such that
these random variates are above or equal to H.
2. Fit a truncated distribution and estimate conditional parameters y^ for each
sample i ¼ 1, 2, . . .I.
3. Estimate the GOF statistic value Di for each sample i ¼ 1, 2, . . .I.
4. Calculate p-value as the proportion of times the sample statistic values exceed
the observed value d of the original sample.
5. Reject H0 if the p-value is smaller than a.
580 A. Chernobai et al.
Fθ(H)
EDF missing
EDF observed
Fitted CDF missing
H Fitted CDF observed
0
missing observed x
1
fitted CDF = (Fθ(x)−Fθ(H)) / (1−Fθ(H))
F(x)
EDF
EDF observed
Fitted CDF observed
H
x
observed
A p-value of, for example, 0.3, would mean that in 30 % of samples of the same
size simulated from the same distribution with the same parameter estimation
procedure applied, the test statistic value was higher than the one observed in the
original sample.
The EDF statistics are based on the vertical differences between the empirical
and fitted (truncated) distribution function. They are divided into two classes:
(1) the supremum class (such as Kolmogorov-Smirnov and Kuiper statistics) and
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 581
(2) the quadratic class (such as Anderson-Darling and Cramér-von Mises statistics).
In this section, we derive the exact computing formulae for a number of EDF
test statistics, modified so that they can be applied to left-truncated loss samples.
For the left-truncated samples, Fy(X) denotes the null distribution function for
left-truncated sample values. The variable Fy(X) is distributed uniformly over the
[0, 1] interval. The variable Fy(X) is distributed uniformly over the [Fy(H), 1]
interval. We reserve some other notations: zH :¼ F^y ðH Þ that was defined earlier and
zj :¼ F^y xðjÞ for truncated samples. In this section we discuss the Kolmogorov-
Smirnov, Kuiper, Anderson-Darling, and the Cramér-von Mises statistics and use
an asterisk (*) to denote their left-truncated sample analog.
and becomes
The quadratic statistics for complete data samples are grouped under the Cramér-
von Mises family as
ð
1
2
Q¼n Fn ðxÞ F^y ðxÞ c F^y ðxÞ dF^y ðxÞ, (20.14)
1
in which the weight function c F^y ðxÞ is assigned to give a certain weight
to different observations, depending on the purpose. For left-truncated
samples, we denote the Cramér-von Mises family as Q* and F^y ðxÞ is replaced by
F^y ðxÞ:
ð
1
2
Q ¼ n Fn ðxÞ F^y ðxÞ c F^y ðxÞ dF^y ðxÞ:
(20.15)
H
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 583
Depending on the form of the weighting function, the sample observations are
1
given a different weight. c F^y ðxÞ ¼ F^y ðxÞ 1 F^y ðxÞ yields the quadratic
Anderson-Darling statistic and c F^y ðxÞ ¼ 1 yields the Cramér-von Mises statistic.
Derivation of the computing formulae makes use of Eq. 20.1 and involves the Proba-
bility Integral Transformation (PIT) technique. For left-truncated samples, this leads to
ð1 ð1
2 n z zH 2 z zH
Q ¼ n ðFn ðz Þ z Þ cðz Þdz ¼ Fn ð z Þ c dz,
1 zH 1 zH 1 zH
0 zH
(20.16)
where Z ¼ F^y ðXÞ
¼ Zz ^
1zH with Fy ðX Þ U ½0; 1 under the null, and so Z ¼ F y ðX Þ
H
with Fy(X) U[Fy(H), 1] under the null. Fn(Z ) ¼ Fn(X) ¼ Fy(Fn(X)) is the
empirical distribution function of Z*. zH ¼ F^y ðH Þ ¼ Fy F^y ðH Þ .
The first version of ADup belongs to the supremum class EDF statistics. For
complete data samples, each observation of the KS statistic is assigned a weight
1
of c F^y ðxÞ ¼ 1 F^y ðxÞ . Under this specification, the observations in the
upper tail are assigned a higher weight than those in the lower tail. Let {X( j)}1 jn
be the vector of the order statistics, such that X(1) X(2) . . . X(n). Then we
define the ADup statistic for complete data samples as
pffiffiffi Fn ðxÞ F^y ðxÞ
ADup
¼ n sup : (20.19)
x 1 F^y ðxÞ
Denoting zj :¼ F^y xðjÞ , the computing formula is derived from
( )
pffiffiffi j
zj
ADþ
up ¼ n sup n
,
j 1 zj
( )
pffiffiffi zj j n 1
AD ¼ n sup ,
up
j 1 zj
and becomes
n o
ADup ¼ max ADþ
up , AD
up : (20.20)
For left-truncated samples, the counterpart of the ADup statistic can be similarly
computed using
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 585
( ) ( )
pffiffiffi Fn xðjÞ F^y xðjÞ pffiffiffi zH þ njð1 zH Þ zj
ADþ ¼ n sup ¼ n sup ,
1 F^y xðjÞ 1 zj
up
j j
( ) ( )
pffiffiffi F^y xðjÞ Fn xðjÞ pffiffiffi zj zH j n 1ð1 zH Þ
AD ¼ n sup ¼ n sup ,
1 F^ xðjÞ 1 zj
up
j y j
and becomes
n o
ADup ¼ max ADþ
up , ADup : (20.21)
For complete data samples Fy(X) U[0, 1] under the null hypothesis. If we denote
zj :¼ F^y xðjÞ , straightforward calculations lead to the following computing formula:
X
n 1X n
1
AD2up ¼ 2 log 1 zj þ ð 1 þ 2ð n j Þ Þ :
j¼1
n j¼1 1 zj
2
It has been shown that the weighting function cðtÞ ¼ ð1 tÞb possesses nice asymptotic
properties only for b¼[0,2) (Deheuvels and Martynov 2003). For b¼2, which is the case consid-
ered in this chapter, the asymptotic distribution of the test statistic has infinite mean. This is indeed
a concern for very large samples (i.e., the asymptotic case n!1). Yet, because firms’ operational
loss data samples are typically relatively small, the asymptotic distribution of the test statistic
should not generate large concerns. Nevertheless, the results of the proposed quadratic class upper-
tail Anderson-Darling test should be treated with caution and with consideration of the properties
described above.
586 A. Chernobai et al.
8 8
Computing formula: 9 8 99
< >
> < jz >
= >
< z j1 > =>
=
pffiffiffi j j
AD ¼ n max sup qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
n
, sup q ffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
n
>
: j >
> : zj 1 zj > ; j > : zj 1 zj > ;;
ADup pffiffiffi Fn ðxÞ F^y ðxÞ
ADup ¼ n sup
x 1 F^y ðxÞ
Computing formula:
( )
pffiffiffi n zj
j
zj j1
ADup ¼ n max sup , sup n
j 1 zj j 1 zj
AD2 ð
1
Fn ðxÞ F^ y ðxÞ
2
AD2 ¼ n d F^y ðxÞ
F^y ðxÞ 1 F^ y ðxÞ
1
Computing formula:
Xn Xn
1 1
AD2 ¼ n þ ð1 2jÞlog zj ð1 þ 2ðn jÞÞ log 1 zj
j¼1 n j¼1 n
W2 ð
1
2
W2 ¼ n Fn ðxÞ F^ y ðxÞ dF^ y ðxÞ
1
Computing formula:
n 1 Xn Xn
W2 ¼ þ ð1 2jÞzj þ z2
j¼1 j
3 n j¼1
AD2up ð
1 2
Fn ðxÞ F^ y ðxÞ
AD2up ¼ n 2 dF^ y ðxÞ
1
1 F^ y ðxÞ
Computing formula:
1 Xn 1 Xn
AD2up ¼ ð1 þ 2ðn jÞÞ þ2 log 1 zj
n j¼1 1 zj j¼1
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 587
Table 20.2 Description of EDF statistics for left-truncated (threshold ¼ H) data samples
ðxÞFy ðHÞ
H0 : Fn(x) ∈ F*y(x) vs. HA : Fn(x) 2 = F*y(x), Fy ðxÞ :¼ Fy1F y ðH Þ
^ ^
Notations: zj :¼ F y xðjÞ , zH ¼ F y ðH Þ, j ¼ 1, 2, . . . , n
Statistic Description and computing formula
pffiffiffi
KS* KS ¼ n supFn ðxÞ F^ y ðxÞ
x
Computing formula: (
pffiffiffi n o
n j
KS ¼ max sup zH þ ð1 zH Þ zj ,
1 zH j n
)
j1
sup zj zH þ ð1 zH Þ
j n
n o n o
V* pffiffiffi
V ¼ n sup Fn ðxÞ F^ y ðxÞ þ sup F^ y ðxÞ Fn ðxÞ
x x
Computing formula:
pffiffiffi n o !
n j j1
V ¼ sup zH þ ð1 zH Þ zj þ sup zj zH þ ð1 zH Þ
1 zH j n j n
AD*
pffiffiffi Fn ðxÞ F^ y ðxÞ
AD ¼ n suprffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
ffi
x
^
^
F y ð xÞ 1 F y ð xÞ
8 8
Computing formula: 9 8 99
>
< > < z þ j ð1 z Þ z > = <z z j1 ð1 z Þ>
> =>
=
p ffiffi
ffi
AD ¼ n max sup qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
H H j j H H
n
ffi , sup qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
n
ffi
>
: j >: zj zH 1 zj > ; j > : zj zH 1 zj > ;>
;
AD*up pffiffiffi F ðxÞ F^ ðxÞ
n
ADup ¼ n sup y
x 1F ^ ðxÞ
y
Computing formula:
( ( ))
pffiffiffi zH þ nj ð1 zH Þ zj zj zH j1
n ð1 zH Þ
ADup ¼ nmax sup , sup
j 1 zj j 1 zj
2
AD2*
ð Fn ðxÞ F^ ðxÞ
1
y
AD2 ¼ n
dF^ y ðxÞ
H
F^y ðxÞ 1 F^y ðxÞ
Computing formula: Xn
AD2 ¼ n þ 2nlogð1 zH Þ 1n ð1 þ 2ðn jÞÞlog 1 zj þ
Xn
j¼1
1 ð1 2jÞlog z z
n j¼1 j H
W2* ð
1
2
W 2 ¼ n Fn ðxÞ F^ y ðxÞ dF^ y ðxÞ
H
Computing formula:
n n zH 1 X n
1 X
n 2
W 2 ¼ þ þ ð1 2jÞzj þ zj zH
3 1 zH nð1 zH Þ j¼1 2
ð1 zH Þ j¼1
(continued)
588 A. Chernobai et al.
up ¼ 2n logð1 zH Þ þ 2
AD2 log 1 zj þ ð1 þ 2ðn jÞÞ 1z
1
j¼1
n j¼1 j
AD2*
up statistic for left-truncated samples of the following form (the derivation is
given in Appendix):
X
n 1 zH X
n
1
up ¼ 2n logð1 zH Þ þ 2
AD2 log 1 zj þ ð1 þ 2ðn jÞÞ :
j¼1
n j¼1 1 zj
Tables 20.1 and 20.2 summarize the EDF statistics and their computing formulae
for complete and left-truncated samples.
In this section we apply the GOF testing procedure to (1) operational loss data,
extracted from an external database, and (2) catastrophe insurance claims data. The
operational loss data set was obtained from Zurich IC Squared (IC2) FIRST Database
of Zurich IC Squared (IC2), an independent consulting subsidiary of Zurich Financial
Services Group. The external database is comprised of operational loss events
throughout the world. The original loss data cover losses in the period 1950–2002.
A few recorded data points were below $1 million in nominal value, so we excluded
them from the analysis, to make it more consistent with the conventional threshold for
external databases of $1 million. Furthermore, we excluded the observations before
1980 because of relatively few data points available (which is most likely due to poor
data recording practices). The final data set for the analysis covered losses for the time
period between 1980 and 2002. It consists of five types of losses: “relationship” (such
as events related to legal issues, negligence, and sales-related fraud), “human” (such
as events related to employee errors, physical injury, and internal fraud), “processes”
(such as events related to business errors, supervision, security, and transactions),
“technology” (such as events related to technology and computer failure and tele-
communications), and “external” (such as events related to natural and man-made
disasters and external fraud). The loss amounts have been adjusted for inflation using
the Consumer Price Index from the U.S. Department of Labor. The numbers of data
points of each type are n ¼ 849, 813, 325, 67, and 233, respectively.
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 589
The insurance claims data set covers claims resulting from natural catastrophe
events occurred in the United States over the time period from 1990 to 1996. It was
obtained from Insurance Services Office Inc. Property Claim Services (PCS). The
data set includes 222 losses. The observations are greater than $25 million in
nominal value.
Left-truncated distributions of four types were fitted to each of the data set:
exponential, lognormal, Weibull, and Pareto (GPD). Table 20.3 presents the observed
statistic values and the p-values for the six data sets (five operational losses and
insurance claims), obtained with the testing procedure described in Sect. 20.2.
The results reported in Table 20.3 suggest that fitting heavier-tailed distribu-
tions, such as Pareto and Weibull, results in lower values of the GOF statistics,
which leads to acceptance of the null for practically all loss types and all
criteria of the goodness of fit, viewed from the high p-values. Since the analysis
of operational losses deals with estimating the operational value at risk
(VaR), it is reasonable to determine the ultimate best fit on the basis of the ADup
and AD2up statistics, introduced in this paper. As can be seen from Table 20.3, these
proposed measures suggest a much better fit of the heavier-tailed distributions.
Moreover, for the Pareto distribution, while the statistics that focus on the center of
the data (Kolmogorov-Smirnov, Kuiper, Cramér-von Mises) do not show a good
fit, the ADup and AD2up statistics indicate that the fit in the upper tail is very good.
It should be noted that the Pareto and Weibull distributions very often suggest
a superior fit in the upper tail to the lognormal distribution. Yet it is the lognormal
distribution that was suggested in 2001 by the Basel Committee (BCBS (2001)).
20.6 Conclusions
In this paper we present a technique for modifying the existing goodness-of-fit test
statistics so that they can be applied to loss models in which the available data set is
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 591
incomplete and is truncated from below. Such left truncation is often present in loss
data when the data are being recorded starting from a fixed amount and the data
below are not recorded at all. Exact computing formulae for the Kolmogorov-
Smirnov, Kuiper, Anderson-Darling, and Cramér-von Mises for the left-truncated
samples are presented.
In risk management, it is often vital to have a good fit of a hypothesized
distribution in the upper tail of the loss data. It is important in loss models that
deal with value at risk, conditional value at risk, and ruin probability. We suggest
using two other versions of the Anderson-Darling statistic (which we refer to as
the upper tail Anderson-Darling statistic) in which the weighting function is
proportional to the weight of only the upper tail of the distribution. Supremum
and quadratic versions of the statistic are proposed. Such statistic is convenient to
use when it is necessary to examine the goodness of fit of a distribution in the right
tail of the data, while the fit in the left tail is unimportant.
The technique is applied to check the goodness of fit of a number of
distributions using operational loss data and catastrophe insurance claims data
sets. From the empirical analysis we conclude that heavier-tailed distributions better
fit the data than Lognormal or thinner-tailed distributions in many instances. In
particular, the conclusion is strongly supported by the upper tail Anderson-Darling
tests.
Appendix
2
ð
þ1
Fn ðxÞ F^y ðxÞ
AD2 ¼ n
d F^y ðxÞ
^ ^
F y ðxÞ 1 F y ðxÞ
H
ð1
ðFn ðz Þð1 zH Þ þ zH zÞ2
PIT n^c dz,
ðz zH Þð1 zÞ
zH
^
Fy F y xðjÞ , j ¼ 1, 2, . . ., n, as
592 A. Chernobai et al.
ð z1 n1 ð zjþ1 j
2 ð1
1 ðz zH Þ2 X nð1 zH Þ þ zH z ð1 zÞ2
AD2 ¼ dz þ dz þ dz :
n^c zH ðz zH Þð1 zÞ j¼1 zj
ðz zH Þð1 zÞ zn ðz zH Þð1 zÞ
|fflfflfflfflfflfflfflfflfflfflfflfflffl
ffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflffl
ffl {zfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl
ffl}
A B C
1 zH X
n1
2 j2 log zj zH log zjþ1 zH
n2 j¼1
1 zH Xn
¼ z1 zn þ ð1 zH Þlogð1 z1 Þ 2
ð1 þ 2ðn jÞÞ log 1 zj þ . . .
n j¼1
1 zH Xn
þ ð1 zH Þlogðzn zH Þ þ 2
ð1 2jÞlog zj zH ;
n j¼1
Summing the terms A, B, and C, multiplying by n^c , and simplifying yields the
final computing formula:
1X n
AD2 ¼ n þ 2n logð1 zH Þ ð1 þ 2ðn jÞÞ log 1 zj þ . . .
n j¼1
1X n
þ ð1 2jÞlog zj zH :
n j¼1
ð
þ1
2
W 2
¼n Fn ðxÞ F^y ðxÞ d F^y ðxÞ
H
ð1
ð Fn ð z Þ ð 1 z H Þ þ z H z Þ 2
PIT n^c dz,
ð1 zH Þ2
zH
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 593
^
Fy F y xðjÞ , j ¼ 1, 2, . . ., n, as
ð z1 n1 ð zjþ1
X 2 ð1
ð1 zH Þ2 2 j
W ¼ ðzH zÞ2 dz þ ð1 zH Þ þ zH z dz þ ð1 zÞ2 dz :
n^c zH j¼1 zj n zn
|fflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflffl}
A B C
ð1 zH Þ2 Xn
¼ ð1 zH Þzn þ zH ð1 zH Þzn ð1 zH Þzn 2 þ ð1 2jÞzj þ .. .
n2 j¼1
1 zH X
n
1 zH X
n
þ zj 2 2zH zj ;
n j¼1 n j¼1
1 zn 3
C ¼ þ zn 2 zn :
3 3
n^c
Summing the terms A, B, and C, multiplying by ð1zH Þ2
, and simplifying yields
the final computing formula:
n nzH 1 X n
1 X
n 2
W 2 ¼ þ þ ð1 2jÞzj þ zj zH :
3 1 zH nð1 zH Þ j¼1 2
ð1 zH Þ j¼1
594 A. Chernobai et al.
ð
þ1 2 ð1
Fn ðxÞ F^y ðxÞ ðFn ðzÞ zÞ2
AD2up ¼n 2 d ^
F y ð xÞ PIT n dz,
1
1 F^y ðxÞ 0
ð1 zÞ2
where Fn(Z) ¼ Fy(Fn(X)) ¼ Fn(X) is the empirical distribution function of Z ¼ F^y ðXÞ
¼ F F^y ðXÞ so that Fy(·) U[0,1].
Using Eq. 20.1, the computing formula is expressed in terms of zj :¼ F^y xðjÞ ¼
F F^y xðjÞ , j ¼ 1, 2, . . ., n, as
ð z1 n1 ð zjþ1
j 2 ð1
1 z2 X z ð1 zÞ2
AD2up ¼ 2
dz þ n
dz þ dz :
n 0 ð1 zÞ j¼1 zj ð1 zÞ2 zn ð1 zÞ2
|fflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflffl}
A B C
1Xn1
2 ðn jÞ log 1 zj log 1 zjþ1
n j¼1
1 1Xn
1 1X n
¼ zn z1 þ 2 ð1 þ 2ðn jÞÞ 2logð1 z1 Þ þ 2 log 1 zj ;
1 z1 n j¼1 1 zj n j¼1
C ¼ 1 zn :
X
n 1X n
1
AD2up ¼ 2 log 1 zj þ ð 1 þ 2ð n j Þ Þ :
j¼1
n j¼1 1 zj
20 Composite Goodness-of-Fit Tests for Left-Truncated Loss Samples 595
Derivation of AD2*
up Computing Formula
F F^y xðjÞ , j ¼ 1, 2, . . ., n as
ð z1 n1 ð zjþ1
j 2 ð1
1 ðz zH Þ2 X ð1 zH Þ þ zH z ð1 zÞ2
AD2 ¼ dz þ n
dz þ dz :
n^c up zH ð1 zÞ 2
j¼1 zj ð1 zÞ 2
zn ð1 zÞ2
|fflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflfflffl} |fflfflfflfflfflfflfflfflffl{zfflfflfflfflfflfflfflfflffl}
A B C
1
A ¼ z1 zH ð1 zH Þ þ ð1 zH Þ2 2ð1 zH Þlogð1 zH Þ
1 z1
þ 2ð1 zH Þlogð1 z1 Þ;
ð1 z H Þ2 X
n1
2 1 1
B ¼ zn z1 ð n j Þ ...
n2 j¼1
1 zj 1 zjþ1
1 zH X
n1
2 ðn jÞ log 1 zj log 1 zjþ1
n j¼1
1 ð1 zH Þ2 Xn
1
¼ zn z1 ð1 zH Þ2 þ ð1 þ 2ðn jÞÞ ...
1 z1 n 2
j¼1
1 zj
1 zH X
n
2ð1 zH Þlogð1 z1 Þ þ 2 log 1 zj ;
n j¼1
C ¼ 1 zn :
596 A. Chernobai et al.
Summing the terms A, B, and C, multiplying by n^c , and simplifying yields the
final computing formula:
Xn 1 zH Xn
1
AD2 ¼ 2nlogð 1 z H Þ þ 2 log 1 z j þ ð1 þ 2ðn jÞÞ :
up
j¼1
n j¼1
1 zj
References
Anderson, T. W., & Darling, D. A. (1952). Asymptotic theory of certain ‘Goodness of Fit’ criteria
based on stochastic processes. The Annals of Mathematical Statistics, 23, 193–212.
Anderson, T. W., & Darling, D. A. (1954). A test of goodness of fit. Journal of the American
Statistical Association, 49, 765–769.
BCBS. (2001). Working paper on the regulatory treatment of operational risk. Basel, Switzerland:
Basel Committee on Banking Supervision. http://www.bis.org
BCBS. (2003). The 2002 loss data collection exercise for operational risk: summary of the data
collected. Basel, Switzerland: Basel Committee on Banking Supervision. http://www.bis.org
D’Agostino, R., & Stephens, M. (1986). Goodness-of-fit techniques. New York/Basel: Marcel
Dekker.
Deheuvels, P., & Martynov, G. (2003). Karhunen-Loève expansions for weighted Wiener pro-
cesses and Brownian bridges via Bessel functions. Progress in Probability, 55, 57–93.
Dufour, R., & Maag, U. R. (1978). Distribution results for modified kolmogorov-smirnov statistics
for truncated or censored samples. Technometrics, 20, 29–32.
Gastaldi, T. (1993). A kolmogorov-smirnov test procedure involving a possibly censored or
truncated sample. Communications in Statistics: Theory and Method, 22, 31–39.
Guilbaud, O. (1998). Exact kolmogorov-type test for left-truncated and/or right-censored data.
Journal of American Statistical Association, 83, 213–221.
Ross, S. (2001). Simulation (3rd ed.). Boston: Academic.
Shorack, G., & Wellner, J. (1986). Empirical processes with applications to statistics.
New York: Wiley.
Effect of Merger on the Credit Rating and
Performance of Taiwan Security Firms 21
Suresh Srivastava and Ken Hung
Contents
21.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 598
21.2 Merger Literature . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 599
21.3 Empirical Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 601
21.3.1 Comprehensive Performance Score . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 601
21.3.2 Test of Merger Synergy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 603
21.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 606
Appendix 1: Variables for Merger Synergy Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 607
Merger Synergy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 608
Appendix 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 611
Principal Component Factor Analysis of Merger Synergies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 611
Variability Percentage Adjustment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 612
Factor Loading Adjustment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 612
Performance Scores . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 612
Appendix 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 613
Wilcoxon Sign Rank Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 613
Test Hypotheses . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 614
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 614
S. Srivastava (*)
University of Alaska Anchorage, Anchorage, AK, USA
e-mail: afscs@uaa.alaska.edu
K. Hung
Division of International Banking & Finance Studies, Texas A&M International University,
Laredo, TX, USA
e-mail: ken.hung@tamiu.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 597
DOI 10.1007/978-1-4614-7750-1_21,
# Springer Science+Business Media New York 2015
598 S. Srivastava and K. Hung
Abstract
The effect of a merger on credit rating is investigated by testing the significance
of change in a firm’s rank based on comprehensive performance score and
synergistic gains. We extract principle component factors from a set of financial
ratios. Percentage of variability explained and factor loadings are adjusted to get
a modified average weight for each financial ratio. This weight is multiplied by
the standardized Z value of the variable, and summed a set of variables to get
a firm’s performance score. Performance scores are used to rank the firm.
Statistical significance of difference in pre- and post-merger rank is tested
using the Wilcoxon sign rank (double end).
We studied the merger of financial firms after the enactment of Taiwan’s
Merger Law for Financial Institution in November 2000 to examine synergies
produced by merger. Synergistic gains affect corporate credit ratings. After
taking into account the large Taiwan market decline from 1999 to 2000, test
results show there is no significant operating, market, and financial synergy
produced by the merger firms. Most likely explanations for the insignificant
rank changes are short observation period and the lack of an adequate sample in
this investigation.
We identify and define variables for merger synergy analysis followed
by principal component factor analysis, variability percentage adjustment,
and performance score calculation. Finally, Wilcoxon sign rank test is used
for hypothesis testing. Reader is well referred to the appendix for details.
Keywords
Corporate merger • Financial ratios • Synergy • Economies of scale • Credit
rating • Variability percentage adjustment • Principle component factors • Firm’s
performance score • Standardized Z • Wilcoxon rank test
21.1 Introduction
1
World leading credit rating agencies are Moody’s Investors Service, Standard & Poor’s Corp,
Fitch Investors Service, Duff & Phelps, Japan Bond Research Institution, Nippon Investors
Service, Japan Credit Rating Agency, China Credit Rating Agency, and Taiwan New Economy
Newspaper (Weston and Mansinghka 1971; Williamson 1981).
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 599
Firms merge with the stated intent of shareholders’ wealth maximization. Value
maximization is achieved by increased profit, reduced risk, or both. Theories
that posit wealth maximization are efficiency theory (synergy), information and
signaling theory, and tax advantage.2 Non-value maximization concerns with
agency theory. In this research we focus primarily on efficiency theory, also called
synergy theory. Synergy theory implies three components: operation synergy,
market synergy, and financial synergy. Operation synergy is produced due to
economies of scale, operating reduction, and sharing of management expertise.
Under market synergy, a merger reduces market competitors and increases market
concentration leading to a monopolization or increased market share. The increased
market share produces superior profits via pricing strategy or corporate collusion.
The merging corporation increases its market share so as to influence product price
2
Theory of merger is discussed in Brigham and Daves (2007), Copeland and Weston (1979),
Gardner (1996), and Rhodes (1983). Tax incentive is discussed in Weston and Chung (1983).
600 S. Srivastava and K. Hung
The effect of a merger on credit rating is investigated by (1) testing the significance
of change in a firm’s rank based on a comprehensive performance score and
(2) examining post-merger synergy. Synergistic gains from a merger could be in
the form of operating (management) synergy, market synergy, and/or financial
synergy. Data for this research cover January 1999 to December 2000. Premerger
analysis covers year 1999, and post-merger analysis covers year 2000.3 Definitions
of financial ratios are presented in Appendix 1. The methodology of factor extrac-
tion and weight assignment is discussed in Appendix 2. Since the analyses are based
on ranks, the usual “t-test” method is unsuitable to examine the significance of
change in operating, financial, or market performance pre- and post-merger. Hence,
we use the Wilcoxon Sign Rank Test that checks whether two sets of ranks, pre- and
post-merger, come from the same sample or two samples (Appendix 3).4
In Table 21.1, we list five acquiring financial institutions and the acquired institu-
tions. The footnote of Table 21.1 lists 13 non-merger firms. These 13 firms are
included in the analysis for comparison.
Principle component factor analysis was used to extract common factors with
factor loading greater than 1. Using 12-month financial ratio covering 1999,
we produced five factors (Table 21.2, Panel A). The variability explained by five
extracted factors is 24.458 %, 23.764 %, 13.268 %, 12 %, and 11.031 %.
3
3-month and 6-month analyses were performed but are omitted as these did not add additional
foresight.
4
References for Statistical and Econometric issues are Johnson and Wichern (2007) and
Woolridge (2009).
602 S. Srivastava and K. Hung
The percentage of variability explained was adjusted to make the total variability
explained equal to 100. The adjusted percentage of variability explained by five
extracted factors is 28.937 %, 28.116 %, 15.698 %, 14.198 %, and 13.051 %. Using
12-month financial ratio covering 2000, six extracted common factors with factor
loading greater than 1 are presented in Table 21.2 (Panel B). The adjusted percent-
age of variability explained by six extracted factors is 24.746 %, 22.407 %,
17.435 %, 13.857 %, 11.724 %, and 9.832 %. The adjusted variability percentage
is multiplied by factor loading and summed over all variables to get the total
loading for a variable. Then factor loadings are adjusted such that total factor
loadings add to 100 (Eq. 21.2). Initial and adjusted weights for each variable are
presented in Table 21.3. Finally, the values of adjusted weights for 1999 and 2000
are averaged and sign modified to reflect a positive or negative variable. It is
presented in the last column of Table 21.3 and used to calculate comprehensive
performance score. The modified average weight for each variable is multiplied by
the standardized Z value of the variable and summed over all the variables to get
a firm’s performance score and comprehensive rank (Table 21.4). The premerger
comprehensive ranks of the five merger firms and 13 non-merger firms are listed in
Table 21.4, Panel A. Panel B of Table 21.4 presents comprehensive post-merger
comprehensive ranks. Rank changes in Table 21.4 can be summarized as JianHong
dropped from rank 1 to 6, YuanDaJinHua rose from rank 3 to 2, YuanFu
rose from rank 7 to 1, FuBang dropped from rank 2 to 4, and BaoLai rose from
rank 16 to 8. The test of significance of the comprehensive rank differences, for
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 603
merger and non-merger firms, was conducted using the Wilcoxon sign rank (double
end) test discussed in Appendix 3. Panel A of Table 21.5 indicates that the merger
firm’s comprehensive rank difference is statistically insignificant. Panel B of
Table 21.5 indicates that the non-merger firm’s comprehensive rank difference is
also statistically insignificant. Plausible explanations for the insignificant rank
changes are a short observation period, the lack of adequate samples, and the
sharp decline in JianHong’s performance score. Another plausible reason may be
that some security firms (such as YuanDaJinHua and YuanFu) have already
achieved high premerger comprehensive performance scores, and any improvement
in post-merger performance did not make a significant difference.
We use 12-month data to examine operating synergy, market synergy, and financial
synergy. If operating synergy exists, then operating cost ratio will be reduced
604 S. Srivastava and K. Hung
significantly; if market synergy exists, then two other ratios (ratio of operating
income to total assets and market share change rate) will increase significantly; and
existence of financial synergy is indicated by the decline in variability of operating
(business) risk.
Some financial variables of security firms are greatly influenced by market
conditions. The operating income, in particular, is greatly influenced by a bull or
bear market. In years 1999 and 2000, the Taiwan stock market fell from over 10,000
points to over 5,000 points. In order to reduce the market impact on variables,
we adjust the operating cost to operating income ratio and the operating income to
total asset ratio. The operating cost ratio and the operating return on assets are
normalized by the industry average of the ratio.
Table 21.6 presents a test of significance of operating synergy for merger and
non-merger firms. The test result shows there are no significant changes of operat-
ing costs ratio for merger firms. But for non-merger firms, there are significant
increases of operating cost ratio. Then operating income is adjusted for market drop
in 2000, and the significance of operating synergy is tested again. The test result is
presented in Table 21.7.
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 605
Table 21.5 Test of significance of merger and non-merger firms’ comprehensive rating
December December
Firm 1999 ranks 2000 ranks Difference Di Class W
Panel A: merger firms
JianHong 1 6 5 5 3 W (+) ¼ 33
YuanDaJinHua 3 2 1 1 1 W () ¼ 45
YuanFu 7 1 6 6 4 W ¼ 33
FuBang 2 4 2 2 2
BaoLai 16 8 8 8 5
Panel B: non-merger firms
TaiYu 12 15 3 3 6.5 W (+) ¼ 33
DaHua 13 11 2 2 3.5 W () ¼ 45
QunYi 5 7 2 2 3.5 W ¼ 33
ZhongXin 11 5 6 6 11
YongChang 10 14 4 4 8.5
TaiZhen 8 10 2 2 3.5
JinDing 15 16 1 1 1
RiShen 6 3 3 3 6.5
DaXin 18 18 0
KangHe 14 12 2 2 3.5
YaZhou 17 13 4 4 8.5
XinBao 9 17 8 8 12
TongYi 4 9 5 5 10
Wilcoxonsign rank (double end) test indicates that rank difference is statistically insignificant,
Eq. 21.6
Table 21.6 Test of significance of merger and non-merger firms’ operating cost ratioa
December December
Firm 1999 score 2000 score Difference Di Class W
Panel A: merged firms
JianHong 71.634 67.332 4.302 4.302 2 W (+) ¼ 3
YuanDaJinHua 58.689 78.167 –19.478 19.478 5 W () ¼ 12
YuanFu 71.857 79.305 –7.448 7.448 4 W¼3
FuBang 70.040 75.378 5.338 5.338 3
BaoLai 92.007 91.503 0.503 0.503 1
Panel B: non-merger firms
TaiYu 76.548 122.296 45.748 45.748 12 W (+) ¼ 3
DaHua 84.481 83.181 1.299 1.299 2 W () ¼ 88
QunYi 71.105 73.431 2.325 2.325 4 W¼3
ZhongXin 79.395 84.665 5.269 5.269 5
YongChang 79.471 111.708 32.238 32.238 11
TaiZhen 81.915 95.024 13.109 13.109 8
JinDing 98.436 107.020 8.584 8.584 6
RiShen 66.200 66.167 0.034 0.034 1
DaXin 121.563 143.634 22.071 22.071 10
KangHe 81.531 93.207 11.676 11.676 7
YaZhou 104.699 106.014 1.316 1.316 3
XinBao 74.861 136.984 62.123 62.123 13
TongYi 63.744 85.364 21.621 21.621 9
a
Operating cost/operating income
Wilcoxon sign rank (double end) test indicates that rank difference is statistically
insignificant, Eq. 21.6
21.4 Conclusion
The effect of a merger on credit rating was examined by testing the significance
of change in firm’s rank based on a comprehensive performance score
and examining post-merger synergy. Synergistic gains from a merger could be
in the form of operating synergy, market synergy, and/or financial synergy.
Our test showed that merger and non-merger firms’ comprehensive performance
rank difference was statistically insignificant. Plausible explanations for the
insignificant rank changes are a short observation period and the lack of adequate
samples. Other plausible explanation may be that some security firms have already
achieved high premerger comprehensive performance scores, and any improvement
in post-merger performance did not make a significant difference.
We used a standardized score of operating cost ratio to rank firms for their
operating synergy. Test results show there is no significant operating synergy for
merger and non-merger security firms. We used standardized scores of ratio of
operating income to total assets and market share change to rank firms for their
market synergy. Test result shows there are significant increases in the ratio of
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 607
Table 21.7 Test of significance of merger and non-merger firms’ adjusted operating cost ratioa
December December
Firm 1999 score 2000 score Difference Di Class W
Panel A: merger firms
JianHong 93.492 79.475 14.017 14.017 4 W (+) ¼ 10
YuanDaJinHua 76.597 92.263 15.667 15.667 5 W () ¼ 5
YuanFu 93.783 93.607 0.176 0.176 1 W¼5
FuBang 91.412 88.972 2.440 2.440 2
BaoLai 120.081 108.005 12.076 12.076 3
Panel B: non-merger firms
TaiYu 99.905 144.351 44.446 44.446 12 W (+) ¼ 32
DaHua 110.259 98.183 12.076 12.076 9 W () ¼ 59
QunYi 92.802 86.674 6.129 6.129 5 W ¼ 32
ZhongXin 103.622 99.933 3.688 3.688 3
YongChang 103.720 131.854 28.134 28.134 11
TaiZhen 106.911 112.161 5.250 5.250 4
JinDing 128.472 126.320 2.152 2.152 1
RiShen 86.400 78.099 8.301 8.301 6
DaXin 158.656 169.537 10.881 10.881 7
KangHe 106.409 110.016 3.608 3.608 2
YaZhou 136.646 125.133 11.513 11.513 8
XinBao 97.704 161.688 63.985 63.985 13
TongYi 83.194 100.759 17.565 17.565 10
a
Operating cost/operating income adjusted for the drop in the market
Wilcoxon sign rank (double end) test indicates that rank difference is statistically
insignificant, Eq. 21.6
operating income to total assets for both merger and non-merger security firms,
indicating a positive market synergy. However, after adjusting the operating
income market decline, the results indicate insignificant differences in the ratio of
operating income to total assets for both merger and non-merger security
firms. We used standardized score of variability of operating risk to rank firms
for their financial synergy. The test results show there is insignificant financial
synergy for merger firms; however, there is significant change in the variability of
operating risk for non-merger security firms.
Post-merger credit rating of the firm depends on the extent of synergy produced by
the merger. Components of merger synergy are operating synergy, market synergy,
and financial synergy.
Each synergy component is determined by firm characteristics: financial
structure, solvency, asset utilization, profitability, cash flow, growth, scale, and
industry-specific ratio. In this study a number of financial ratios are used to assess
608 S. Srivastava and K. Hung
Table 21.8 Test of significance of merger and non-merger firms’ operating return on assetsa
December December
Firm 1999 score 2000 score Difference Di Class W
Panel A: merger firms
JianHong 16.868 17.171 0.303 0.303 1 W (+) ¼ 0
YuanDaJinHua 13.591 22.133 8.542 8.542 3 W () ¼ 15
YuanFu 14.667 26.617 11.949 11.949 4 W¼0
FuBang 14.245 16.876 2.630 2.630 2
BaoLai 12.429 25.243 12.814 12.814 5
Panel B: non-merger firms
TaiYu 12.881 15.878 2.997 2.997 12 W (+) ¼ 10
DaHua 15.783 12.746 3.037 3.037 2 W () ¼ 81
QunYi 13.893 18.924 5.031 5.031 4 W ¼ 10
ZhongXin 13.214 26.546 13.332 13.332 5
YongChang 16.797 19.702 2.905 2.905 11
TaiZhen 12.783 14.188 1.405 1.405 8
JinDing 11.884 15.938 4.054 4.054 6
RiShen 18.111 24.489 6.379 6.379 1
DaXin 15.232 15.240 0.008 0.008 10
KangHe 14.146 16.561 2.414 2.414 7
YaZhou 16.653 16.268 0.385 0.385 3
XinBao 13.316 15.023 1.706 1.706 13
TongYi 13.732 21.170 7.438 7.438 9
a
Operating income/total assets
Wilcoxon sign rank (double end) test indicates that ratio difference is statistically significant,
Eq. 21.6
operating performance of the firm before and after the merger. The following
section outlines firm characteristics and variables to measure operating
performance.
Merger Synergy
Operating Synergy
A. – Financial structure
Xi1 – Debt ratio
Xi2 – Ratio of long-term capital to fixed assets
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 609
Table 21.9 Test of significance of merger and non-merger firms’ adjusted operating return on
assetsa
December December
Firm 1999 score 2000 score Difference Di Class W
Panel A: merger firms
JianHong 0.0190 0.0154 0.0035 0.0035 2 W (+) ¼ 3
YuanDaJinHua 0.0153 0.0199 0.0046 0.0046 3 W () ¼ 12
YuanFu 0.0165 0.0239 0.0074 0.0074 4 W¼3
FuBang 0.0160 0.0152 0.0008 0.0008 1
BaoLai 0.0140 0.0227 0.0087 0.0087 5
Panel B: non-merger firms
TaiYu 0.0145 0.0143 0.0002 0.0002 1 W (+) ¼ 53
DaHua 0.0177 0.0115 0.0063 0.0063 12 W () ¼ 38
QunYi 0.0156 0.0170 0.0014 0.0014 5 W ¼ 33
ZhongXin 0.0149 0.0239 0.0090 0.0090 13
YongChang 0.0189 0.0177 0.0012 0.0012 4
TaiZhen 0.0144 0.0128 0.0016 0.0016 7
JinDing 0.0134 0.0143 0.0010 0.0010 2
RiShen 0.0204 0.0220 0.0017 0.0017 8
DaXin 0.0171 0.0137 0.0034 0.0034 9
KangHe 0.0159 0.0149 0.0010 0.0010 3
YaZhou 0.0187 0.0146 0.0041 0.0041 11
XinBao 0.0150 0.0135 0.0015 0.0015 6
TongYi 0.0154 0.0190 0.0036 0.0036 10
a
(Operating income/total assets) adjusted for the drop in the market
Wilcoxon sign rank (double end) test indicates that ratio difference is statistically insignificant, Eq. 21.6
B. Solvency
Xi3– Current ratio
Xi4– Quick ratio
C. Asset utilization
Xi5 – Operating return on assets
Xi6 – Net worth turnover ratio
D. Profitability
Xi7 – Return on assets
Xi8 – Return on equity
Xi9 – Profit margin
Xi10 – Earnings per share
E. Cash flow
Xi11 – Cash flow to short-term liability
Xi12 – 5-year cash flow to debt obligations
Xi13 – Retention ratio
F. Growth
Xi14 – Growth in revenue
Xi15 – Growth in earnings
610 S. Srivastava and K. Hung
Table 21.10 Test of significance of merger and non-merger firms’ change in market share
December December
Firm 1999 score 2000 score Difference Di Class W
Panel A: merger firms
JianHong 6.169 5.255 0.914 0.914 1 W (+) ¼ 1
YuanDaJinHua 11.463 17.270 5.807 5.807 5 W () ¼ 14
YuanFu 5.992 7.497 1.505 1.505 2 W¼1
FuBang 6.007 7.698 1.692 1.692 3
BaoLai 5.745 8.076 2.331 2.331 4
Panel B: non-merger firms
TaiYu 1.982 1.475 0.507 0.507 5 W (+) ¼ 80
DaHua 9.589 5.589 4.000 4.000 13 W () ¼ 11
QunYi 8.389 8.197 0.192 0.192 1 W ¼ 11
ZhongXin 4.932 6.402 1.470 1.470 11
YongChang 3.090 2.559 0.531 0.531 6
TaiZhen 5.462 4.929 0.532 0.532 7
JinDing 5.128 4.678 0.450 0.450 4
RiShen 9.547 7.487 2.061 2.061 12
DaXin 3.109 1.830 1.280 1.280 10
KangHe 2.580 2.209 0.371 0.371 3
YaZhou 2.300 1.462 0.838 0.838 8
XinBao 1.431 1.178 0.253 0.253 2
TongYi 7.085 6.210 0.876 0.876 9
Wilcoxon sign rank (double end) test indicates that ratio difference is statistically significant,
Eq. 21.6
G. Size
Xi16 – Total assets
Xi17 – Net worth
H. Industry-specific ratios
Xi18 – Consignment to current assets
Xi19 – Long-term financing to net worth
Market Synergy
Financial Synergy
Table 21.11 Test of significance of merger and non-merger firms’ operating risk
December December
Firm 1999 score 2000 score Difference Di Class W
Panel A: merger firms
JianHong 5250.000 8.178 5258.178 5258.178 5 W (+) ¼10
YuanDaJinHua 42.250 29.701 71.951 71.951 2 W () ¼ 5
YuanFu 10.324 14.973 4.649 4.649 1 W¼5
FuBang 1017.647 17.895 1035.542 1035.542 3
BaoLai 93.671 1010.000 1103.671 1103.671 4
Panel B: non-merge firms
TaiYu 1274.074 205.391 1479.465 1479.465 11 W (+) ¼ 70
DaHua 127.941 54.839 182.780 182.780 7 W () ¼ 21
QunYi 1657.895 1.198 1656.697 1656.697 12 W ¼ 21
ZhongXin 59.877 3.089 62.965 62.965 3
YongChang 249.479 216.376 33.103 33.103 1
TaiZhen 230.769 96.078 134.691 134.691 6
JinDing 101.587 7566.667 7465.079 7465.079 13
RiShen 29.612 67.241 96.853 96.853 4
DaXin 58.108 127.016 185.124 185.124 8
KangHe 228.378 72.840 301.218 301.218 9
YaZhou 50.000 0.000 50.000 50.000 2
XinBao 413.333 292.208 705.541 705.541 10
TongYi 64.151 37.701 101.852 101.852 5
Wilcoxon sign rank (double end) test indicates that ratio difference is statistically insignificant,
Eq. 21.6
Appendix 2
where F1 . . ., Fq are common factors, ei is specific factor for variable Xi, and Lij is
the factor loading of variable Xi and common factor Fj. Factors extracted are
independent and also analysis preserves the information in the original variables.
There is no overlap of information among principle components. The model should
be parsimonious, in the sense that a few principle components should be able to
replace the original set of variables.
All principle components with factor loading greater than 1 are selected. Then
weights are assigned to different variables based on percentage of variability
explained. The variability percentage adjustment is adjusted as follows:
V adj
j ¼ Vj =Sj V j (21.2)
The adjusted variability percentage Vadj j is multiplied by factor loading Lij and
summed over all variables (j ¼ 1,2. . .q) to get the total loading fo variable Xi. Then
adjust Li such that total factor loadings add to 100.
X
Li ¼ V adj
j Lij
X (21.3)
Ladj
i ¼ 100 Li = Li
Every variable has a weight, Ladji before and after the estimation period; we average
it to obtain weight Wi. For some variables, Xi, the greater the value of the variable,
the better it is for the operating, financial, or merger synergy. Those variables are
classified as positive variables. If opposite is true, then those variables are classified
as negative variable. Hence adjusted variable weights are redesigned to correctly
reflect operating synergy score: Wi* ¼ {Wi or –Wi} for positive and negative
variables, respectively.
Performance Scores
All variables are not measured in the same unit, so they are standardized as
Xi ¼ {Xi–Ave(Xi)}/s, where Ave(Xi) is the average and s is the standard
deviation of variable Xi (i ¼ 1,2. . .p). The adjusted variable weight Wi* multiplied
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 613
P
S¼ W i : Z i (21.4)
where the summation is over the first 19 variables listed in Appendix 1. Then firms
are ranked in terms of their respective comprehensive performance score. The
greater the total score is, the better is the comprehensive performance rating. On
the other hand, the smaller the total score, the worse is the performance rating and
lower is its rank.
Appendix 3
Di ¼ Xi Yi i ¼ 1, 2, n
Ri ¼ rankðjDi jÞ i ¼ 1, 2, n
X
Wð þ Þ ¼ Ri Xi Yi > 0 (21.5)
X
Wð Þ ¼ Ri Xi Yi < 0
W ¼ minðW þ , WÞ
where W(+) is the total of absolute value of serial numbers of positive rank changes
and W() is the total of absolute value serial numbers of negative rank changes.
W is the test statistic.
614 S. Srivastava and K. Hung
Test Hypotheses
W ðWða=2ÞÞ or
(21.6)
W ðWða=2ÞÞ
This means that merger has produced significant synergy and performance
(and credit) rating has affected. The appropriate variables for operating, market,
and financial variables are operating cost ratio, ratio of operating income to total
assets and market share, and variability of operating risk, respectively.
References
Ansoff, H. I., Brandenburg, R. G., Portner, F. E., & Radosevich, R. (1971). Acquisition behavior of
U.S. manufacturing firms (pp. 1946–1965). Nashville: Vanderbilt University Press.
Beattie, D. L. (1980). Conglomerate diversification and performance: A survey and applied
economics. Applied Economics, 12, 251–273.
Brigham, E. F., & Daves, P. (2007). Intermediate financial management. Mason: Thomson South
West.
Copeland, T. E., & Weston, J. F. (1979). Financial theory and corporate policy. Reading:
Addison-Wesley.
Fluck, Z., & Lynch, A. W. (1999). Why do firms merge and then divest? A theory of financial
synergy. Journal of Business, 72, 319–346.
Gardner, D. C. (1996). Introduction to mergers and acquisitions. London: Pitman Publishing.
Healy, P. M., Palepu, K. G., & Ruback, R. S. (1992). Does corporate performance improved after
mergers? Journal of Financial Economics, 31, 135–175.
Higgins, R. C., & Schall, L. D. (1975). Corporate bankruptcy and conglomerate merger.
The Journal of Finance, 30, 93–119.
Hoshino, Y. (1982). The performance of corporate mergers in Japan. Journal of Business
Financial and Accounting, 9, 145–163.
Johnson, R., & Wichern, D. (2007). Applied multivariate statistical analysis (6th ed.). Upper
Saddle River: Pearson Education.
Levy, H., & Sarnat, M. (1970). Diversification, portfolio analysis and the uneasy case for
conglomerate mergers. The Journal of Finance, 25, 795–802.
Lewellen, W. G. (1971). A pure financial rationable for conglomerate merger. The Journal
of Finance, 26, 521–537.
21 Effect of Merger on the Credit Rating and Performance of Taiwan Security Firms 615
Muller, D. C. (1985). Mergers and market share. The Review of Economics and Statistics, 67,
66–74.
Rhodes, S. A. (1983). Power, empire building, and mergers. Lexington: Lexington Books.
Scherer, F. M. (1980). Industrial market structure and economic performance. Boston: Houghton
Mifflin.
Seth, A. (1990). Value creation in acquisitions: A re-examination of performance issues. Strategic
Management Journal, 11, 90–115.
Singh, P., & Montegomery, T. (1987). Using simulated mergers to evaluate corporate diversifica-
tion strategies. Strategic Management Journal, 8, 377–386.
Standard & Poor’s Corporation. (1996). Standard & Poor’s corporation ratings criteria.
New York: McGraw Hill.
Weston, J. F., & Chung, K. S. (1983). Do merger make money? A reach summary. Mergers &
Acquisitions, 18, 40–48.
Weston, J. F., & Mansinghka, S. K. (1971). Test of the efficiency: Performance of conglomerate
firms. The Journal of Finance, 26, 919–939.
Williamson, O. E. (1981). The modern corporation: Origins, evolution, attributes. Journal
of Economic Literature, 14, 1537–1568.
Woolridge, J. (2009). Introductory econometrics: A modern approach (4th ed.). Mason: Cengage
Learning.
On-/Off-the-Run Yield Spread Puzzle:
Evidence from Chinese Treasury Markets 22
Rong Chen, Hai Lin, and Qianni Yuan
Contents
22.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 618
22.2 Bond Pricing Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 620
22.2.1 On-the-Run Bond Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 620
22.2.2 Off-the-Run Bond Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 621
22.3 Data and Empirical Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 621
22.3.1 Data Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 621
22.3.2 Empirical Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 623
22.3.3 Estimation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 624
22.3.4 Regression Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 625
22.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 629
Appendix 1: Nonlinear Kalman Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 629
Appendix 2: Matlab Codes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 632
Codes for the On-the-Run Bonds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 632
Codes for the Off-the-Run Bonds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 635
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 637
Abstract
In this chapter, we document a negative on-/off-the-run yield spread in Chinese
Treasury markets. This is in contrast with a positive on-/off-the-run yield spread
in most other countries and could be called an “on-/off-the-run yield spread
puzzle.” To explain this puzzle, we introduce a latent factor in the pricing of
Chinese off-the-run government bonds and use this factor to model the yield
We thank the National Natural Science Foundation of China (Grant No. 71101121 and
No. 70971114) for financial support.
R. Chen (*) • Q. Yuan
Department of Finance, Xiamen University, Xiamen, China
e-mail: aronge@xmu.edu.cn; qnyuan@gmail.com
H. Lin
School of Economics and Finance, Victoria University of Wellington, Wellington, New Zealand
e-mail: hai.lin@vuw.ac.nz; cfc1080@gmail.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 617
DOI 10.1007/978-1-4614-7750-1_22,
# Springer Science+Business Media New York 2015
618 R. Chen et al.
Keywords
On-/off-the-run yield spread • Liquidity • Disposition effect • CIR model •
Nonlinear Kalman filter • Quasi-maximum likelihood
22.1 Introduction
1
Specialness refers to the phenomenon that loans collateralized by on-the-run bonds offer lower
interest rates than their off-the-run counterparts in repo markets.
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 619
costs (Vayanos and Weill 2008), and market frictions of information heterogeneity
and imperfect competition among informed traders (Pasquariello and Vega 2009).
No matter what arguments are proposed, however, the important role that liquidity
plays in the positive on-/off-the-run yield spread and the liquidity premium
hypothesis (Amihud and Mendelson 1986) has never been denied. It is widely
accepted, by both practitioners and academics, that off-the-run bonds with a lower
liquidity level tend to have a higher yield than otherwise similar, yet more liquid,
on-the-run bonds.
In this chapter, we document a negative on-/off-the-run yield spread in Chinese
Treasury markets, which is contrary to the usual on-the-run phenomenon in other
countries and could be called the on-/off-the-run yield spread puzzle in China. Guo and
Wu (2006) and Li and He (2008) report that on-/off-the-run yield spread in Chinese
Treasury markets is significantly positive, but they did not match the on-the-run and
off-the-run bonds correctly. For example, they compare the yield of a just-issued
7-year government bond with that of a previously issued 7-year government bond
that has a different maturity. This is not consistent with the calculation of the usual on-/
off-the-run yield spread. In order to calculate the spread correctly, we need to match
the bonds in terms of maturity date. For example, we compare a just-issued 1-year
government bond and a previously issued government bond maturing on similar dates.
The maturities are both about 1 year and the durations are close to each other.
To explain on-/off-the-run yield spread puzzle, we introduce a latent factor in the
pricing of Chinese off-the-run bonds. This latent factor is used to model the yield
difference between on-the-run bonds and off-the-run bonds. We employ a nonlinear
Kalman filter to estimate the model and examine the temporal properties of the latent
factor. We find that the liquidity premium hypothesis still holds in Chinese Treasury
markets. In particular, the change of the latent factor is positively related to the
liquidity difference between off-the-run and on-the-run bonds and positively related
to the market-wide liquidity condition. Both findings are consistent with the liquidity
premium hypothesis. On the other hand, disposition effect (unwillingness to sell old
bonds in bear markets) dramatically changes the sign of the yield spread and causes
the puzzle. The change of latent factor in Chinese Treasury markets is negatively
related to 7-day repo rates. When interest rates go up and the returns of the bond
markets are negative, the holders of off-the-run bonds are reluctant to realize loss and
will not sell their bonds, which consequently leads to a relatively low yield level of
off-the-run bonds and a negative on-/off-the-run yield spread.
Our article makes several contributions to the literature. We document
a negative on-/off-the-run yield spread in China. We introduce a latent factor to
explain the yield difference between on-the-run bonds and off-the-run bonds and
employ the nonlinear Kalman filter in estimation. Our basic ideas are in line with
Longstaff et al. (2005) and Lin et al. (2011). We also provide evidence of irrational
investor behavior that is quite specific in emerging Treasury markets such as China.
In China, the liquidity premium hypothesis still holds, whereas the existence of
disposition effect causes the puzzle.
The chapter is organized as follows. In Sect. 22.2, we present a pricing model of
government bonds that introduces a latent factor for the off-the-run bonds.
620 R. Chen et al.
In Sect. 22.3, we describe the data, report the estimation results, and perform
a variety of regression analyses. We present our conclusions in Sect. 22.4.
We use the Cox et al. (1985, CIR) model to price the Chinese on-the-run govern-
ment bonds. The CIR model has been a benchmark interest rate model because of
its analytical tractability and other good properties. In this model, the risk-free short
rate, rt, is assumed to follow a square-root process as
pffiffiffiffi
dr t ¼ kðy r t Þdt þ sr r t dW r, t , (22.1)
under the risk-neutral measure Q. k is the speed of mean reversion, y is the long-
term mean value, sr is the volatility parameter of rt, and Wr denotes a standard
Brownian motion under Q. Such specification allows for both mean reversion and
conditional heteroskedasticity and guarantees that interest rates are nonnegative.
At time t, the price of an on-the-run government bond maturing at tM could be
written as
" ð tm #
X
M
t ¼ Et
Pon Cm exp r s ds ,
Q
(22.2)
m¼1 t
where Ptonis the on-the-run bond price, Cm is the cash flow payments at time tm, and
M is the total number of cash flow payments. That is, the price of a government bond is
the expected present value of the cash flow payments under the risk-neutral measure.
Solving Eq. 22.2 gives
X
M
t ¼
Pon Cm Am, t exp Bm, t r t , (22.3)
m¼1
where
2ky=s2r
2h exp fðk þ hÞðtm tÞ=2g
Am , t ¼ ,
2h þ ðk þ hÞðexp fðtm tÞhg 1Þ
and
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
h¼ k2 þ 2s2r :
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 621
" ð tm #
X
M
Poff
t ¼ EQ
t Cm exp ðr s þ ls Þds , (22.4)
m¼1 t
where Ptoff is the off-the-run bond price. Similar to Longstaff et al. (2005) and Lin
et al. (2011), we assume that under the risk-neutral measure Q,
dlt ¼ sl dW l, t , (22.5)
X
M
Poff
t ¼ Cm Am, t exp Dm, t Bm, t r t ðtm tÞlt , (22.6)
m¼1
Our sample period is from December 2003 to February 2009. We use monthly
data and choose the actively traded government bonds with 1 year and 3 years to
maturity. Thus, for each month in our sample period, a most recently issued 1-year
and 3-year government bonds are selected as the on-the-run bonds, and we get
another old bond maturing on similar dates to match each on-the-run bond.2
Altogether, 63 matched pairs of 1-year government bonds and 63 matched pairs
of 3-year government bonds are included in the final sample.
Table 22.1 reports the summary statistics for the sample bonds. As expected, the
modified durations of off-the-run and on-the-run bonds are quite close to each other.
This means if interest rate risk is the only risk factor, these bonds should be traded at
similar yields. To examine whether the same on-the-run phenomenon exists in
Chinese Treasury markets, we compute the on-/off-the-run yield spread as
DyM, t ¼ yoff
M, t yM, t ,
on
(22.7)
where yoff on
M;t and yM;t are the time t yield of the off-the-run bond and the on-the-run
bond maturing at tM, respectively. In our sample, tM t is equal to 1 year or 3 years.
Figure 22.1 plots the time series of DyM,t of the 1-year bond and 3-year bond.
Table 22.1 also reports the means of on-/off-the-run yield spreads and their
statistical significance. Both the 1-year on-/off-the-run yield spread and the 3-year
on-/off-the-run yield spread are negative, which is inconsistent with the findings in
other markets. Moreover, the 3-year on-/off-the-run yield spread is significantly
2
The main reason for the data selection comes from the concern of trading activity. Trading in
Chinese Treasury markets is not active, especially in the earlier period.
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 623
2.00%
1.00%
0.50%
0.00%
−0.50%
−1.00%
−1.50%
−2.00%
−2.50%
Fig. 22.1 Time series of Chinese on-/off-the-run yield spread. This figure plots the yield
difference between the off-the-run and on-the-run Chinese government bonds between December
2003 and February 2009
To explain the on-/off-the-run yield spread puzzle in China, we use the CIR
model to price the on-the-run issues and the CIR model with the latent factor to
price the off-the-run issues. We first estimate the parameters of the CIR model
using all on-the-run bonds. Given the parameters of the CIR process, we further
estimate the parameters of the latent factor using off-the-run bonds. Thus, the latent
factor represents the yield difference between on-the-run and off-the-run bonds.
In our empirical study, we employ the Kalman filter to estimate the parameters.3
The standard Kalman filter is not appropriate here because it requires linear state
functions and the measurement functions, while Eqs. 22.3 and 22.6 are nonlinear.
3
See Hamilton (1994) for an explanation of Kalman filter.
624 R. Chen et al.
function
k 0.089(6.403)a
y 0.023(6.486)a
sr 0.063(247.921)a
sl 0.009(8.530)a
var(et) 0.000008
var(t) 0.000007
One year Three years One year Three years
var(vont ) 0.000035 0.000020
var(voff
t ) 0.000011 0.0000006
RMSE 0.0045 0.0038 0.0062 0.0058
MAD 0.0596 0.0517 0.0705 0.0672
This table reports the estimate results of pricing models for on-the-run and off-the-run
Chinese government bonds. The parameters of the CIR model are estimated from monthly
on-the-run bond data. These parameters are then used to estimate the latent factor in the monthly
off-the-run bond " data. # We use a nonlinear Kalman" filter# approach to estimate the parameters.
on off on off
y1, t y 1, t o1, t o 1, t
yon off
t = yon , yt = t = oon , and vt =
, von off
, where subscript 1 and 3 denote 1-year
3, t yoff
3, t 3, t ooff
3, t
and 3-year bonds and superscript on and off denote on-the-run and off-the-run bonds, respectively.
The numbers in parentheses are t values
a
indicates statistical significance at the 1 % level respectively. RMSE and MAD are root mean
square error and mean absolute deviation, respectively
We therefore use the nonlinear Kalman filter for estimation. The details of the
nonlinear Kalman filter with its Matlab codes are reported in the Appendix.
After we estimate the parameters, we then study the dynamic of the latent
component and examine its temporal properties to explore the explanations of the
on-/off-the-run yield spread puzzle in China.
5.00%
r 7-day Repo Rate
4.50%
4.00%
3.50%
3.00%
2.50%
2.00%
1.50%
1.00%
0.50%
0.00%
Fig. 22.2 Time series of implied risk-free interest rate and 7-day repo rate in Chinese interbank
market. This figure plots the time series of implied risk-free interest rate estimated from the CIR
model using Chinese on-the-run government bonds, and the time series of 7-day repo rate in
Chinese interbank market
Figure 22.2 plots the time series of r estimated from the model. Most of the time,
the value of r is in the interval between 2 % and 4 %. We also plot the time series of
the 7-day repo rate in the Chinese interbank market. Similar trends in these two
curves suggest that r does capture the change of the market interest rate.
The analysis so far reveals that on average, the off-the-run bonds are traded at
a higher price or lower yield than the on-the-run bonds in Chinese Treasury
markets, which is hard to explain rationally. We next explore the information
contained in this negative yield spread by examining the temporal properties of
the latent component, l.
4
We also estimate the parameters of r and l jointly using the on-the-run and off-the-run data
together and find the results are quite similar.
626 R. Chen et al.
0.40%
0.20%
0.00%
−0.20%
−0.40%
−0.60%
−0.80%
Fig. 22.3 Time series of latent factor. This figure plots the time series of latent factor estimated
from the Chinese off-the-run government bonds
TRoff
1, t TR on
1, t þ TR off
3, t TR on
3, t
TRt ¼ 105 ,
2
where TR is the turnover ratio, the subscript 1 and 3 denote 1-year bonds and 3-year
bonds, and the superscript on and off denote on-the-run bonds and off-the-run bonds.
Pasquariello and Vega (2007, 2009) find that the release of macroeconomic news
changes liquidity, and hence the on-/off-the-run yield spread, in the US Treasury
market. For example, when macroeconomic news brings more funds into the bond
market, market-wide liquidity conditions will be better, investors might trade old
bonds more actively, and the yield difference between off-the-run bonds and
on-the-run bonds might decrease, and vice versa. Similarly, we introduce the
percentage change of a broad money supply measure, DM2, as a proxy of
market-wide liquidity conditions to examine whether there is covariation of the
latent component with changes in market-wide liquidity conditions. We use one
lagged DM2 to examine the impact of macroeconomic conditions on Dlt.
The last factor we investigate is investors’ behavior. It is observed that in
Chinese bond markets, there exists a “disposition effect.” Bond holders are
reluctant to realize loss and will not sell old bonds if they have a loss from the
investment. Consequently, old bonds might be traded at a lower yield than new
bonds. In China, the 7-day repo market is one of the most active bond markets, and
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 627
the change of the 7-day repo rate is a good measure of market conditions. When the
7-day repo rate goes up, the bond investment will generate a loss for the investors
and the disposition effect might occur. We use the interbank 7-day repo rates as the
proxy of the market interest rate to investigate whether the on-/off-the-run yield
spread puzzle in China is related to this irrational behavior. If the investors are
rational and the liquidity premium hypothesis holds, the liquidity of the whole
market will decrease in the bear market, and the old bonds will be traded at a higher
yield. Thus, detecting the response of the on-/off-the-run yield spread to the
change of the 7-day repo rate could help us distinguish whether the disposition
effect or the liquidity factor dominates. Similarly, we use one lagged 7-day repo
rate in the time series regression.
In what follows, we first run univariate time series regressions of Dlt against each
variable and then conduct multivariate regression analysis against all the three
factors. The regression models are specified as follows:
Univariate regression of Dlt on the turnover ratio difference, TRt:
Dlt ¼ b0 þ b1 TRt þ et :
Univariate regression of Dlt on the lagged percentage change of the money supply,
DM2t1:
Dlt ¼ b0 þ b1 DM2t1 þ et :
Univariate regression of Dlt on the lagged 7-day repo rate, Rt1:
Dlt ¼ b0 þ b1 Rt1 þ et :
628 R. Chen et al.
The adjusted R square of the multivariate is about 16 %. This suggests that the
on-/off-the-run spread could be partly explained by the change in market-wide
liquidity conditions and the disposition effect.
22.4 Conclusion
Let yon
M;t represent the time t yield of an on-the-run government bond maturing at tM.
Equation 22.3 could be written as
XM X M
t ¼
Pon Cm Am, t exp Bm, t r t ¼ Cm exp yon
M, t ðtm tÞ : (22.8)
m¼1 m¼1
As shown, yon
M;tis a nonlinear function of rt, which is inconsistent with the
requirements of the standard Kalman filter that state functions and measurement
functions should be linear. So we use the extended (nonlinear) Kalman filter to
linearize nonlinear functions. The idea is to employ the Taylor expansions around
the estimate at each step. That is, we express yon
M;t as
@yon
M, t
yon ðr
M, t t Þ y on
^
M, t tjt1 þ
r jrt ¼^r tjt1 r t r^tjt1 , (22.9)
@r t
where r^tjt1 is the estimate of rt at time t–1.
630 R. Chen et al.
@yon
M, t
To get @rt , we calculate the first-order derivative of Pon
t with respect to rt,
Thus, we have
X
M
Cm Am, t Bm, t exp Bm, t r t
@yon
M, t m¼1
¼
: (22.11)
@r t X
M
Cm ðtm tÞ exp yM, t ðtm tÞ
on
m¼1
Given r^tjt1, we can use Eq. 22.8 to calculate yon
M, t r^tjt1 and then use Eq. 22.11
@yon
to get M, t
@rt
Finally, the linearized measurement model for the on-the-run issues at time t is
t ¼at þbt r t þ vt
yon on on on
(22.12)
where
yon
1, t
t ¼
yon
yon
,
3, t
2 3
@yon 1, t
on ^
6 y1, t ^
r tjt1 r tjt1 jr ¼^r
aon ¼6 @r t t tjt1 7 7,
t 4 @y3, t
on 5
y3, t r tjt1 r^tjt1
on ^
jrt ¼^r tjt1
@r
2 on 3
@y1, t
j
6 @r rt ¼^r tjt1 7
bon 6 7,
t ¼ 4 @yon
t
3, t
5
jrt ¼^r tjt1
@r t
and von
t is the error term,
oon
1, t
von ¼ ,
t oon
3, t
where the subscript of 1 and 3 represent the 1-year bonds and 3-year bonds, while
the superscript on refers to on-the-run bonds.
After we get the measurement function, the third step is to rewrite (22.1) as
a discrete state function,
r t ¼ g þ fr t1 þ et , (22.13)
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 631
where
g ¼ yð1 expðk DtÞÞ,
f ¼ expðk DtÞ,
and et is the error term of rt and Dt is the size of the time interval in the discrete
sample. In our study, Dt ¼ 0.0833. The conditional mean and conditional variance
of rt are
r^tjt1 ¼ yð1 expðkDtÞÞ þ expðkDtÞ r t1
2 1 expðkDtÞ 1
Var r tjt1 ¼ sr yð1 expðkDtÞ þ expðkDtÞ r t1 Þ :
k 2
(22.14)
Similarly, the state functions of the off-the-run issues are
r t ¼ g þ fr t1 þ et
(22.15)
lt ¼ lt1 þ sl et :
The conditional mean and conditional variance of lt are lt1 and s2l Dt,
respectively.
The corresponding measurement function is
t ¼ at þ bt r t þ jt lt þ vt ,
off
yoff off off off
(22.16)
" #
ooff
1, t
voff ¼ ,
t
ooff
3, t
X
M
Cm Am, t Bm, t exp Dm, t Bm, t r t ðtm tÞlt
@yoff
M, t
¼ m¼1
,
@r t X
M
Cm ðtm tÞ exp yoff
M, t ðtm tÞ
m¼1
X
M
Cm Am ðtm tÞ exp Dm, t Bm, t r t ðtm tÞlt
@yoff
M, t
¼ m¼1
,
@lt X
M
Cm ðtm tÞ exp yoff
M, t ðtm tÞ
m¼1
H¼[pz1(i) pz3(i)]’;
%C1¼[zm1 zm3]’-H*r_(i);
%zm(i,:)¼ C1+H*r_(i);
%zm(i,:)¼ C1+H*r(i);
zm(i,:)¼[zm1 zm3]’; %the prediction of YTMs
v¼z(i,:)’-zm(i,:)’; %the error of measurement functions
v1(i,:)¼v’; % the error between the prediction and the
real value
F¼H*P_*H’+R; %the kalman gain
if det(F)<¼0
logfun¼0;
return
end
rr(i,:)¼r_(i);
zz(i,:)¼zm(i,:);
r(i)¼r_(i)+P_*H’*inv(F)*v; %update r
P¼P_-P_*H’*inv(F)*H*P_; %update P
ll¼-0.5*log(det(F))-0.5*v’*inv(F)*v; %likelihood
function
logfun¼logfun+ll;
r_(i+1)¼A*r(i)+C; %predict r
P_¼A*P*A’+Q; %predict P
end
QQ¼Q;
RR¼R;
logfun¼-logfun;
function covv¼covirance(param)
covv¼zeros(4,4);
for i¼1:4
for j¼1:4
parama¼param;
paramb¼param;
paramab¼param;
parama(i)¼param(i)*1.01;
paramb(j)¼param(j)*0.99;
paramab(i)¼param(i)*1.01;
paramab(j)¼paramab(j)*0.99;
ua¼kalfun(parama);
db¼kalfun(paramb);
udab¼kalfun(paramab);
kk¼kalfun(param);
covv(i,j)¼(ua+db-kk-udab)/((0.01*param(i))*
(0.01*param(j)));
end
end
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 635
for i¼1:63
Q¼sigm3^2*dt; %the conditional variance of state
functions
pz1(i)¼(c1(i)*exp(a(1)-b(1)*r(i)+e(1)-L_(i)*1))/c1(i)
*exp(-z1(i)*1); %the partial derivative of one year z against r
pz3(i)¼sum(c3(i,:)’.*tao.*exp(a-b*r(i)+e-L_(i)
*tao))/sum(c3(i,:)’.*tao.*exp(-z3(i)*tao)); %the par-
tial derivative of three year z against r
P1¼c1(i)*exp(a(1)-b(1)*r(i)+e(1)-L_(i)*1); %the pre-
diction price of one-year bond
P3¼sum(c3(i,:)’.*exp(a-b*r(i)+e-L_(i)*tao)); % the
prediction price of three-year bond
%zm1¼bndyield(P1,c1(i),’20-Jan-1997’,’20-Jan-
1998’,1);
zm1¼-log(P1/c1(i)); %nonlinear measurement function
for one-year bonds
zm3¼bndyield(P3,couponrate(i),’20-Jan-1997’,’20-Jan-
2000’,1); %nonlinear measurement function for three-year
bonds
H¼[pz1(i) pz3(i)]’;
%C1¼[zm1 zm3]’-H*r_(i);
%zm(i,:)¼ C1+H*r_(i);
%zm(i,:)¼ C1+H*r(i);
zm(i,:)¼[zm1 zm3]’; %the prediction of YTMs
v¼z(i,:)’-zm(i,:)’; %the error of measurement
functions
v1(i,:)¼v’; % the error between the prediction and
the real value
F¼H*P_*H’+R; %the kalman gain
if det(F)<¼0
logfun¼0;
return
end
LL(i,:)¼L_(i);
zz(i,:)¼zm(i,:);
L(i)¼L_(i)+P_*H’*inv(F)*v; %update r
P¼P_-P_*H’*inv(F)*H*P_; %update P
ll¼-0.5*log(det(F))-0.5*v’*inv(F)*v;
logfun¼logfun+ll;
L_(i+1)¼L(i); %predict r
P_¼P+Q; %predict P
end
QQ¼Q;
RR¼R;
logfun¼-logfun;
22 On-/Off-the-Run Yield Spread Puzzle: Evidence from Chinese Treasury Markets 637
function covvL¼coviranceL(paramL)
covvL¼zeros(2,2);
for i¼1:2
for j¼1:2
parama¼paramL;
paramb¼paramL;
paramab¼paramL;
parama(i)¼paramL(i)*1.0000001;
paramb(j)¼paramL(j)*0.9999999;
paramab(i)¼paramL(i)*1.0000001;
paramab(j)¼paramab(j)*0.9999999;
ua¼kalfunL(parama);
db¼kalfunL(paramb);
udab¼kalfunL(paramab);
kk¼kalfunL(paramL);
covvL(i,j)¼(ua+db-kk-udab)/((0.0000001*paramL(i))*
(0.0000001*paramL(j)));
end
end
References
Amihud, Y., & Mendelson, H. (1986). Asset pricing and the bid-ask spread. Journal of Financial
Economics, 17, 223–249.
Amihud, Y., & Mendelson, H. (1991). Liquidity, maturity, and the yields on US treasury
securities. The Journal of Finance, 46, 1411–1425.
Boudoukh, J., & Whitelaw, R. (1991). The benchmark effect in the Japanese government bond
market. The Journal of Fixed Income, 1, 52–59.
Boudoukh, J., & Whitelaw, R. (1993). Liquidity as a choice variable: A lesson from the Japanese
government bond market. Review of Financial Studies, 6, 265–292.
Cox, J., Ingersoll, J. E., & Ross, S. A. (1985). A theory of the term structure of interest rates.
Econometrica, 53, 385–407.
Furfine, C., & Remolona, E. (2002). What’s behind the liquidity spread? On-the-run and off-the-
run US treasuries in autumn 1998, BIS Quarterly Review, 6, 51–58.
Goldreich, D., Hanke, B., & Nath, P. (2005). The price of future liquidity: Time-varying liquidity
in the US treasury market. Review of Finance, 9, 1–32.
Guo, H., & Wu, K. (2006). Liquidity premium analysis of treasury bonds market in Shanghai stock
exchange. Finance and Economics (In Chinese), 217, 23–29.
Hamilton, J. D. (1994). Time series analysis. Princeton: Princeton University Press.
Hong, Y., Lin, H., & Wang, S. (2010). Modeling the dynamics of Chinese spot interest rates.
Journal of Banking and Finance, 34, 1047–1061.
Kamara, A. (1994). Liquidity, taxes, and short-term treasury yields. Journal of Financial and
Quantitative Analysis, 29, 403–417.
Krishnamurthy, A. (2002). The bond/old-bond spread. Journal of Financial Economics, 66, 463–506.
Li, D., & He, Y. (2008). Study on the price gap between on-the-run and off-the-run treasury
securities of Shanghai stock exchange. Journal of Shanxi Finance and Economics University
(In Chinese), 30, 80–85.
638 R. Chen et al.
Lin, H., Liu, S., & Wu, C. (2011). Dissecting corporate bond and CDS spreads. The Journal of
Fixed Income, 20, 7–39.
Longstaff, F., Mithal, S., & Neis, E. (2005). Corporate yield spreads: Default risk or liquidity?
New evidence from the credit default swap market. The Journal of Finance, 60, 2213–2253.
Mason, R. (1987). The 10-year bond markets. Credit Suisse First Boston, CSFB Research.
Pasquariello, P., & Vega, C. (2007). Informed and strategic order flow in the bond markets. Review
of Financial Studies, 20, 1975–2019.
Pasquariello, P., & Vega, C. (2009). The on-the-run liquidity phenomenon. Journal of Financial
Economics, 92, 1–24.
Strebulaev, I. (2002). Liquidity and asset pricing: Evidence from the US Treasury securities
market (Working paper). London Business School, SSRN.
Vayanos, D., & Weill, P. (2008). A search-based theory of the on-the-run phenomenon. The
Journal of Finance, 63, 1361–1398.
Warga, A. (1992). Bond returns, liquidity, and missing data. Journal of Financial and Quantitative
Analysis, 27, 605–617.
Factor Copula for Defaultable Basket
Credit Derivatives 23
Po-Cheng Wu, Lie-Jane Kao, and Cheng-Few Lee
Contents
23.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 640
23.2 Factor Copula with Issuer Default Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 641
23.3 Pricing a Defaultable Basket Credit Linked Note . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 644
23.4 Numerical Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 646
23.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 650
Appendix 1: Factor Copula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 651
One-Factor Gaussian Copula Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 651
Law of Iterated Expectations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 652
Appendix 2: Cholesky Decomposition and Correlated Gaussian Random Numbers . . . . . . . . . 653
Cholesky Decomposition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 653
Correlated Gaussian Random Numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 653
Appendix 3: Monte Carlo Simulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 654
Weak Law of Large Numbers (WLLN) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 654
Strong Law of Large Numbers (SLLN) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 654
Uniform and Nonuniform Random Numbers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 654
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 655
Abstract
In this article, we consider a factor copula approach for evaluating basket credit
derivatives with issuer default risk and demonstrate its application in a basket
credit linked note (BCLN). We generate the correlated Gaussian random
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 639
DOI 10.1007/978-1-4614-7750-1_23,
# Springer Science+Business Media New York 2015
640 P.-C. Wu et al.
numbers by using the Cholesky decomposition, and then the correlated default
times can be decided by these random numbers and the reduced-form model.
Finally, the fair BCLN coupon rate is obtained by the Monte Carlo simulation.
We also discuss the effect of issuer default risk on BCLN. We show that the
effect of issuer default risk cannot be accounted for thoroughly by considering
the issuer as a new reference entity in the widely used one-factor copula model,
in which constant default correlation is often assumed. A different default
correlation between the issuer and the reference entities affects the coupon rate
greatly and must be taken into account in the pricing model.
Keywords
Factor copula • Issuer default • Default correlation • Reduced-form model •
Basket credit derivatives • Cholesky decomposition • Monte Carlo simulation
23.1 Introduction
Structural and reduced-form models are the two main approaches for modeling
default risk. The structural model (Merton 1974) defines default as occurring when
a firm’s asset value falls below its debt. The reduced-form model (Jarrow and
Turnbull 1995), also known as the intensity model, views the default event as an
unexpected exogenous stochastic event. It estimates the intensity of the default
occurrence by using market data.
However, whether by the structural or reduced-form model, obtaining the joint
distribution of default times among a set of assets will be very complicated. Li (1999,
2000) first introduces the copula function (Sklar 1959) to simplify the estimation of
the joint distribution. Li assumes that the default events of reference entities follow
a Poisson process and sets the dependence structure as a Gaussian copula function.
Finally, he performs Monte Carlo simulation to obtain the default times. The copula
approach is the main approach for multi-name credit derivatives pricing in the last
decade. Mashal and Naldi (2003) use the copula approach to analyze how the default
probabilities of the protection sellers and buyers affect basket default swap (BDS)
spreads. While pricing the single-name credit default swap (CDS) with counterparty
risk based on the continuous-time Markov model, Walker (2006) indicates that using
a time-dependent correlation coefficient can improve the market-standard Gaussian
copula approach. By connecting defaults through a copula function, Brigo and
Chourdakis (2009) find that when the counterparty risk is involved, both the default
correlation and credit spread volatility impact the contingent CDS value.
In the implementation of the Gaussian copula using Monte Carlo simulation, the
computational complexity increases with the number of reference entities. Thus the
factor copula method, which makes the default event conditional on independent
state variables, is introduced to deal with these problems. Andersen et al. (2003)
find that one or two factors provide sufficient accuracy for the empirical correlation
matrices one encounters in credit basket applications. Hull and White (2004)
employ a multifactor copula model to price the kth-to-default swap and
23 Factor Copula for Defaultable Basket Credit Derivatives 641
collateralized debt obligation (CDO). Moreover, Laurent and Gregory (2005) use
one-factor Gaussian copula to simplify the dependence structure of reference
entities and apply this approach to price BDS and CDO. Wu (2010) develops
three alternative approaches to price the basket credit linked note (BCLN) with
issuer default risk using only one correlation parameter. Wu et al. (2011) analyze
how issuer default risk impacts the BCLN coupon rate by an implied default
correlation between the issuer and the reference entities.
On the other hand, acceleration techniques such as the importance sampling method
and others are used to improve the simulation efficiency. Chiang et al. (2007) and Chen
and Glasserman (2008) apply the Joshi-Kainth algorithm (Joshi and Kainth 2004), and
Bastide et al. (2007) use the Stein method (Stein 1972) for the multi-name credit
derivative pricing to reduce variance of the simulation results.
This article constructs a factor copula framework to evaluate defaultable basket
credit derivatives. The effect of the default correlation between the issuer and the
reference entities is considered in the proposed model. Its application in BCLN with
issuer default risk is also demonstrated. This study shows that the default correla-
tion between the issuer and the reference entities plays an important role in the
decision of fair BCLN coupon rate.
The remainder of this article is organized as follows. Section 23.2 reviews the
factor copula model and shows the proposed basket credit derivative pricing
model with issuer default event. Subsequently, Section 23.3 introduces the BCLN
and demonstrates how to price it when issuer default risk exists. Section 23.4 presents
the results of numerical analysis. Conclusions are finally drawn in Section 23.5.
The most widely used copula function is the Gaussian copula and its definition is as
follows:
1 1 1
CGa ðu1 , u2 , . . . , uN Þ ¼ FR f ðu1 Þ, f ðu2 Þ, . . . , f ðuN Þ (23.1)
where li is the hazard rate of the reference entity i. Because Fi(t) U(0,1), applying
the Gaussian copula obtains the multivariate joint distribution of default times, as
follows:
642 P.-C. Wu et al.
Fðt1 , t2 , . . . , tN Þ ¼ FR f1 ðF1 ðt1 ÞÞ, f1 ðF2 ðt2 ÞÞ, . . . , f1 ðFN ðtN ÞÞ
(23.3)
Let Xi represent the Gaussian random variable corresponding to the default time
of the reference entity i. In the one-factor copula model, the default time of
reference entity i depends on a common factor eY and a firm specific risk factor
exi. Both eY and eXi are independent standard Gaussian variables. The details of
one-factor copula model are given in Appendix 1. Thus Xi can be created via
Cholesky decomposition, as follows:
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
X i ¼ r X i Y eY þ 1 rX i Y 2 e X i , i ¼ 1, 2, . . . , N (23.4)
where rXiY denotes the correlation coefficient between the reference entity Xi and
the common factor eY. How to apply the Cholesky decomposition for generating
correlated random variables is shown in Appendix 2.
One-factor Gaussian copula model with constant pairwise correlations has
become the standard market model. In the standard market model, all rXi Y in
Eq. 23.4 are equal to r, then the constant pairwise correlation rXi Xj (i 6¼ j) will
be r2. Let X1 ¼ f1(F1(t1)), X2 ¼ f1(F2(t2)), . . . , XN ¼ f1(FN(tN)), by mapping
ti and Xi, we can simulate the default time of the reference entity i using the
following equation:
lnð1 fðXi ÞÞ
ti ¼ F1
i ðfðXi ÞÞ ¼ , i ¼ 1, 2, . . . , N (23.5)
li
When issuer default risk is involved in the basket credit derivative, a natural
way is to view it as one reference entity of the derivative holder’s credit portfolio,
as discussed in Wu (2010). Wu (2010) assumes the issuer default time is deter-
mined by a Gaussian random variable Z. Like the reference entity variable Xi, Z is
decided by the common factor eY and the issuer’s specific risk factor eZ. Both eY
and eZ are independent standard Gaussian variables. Because the issuer is
viewed as one additional reference name in the portfolio, the correlation coeffi-
cient between Z and eY is also r. In this approach, Z and Xi are formulated as
follows:
pffiffiffiffiffiffiffiffiffiffiffiffiffi
Z ¼ reY þ 1 r2 e Z (23.6)
pffiffiffiffiffiffiffiffiffiffiffiffiffi
Xi ¼ reY þ 1 r2 e X i , i ¼ 1, 2, . . . , N (23.7)
In the above approach, the default correlation between the issuer and the reference
entities will be fixed to r2, which is always positive. Thus, it is not flexible enough
to deal with the default correlation. The default correlation between the issuer and
23 Factor Copula for Defaultable Basket Credit Derivatives 643
Reference
Entities Issuer
Xi, i = 1,2,…,N Z
rXZ
the reference entities has a different impact on the fair coupon rate, and this needs to
be considered in credit derivative pricing.
Suppose that the default correlation between the issuer and the reference entities is
rXZ. The relationship between the issuer, reference entities, and the common factor in
the proposed model is shown in Fig. 23.1. Given that the three random variables eY,
eZ, and eXi are independent of each other, Xi, which is the Gaussian random variable
corresponding to the default time of the reference entity i, can be obtained by the
Cholesky decomposition. Thus, when the default correlation between the issuer and
the reference entities is incorporated into the pricing model, Z and Xi should be
formulated as follows:
pffiffiffiffiffiffiffiffiffiffiffiffiffi
Z ¼ reY þ 1 r2 e Z (23.8)
vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
!2
u
rXZ r 2 u r r 2
t
Xi ¼ reY þ pffiffiffiffiffiffiffiffiffiffiffiffiffi eZ þ 1 r2 pffiffiffiffiffiffiffiffiffiffiffiffiffi eXi
XZ
(23.9)
1 r2 1 r2
To obtain a real number value of Xi, the following criteria must be satisfied.
!2
r r2
r2 þ pXZffiffiffiffiffiffiffiffiffiffiffiffiffi 1 (23.10)
1 r2
According to the above settings, the correlation coefficient between the reference
entity Xi and Xj will be
644 P.-C. Wu et al.
Cov Xi ; Xj
rXi Xj ¼
sXi sXj
0 1
rXZ r2
B re Y þ p ffiffiffiffiffiffiffiffiffiffiffiffiffi eZ C
1 r2
B vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi C
B u !2 C
B u C
B t rXZ r2 C
B þ 1 r pffiffiffiffiffiffiffiffiffiffiffiffiffi eXi ,
2 C
B 1 r2 C
B C
¼ CovB C
B r r 2
C
B reY þ pXZffiffiffiffiffiffiffiffiffiffiffiffiffi eZ C (23.12)
B 1 r2 C
B vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi C
B u !2 C
B u rXZ r2 C
@ t A
þ 1 r pffiffiffiffiffiffiffiffiffiffiffiffiffi eXj
2
1 r2
2
ðrXZ r2 Þ
¼ r2 þ
1 r2
r þ rXZ 2rXZ r2
2 2
¼
1 r2
2 3
XT
rti
6c e I ð t i < tk Þ 7
6 7
BCLN ¼ EQ 6 i¼1 7 (23.13)
4 þ dk ertk I ðtk tT Þ 5
þ ertT I ðtk > tT Þ
where the coupon rate is c and is paid annually. The coupon payment dates are
ti,i ¼ 1,2, ,T. The maturity date is tT. Furthermore, tk is the kth default time, and
t1 < t2 < < tN. dk is the recovery rate of the kth default reference entity. Thus dk
denotes the recovery value, which the issuer pays to the BCLN holder on the kth
default. The discount rate is r%. Finally, Q denotes the risk-neutral probability
measure, and I(.) is an indicator function.
Let the above equation equals one, the equation can be rewritten as:
" #
X
T
cE Q
erti I ðti < tk Þ
i¼1 rtk
(23.14)
Q 1 dk e I ð tk t T Þ
¼E
ertT I ðtk > tT Þ
XW
1 dsk ertk I tsk tT
s
s¼1 i¼1
where W represents the number of simulation runs. tks represents the kth default
time, and dks denotes the recovery rate of the kth default reference entity at the sth
simulation, respectively.
When the issuer default risk is involved, whether the issuer default occurs before
or after the kth default must be taken into account. This article defines ^t as the issuer
default time and ^d as the issuer recovery rate. The BCLN holder gets back the
recovered value of the reference obligation if the kth default occurs before both the
issuer default time ^t and maturity date tT. If the issuer default occurs before the kth
default and maturity date, the issuer will not provide the BCLN holder with the
redemption proceeds and stop the coupon payments. In this situation, the notional
principal multiplied by the issuer recovery rate is returned to the BCLN holder. To
obtain all of the notional principal back, both the kth default time and the issuer
default time must be later than the contract maturity date. Thus, the value of a kth-
to-default BCLN with issuer default risk is modified as follows:
646 P.-C. Wu et al.
2 3
X T
rti
6 c e I ð ti < min ð t k , ^
t Þ Þ 7
6 7
6 i¼1
7
6 7
BCLN ¼ EQ 6 þ d ertk I ðt < minð^t , t ÞÞ 7 (23.16)
6 k k T 7
6 7
4 þ ^d er^t I ð^t < minðtk , tT ÞÞ 5
þ ertT I ðtT < minðtk , ^t ÞÞ
Therefore, the fair value of the coupon rate c with issuer default risk is
2 3
1 dsk ertk I tsk < minð^t s , tT Þ
s
X
W 6 7
6 7
4 ^d er^t s I ^t s < min ts , tT 5
s¼1
rtT
s s k
e I tT < min tk , ^t
c¼ " # (23.17)
XW XT
erti I ti < min tsk , ^t s
s¼1 i¼1
The numerical example presented here is a 5-year BCLN with three reference
entities; all of which are with notional principal one dollar, hazard rate 5 %, and
recovery rate 30 %. Furthermore, the coupon is paid annually; the hazard rate and
recovery rate of the issuer are 1 % and 30 %, respectively. Sixty thousand runs of
Monte Carlo simulation are executed to calculate the coupon rates, and the results
are shown in Tables 23.1, 23.2 and 23.3.
As we can see in Tables 23.1, 23.2 and 23.3, when issuer default risk
is considered by viewing it as one reference entity of the credit portfolio
(column II), the BCLN coupon rate increases compared to those without issuer
default risk (column I) for k ¼ 1–3. This is reasonable because the existence of
issuer default risk increases the risk of holding a BCLN; thus, the holder will
demand a higher coupon rate.
When the default correlations between the issuer and the reference entities are
considered as in the proposed model, the BCLN coupon rates with issuer default risk
(column III) are greater than those without issuer default risk (column I) for k ¼ 2 and 3
in Tables 23.2 and 23.3. However, when k ¼ 1 in Table 23.1, most of the BCLN
coupon rates with issuer default risk are lower than those without issuer default risk,
especially when the issuer and the reference entities are highly negatively or positively
correlated. This result shows that the BCLN coupon rates with issuer default risk are
not necessarily greater than those without issuer default risk. Moreover, from Figs. 23.2,
23
Table 23.1 First-to-default BCLN coupon rates without and with issuer default risk. (I) Default free: Issuer default risk is not included in the pricing model.
(II) In credit portfolio: The issuer is viewed as one reference entity of the credit portfolio. The default correlations between the issuer and the reference entities
are fixed to r2, which is always positive. (III) The proposed model: The default correlation between the issuer and the reference entities is rXZ, which may be
positive or negative
(I) (II) (III) The proposed model
rXZ
r Default free In credit portfolio 0.9 0.6 0.3 0 0.3 0.6 0.9
0.9 8.0195 % 8.0346 % – – – – – – 7.6700 %
0.8 9.2625 % 9.3471 % – – – – 7.0292 % 9.3593 % 7.8852 %
0.7 10.2417 % 10.4279 % – – – 7.2895 % 10.2124 % 10.1562 % 7.9327 %
0.6 11.0469 % 11.3501 % – – – 10.5546 % 11.4061 % 10.5488 % 7.9452 %
0.5 11.7604 % 12.1928 % – – 10.2184 % 12.0213 % 12.0974 % 10.7840 % 7.9650 %
0.4 12.3336 % 12.8846 % – 8.8110 % 11.8856 % 12.9609 % 12.5139 % 10.9275 % 7.9665 %
0.3 12.7890 % 13.4509 % – 10.6576 % 12.9594 % 13.5479 % 12.7852 % 11.0273 % 7.9659 %
0.2 13.1304 % 13.8645 % 7.5062 % 11.7035 % 13.5910 % 13.9236 % 12.9839 % 11.0791 % 7.9894 %
0.1 13.3336 % 14.1181 % 8.7846 % 12.2398 % 13.9263 % 14.1327 % 13.1152 % 11.1038 % 7.9758 %
0 13.3875 % 14.1785 % 9.0738 % 12.3865 % 14.0210 % 14.1785 % 13.1205 % 11.1105 % 7.9933 %
Factor Copula for Defaultable Basket Credit Derivatives
0.1 13.3547 % 14.1215 % 8.7274 % 12.1743 % 13.8962 % 14.1373 % 13.0764 % 11.1036 % 8.0044 %
0.2 13.1236 % 13.8336 % 7.4400 % 11.5964 % 13.4923 % 13.8863 % 12.9678 % 11.0703 % 7.9941 %
0.3 12.7744 % 13.4073 % – 10.5344 % 12.8035 % 13.4871 % 12.7940 % 10.9995 % 7.9702 %
0.4 12.2933 % 12.8209 % – 8.6831 % 11.7668 % 12.8436 % 12.5034 % 10.9036 % 7.9704 %
0.5 11.7091 % 12.1252 % – – 10.1270 % 11.9387 % 12.0415 % 10.7603 % 7.9715 %
0.6 10.9854 % 11.2806 % – – – 10.4846 % 11.3343 % 10.5102 % 7.9448 %
0.7 10.1941 % 10.3750 % – – – 7.1962 % 10.1661 % 10.1106 % 7.9210 %
0.8 9.2061 % 9.2802 % – – – – 6.9644 % 9.2933 % 7.8398 %
0.9 7.9788 % 7.9892 % – – – – – – 7.6319 %
647
648
Table 23.2 Second-to-default BCLN coupon rates without and with issuer default risk. (I) Default free: Issuer default risk is not included in the pricing
model. (II) In credit portfolio: The issuer is viewed as one reference entity of the credit portfolio. The default correlations between the issuer and the reference
entities are fixed to r2, which is always positive. (III) The proposed model: The default correlation between the issuer and the reference entities is rXZ, which
may be positive or negative
(I) (II) (III) The proposed model
rXZ
r Default free In credit portfolio 0.9 0.6 0.3 0 0.3 0.6 0.9
0.9 5.2169 % 5.2585 % – – – – – – 5.2870 %
0.8 4.9714 % 5.1490 % – – – – 5.9481 % 5.1998 % 5.2213 %
0.7 4.7203 % 5.0406 % – – – 6.1463 % 5.3390 % 4.9617 % 5.2041 %
0.6 4.4753 % 4.9077 % – – – 5.5463 % 4.9738 % 4.8270 % 5.2135 %
0.5 4.2476 % 4.7844 % – – 5.8264 % 5.1061 % 4.7372 % 4.7592 % 5.1736 %
0.4 4.0568 % 4.6587 % – 6.1656 % 5.3827 % 4.8129 % 4.6044 % 4.7055 % 5.1541 %
0.3 3.8888 % 4.5440 % – 5.7880 % 5.0664 % 4.6078 % 4.4986 % 4.6566 % 5.1422 %
0.2 3.7766 % 4.4490 % 6.3028 % 5.5060 % 4.8417 % 4.4584 % 4.4238 % 4.6153 % 5.1375 %
0.1 3.6917 % 4.3717 % 6.1079 % 5.3332 % 4.7028 % 4.3759 % 4.3711 % 4.6020 % 5.1434 %
0 3.6646 % 4.3412 % 6.0386 % 5.2612 % 4.6477 % 4.3412 % 4.3441 % 4.5867 % 5.1507 %
0.1 3.6833 % 4.3559 % 6.0959 % 5.3033 % 4.6836 % 4.3606 % 4.3478 % 4.5939 % 5.1636 %
0.2 3.7480 % 4.4025 % 6.2475 % 5.4584 % 4.7916 % 4.4244 % 4.3710 % 4.5969 % 5.1732 %
0.3 3.8817 % 4.5070 % – 5.7235 % 4.9942 % 4.5580 % 4.4430 % 4.6288 % 5.1900 %
0.4 4.0242 % 4.5995 % – 6.0943 % 5.3165 % 4.7513 % 4.5276 % 4.6679 % 5.1908 %
0.5 4.2155 % 4.7249 % – – 5.7533 % 5.0596 % 4.6854 % 4.7051 % 5.2016 %
0.6 4.4092 % 4.8338 % – – – 5.4952 % 4.9002 % 4.7565 % 5.2005 %
0.7 4.6447 % 4.9559 % – – – 6.0639 % 5.2704 % 4.8851 % 5.1929 %
0.8 4.9138 % 5.0873 % – – – – 5.8734 % 5.1404 % 5.2065 %
0.9 5.1952 % 5.2336 % – – – – – – 5.2602 %
P.-C. Wu et al.
23
Table 23.3 Third-to-default BCLN coupon rates without and with issuer default risk. (I) Default free: Issuer default risk is not included in the pricing model.
(II) In credit portfolio: The issuer is viewed as one reference entity of the credit portfolio. The default correlations between the issuer and the reference entities
are fixed to r2, which is always positive. (III) The proposed model: The default correlation between the issuer and the reference entities is rXZ, which may be
positive or negative
(I) (II) (III) The proposed model
rXZ
r Default free In credit portfolio 0.9 0.6 0.3 0 0.3 0.6 0.9
0.9 3.5790 % 3.7167 % – – – – – – 3.8150 %
0.8 3.0115 % 3.3632 % – – – – 5.1041 % 3.4122 % 3.6937 %
0.7 2.6429 % 3.1441 % – – – 5.2536 % 3.3862 % 3.1083 % 3.6545 %
0.6 2.3993 % 2.9853 % – – – 3.4885 % 3.0143 % 2.9859 % 3.6223 %
0.5 2.2319 % 2.8735 % – – 3.7706 % 3.0416 % 2.8644 % 2.9228 % 3.6225 %
0.4 2.1107 % 2.7835 % – 4.5560 % 3.2013 % 2.8419 % 2.7752 % 2.8858 % 3.6088 %
0.3 2.0280 % 2.7177 % – 3.6782 % 2.9511 % 2.7344 % 2.7243 % 2.8614 % 3.5933 %
0.2 1.9809 % 2.6597 % 5.3797 % 3.3040 % 2.8003 % 2.6635 % 2.6743 % 2.8273 % 3.5886 %
0.1 1.9499 % 2.6251 % 4.5085 % 3.1384 % 2.7242 % 2.6247 % 2.6480 % 2.8155 % 3.5889 %
0 1.9359 % 2.6020 % 4.3246 % 3.0819 % 2.6916 % 2.6020 % 2.6360 % 2.8013 % 3.5745 %
Factor Copula for Defaultable Basket Credit Derivatives
0.1 1.9470 % 2.6101 % 4.5012 % 3.1331 % 2.7072 % 2.6095 % 2.6356 % 2.7996 % 3.5754 %
0.2 1.9719 % 2.6320 % 5.3262 % 3.2787 % 2.7677 % 2.6357 % 2.6482 % 2.8099 % 3.5761 %
0.3 2.0171 % 2.6668 % – 3.6289 % 2.8962 % 2.6780 % 2.6788 % 2.8205 % 3.5859 %
0.4 2.0883 % 2.7214 % – 4.4910 % 3.1651 % 2.7836 % 2.7186 % 2.8426 % 3.5879 %
0.5 2.2105 % 2.8231 % – – 3.7405 % 2.9960 % 2.8098 % 2.8894 % 3.6007 %
0.6 2.3793 % 2.9471 % – – – 3.4615 % 2.9825 % 2.9468 % 3.6257 %
0.7 2.6279 % 3.1099 % – – – 5.2006 % 3.3573 % 3.0784 % 3.6345 %
0.8 2.9783 % 3.3144 % – – – – 5.0265 % 3.3628 % 3.6818 %
0.9 3.5583 % 3.6858 % – – – – – – 3.7992 %
649
650 P.-C. Wu et al.
12% 0.5
11% 0
10% −0.5
−0.9
9%
ND
8%
7%
−1 −0.5 0 0.5 1
r
Fig. 23.2 First-to-default BCLN coupon rates under various default correlations between the
common factor and reference entities/issuer (r)
rxz
6%
0.9
Coupon Rate
0.5
5% 0
−0.5
4% −0.9
ND
3%
−1 −0.5 0 0.5 1
r
Fig. 23.3 Second-to-default BCLN coupon rates under various default correlations between the
common factor and reference entities/issuer (r)
23.3 and 23.4, we find that when the correlation between the issuer and the reference
entities approaches a strongly positive correlation (rXZ ¼ 0.9), the BCLN coupon rate
curve becomes flatter and less sensitive to the common factor.
23.5 Conclusion
This article applies a factor copula approach for evaluating basket credit derivatives
with issuer default risk. The proposed model considers the different effects of
23 Factor Copula for Defaultable Basket Credit Derivatives 651
5% rxz
0.9
Coupon Rate
4% 0.5
0
3%
−0.5
−0.9
2%
ND
1%
−1 −0.5 0 0.5 1
r
Fig. 23.4 Third-to-default BCLN coupon rates under various default correlations between the
common factor and reference entities/issuer (r)
the default correlation between the issuer and the reference entities.
A numerical example of the proposed model on BCLN is demon-
strated and discussed in the article. The example shows that viewing the issuer
default as a new reference entity cannot reflect the effect of issuer default risk
thoroughly. The different default correlation between the issuer and the reference
entities affects the coupon rate greatly and must be taken into account in credit
derivative pricing.
Let Si(t) ¼ P(ti > t) and Fi(t) ¼ P(ti t) be the marginal survival and marginal
default distributions, respectively. Let Y be the common factor and f its density
function. Assume the default times are conditionally independent, given the
ijY ijY
common factor Y. qt ¼ P(ti > t | Y) and pt ¼ P(ti t | Y) are the conditional
survival and conditional default distributions, respectively. According to the law
of iterated expectations, the joint survival and default distribution functions are
as follows:
652 P.-C. Wu et al.
Sðt1 ; t2 ; . . . ; tn Þ
¼ Pðti > t1 , t2 > t2 , . . . , tn > tn Þ
ðY n (23.18)
ijy
¼ qt f ðyÞdy
i¼1
Fðt1 ; t2 ; . . . ; tn Þ
¼ Pðti t1 , t2 t2 , . . . , tn tn Þ
ðY n (23.19)
ijy
¼ pt f ðyÞdy
i¼1
where Y and eXi are two independent standard Gaussian random variables.
We can get the default time ti ¼ F1
i (f(Xi)), where f() is the cumulative density
function of a standard Gaussian variable. Then the conditional distribution of ti,
given the common factor Y, is
0 1
1
ijY Bf ðFi ðtÞÞ rXi Y Y C
pt ¼ f @ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi A, (23.21)
1 r2Xi Y
where f(y) is the standard Gaussian density. The copula of default times is
a Gaussian copula.
The law of iterated expectations states that E(Y) ¼ E(E(Y|X)). The proof is as follows:
23 Factor Copula for Defaultable Basket Credit Derivatives 653
ð
EðEðY jXÞÞ ¼ EðY jXÞf X ðxÞ dx
ð ð
¼ y f XjY ðyjxÞ dy f X ðxÞ dx
ðð
¼ y f X, Y ðx; yÞ dxdy (23.23)
ð ð
¼ y f X, Y ðx; yÞ dx dy
ð
¼ y f ðyÞ dy ¼ EðY Þ
Cholesky Decomposition
A ¼ LLT (23.24)
where L is a lower triangular matrix with strictly positive diagonal entries and LT is
the transpose of L. This format is the Cholesky decomposition and it is unique. The
entries of L are as follows:
vffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
u
u Xj1
Lj, j ¼ tAj, j L2j, k (23.25)
k¼1
!
1 X
j1
Li, j ¼ Ai , j Li, k Lj, k , for i>j (23.26)
Lj , j k¼1
2 3 2 3
X1 Z1
6 X2 7 6 Z2 7
X¼6 7 6 7
4 ⋮ 5, Z ¼ 4 ⋮ 5:
Xn Zn
Let X1, X2, , Xn be an iid sequence of random variables for which E[X] < 1.
Then X1 þX2 þþX
P
n
n
! E½X as n ! 1. The WLLN implies that the Monte
Carlo method converges to E[X].
Let X1, X2, , Xn be an iid sequence of random variables for which E[X] < 1.
Then X1 þX2nþþXn ! E½X as n ! 1. The SLLN implies that the Monte
a:s:
Usually, the computer has a uniform random number generator, which can generate
a sequence of iid uniform random numbers U1, U2, on [0,1]. Then, how do we
get the nonuniform random numbers? Usually, these can be achieved by inversion.
Given a distribution function F(∙), there is a one-to-one mapping from U(∙) to F(∙).
Since F(∙) is nondecreasing, F-1(x) ¼ inf{y : F(y) x}. If F is continuous and
strictly increasing, the inverse of the function is F(F1(x)) ¼ F1(F(x)) ¼ x. Thus,
the nonuniform random numbers with distribution F can be obtained by x ¼ F1(u),
where u is a uniform random number on [0,1].
23 Factor Copula for Defaultable Basket Credit Derivatives 655
References
Andersen, L., Sidenius, J., & Basu, S. (2003). All your hedges in one basket. Risk, 16(11), 67–72.
Bastide, D., Benhamou, E., & Ciuca, M. (2007). A comparative analysis of basket default swaps
pricing using the Stein method. Working paper, pricing partners.
Brigo, D., & Chourdakis, K. (2009). Counterparty risk for credit default swaps: Impact of spread
volatility and default correlation. International Journal of Theoretical and Applied Finance,
12(7), 1007–1026.
Chen, Z., & Glasserman, P. (2008). Fast pricing of basket default swaps. Operations Research,
56(2), 286–303.
Chiang, M. H., Yueh, M. L., & Hsieh, M. H. (2007). An efficient algorithm for basket default swap
valuation. Journal of Derivatives, 15(2), 1–22.
Hull, J., & White, A. (2004). Valuation of a CDO and an nth to default CDS without Monte Carlo
simulation. Journal of Derivatives, 12(2), 8–23.
Jarrow, R., & Turnbull, S. (1995). Pricing derivatives on financial securities subject to credit risk.
Journal of Finance, 1, 53–85.
Joshi, M. S., & Kainth, D. (2004). Rapid and accurate development of prices and Greeks for Nth to
default credit swaps in the Li model. Quantitative Finance, 4, 266–275.
Laurent, J. P., & Gregory, J. (2005). Basket default swaps, CDOs and factor copulas. Journal of
Risk, 7(4), 103–122.
Li, D. X. (1999). The valuation of basket credit derivatives. CreditMetrics Monitor, 34–50.
Li, D. X. (2000). On default correlation: A copula function approach. Journal of Fixed Income, 9,
43–54.
Mashal, R., & Naldi, M. (2003). Pricing multi-name default swaps with counterparty risk.
Quantitative Credit Research, 1–16.
Merton, R. C. (1974). On the pricing of corporate debt: The risk structure of interest rates. Journal
of Finance, 29, 449–470.
Sklar, A. (1959). Fonctions de repartition a n dimensions et leurs marges. Public Institute Statistic
University of Paris, 8, 229–231.
Stein, C. (1972). A bound for the error in the normal approximation to the distribution of a sum of
dependent random variables. In Proceedings of the sixth Berkeley symposium on mathematical
statistics and probability (pp. 583–602). Berkeley: University of California Press.
Walker, M. B. (2006). Credit default swaps with counterparty risk: A calibrated Markov model.
Journal of Credit Risk, 2(1), 31–49.
Wu, P. C. (2010). Applying a factor copula to value basket credit linked notes with issuer default
risk. Finance Research Letters, 7(3), 178–183.
Wu, P. C., Kao, L. J., & Lee, C. W. (2011). How issuer default risk affects basket credit linked note
coupon rate. International Journal of Information and Management Sciences, 22(1), 59–71.
Panel Data Analysis and Bootstrapping:
Application to China Mutual Funds 24
Win Lin Chou, Shou Zhong Ng, and Yating Yang
Contents
24.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 658
24.2 Wild-Cluster Bootstrap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 660
24.3 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 661
24.3.1 Data and Definitions of the Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 661
24.3.2 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 661
24.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 665
Appendix 1: Wild-Cluster Bootstrap Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 666
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 667
Abstract
Thompson (Journal of Financial Economics 99, 1–10, 2011) argues that double
clustering the standard errors of parameter estimators matters the most when the
number of firms and time periods are not too different. Using panel data of
similar number in firms and time periods on China’s mutual funds, we estimate
double- and single-clustered standard errors by wild-cluster bootstrap procedure.
To obtain the wild bootstrap samples in each cluster, we reuse the regressors
(X) but modify the residuals by transforming the OLS residuals with weights
which follow the popular two-point distribution suggested by Mammen (Annals
of Statistics 21, 255–285, 1993) and others. We then compare them with other
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 657
DOI 10.1007/978-1-4614-7750-1_24,
# Springer Science+Business Media New York 2015
658 W.L. Chou et al.
Keywords
Asset-pricing regression • Bootstrapped critical values • Cluster standard
errors • Double clustering • Firm and time effects • Finance panel data • Single
clustering • Wild-cluster bootstrap
24.1 Introduction
Researchers using finance panel data have increasingly realized the need to account
for the residual correlation across both firms and/or time in estimating standard
errors of regression parameter estimates. Ignoring such clustering can result in
biased OLS standard errors. Two forms of residual dependence that are common in
finance applications are time-series dependence and cross-sectional dependence.
The former is called a firm effect, whereas the latter a time effect. The usual
solution to account for the residual dependence is to compute clustered standard
errors. The notable examples are Petersen (2009) and Thompson (2011).
Using the Monte Carlo-simulated panel data, Petersen (2009) compares the
performance of many different standard error estimation methods surveyed in the
literature. These methods include White’s heteroskedasticity-robust standard
errors, single clustering (by firm or by time), and double clustering (by both firm
and time). His findings suggest that the performance of different methods depends
on the forms of residual dependence. For example, in the presence of a firm effect,
the clustered standard errors are unbiased and can produce correctly sized
confidence intervals while those estimated by OLS, White, or Fama-MacBeth
method are biased.
Much of the analysis in Petersen (2009) is based on the simulated panel data set
whose data structure is certain. With the simulated panel data set, it is easier to
choose among the estimation methods. This paper chooses an alternative method,
namely, bootstrapping, to investigate the performance of standard errors estimated
by White’s OLS and single- and double-clustering methods with actually observed
data. The use of the bootstrap method is motivated by Kayhan and Titman (2007)
who show that bootstrapped standard errors are robust to heteroskedasticity, and
serial correlation problems in panel finance data applications. Moreover, despite
the wide use of the bootstrap in statistical and econometric applications, the
survey finding of Petersen (2009) found that the bootstrap applications are
relatively scarce in the finance literature. Hence, it may be of some interest to
investigate the bootstrapping application to a set of panel finance data on
Chinese mutual funds.
24 Panel Data Analysis and Bootstrapping: Application to China Mutual Funds 659
Table 24.1 Summary statistics for variables used in China’s mutual fund regressions
(Sept 2002–Aug 2006)
Std. Normality test stat.
Variable Mean deviation Skewness Kurtosis W-sq. A-sq.
Mutual funds 0.0006 0.0580 0.1313 0.1489 0.1809*** 0.2062***
returns
Market 0.0025 0.0465 0.1554 0.7801 2.4993*** 18.3650***
excess
returns
Notes: Sample size N ¼ 2,592. Individual mutual funds return is the dependent variable, while the
market excess return is the independent variable. W-sq or W2 ¼ Cramer-von Mises test statistic, and
A-sq or A2 ¼ Anderson-Darling test statistic. Both test statistics are empirical distribution function (EDF)
statistics. The computing formulas for W2 and A2 statistics are available in Stephens (1974), p. 731
***
denotes statistical significance at the 1 % level
The bootstrap method is applied to a panel data set on the monthly returns for
54 Chinese mutual funds over the period of September 2002–August 2006. The data
set is applied to a set of asset pricing regressions. Table 24.1 contains summary
statistics such as sample skewness, sample excess kurtosis, and two test statistics for
normality for the variables used in the asset pricing regressions. They suggest that
normality does not characterize the variables. Additionally, since the time-series
and/or cross-sectional independence assumption is most likely to be violated in panel
data sets, ignoring these dependence could result in biased estimates of the standard
errors. As evidenced in Kayhan and Titman (2007), bootstrapping is a possible
alternative to handle this dependence issue.
In this paper, we are particularly interested in the performance of the bootstrapped
double-clustered standard error estimates, because Thompson (2011) has argued that
double clustering matters most when the number of firms and time periods are not too
different. Given the panel data set we have collected which consists of 54 China mutual
fund returns for 48 months with data exhibiting firm and time effects, double clustering
is likely to show a significant difference. Our findings show that the bootstrapped
standard errors from double clustering leads to more significant test results. We also
demonstrate the importance of using bootstrapped critical values in hypothesis testing.
A number of bootstrap procedures are available in the literature. The
bootstrap procedure we consider in this paper is the wild-cluster bootstrap procedure,
which is an extended version of the wild bootstrap proposed by Cameron et al. (2008)
in a cluster setting. This procedure has been shown by Cameron et al. (2008) to perform
very well in practice, despite the fact that the pairs cluster bootstrap works well in
principle. In this paper, our comparison of the finite-sample size of the bootstrapped
t-statistics resulting from the pairs cluster bootstrap and wild-cluster bootstrap also
indicates that the wild-cluster bootstrap performs better.
The rest of the paper is organized as follows. Section 24.2 presents the wild-cluster
bootstrap procedure. Section 24.3 discusses the empirical results and the last section
gives conclusions.
660 W.L. Chou et al.
1
For example, in Cameron et al. (2008), ag takes the value +1 with probability 0.5, or the value 1
with probability 1–0.5.
24 Panel Data Analysis and Bootstrapping: Application to China Mutual Funds 661
The Data. The sample consists of the returns on 54 publicly traded closed-end
mutual funds that are gathered for 48 months from September 2002 to August 2006,
a total of 2,592 observations. The mutual fund data set is purchased from the GTA
Information Technology Company, Shenzhen, China. For simplicity, we divide the
mutual funds investment objectives into equity growth and nongrowth funds. Our
sample consists of 37 (68.5 %) growth funds and 17 (31.5 %) nongrowth funds.
Although the first closed-end fund in China was sold to the public in April 19982,
complete data for all 54 mutual funds are collected from September 2002.
The summary statistics for the variables used in China’s mutual fund return
regressions are displayed in Table 24.1, and the test statistics for normality for these
variables suggest that the variables are non-normal.
Definitions of the variables. The following are the definitions of the variables
used in the estimation procedures: Ri, t ¼ the return of a mutual fund i in excess of
the risk-free rate in month t. Savings deposit rate of the People’s Bank of China3 is
used as the proxy for the risk-free rate. Rm, t ¼ is the market return in excess of the
risk-free rate in month t and is calculated4 as follows:
where R1 is the monthly return on Shanghai Stock Exchange index, R2 the monthly
return on Shenzhen Stock Exchange index, and 0.06 % is the monthly return on
savings deposits.
24.3.2 Results
Table 24.2 presents the results from a set of asset pricing regressions of China
mutual fund returns on its market returns. The firm and time effect in OLS residuals
and data can be seen graphically in Fig. 24.1. Figure 24.1, Panel A, shows the
within-firm autocorrelations in OLS residuals and independent variable, respec-
tively, for lags 1–12. Panel B of Fig. 24.1 displays the within-month autocorrela-
tions for residuals for lags 1–12. As the independent variable is a monthly series
without cross-sectional units, we cannot calculate its within-month autocorrela-
tions, thus no within-month plot is available for the independent variable.
Figure 24.1 suggests that the residuals exhibit both firm and time effects, whereas
2
Chen and Lin (2006), p. 384
3
Data are taken from the website of the People’s Bank of China: http://www.pbc.gov.cn/publish/
zhengcehuobisi/627/index.html.
4
The calculation follows that of Shen and Huang (2001), p. 24.
662 W.L. Chou et al.
independent variable shows firm effects. Given Thompson’s (2011) argument that
double clustering matters most when the number of firms and time periods are not
too different, our data set which has the number of mutual funds (54) similar to
the number of months (48) is expected to imply that double clustering is important
in our analysis.
In Table 24.2, the second column presents the OLS parameter estimates, whereas
the remaining columns report their corresponding t-statistics by dividing each param-
eter estimate by its corresponding standard error. These t-statistics indicate all beta
coefficients are statistically significant at the 1 % level, whereas the intercept is only
significant in one case (under column II) when the standard error computed by single
clustering by firm is used. More importantly, the t-statistics in Table 24.2 enable us to
compare the clustered standard errors constructed from double and single clustering
with the OLS White estimate. Notice that the t-statistic for beta coefficient
obtained from double clustering (SÊboth) is 9.10 (column IV) which is much smaller
than 46.9 calculated using the White method (column I), indicating the presence of
firm and time effects. It also means that the double-clustering standard errors are
much larger. A comparison of t-statistics in columns III and IV implies the SÊboth of
the beta coefficient (9.10) is similar to the standard error clustered by time which is
9.01. This means that the firm effects do not matter much. The comparison reveals
that OLS White standard errors are underestimated when residuals exhibit both firm
and time effects.
24 Panel Data Analysis and Bootstrapping: Application to China Mutual Funds 663
0.1
−0.1 1 2 3 4 5 6 7 8 9 10 11 12
−0.2
−0.3
Lag
0.05
0
1 2 3 4 5 6 7 8 9 10 11 12
−0.05
−0.1
−0.15
−0.2
−0.25
Lag
Fig. 24.1 The autocorrelations of residuals and the independent variable are plotted for 1–12
lags. The solid lines in Panel A and B show, respectively, within-firm and within-month autocor-
relations in residuals, whereas the dashed line in Panel A shows the within-firm autocorrelations in
the independent variable
Turning to the results obtained by the wild-cluster bootstrapping, the story changes.
The bootstrapped t-statistics of the beta coefficient estimates displayed in columns
V–VIII of Table 24.2 differ quite significantly from those in columns I–IV. The
t-statistic is 47.2 when the bootstrapped White standard error is used and 66.8 if the
bootstrapped standard error clustered by the firm is used. This means the firm effect is
significant in the data. By a similar comparison, the bootstrapped t-statistic is 73.7 when
the double-clustered standard error is used, meaning both the time and firm effects are
strong in the residuals. The bootstrapped t-statistic is 50.9 when the bootstrapped
standard error clustered by time is used implying the time effect exists in the data.
These comparisons suggest both the firm and time effects matter in the computation of
the bootstrapped standard errors using double as well as single clustering. Their
implication is that we might follow what Kayhan and Titman (2007) have done in
their study to simply compute the bootstrapped standard errors with our panel data set.
664 W.L. Chou et al.
reflected by the close to 5 % size values. However, for the tests based on the
standard asymptotic t-test critical values, all tests suffer from serious size distortion.
For example, those based on the OLS White and clustered by time are undersized,
while others oversized (see Table 24.5).
24.4 Conclusion
Table 24.6 Rejection rates using pairs cluster bootstrapped critical values
OLS White Clustered by firm Clustered by time Clustered by firm and time
Rm,t 0.043 0.040 0.040 0.040
Intercept 0.062 0.048 0.059 0.060
Note: The number of replications is 1,000
The following steps are used to obtain the wild-clustered bootstrapped standard
errors and critical values:5
(i) Define firm effects, and time effects. The asset pricing model using individual
variables Ri,t and Rm,t is specified as
where the variables Ri,t and Rm,t are, respectively, the return of mutual fund
i and the market return in excess of risk-free rate in month t. b0 and b1
are unknown parameters. The construction of the variables is detailed in
Sect. 24.3.1. eit is the error term. It may be heteroskedastic but is assumed to
be independent of the explanatory variable E(eit | Rm,t) ¼ 0.
Following Thompson (2011), we make the following assumptions on the
correlations between errors, eit:
(a) Firm effects: The errors may be correlated across time for a particular firm,
that is, E(eit, eik | Rm,t,Rm,k) 6¼ 0 for all t 6¼ k.
(b) Time effects: The errors may be correlated across firms within the same time
period, that is,
E eit , ejt Rm, t 6¼ 0 for all i 6¼ j:
Rig ¼ b0 þ Rm b1 þ eig , i ¼ 1, . . . , Ng g ¼ 1, . . . , G,
(24.3)
Rg ¼ db0 þ Rmg b1 þ eg , g ¼ 1, . . . , G,
where Rig, Rm, and eig are scalars; Rg, Rmg, and eg are Ng 1 vectors; and
d is Ng 1 vector with all elements equal to 1.
(ii) Fit data to model. We fit model (Eq. 24.3) to the observed data using OLS
and obtain the parameter estimates b ^ and b ^ together with the OLS residuals
0 1
êg, g ¼ 1,. . ., G
5
See also Cameron et al. (2008) for details
24 Panel Data Analysis and Bootstrapping: Application to China Mutual Funds 667
Rg ¼ db ^ þ e :
^ þ Rmg b (24.4)
0 1 g
(iv) With each pseudo sample generated in step (iii), we estimate the parameters b ^
0
^ by OLS, White standard errors (SÊwhite)which are OLS standard errors
and b 1
robust to heteroskedasticity, as well as standard errors clustered by firm (SÊfirm), by
time (SÊtime), and by both firm and time (SÊboth). The simulations are performed
using GAUSS 9.0. The standard error formulas can be found in Thompson (2011).
(v) Construct bootstrapped test statistics by taking ratios of b ^ ði = 0, 1Þ obtained
i
by OLS to its corresponding SÊwhite, SÊfirm, SÊtime, and SÊboth obtained in
step (iv). More specifically, the bootstrapped test statistics are expressed as
follows:
^ b
b ^
wi ¼ i i , i ¼ 0, 1, (24.5)
^ b
SE ^
i
where SE ^ b^ can either be SÊwhite, SÊfirm, SÊtime, or SÊboth.
i
(vi) Obtain the empirical distribution of the individual test statistics by sorting the
1,000 test statistics computed in step (v) in an ascending order. Bootstrapped
critical values are then obtained from this empirical distribution at the following
quantiles: 0.5 %, 2.5 %, 5 %, 95 %, 97.5 %, and 99.5 %, respectively.
References
Brownstone, D., & Valletta, R. (2001). The bootstrap and multiple imputations: Harnessing
increased computing power for improved statistical tests. Journal of Economic Perspectives,
15, 129–141.
Cameron, A. C., Gelbach, J. B., & Miller, D. L. (2008). Bootstrap-based improvements for
inference with clustered errors. Review of Economics and Statistics, 90, 414–427.
Chen, Z., & Lin, R. (2006). Mutual fund performance evaluation using data envelopment analysis
with new risk measures. OR Spectrum, 28, 375–398.
Davidson, R., & Flachaire, E. (2008). The wild bootstrap, tamed at last. Journal of Econometrics,
146, 162–169.
Kayhan, A., & Titman, S. (2007). Firms’ histories and their capital structures. Journal of Financial
Economics, 83, 1–32.
668 W.L. Chou et al.
Liu, R. Y. (1988). Bootstrap procedures under some non-IID models. Annals of Statistics, 16,
1696–1708.
Mammen, E. (1993). Bootstrap and wild bootstrap for high dimensional linear models. Annals of
Statistics, 21, 255–285.
Petersen, M. A. (2009). Estimating standard errors in finance panel data sets: Comparing
approaches. The Review of Financial Studies, 22, 435–480.
Shen, W. T., & Huang, S. L. (2001). An empirical analysis of China’s mutual funds: Performance
and evaluation. Economic Research, 9, 22–30 (in Chinese).
Stephens, M. A. (1974). EDF statistics for goodness of fit and some comparisons. Journal of the
American Statistical Association, 69, 730–737.
Thompson, S. B. (2011). Simple formulas for standard errors that cluster by both firm and time.
Journal of Financial Economics, 99, 1–10.
Wu, C. F. G. (1986). Jackknife, bootstrap and other resampling methods in regression analysis.
Annals of Statistics, 14, 1261–1295.
Market Segmentation and Pricing of
Closed-End Country Funds: 25
An Empirical Analysis
Dilip K. Patro
Contents
25.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 670
25.2 Theoretical Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 673
25.2.1 Pricing of Country Funds in the Context of International CAPM . . . . . . . . . . . . 673
25.2.2 Cross-Sectional Variations in Country Fund Premiums . . . . . . . . . . . . . . . . . . . . . . 674
25.2.3 Conditional Expected Returns and Pricing of Country Funds . . . . . . . . . . . . . . . . 676
25.2.4 Conditional Expected Returns and Time-Varying Country
Fund Premiums . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 677
25.3 Data and Descriptive Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 677
25.4 Empirical Results: Pricing of Country Funds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 684
25.4.1 Unconditional Risk Exposures of Country Funds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 684
25.4.2 Pricing of Country Funds in the Context of International CAPM . . . . . . . . . . . . 689
25.4.3 Cross-Sectional Variations in Country Fund Premiums . . . . . . . . . . . . . . . . . . . . . . 690
25.4.4 Predictability of Closed-End Country Fund Returns . . . . . . . . . . . . . . . . . . . . . . . . . 693
25.4.5 Conditional Expected Returns and Pricing of Country Funds . . . . . . . . . . . . . . . . 694
25.4.6 Conditional Expected Returns and Time-Varying Country
Fund Premiums . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 700
25.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 701
Appendix 1: Generalized Method of Moments (GMM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 701
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 703
Abstract
This paper finds that for closed-end country funds, the international CAPM can
be rejected for the underlying securities (NAVs) but not for the share prices. This
finding indicates that country fund share prices are determined globally where as
the NAVs reflect both global and local prices of risk. Cross-sectional variations
in the discounts or premiums for country funds are explained by the differences
The views expressed in this paper are strictly that of the author and not of the OCC or the US
department of Treasury.
D.K. Patro
RAD, Office of the Comptroller of the Currency, Washington, DC, USA
e-mail: dilip.patro@occ.treas.gov
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 669
DOI 10.1007/978-1-4614-7750-1_25,
# Springer Science+Business Media New York 2015
670 D.K. Patro
in the risk exposures of the share prices and the NAVs. Finally, this paper shows
that the share price and NAV returns exhibit predictable variation and country
fund premiums vary over time due to time-varying risk premiums. The paper
employs generalized method of moments (GMM) to estimate stochastic
discount factors and examines if the price of risk of closed-end country fund
shares and NAVs is identical. GMM is an econometric method that was a
generalization of the method of moments developed by Hansen (Econometrica
50, 1029–1054, 1982). Essentially GMM finds the values of the parameters so
that the sample moment conditions are satisfied as closely as possible.
Keywords
Capital markets • Country funds • CAPM • Closed-end funds • Market
segmentation • GMM • Net asset value • Stochastic discount factors • Time-
varying risk • International asset pricing
25.1 Introduction
1
Hence forth both premiums and discounts are referred to as premiums. Therefore, a discount is
treated as a negative premium.
672 D.K. Patro
The sample consists of 40 closed-end country equity funds. Twenty of the funds
are from developed markets and 20 funds are from emerging markets.2 The main
empirical findings of the paper are as follows. For country fund share prices, the
hypothesis that the unconditional international CAPM is a valid model cannot be
rejected. However, for the NAVs the international CAPM can be rejected.
This finding suggests that country fund share prices and NAVs are not priced
identically. The share prices reflect the global price of risk only, but the NAVs
reflect both the global and the local prices of risk. It is shown that the differences in
risk exposure to the world market index of share prices and NAVs can explain up to
18.7 % of the cross-sectional variations in premiums for developed market funds,
but only 1.9 % of the variation for emerging market funds.
When conditioning on information is allowed, the international CAPM explains
country fund share returns and NAV returns for both developed and emerging
markets. However, the hypothesis that the price of risk is identical between
closed-end fund shares and NAVs can be rejected for alternate stochastic discount
factors for majority of the markets. This finding is consistent with market segmen-
tation. Therefore, differential pricing of the fund shares and the underlying portfolio
causes expected returns to be different and explains the existence of premiums and
their variation over time. Finally, it is shown that the country fund premiums vary
over time due to differential conditional risk exposures of the share prices
and NAVs.
The principal contributions of this paper are as follows. Existence of premiums
on closed-end funds has been a long-standing anomaly. In the domestic setting,
negative premiums have been the norm, which has been attributed to taxes,
management fees, illiquid stocks, or irrational factors (e.g., noise trading).
Although these factors may also be important in explaining country fund premiums,
unlike domestic closed-end funds, country fund share prices and NAVs are deter-
mined in two different market segments. This paper provides a rational explanation
for the premiums on country funds based on differential risk exposures of the share
prices and NAVs and market segmentation. Unlike the noise trading hypothesis
which assumes the presence of “sentiment” or an additional factor for the share
prices but not the NAVs, this paper shows that the same factor may be priced for the
share prices and the NAVs, but priced differently. The differential risk exposures
are shown to be among the significant determinants of cross-sectional variations in
the premiums. Further, this paper examines the role of time variations in expected
returns for country fund returns and attributes the time-varying country fund
premiums to time-varying country fund returns.
The paper is organized as follows. Section 25.2 presents the theoretical motiva-
tion and the hypotheses. Section 25.3 presents the data and the descriptive statistics.
Section 25.4 provides the empirical results for pricing of country funds. Concluding
remarks are presented in the last section.
2
In 1993, the World Bank defined an emerging market as a stock market in a developing country
with a GNP per capita of $8,625 or less. This is the definition of an emerging market in this paper.
25 Market Segmentation and Pricing of Closed-End Country Funds 673
This section first focuses on the theoretical motivation for the empirical testing of the
unconditional mean-variance efficiency of the world market index in the context of
country funds. The following subsections present tests for cross-sectional variations in
country fund premiums and the methodology for pricing country funds using stochas-
tic discount factors. Finally, this section outlines the methodology for explaining the
time variations in country fund premiums attributable to time-varying risk premium.
In a global pricing environment, the world market portfolio surrogates the market
portfolio. If purchasing power parity is not violated and there are no barriers to
investment, the world market portfolio is mean-variance efficient (see, e.g., Solnik
1996; Stulz 1995) and expected returns are determined by the international CAPM.
The international CAPM implies that the expected return on an asset is proportional
to the expected return on the world market portfolio:
where, for any asset i, E[ri] is the expected excess return on the asset and E[rw] is the
expected excess return on the world market portfolio. The excess returns are
computed in excess of the return on a risk-free asset. Following a standard practice
(see Roll 1977), mean-variance efficiency of a benchmark portfolio can be
ascertained by estimating a regression of the form
where ri,t is the excess return on a test asset and rw,t is the return on a benchmark
portfolio.
The GMM-based methodology as outlined in Mackinlay and Richardson (1991) is
employed to test mean-variance efficiency of the world market index in the context of
country funds, using just a constant and the excess return on the world market index as
the instruments.3 The hypothesis that ai ¼ 0, for all i ¼ 1, . . ., N, where N is the
number of assets, is tested. This is the restriction implied by the international CAPM.
The joint hypothesis is tested by a Wald test for country fund share prices and NAVs
returns. Since country fund’s share prices are determined with complete access to
their markets of origin, their expected returns are expected to be determined via the
3
Although the evidence in favor of the domestic CAPM is ambiguous, many studies such as
Cumby and Glen (1990) and Chang et al. (1995) do not reject the mean-variance efficiency of the
world market index. Interestingly, although Cumby and Glen (1990) do not reject mean-variance
efficiency of the world market index, they reject mean-variance efficiency of the US market index.
674 D.K. Patro
international CAPM (see, e.g., Errunza and Losq 1985). However, if the country from
which the fund originates has restrictions on foreign equity investments, the expected
returns of the NAVs will be determined via their covariances with both the world
market portfolio and the part of the corresponding local market portfolio which is not
spanned by the world market portfolio (see Errunza et al. 1998).
for the N assets, the residuals from the above equation can be stacked into
a vector et+1. The model implies that E[ei,t+1jZt] ¼ 0, for 8 i and 8 t. Therefore, E
[et+1 Zt] ¼ 0 for 8 t, where Zt is a set of predetermined instruments. GMM
estimation is based on minimizing the quadratic form f0 Of, where the sample
counterpart of E[et+1 Zt] is given by f ¼ {1/T[∑E[et+1 Zt]} and O is an optimal
weighting matrix in the sense of Hansen (1982).
If the mean-variance efficiency of the world market portfolio is not rejected, then
expected returns of the test assets are proportional to their “betas” with respect to
the world market portfolio. Therefore,
where E(rp) and E(rn) are the expected excess returns on the share prices and NAVs,
respectively, and lw is the risk premium on the world market index. Stulz and
Wasserfallen (1995) demonstrate that the logarithm of two share prices can be
written as a linear function of the differences in their expected returns.4 Therefore,
Prem ¼ logðP=NAVÞ ¼ Eðrn Þ E rp (25.5)
where Prem is the premium on a fund calculated as the logarithm of the price-NAV
ratio. Combining Eqs. 25.4 and 25.5 leads to the testable implication
Prem ¼ lw bn, w bp, w (25.6)
The above equation assumes that the international CAPM is valid and world
capital markets are integrated. In reality, the world markets may be segmented.
Therefore, the effect of market segmentation is captured by introducing additional
4
Stulz and Wasserfallen (1995) use the log-linear approximation of Campbell and Ammer to write
the logarithm of the price of a stock as ln(P) ¼ ESZi[(1 Z)dt+j+1 rt+j+1] + y, where Z is
a log-linear approximation parameter and rt+j+1 is the return from t+j to t+j+1. Assuming that Z is
the same for prices and NAVs, and the relation holds period by period, the premium can be written
as above.
25 Market Segmentation and Pricing of Closed-End Country Funds 675
variables based on prior research of Errunza et al. (1998) who document that measures
of access, spanning, and substitutability of the prices and NAVs have explanatory
power for the cross-sectional variation of premiums. The spanning measure (SPN) is
the conditional variance of the NAV return of a fund, unspanned by the US market
return and the fund’s share price returns, with specification as follows:
rn, t ¼ ai þ bi rUS, t þ bi rp, t þ ei, t (25.7)
where ei,t has a GARCH (1, 1) specification. The conditional volatility of ei,t is the
measure of spanning. The measure of substitution (SUB) is the ratio of conditional
volatilities of the share price and NAV returns of a fund not spanned by the US
market. The specifications for the expected return equations are
where the error terms en,t and ep,t have GARCH (1, 1) specifications.5 The ratio of
the conditional volatilities of the residuals in Eqs. 25.8 and 25.9 is used as the
measure of substitutability of the share prices and NAVs.
Since it is difficult to systematically classify countries in terms of degree of access,
a dummy variable is used to differentiate developed and emerging markets. The
dummy variable takes value of one for developed markets. Another measure of
access, which is the total purchase of securities from a country by US residents as
a proportion of total global purchase, is also used as a measure of access to
a particular market. It is expected that the premiums are lower for countries with
easy access to their capital markets. Therefore, the coefficient on the access
variable is expected to be negative. The measure of spanning is interpreted as the
degree of ease with which investors could obtain substitute assets for the NAVs. As
noted earlier, even if there are barriers to foreign equity ownership, availability of
substitute assets can overcome the barriers. Since the spanning variable is
the volatility of the residual as specified in Eq. 25.7, it is expected to be positive.
The measure of imperfect substitutability of the share prices and NAVs is the ratio of
the variances of the share price and NAV returns. It is expected that the premium is
inversely related to this measure. Therefore, the extended cross-sectional equation is
where:
DIFi ¼ the difference of the betas on the world market index for the NAVs and the
share prices.
SPNi ¼ the conditional residual volatility from Eq. 25.7 used as measures of spanning.
5
Note that unlike here, Errunza et al. (1998) use returns on industry portfolios to proxy for the US
market.
676 D.K. Patro
SUBi ¼ the imperfect substitutability of the share prices and the NAVs proxied by
the ratio of conditional volatilities in Eqs. 25.8 and 25.9.
ACC ¼ a measure of access to a market proxied by a dummy variable which is
1 for developed markets [ACC(d)] or the total purchase of securities from the US
residents from that country as a fraction of global security purchases [ACC(cf)].
The empirical specification in Eq. 25.10 or its variations is used in the empirical
analysis for explaining cross-sectional variations in premiums. The higher the
difference of risk exposures, the higher is the premium. Therefore, the sign of d1
is expected to be positive. However, as the level of unspanned component is higher,
the level of premiums should be higher. Therefore the sign of d2 is expected to be
positive. Also, the higher the substitutability of the share prices and the NAVs, the
lower should be the premium. Therefore, d3 is expected to be negative. Since higher
access is associated with lower premiums, d4 is also expected to be negative.
A two-factor SDF, where the second factor is the return on a regional index
(specifically for Asia or Latin America), is also estimated. Such a model is implied
by the models of Errunza and Losq (1985).7 Specifically, the SDFs is:
Two-factor model: M ¼ l0 þ l1 Rw, tþ1 þ l2 Rh, tþ1 (25.12)
6
Ferson and Foerster (1994) show that an iterated GMM approach has superior finite sample
properties. Therefore the iterated GMM approach is used in the estimations.
7
For such conditional representation, see, for example, Ferson and Schadt (1996).
25 Market Segmentation and Pricing of Closed-End Country Funds 677
Once an SDF is estimated for a group of share prices and NAVs, the coefficients
of the estimated SDFs are compared to test the hypothesis that for a given SDF, the
share price and NAV are priced identically.8
To examine the role of time-varying risk premiums in explaining the time variation
in country fund premiums, the following procedure is used. Similar to Ferson and
Schadt (1996), a conditional international CAPM in which the betas vary over time
as a linear function of the lagged instrumental variables is used. The following
equations are estimated via GMM:
where the excess return on the world market index is scaled by a set of instrumental
variables. Using Eq. 25.6 and the above time-varying betas, the premium can be
written as
Premtþ1 ¼ g0 þ g1 bn, w ðZt Þ bp, w ðZt Þ þ eprem, tþ1 (25.15)
The empirical specification in Eq. 25.15 is used for explaining the time-varying
premiums. If the time-varying betas are significantly different and their difference
can explain the time variation of country fund premiums, the coefficient g1 is
expected to be positive and significant.
The sample includes all single-country closed-end country equity funds publicly
trading in New York as of 31 August 1995. The test sample is limited to country
funds with at least 100 weeks of weekly NAV observations.9 An overview of the
sample is presented in Table 25.1. The funds are classified as developed market or
8
See Cochrane (1996) for such specifications. Cochrane (1996) calls such models as “scaled factor
models.” See Bansal et al. (1993) for a nonlinear specification of stochastic discount factors.
9
Unlike industrial initial public offerings (IPOs), closed-end fund IPOs are “overpriced” (Weiss
1989). Peavy (1990) finds that new funds show significant negative returns in the aftermarket.
Hanley et al. (1996) argue that closed-end funds are marketed to a poorly informed public and
document the presence of flippers – who sell them in the immediate aftermarket. They also
document evidence in support of price stabilization in the first few days of trading. Therefore,
the first 6 months (24 weeks) of data for each fund is excluded in the empirical analysis.
678 D.K. Patro
emerging market funds using the World Bank’s definition of an emerging market.
The three China funds are classified as developed market funds because the majority
of their assets are in Hong Kong. The weekly (Friday closing) prices, NAVs, and
corresponding dividends for each fund are obtained from Barrons and Bloomberg. 10
The returns are adjusted for stock splits and dividends. The 7-day Eurodollar deposit
rate, provided by the Federal Reserve, is used as the risk-free benchmark.
10
For some of the funds, such as the India growth fund, the prices and net asset values are as of
Wednesday closing. This may lead to nonsynchronous prices and NAVs. However, as Bodurtha
et al. (1995) and Hardouvelis et al. (1993) show, the effects of nonsynchronous trading are not
pervasive and do not affect the analysis.
680 D.K. Patro
Table 25.2 reports the descriptive statistics for the sample of funds. The IFC
global indices are available weekly from December 1988. Therefore, for emerging
market funds which were launched before December 1988, the analysis begins in
December 1988. To enable comparison across funds, a common time frame of
January 1991–August 1995 is chosen. Since, the data for the first 6 months are not
used in the analysis, only 30 funds listed on or before July 1990 are considered.
The mean premium on the index of developed market funds is 8.31 %, whereas
the mean premium on the emerging market funds is 0.44 %. A t-test of the
difference of means indicates that the mean premium of the index of the emerging
market funds is significantly higher than the mean premiums on the index of
developed market funds at the 1 % level of significance.
Figure 25.1 plots the time series of the premiums for the index of developed and
emerging market funds. The figure clearly indicates that the premiums on emerging
market funds are usually positive and higher than the premiums on developed market
funds, highlighting the effects of market segmentation. Table 25.2 also presents the
ratio of the variance of the excess price returns over the variance of the excess NAV
returns. For 27 funds out of the 30 funds, the ratios are significantly higher than one.11
A set of predetermined instrumental variables similar to instruments used in
studies of predictability of equity returns for developed markets (Fama and French
1989; Ferson and Harvey 1993) and emerging markets (Bekaert 1995) are used in
the empirical analysis when testing conditional asset-pricing models. The instru-
ments are the dividend yield on the Standard and Poor’s 500 Index, calculated as the
last quarter’s dividend annualized and divided by the current market price (DIVY);
the spread between 90-day Eurodollar deposit rate and the yield on 90-day US
treasury bill (TED); and the premium on an index of equally weighted country
funds (FFD). Only three instruments are used in order to be parsimonious repre-
sentation of the conditional asset-pricing models.
TED is calculated using data from the Federal Reserve of Chicago and DIVY is
obtained from Barrons. FFD is constructed using the sample of 30 funds listed in
Table 25.2. Table 25.3 reports the sample characteristics of these instruments.
As Fama and French (1989) show, the dividend yields and interest rates track the
business cycle, and expected returns vary as a function of these instruments.
Table 25.4 reports the descriptive statistics for the returns on the market indices
and the correlations of returns on the market indices of various countries with the
returns on the US index. The developed market indices are from Morgan Stanley
Capital International and the emerging market indices are from the International
Finance Corporation. All the developed market indices have significantly higher
correlations with the US market index, compared to the emerging markets, which
implies that there are potential diversification benefits from investing in emerging
markets.
11
Assuming that the price returns and NAV returns are from two normal populations, the test
statistic which is the ratio of the variances has an F distribution (if the two variances are estimated
using the same sample size). The null hypothesis that the variance of price returns is greater than
the variance of NAV returns is tested using this statistic at the 5 % level of significance.
682 D.K. Patro
20
10
5
Premium (%)
0
910104
910315
910517
910726
911004
911213
920221
920417
920626
920904
921113
930108
930319
930521
930730
931008
931217
940225
940506
940715
940923
941202
950210
950421
950630
−5
−10
−15
−20
Date
This section presents the empirical results for pricing of country funds. The results
from the unconditional tests indicate that, in general, developed market closed-end
fund returns have significant risk exposures to the world market index, while
emerging market closed-end fund returns have significant risk exposures to both
the world market index and the corresponding local market index. Second, the
hypothesis of unconditional mean-variance efficiency of the world market
index cannot be rejected using the share price returns of either developed or
emerging market funds and NAV returns of some of the developed market funds.
However, the hypothesis of unconditional mean variance of the world market index
can be rejected for emerging market NAVs and some developed market NAVs.
This finding indicates that while the share prices reflect the global price of risk, the
NAVs may reflect both the global and the respective local prices of risk. Tests of
predictability using global instruments indicate that country fund share price and
NAVs exhibit significant predictable variation. When conditional asset-pricing
restrictions are examined using alternate stochastic discount factors, the results
indicate that the share prices and NAVs of the closed-end country funds are not
priced identically for some developed and all the Asian market funds. This finding
is consistent with market segmentation. Finally, it is shown that the time-varying
premiums for country funds are attributable to time-varying risk premiums.
The detailed results are discussed below.
To ascertain the risk exposures of the share price and NAV returns of the sample of
country funds, the following econometric specification is employed:
where, for any country fund “i” (the subscript “i” has been dropped for
convenience):
rp,t ¼ the excess total return on the share price (including dividends) of a country
fund between t1 and t.
rn,t ¼ the excess return on the NAV of a country fund between t1 and t.
rw,t ¼ the excess return on the Morgan Stanley Capital International (MSCI) world
market index.
rh,t ¼ the excess return on the MSCI or International Finance Corporation (IFC)
global index, corresponding to the country of origin of the fund.
The coefficients bp,w, bp,h are the risk exposures on the world and home market
portfolios for the price returns, and bn,w,bn,h are the risk exposures on the world and
25 Market Segmentation and Pricing of Closed-End Country Funds 685
home market portfolios for the NAV returns. Two hypotheses are tested. The
hypotheses are share price returns have significant risk exposures to the world market
index and the NAV returns have significant risk exposures to both the world market
index and the corresponding local market index. These hypotheses are implied by the
international CAPM and its extension under market segmentation (see Errunza and
Losq 1985 and Diwan et al. 1995). If assets are priced in a global environment,
the country fund share price as well as NAV returns should have significant
risk exposure to the world market index only. However, when there are barriers
to investment, the local market factor may be a source of systematic risk for
the NAVs.
The risk exposures on the prices and NAVs could be different when the prices
and NAVs are imperfect substitutes. By jointly estimating Eqs. 25.16 and 25.17,
the hypothesis that the risk exposures on the world market and the home market
indices are identical for the price and NAV returns is tested via a Wald
test which is w2 distributed, with degrees of freedom equal to the number of
restrictions.12 Since the local indices are significantly correlated with the world
index, for ease of interpretation, they are made orthogonal to each other by
regressing the local index returns on a constant and the world index return and
using the residuals as the local index return. Therefore, the risk exposure on
the local or regional index is the marginal exposure in the presence of the world
market index.
The results from the regressions to estimate risk exposures are reported in
Table 25.5. Panel A of Table 25.5 presents the results for developed market funds
and panel B presents the results for the emerging market funds. The results
presented in Table 25.5 indicate that 15 of the 15 developed market funds and
12 of the 15 emerging market funds price returns have significant exposure to the
world index at the 5 % level of significance. Also, 12 of the developed market funds
and 14 of the emerging market funds price returns have significant exposure to their
corresponding local market index, in the presence of the orthogonal world
market index. For the NAVs, all 15 of the developed funds and ten of the emerging
market funds returns have significant exposure to the world market index. More-
over, 14 of the developed and 13 of the emerging market NAV returns have
significant exposure to their local market index, in the presence of the world market
index. The adjusted R-squares across all the regressions vary from a low of zero to
a high of 92 %.13
12
For a set of linear restrictions, the Wald test statistic is given by W ¼ ½Rb-r0 ½R VarðbÞ R0 ½Rb-r,
where b is the vector of estimated parameters and Rb ¼ r is the set of linear restrictions.
13
Risk exposures of the price and NAV returns were also estimated using regional indices in place
of the local market indices for the funds from Europe, Latin America, and Asia. These results
indicate that, out of the 12 European funds, only two funds’ price returns have significant risk
exposure to the MSCI Europe index in the presence of the MSCI world index. Also, out of the
seven Latin American funds, all seven price returns and six NAV returns have significant exposure
to the Latin American index. Also, out of the 11 Asian funds, nine price returns and eight NAV
returns have significant risk exposures to the Asian index.
Table 25.5 Risk exposures of country funds
686
IRL 0.00 0.08 0.81 0.13 0.00 0.75 0.65 0.78 1.04 48.12
(0.75) (0.95) (5.73)* (0.33) (21.43)* (12.18)* [0.30] [0.00]
FRA 0.00 0.56 0.79 0.21 0.00 0.84 0.68 0.79 0.63 6.13
(0.20) (5.46)* (6.02)* (1.47) (26.34)* (17.95)* [0.42] [0.01]
SGP 0.00 0.63 0.93 0.15 0.00 0.65 0.43 0.25 11.27 0.01
(0.26) (4.18)* (5.65)* (0.14) (6.35)* (4.61)* [0.00] [0.90]
Panel B: Emerging market funds
Chi-squares for Wald tests
Coefficients on price returns (t-stats) Coefficients on NAV returns (t-stats) [p-values]
Symbol ap bp,h bp,w Adj. R2 an bn,h bn,w Adj. R2 w2local w2world
MEX 0.00 0.42 0.84 0.02 0.00 0.39 0.96 0.01 0.07 0.16
(1.36) (2.86)* (2.60)* (1.12) (2.01)* (2.35)* [0.78] [0.68]
KOR 0.00 0.20 0.09 0.01 0.00 0.00 0.01 0.00 3.09 0.14
(0.71) (1.77) (0.58) (0.73) (0.05) (0.09) [0.07] [0.70]
TWN 0.00 0.49 0.95 0.19 0.00 0.38 0.30 0.22 1.49 8.57
(0.26) (4.85)* (4.34)* (0.33) (6.91)* (2.26)* [0.22] [0.00]
MYS 0.00 0.62 1.01 0.20 0.00 0.92 0.74 0.63 8.35 0.83
(0.60) (5.98)* (3.74)* (0.89) (23.92)* (7.15)* [0.00] [0.36]
THA 0.00 0.56 1.02 0.27 0.00 0.87 0.72 0.82 18.04 3.07
(0.26) (7.80)* (5.71)* (2.27)* (23.26)* (13.28)* [0.00] [0.07]
Market Segmentation and Pricing of Closed-End Country Funds
BRA 0.00 0.50 0.90 0.31 0.00 0.64 1.07 0.58 7.38 0.29
(1.89) (10.84)* (2.98)* (2.41)* (17.65)* (5.64)* [0.00] [0.59]
INDG 0.00 0.34 0.06 0.07 0.00 0.35 0.06 0.14 0.00 0.20
(0.77) (2.15)* (0.27) (0.26) (5.04)* (0.45)* [0.98] [0.64]
RTWN 0.00 0.56 1.01 0.30 0.00 0.53 0.42 0.62 0.12 10.87
(0.10) (6.47)* (5.87)* (1.58) (12.03)* (6.17)* [0.71] [0.00]
(continued)
687
688
These results confirm the hypothesis that returns on prices and NAVs are
generated as a linear function of the returns on the world market and possibly the
corresponding local market index. The fact that 27 out of the 30 funds have
significant risk exposure to the world market index indicates that the share prices
may be determined via the global price of risk. The higher adjusted R-squares for the
NAVs indicate that the world market index and the corresponding local market index
have higher explanatory power for the time-series variations of the returns. The
intercepts in the univariate regressions are not significantly different from zero,
which suggests that the return on the world market index and an orthogonal return
on the local market index are mean-variance efficient for the price and NAV returns.
When Wald tests are performed to test the hypothesis that the risk exposures on
the prices and NAVs are identical, the results presented in the last two columns of
Table 25.5 indicate that for 21 (14 from developed) out of the 30 funds, the null
hypothesis of the same betas on the world index is rejected at the 10 % level of
significance. Similarly, for eight funds the null hypothesis of the same betas on the
local index is rejected at the 5 % level. This is a very important finding since it
indicates that the systematic risks of the shares and NAVs are different. The fact
that closed-end funds may have different risk exposures for the share and NAVs has
been documented for domestic funds by Gruber (1996). Also different risk expo-
sures for restricted and unrestricted securities have been documented by Bailey and
Jagtiani (1994) for Thai securities and by Hietala (1989) for Finnish securities.
These results indicate that country fund share prices and NAVs may have
differential risk exposures. The different risk exposures will result in different
expected returns for the share prices and NAVs, which is one of the sources of
the premiums. This section clearly shows that country fund share price and NAV
returns have significant risk exposure to the world market index. The different risk
exposures will result in different expected returns for the share prices and NAVs,
which is one of the sources of the premiums. The issue of whether the country funds
are priced in equilibrium via the international CAPM is the focus of next section.
The results of the mean-variance efficiency of the MSCI world index, using
Eq. 25.2, are presented in Table 25.6. The table reports the w2 and the p-values
from a Wald test for the hypothesis that a set of intercepts are jointly zero.14 This
hypothesis is the restriction implied by unconditional international CAPM. Failure
to reject this hypothesis would imply that the world market index is mean-variance
efficient and expected returns are proportional to the expected returns on the world
14
Other studies such as Cumby and Glen (1990) fail to reject the mean-variance efficiency of the
MSCI world market portfolio for national indices of developed markets. Also, Chang et al. (1995)
fail to reject the mean-variance efficiency of the MSCI world market index for a sample of
developed as well as emerging market indices.
690 D.K. Patro
market index. The hypothesis is tested using GMM for an exactly identified system
as shown in section IA. To ensure a longer time series, the tests are conducted using
funds listed before December 1990.
The null hypothesis that the intercepts are jointly zero cannot be rejected for the
share prices of the 15 developed market funds as well as the 15 emerging market
funds. But, the null hypothesis is rejected at the 5 % level for the NAVs of both
developed and emerging market funds. The tests are also conducted on subsets of
funds to check the robustness of the results. Since the classification of developed
and emerging markets is based on national output, it may not always capture
whether a country’s capital market is well developed. The subset consists of
funds with zero intercepts for the NAV returns, on an individual basis. For this
subset of the developed market funds consisting of 11 funds, though not reported
here, the mean-variance efficiency of the world index cannot be rejected for both
the share prices and NAVs. For emerging market funds NAVs, even for subsets of
funds, the MSCI world market index is not mean-variance efficient. These results
indicate that the world market index is an appropriate benchmark for the share
prices but not necessarily for the NAVs.
These results have important implications for capital market integration. Since
for the NAVs, the mean-variance efficiency of the world market index is rejected, it
implies that the share prices and NAVs are priced differently. This differential
pricing is sufficient to generate premiums on the share prices. Also, the fact that
mean-variance efficiency of the world market index cannot be rejected for share
prices of both developed and emerging markets is consistent with the theoretical
prediction of Diwan et al. (1995). Since the country fund share prices are priced
with complete access to that market, in equilibrium, their expected returns should
be proportional to their covariances with the world market portfolio. If the NAVs
are priced with incomplete access, the world market portfolio is not mean-variance
efficient with respect to those returns. The results in this section are consistent
with this notion. As the earlier section shows, the risk exposures of the share prices
and NAVs may differ. If the international CAPM is a valid model for the share
prices and NAVs, as outlined in section IB, this differential risk exposure can
explain cross-sectional variations in premiums. The next section analyzes the effect
of the differential risk exposures of the share prices and NAVs on the country fund
premiums.
This section presents the results for tests of the hypothesis that cross-sectional
variation in country fund premiums is positively related to the differences in the risk
exposures of the share prices and NAVs. The theoretical motivation for this was
presented in section IC. Equation 25.10 is estimated for each week during the
last 75 weeks of the sample and the parameter estimates are averaged. The results
indicate that cross-sectional variation in country fund premiums is explained by the
25 Market Segmentation and Pricing of Closed-End Country Funds 691
differential risk exposures of a closed-end country fund share price and NAV returns.
The estimation proceeds are as follows: after the betas are estimated using 75 weeks
of data prior to the last 75 weeks of the sample, the difference in the betas on the world
index is used as an explanatory variable in regressions with the premiums as the
dependent variable.
The results are reported in Table 25.7. The results indicate that, for the full
sample, the difference in risk exposure is significant and the adjusted R-square is
9.5 %. For the developed market funds, the results in panel B indicate that the
differences in the betas on the world index have significant explanatory power for
the cross-section of premiums for the developed market funds, with an adjusted
R-square of 18.7 %. For the emerging market funds, however, as reported in
panel C, the differences in risk exposures are not significant, and the adjusted
R-square is only 1.9 %. This result is not surprising, given the result of the previous
section that the world market index is not mean-variance efficient for the emerging
market NAVs. When a dummy variable which takes value of one for the developed
markets is added, it is highly negatively significant, indicating that the premiums
for developed market funds are significantly lower than the emerging market fund
premiums. The differential risk exposure is however significant only at the 10 %
level, when the dummy variable is added.
As an extended specification, measures of spanning, integration, and substitution
(based on prior research of Errunza et al. 1998) are used as additional explanatory
variables. When measures of spanning, substitution, and access are used as addi-
tional explanatory variables, the adjusted R-squares go up to 39.5 % for the full
sample. The measure of spanning has a positive sign as expected. The measure of
spanning is also significant across the developed market funds. However, contrary
692 D.K. Patro
To assess the predictability of price and NAV returns, 4-week cumulative returns
(the sum of returns over a 4-week period beginning the next week) are used, since
most studies of predictability have used a monthly horizon. Predictability of the
returns would imply that expected returns vary over time due to rational variation in
694 D.K. Patro
risk premiums. To test the hypothesis of predictability, the returns for both
share prices and NAVs are regressed on a set of global and fund-specific
instruments, to ascertain the predictive power of these instruments. The global
instruments are TED and DIVY. The fund-specific instrument is the lagged
premium on an equally weighted index of all the funds in the sample (FFD).
The lagged premium was found to have predictive power for price returns by
Bodurtha et al. (1995). Also, Errunza et al. (1998) find that when time series of
premiums is regressed on a global fund premium, it is highly significant. However,
unlike Errunza et al. (1998) here the lagged premium is used. Significance of this
variable would indicate that investors use information about past premiums to form
expectations about future prices. The regression is of the form
X
t¼ tþ4
ra, t ¼ Z0 þ Z1 FFDt þ Z2 TEDt þ Z3 DIVYt þ ea, t (25.18)
t ¼ tþ1
where ra, t is the excess return on the price or NAV of an equally weighted portfolio
of developed or emerging market funds.
The results from the regressions are presented in Table 25.8. Wald tests
of no predictability, which test the hypothesis that a set of coefficients are jointly
zero, indicate that the fund factor predicts price returns for 4 funds only and it
also predicts 15 funds’ NAV returns. The predictive power of the lagged premiums
for price returns has been attributed to noise trading (see Bodurtha et al. 1995).
The fact that it predicts NAV returns may indicate that noise trading is not an
important determinant of expected returns of share prices of closed-end funds.
When global instruments are used, the Wald test of no predictability can be rejected
at the 5 % level for only five funds’ price returns and 15 funds’ NAV returns.
Overall, using all the instruments, the null hypothesis of no predictability can be
rejected for six share prices and 20 NAVs. The poor predictability of the share
prices is puzzling, since the share prices are determined in the USA and the
instruments used have been shown to predict returns in the USA.
In summary, closed-end country fund price and NAV returns exhibit
considerable predictability. The existence of predictable variation in returns is
interpreted as evidence of time-varying expected returns. The differences
between the predictability of prices and NAVs also imply that they are not priced
identically, which is further examined in the next section using time-varying
expected returns.
Table 25.9 reports the results of estimating the SDFs specified in Eqs. 25.11
and 25.12. The models in panels A and B correspond to the international
CAPM and a two-factor model, with the world market return and the regional
returns for Latin America or Asia as the factors. For both the models,
25 Market Segmentation and Pricing of Closed-End Country Funds 695
the conditional restrictions are tested. As noted earlier, if expected returns vary over
time due to rational variation in risk premiums, conditional expected returns should
be used. The conditional restrictions are tested by using a set of lagged instrumental
variables TED and DIVY as predictors of excess returns. The tests use sets of
assets – two sets of four developed market funds, a set of three Latin American
funds, and two sets of three Asian funds. To ensure an adequate time series,
25 Market Segmentation and Pricing of Closed-End Country Funds 697
the funds selected are the ones listed prior to December 1989. Also, using too many
assets would result in too many moment conditions.15
The funds selected are listed in Table 25.9. After estimating a model using
GMM, its goodness-of-fit is ascertained by the J-statistic. The J-statistic is a test of
the overidentifying restrictions of the model. The p-values given below the
J-statistic indicate that none of the models are rejected for any test assets, at
conventional significance levels. This finding is consistent with the idea that for
a group of assets, there may be a number of SDFs that satisfy the pricing relation.
Also, failure to reject the conditional international CAPM even for the NAVs
indicates that while the unconditional international CAPM is not a valid model for
emerging market funds and some developed market funds, the conditional interna-
tional CAPM may be a valid model. This finding is consistent with the finding of
Buckberg (1995) who fails to reject the conditional international CAPM for a set of
emerging market indices. It must be noted that, unlike traditional conditional models,
using stochastic discount factors implies a nonlinear relation between asset returns
and the returns on the market index. Failure to reject such specifications may indicate
that nonlinear specifications perform better in explaining equity returns.
Table 25.9 also reports the results for tests of hypothesis that the coefficients of
the SDFs of the price returns and the NAV returns are identical. The p-values for the
Wald tests indicate that the null hypothesis can be rejected at all conventional levels
for a subset of the developed market funds and all the Asian funds. This is a striking
result, since it indicates that, although the same factors may be priced for the share
prices and NAVs, the factors are priced differently. In traditional asset-pricing
framework, it is equivalent to saying that the risk premiums are different.
This result is consistent with the previous finding of Harvey (1991) who reports
that for a sample of industrial markets, although the conditional international
CAPM cannot be rejected, the price of risk varies across countries. Also, De Santis
(1995) finds that stochastic discount factors that can price developed market indices
cannot price emerging market indices. The difference in the price of risk is one of the
sources of the premiums for the country funds. When the two-factor model is used,
again the tests reject the hypothesis that the coefficients are identical for the Asian
funds. However, the hypothesis of identical coefficients for the price and NAV returns
cannot be rejected for a set of developed market funds and the Latin American funds.
The above results imply that a subset of the developed markets and the Asian
market are segmented from the US market. However, a subset of the developed
markets and the Latin American markets are integrated with the US market. The
results for Latin America are not affected, when tests are conducted excluding the
time period surrounding the Mexican currency crisis. Therefore, differential risk
exposures are more important in explaining the premiums for the Latin American
markets and the developed markets that are integrated with the world market.
The results in this section clearly show that while some markets may be integrated
15
Cochrane (1996) shows that iterated GMM estimates behave badly when there are too many
moment conditions (37 in his case).
698
Twofactor model
Asian funds
KOR, MYS, 2.35 3.54 (4.06)* 1.16 3.10 4.10 (4.80)* 1.19 18 11.51 [0.87] 3.12 [0.26] 2.94 [0.08]
TWN (2.92)* (1.48) (3.63)* (1.74)
THA, INDG, 3.00 (5.93)* 2.00 1.35 1.88 (3.86)* 0.88 1.50 18 10.66 [0.90] 8.16 [0.04] 7.65 [0.00]
FPHI (3.95)* (1.63) (1.81) (1.84)
Latin American funds
BRA, CHL, 0.70 (0.57) 0.29 (0.24) 1.09 0.83 (0.98) 0.16 (0.19) 0.76 18 12.64 [0.81] 2.27 [0.51] 0.02 [0.88]
MEX (3.27)* (3.97)*
Estimation of stochastic discount factors, based on the equation E[M(t,1) R(t,1)jZt] ¼1 via GMM. The sample covers January 1990–August 1995.
The instrument set Zt includes a constant, the lagged spread on 90-day Eurodollar deposits and 90-day US Treasury yields (LTED), and the dividend yield
on Standard and Poor’s 500 index (LDIVY). The estimations with the returns on regional indices (Rh) also use the lagged regional return (for Asia or Latin
America) as an instrument
The stochastic discount factors estimated are
ICAPM: M ¼ l0 + l1Rw
Two-factor: M ¼ l0 + l1Rw + l2Rh
DF is the degrees of freedom for Hansen’s test of the overidentifying restrictions (the J-statistic), which has an w2 distribution. The t-stats for the coefficients
are given in parenthesis (an asterisk indicates significance at the 5 % level) and the p-values for the J-stat is given in brackets. The last two columns report
Chi-squares and p-values for tests of hypotheses that the coefficients of the price returns and NAV returns are identical. J1 is for a joint test for all the
coefficients and J2 is a test for the coefficients on the world market index
Market Segmentation and Pricing of Closed-End Country Funds
699
700 D.K. Patro
with the world capital market, others are not. If capital markets are segmented, that
is sufficient to generate the premiums on the share prices. However, if markets are
integrated, differential risk exposures may explain the premiums. This finding is
different from the existing literature on country fund pricing, which has attributed
the existence of premiums on irrational factors such as noise trading. The preceding
analysis clearly shows that differential pricing of the same factors or differential
risk exposures on the same factor may lead to premiums on the share prices.
The next section provides an analysis of the effect of time-varying expected returns
on country fund premiums.
Table 25.9 reports the results of regressing the premiums on the country funds on
the differences in conditional risk exposures of share price returns and NAV returns
estimated using Eqs. 25.13, 25.14, and 25.15. The conditional risk exposures are
estimated as a linear function of lagged dividend yields and the term premium. The
null hypothesis is that the coefficient of the difference in the conditional betas is
significant. The results indicate that the difference in conditional betas is highly
significant in explaining the time-varying premiums on country funds. For 24 out of
the 30 funds, the coefficient is significant. Also, the adjusted R-squares are high,
especially for the developed market funds. For many of the funds, however, the
coefficient is negative, implying that using the world market index alone is not
sufficient to explain the return-generating process and the returns on the NAVs may
reflect the local price of risk.
The adjusted R2 values are higher for the developed market funds. This result is
consistent with earlier results which show that the world market index is not an
appropriate benchmark for emerging market NAVs. Also, majority of the emerging
market funds are from Asia. The results in Table 25.9 indicated that these markets
are segmented from the world market. Therefore, different risk exposures to the
world market index do not explain much of the variation in the premiums for the
emerging markets.
This is an important finding since the existing literature on closed-end funds
has attributed the existence of premiums to irrational factors, such as noise trading.
The preceding analysis shows clearly that segmented capital markets in which risk
premiums vary over time are sufficient to generate two different prices for the same
set of cash flows. Also, the differences in prices may vary over time as expected
returns vary over time due to rational variations in expected returns. If the price of
risk is different across markets, the same security will have different expected
returns. The difference in these expected returns results in the premium on the share
price. If the expected returns vary over time because of rational variation in risk
premiums, the premiums will also vary over time as a function of the differential
expected returns (Table 25.10).
25 Market Segmentation and Pricing of Closed-End Country Funds 701
25.5 Conclusions
The law of large numbers ensures that the RHS of above equation is the same as
E½f ðZt ; y0 Þ: (25.21)
702 D.K. Patro
yt ¼ Xt 0 b þ et (25.23)
The White robust covariance matrix may be used for O as discussed in appendix
C when heteroskedasticity is present. Using this approach, we estimate GMM with
White heteroskedasticity consistent t-stats.
References
Bailey, W., & Jagtiani, J. (1994). Foreign ownership restrictions and stock prices in the Thai
Capital Market. Journal of Financial Economics, 36, 57–87.
Bailey, W., & Lim, J. (1992). Evaluating the diversification benefits of the new country funds.
Journal of Portfolio Management, 18, 74–80.
Bansal, R., Hsieh, D. A., & Viswanathan, S. (1993). A new approach to international arbitrage
pricing. Journal of Finance, 48, 1719–1747.
Bekaert, G. (1995). Market integration and investment barriers in equity markets. World Bank
Economic Review, 9, 75–107.
704 D.K. Patro
Bekaert, G., & Urias, M. (1996). Diversification, integration and emerging market closed-end
funds. Journal of Finance, 51, 835–869.
Bodurtha, J. N., Jr., Kim, D.-S., & Lee, C. M. C. (1995). Closed-end country funds and U.S. market
sentiment. Review of Financial Studies, 8, 881–919.
Bonser-Neal, C., Brauer, G., Neal, R., & Wheatly, S. (1990). International investment restrictions
and closed-end country fund prices. Journal of Finance, 45, 523–547.
Chang, E., Eun, C., & Kolodny, R. (1995). International diversification through closed-end
country funds. Journal of Banking and Finance, 19, 1237–1263.
Cochrane, J. H. (1996). A cross-sectional test of an investment-based asset pricing model. Journal
of Political Economy, 104(3), 572–621.
Cumby, R. E., & Glen, J. D. (1990). Evaluating the performance of international mutual funds.
Journal of Finance, 45(2), 497–521.
De Santis, G. (1995). Volatility bounds for stochastic discount factors: Tests and implications from
international financial markets. Working paper, University of Southern California, Jan.
Diwan, I., Vihang, E., & Senbet, L.W. (1995). Diversification benefits of country funds. In
M. Howell (Ed.), Investing in emerging markets. Washington, DC: Euromoney/World Bank.
Errunza, V., & Losq, E. (1985). International asset pricing under mild segmentation: Theory and
test. Journal of Finance, 40, 105–124.
Errunza, V., Senbet, L. W., & Hogan, K. (1998). The pricing of country funds from emerging
markets: Theory and evidence. International Journal of Theoretical and Applied Finance,
1, 111–143.
Fama, E. F., & French, K. R. (1989). Business conditions and expected returns on stocks and
bonds. Journal of Financial Economics, 25, 23–49.
Ferson, W. (1995). Theory and empirical testing of asset pricing models. In R. Jarrow,
V. Maksimovic, & W. T. Ziemba (Eds.), Handbooks in operation research and management
science (Vol. 9). Amsterdam: North-Holland.
Ferson, W., & Foerster, S. (1994). Finite sample properties of generalized methods of moments in
tests of conditional asset pricing models. The Journal of Financial Economics, 36, 29–55
Ferson, W. E., & Harvey, C. R. (1993). The risk and predictability of international equity returns.
Review of Financial Studies, 6(3), 527–566.
Ferson, W. E., & Schadt, R. (1996). Measuring fund strategy and performance in changing
economic conditions. Journal of Finance, 51(2), 425–461.
Greene, W. (2008). Econometric analysis. Washington, DC: Pearson.
Gruber, M. J. (1996). Presidential address: Another puzzle: The growth in actively managed
mutual funds. Journal of Finance, 51, 783–810.
Hanley, K. W., Lee, C., & Seguin, P. (1996). The marketing of closed-end fund IPOs. Journal of
Financial Intermediation, 5, 127–159.
Hansen, L. P. (1982). Large sample properties of generalized method of moments estimators.
Econometrica, 50, 1029–1054.
Hardouvellis, G., Wizman, T., & La Porta, R. (1993). What moves the discount on country equity
funds? Working paper 4571, NBER.
Harvey, C. (1991). The world price of covariance risk. Journal of Finance, 46, 111–159.
Hietala, P. T. (1989). Asset pricing in partially segmented markets: Evidence from the finnish
market. Journal of Finance, 44(3), 697–718.
Lee, C. M. C., Shleifer, A., & Thaler, R. (1991). Investor sentiment and the closed-end fund
puzzle. Journal of Finance, 46(1), 75–109.
Mackinlay, C., & Richardson, M. (1991). Using generalized method of moments to test mean-
variance efficiency. Journal of Finance, 46(2), 511–527.
Peavy, J. W., III. (1990). Returns on initial public offerings of closed-end funds. Review of
Financial Studies, 3, 695–708.
Pontiff, J. (1997). Excess volatility and closed-end funds. American Economic Review,
87(1), 155–169.
25 Market Segmentation and Pricing of Closed-End Country Funds 705
Roll, R. (1977). A critique of the asset pricing theory’s tests. Journal of Financial Economics,
4, 129–176.
Solnik, B. (1996). International investments (3rd ed.). New York: Addison-Wesley.
Stulz, R. (1995). International portfolio choice and asset pricing: An integrative survey.
In R. Jarrow, V. Maksimovic, & W. T. Ziemba (Eds.), Handbooks in operation research and
management science (Vol. 9). Amsterdam: North-Holland.
Stulz, R., & Wasserfallen, W. (1995). Foreign equity investment restrictions, capital flight, and
shareholder wealth maximization: Theory and evidence. Review of Financial Studies,
8, 1019–1057.
Weiss, K. (1989). The post-offering price performance of closed-end funds. Financial Manage-
ment, 18, 57–67.
A Comparison of Portfolios Using
Different Risk Measurements 26
Jing Rung Yu, Yu Chuan Hsu, and Si Rou Lim
Contents
26.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 708
26.2 Portfolio Selection Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 709
26.2.1 The Mean-Variance Model (MV) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 710
26.2.2 The Mean Absolute Deviation Model (MAD) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 710
26.2.3 The Downside Risk Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 711
26.3 The Proposed Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 711
26.3.1 The MV Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 712
26.3.2 The MAD Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 712
26.3.3 The DSR Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 713
26.3.4 The MV_S Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 713
26.3.5 The MAD_S Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 714
26.3.6 The DSR_S Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 715
26.4 Experimental Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 716
26.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 722
Appendix 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 722
The Linearization of the MAD Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 722
Appendix 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 724
The Linearization of the DSR Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 724
Appendix 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 726
Multiple Objective Programming . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 726
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 727
Abstract
In order to find out which risk measurement is the best indicator of efficiency in
a portfolio, this study considers three different risk measurements: the mean-
variance model, the mean absolute deviation model, and the downside risk
model. Meanwhile short selling is also taken into account since it is an important
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 707
DOI 10.1007/978-1-4614-7750-1_26,
# Springer Science+Business Media New York 2015
708 J.R. Yu et al.
strategy that can bring a portfolio much closer to the efficient frontier by
improving a portfolio’s risk-return trade-off. Therefore, six portfolio rebalancing
models, including the MV model, MAD model, and the downside risk model,
with/without short selling, are compared to determine which is the most
efficient. All models simultaneously consider the criteria of return and risk
measurement. Meanwhile, when short selling is allowed, models also consider
minimizing the proportion of short selling. Therefore, multiple objective
programming is employed to transform multiple objectives into a single objec-
tive in order to obtain a compromising solution. An example is used to perform
simulation, and the results indicate that the MAD model, incorporated with
a short selling model, has the highest market value and lowest risk.
Keywords
Portfolio selection • Risk measurement • Short selling • MV model • MAD
model • Downside risk model • Multiple objective programming • Rebalancing
model • Value-at-risk • Conditional value-at-risk
26.1 Introduction
Determining how to maximize the profit and minimize the risk of a portfolio is an
important issue in portfolio selection. The mean-variance (MV) model of portfolio
selection is based on the assumptions that investors are risk averse and the return of
assets is normally distributed (Markowitz 1952). This model is regarded as the basis
of modern portfolio theory (Deng et al. 2005).
However, the MV model is limited in that it only leads to optimal decisions if the
investor’s utility functions are quadratic or if investment returns are jointly elliptically
distributed (Grootveld and Hallerbach 1999; Papahristoulou and Dotzauer 2004). Thus,
numerous researches have focused on risk, return, and diversification in the development
of investment strategies. Also, a large number of researches have been proposed to
improve the performance of investment portfolios (Deng et al. 2000; Yu and Lee 2011).
Konno and Yamazaki (1991) proposed a linear mean absolute deviation (MAD)
portfolio optimization model. The MAD model replaces the variance of objective
function in the MV model with the mean absolute deviation. The major advantage
of the MAD model is that the estimation of the covariance matrix of asset returns is
not needed. Also, it is much easier to solve large-scale problems with linear
programming than with quadratic approaches (Simaan 1997).
Portfolio selection under shortfall constraints originated from Roy’s (1952)
safety-first theory. Economists have found that investors care about downside
losses more than they care about upside gains. Therefore, Markowitz (1959)
suggested using semi-variance as a measure of risk, instead of variance, because
semi-variance measures downside losses rather than upside gains. The use of
downside risk (DSR) measures is proposed due to the problems encountered in
using a conventional mean-variance analysis approach in the presence of
non-normality in the emerging market data. Unlike the mean-variance framework,
26 A Comparison of Portfolios Using Different Risk Measurements 709
the downside risk measure does not assume that the return distributions of assets are
normal. In addition, the increasing emphasis of investors on limiting losses might
make the downside risk measure more intuitively appealing (Stevenson 2001; Ang
et al. 2006). The downside risk measure can help investors make proper decisions
when returns are non-normally distributed, especially for emerging market data or
for an international portfolio selection (Vercher et al. 2007).
Another type of shortfall constraint is value-at-risk (VaR), which is a percentile-
based metric system for risk measurement purposes (Jorion 1996). It is defined as the
maximum loss that a portfolio can suffer at a given level of confidence and at a given
horizon (Fusai and Luciano 2001). However, VaR is too weak to handle the situation
when losses are not “normally” distributed, as loss distribution tends to exhibit “fat tail”
or empirical discreteness. Conditional value-at-risk (CVaR) is an alternative measure
that quantifies losses that might be encountered in the tail of loss distribution
(Rockafellar and Uryasev 2002; Topaloglou et al. 2002). A confidence level is required
when employing the measurements of VaR and CVaR. The allocation of a portfolio
varies with the varying confidence level. Therefore, neither measure is included for
comparison in this chapter.
Generally, short selling has good potential to improve a portfolio’s risk-return
trade-off (White 1990; Kwan 1997) and is considered by most investors to obtain
interest arbitrage; however, it comes with high risk (Angel et al. 2003). Since high
risks should be avoided, the role of short selling is minimized in this chapter. Instead,
three kinds of portfolio models with and without short selling are compared.
Section 26.2 introduces the mean-variance, mean absolute, and downside risk
models. In Sect. 26.3, the rebalancing models with/without short selling are pro-
posed. In Sect. 26.4, the performances of three different risk measurements with/
without short selling are compared by using the historical data of 45 stocks listed in
the TSE50 index. Finally, Sect. 26.5 presents the conclusions along with sugges-
tions for future research.
In this section, the mean-variance, the mean absolute deviation, and the downside
risk models are introduced separately. First, the notations are defined as follows:
n is the number of available securities.
wi is the investment portion in securities i for i ¼ 1, . . ., n.
ri is the return on securities i.
m is the expected portfolio return.
si2 is the variance of the return on securities i.
sij2 is the covariance between the returns of securities i and j.
rit is the return on securities i in period t for t ¼ 1, . . ., T, which is assumed to be
available through historical data.
XT
Ri is equal to T1 r it .
t¼1
dt is the deviation between the return and the average return.
710 J.R. Yu et al.
The MV model uses the variance of the return as the measure of risk and formulates
the portfolio optimization problem as the following quadratic programming prob-
lem (Markowitz 1952):
X
n Xn X
n
Min sp ¼ w2i s2i þ sij wi wj
i¼1 i¼1 j¼1ði6¼jÞ
(26.1)
X
n
s:t: r i wi m,
i¼1
X
n
wi ¼ 1, (26.2)
i¼1
wi 0, (26.3)
for i ¼ 1, . . ., n.
Constraint (26.1) expresses the requirements m of a portfolio return, and con-
straint (26.2) is the budget constraint. The model is known to be valid if an investor
is risk averse in the sense that he prefers less standard deviation of the portfolio
rather than more. Since wi 0, a short sale is not allowed here.
1X T
Min dt
T t¼1
X
n
dt þ ðr it Ri Þwi 0, t ¼ 1, . . . , T, (26.4)
i¼1
X
n
dt ðr it Ri Þwi 0, t ¼ 1, . . . , T: (26.5)
i¼1
If the return is lower than the average return, constraint (26.4) is a binding
Xn
constraint which means d t ¼ ðr it Ri Þwi for t ¼ 1, . . .,T. Otherwise, constraint
i¼1
X
n
(26.5) is a binding constraint which means dt ¼ ðr it Ri Þwi for t ¼ 1, . . .,T. For
i¼1
more details on the reformulation, please refer to Appendix 1.
Apparently, the MAD model does not require the covariance matrix of asset returns,
and consequently its estimation is not needed. Large-scale problems can be solved faster
and more efficiently because the MAD model has a linear rather than quadratic nature.
Vercher et al. (2007) consider the equivalent formulation of the portfolio selection
problem (Speranza 1993) and reformulate the following linear optimization model
by considering downside risk measurement:
1X T
Min d
T t¼1 t
X
n
d
t þ ðr it Ri Þwi 0, t ¼ 1, . . . , T (26.6)
i¼1
Multiple period portfolio selection models with rebalancing mechanisms have become
attractive in the financial field in order to get desired returns in situations that are
subject to future changes (Yu et al. 2010). To reflect a changing situation in the models,
the rebalancing mechanism is adopted for multiple periods (Yu and Lee 2011).
712 J.R. Yu et al.
Six rebalancing models are introduced. These are MV, MAD, DSR, MV_S,
MAD_S, and DSR_S. The first three models lack short selling, and the other three
have short selling. The model notations are denoted as follows: w+i,0 is the weight of
security i held in the previous period, w i,0 is the weight of securities i sold short in
the previous period, wþ i is the total weight of securities i bought after rebalancing,
and w i is the total weight of securities i sold short after rebalancing.
With each rebalancing, lþ
i is the weight of securities i bought in this period, li is
þ
the weight of securities i sold in this period, si is the weight of securities i sold short
in this period, si is the weight of securities i repurchased in this period, ui is the
binary variable that indicates whether the securities i are selected for buying, vi is
the binary variable that indicates whether the securities i are selected for selling
short, and k is the initial margin requirement for short selling.
Min sp
for i ¼ 1, 2, . . ., n.
Constraint (26.7) is the rebalancing constraint; it shows the current weight for
the ith security according to the previous period. Constraint (26.8) is the required
range of weights for each security in buying. For simplexity, the upper and lower
bounds of each weight are set 0.2 and 0.05, respectively.
The objectives of the MAD model include the maximization of return and minimi-
zation of the mean absolute deviation, which is transformed into a linear deviation
as follows:
Xn
Max Ri w i
i¼1
26 A Comparison of Portfolios Using Different Risk Measurements 713
1X T
Min dt
T t¼1
The DSR model considers the objectives of maximizing the return and minimizing
the downside risk, which is transformed into a linear risk as follows:
X
n
Max Ri w i
i¼1
1X T
Min d
T t¼1 t
X
n
Max Ri w þ
i wi
i¼1
Min sp
X
n
Min w
i
i¼1
X
n (26.9)
s:t: wþ
i þ kwi ¼ 1,
i¼1
þ
wþ þ
i ¼ wi, 0 þ ‘i ‘i , (26.10)
w þ
i ¼ wi, 0 þ si si , (26.11)
0:05ui wþ
i 0:2ui , (26.12)
714 J.R. Yu et al.
0:05vi w
i 0:2vi , (26.13)
ui þ v i ¼ y i , (26.14)
for i ¼ 1, 2, . . ., n.
Unlike the MV model, the MV_S model has an extra objective, namely, mini-
mizing the short selling proportion. The total budget is including the cost of buying
and short selling as shown in constraint (26.9). k is the initial margin requirement
for short selling. Constraint (26.10) indicates the current weight for the ith securities
based on the previous period. Constraint (26.11) indicates the current short selling
proportion of the ith security adjusted by the previous period. Constraints (26.12)
and (26.13) limit the upper and lower bounds, respectively, of the long and short
selling proportion for the ith security.
Based on the MAD model (Konno and Yamazaki 1991), the MAD_S model
replaces the objective function of minimizing the variance in the MV_S model
(Yu and Lee 2011) with the objective of minimizing the absolute deviation of
average return, as follows:
X
n
Min wþ w r it wþ w Ri
i i i i
i¼1
The objective of the mean absolute deviation can be transformed into a linear
problem:
1X T
Min dt
T t¼1
X
n þ
s:t: dt þ wþ
i wi r it wi wi Ri 0, (26.15)
i¼1
X
n þ
dt wþ
i wi r it wi wi Ri 0, (26.16)
i¼1
for t ¼ 1, . . ., T.
The following is the MAD_S model:
X
n
Max Ri wþ
i wi
i¼1
26 A Comparison of Portfolios Using Different Risk Measurements 715
1X T
Min dt
T t¼1
X
n
Min w
i
i¼1
The DSR_S model focuses on the deviation when the return falls below the average
return, as follows:
X
n
Max Ri w þ
i wi
i¼1
1X T
Min d
T t¼1 t
X
n
Min w
i
i¼1
X
n þ
d
t þ wþ
i wi r it wi wi Ri 0, (26.17)
i¼1
for t ¼ 1, . . ., T.
Apparently, all six models have multiple objectives. Therefore, multiple
objective programming (Zimmermann 1978, and Lee and Li 1993) is adopted to
transform the multiple objectives into a single objective. For more details, please
refer to Appendix 3. Taking the MAD_S model as an example, we can reformulate
the multiple objectives as follows:
Max l
ðr r l Þ (26.18)
s:t: l ,
rg rl
ð s sl Þ
l , (26.19)
s g sl
716 J.R. Yu et al.
w w
l l
, (26.20)
w
g wl
X
n
r ¼ Ri w þ
i wi , (26.21)
i¼1
1X T
s ¼ d , (26.22)
T t¼1 t
X
n
w ¼ w
i , (26.23)
i¼1
Constraints (26.9) (26.16).
For multiple objective programming (Lee and Li 1993), r is the return of the
portfolio, rl is the anti-ideal return of the portfolio, rg is the ideal return of the
portfolio that maximizes the objective, s is the inherent risk of the portfolio, sl is
the anti-ideal risk of the portfolio, sg is the ideal risk of the portfolio, w is the
short selling proportion of the portfolio, wl is the anti-ideal short selling proportion
of the portfolio, and w g is the ideal short selling proportion of the portfolio.
The constraints (26.18–26.20) are the achievements for maximizing the return,
minimizing the absolute deviation and objectives minimizing the short selling
problem of the corresponding portfolio, which are less than or equal to the whole
achievement (l). The whole achievement (l) should be maximized.
In the same way, the other five multiple objective models can be reformulated in
turn as a single-objective model. The details of the transformation are introduced in
Appendix 3.
Forty-five stocks listed on the Taiwan Stock Exchange were adopted and used to
compare the six models discussed above in order to determine which one is the best.
The benchmark is the Taiwan 50 Index (TSE50).
The exchange codes of the 45 stocks are listed in Table 26.1. The duration of the
analyzed data is from November 1, 2006, to November 24, 2009. The historical data
of the first 60 transaction days are used to build the initial models. For the monthly
updates, 20 transaction days are set as a sliding window. This study assumes
a budget of $1 million NTD is invested in the Taiwan Stock Market. For invest-
ments based on the weights generated by the initial models, the first transaction day
is January 25, 2007, and there are 34 rebalancing times in total. The models are
executed on an Intel Pentium Dual CPU E2200 2.20GHz and 2G RAM computer,
with Lingo11.0, an optimizing software.
From Table 26.2, it is apparent that the MAD and DSR models are more efficient
than the MV model because they both use linear transformation for problem
solving. This is much faster and more efficient when handling a large-scale
26
1.9
MV 14,000
1.7 MAD
DSR
1.3
10,000
1.1
8,000
0.9
6,000
0.7
0.5 4,000
1/24/07
3/24/07
5/24/07
7/24/07
9/24/07
11/24/07
1/24/08
3/24/08
5/24/08
7/24/08
9/24/08
11/24/08
1/24/09
3/24/09
5/24/09
7/24/09
9/24/09
Fig. 26.1 The market value of three risk measurements without short selling, the TSE50, and
TAIEX
problem. As one can see in Table 26.2, the MV model takes 9 min and 6 sec to
compute the result. However, the MAD and DSR models only take 1 min or slightly
more to solve the same data. Moreover, it is not necessary to calculate the covari-
ance matrix to set up the MAD and DSR models; this makes it very easy to update
the models when new data are added (Konno and Yamazaki 1991).
Figures 26.1, 26.2 and 26.3 show the comparisons of the MV, MAD, and DSR
models, respectively. These are the models without short selling. Figures 26.4, 26.5
and 26.6 show the comparisons of the MV_S, MAD_S, and DSR_S models,
respectively. These are the models with short selling. The TSE50 is used as the
benchmark to these models. As shown in Fig. 26.1, the market value of the MAD
26 A Comparison of Portfolios Using Different Risk Measurements 719
0.016
MV
0.014
MAD
DSR
0.012
0.01
Expected Return
0.008
0.006
0.004
0.002
−1E-17
1/24/07
3/24/07
5/24/07
7/24/07
9/24/07
11/24/07
1/24/08
3/24/08
5/24/08
7/24/08
9/24/08
11/24/08
1/24/09
3/24/09
5/24/09
7/24/09
9/24/09
−0.002
Fig. 26.2 The expected return of three risk measurements without short selling
0.035
MV
MAD
0.03 DSR
0.025
Risk
0.02
0.015
0.01
1/24/07
3/24/07
5/24/07
7/24/07
9/24/07
11/24/07
1/24/08
3/24/08
5/24/08
7/24/08
9/24/08
11/24/08
1/24/09
3/24/09
5/24/09
7/24/09
9/24/09
Fig. 26.3 The risk of three risk measurements without short selling
720 J.R. Yu et al.
2.3 18,000
MV_S
2.1 MAD_S
16,000
DSR_S
1.9 TSE50
TAIEX 14,000
Market Value (million)
1.7
1.5 12,000
1.3 10,000
1.1
8,000
0.9
6,000
0.7
0.5 4,000
1/24/07
3/24/07
5/24/07
7/24/07
9/24/07
11/24/07
1/24/08
3/24/08
5/24/08
7/24/08
9/24/08
11/24/08
1/24/09
3/24/09
5/24/09
7/24/09
9/24/09
Fig. 26.4 The market value of three risk measurements with short selling, the TSE50, and TAIEX
model is always greater than that of the other models and the benchmark.
Figures 26.2 and 26.3 display the expected return and risk of the portfolios
constructed with three risk measurements without short selling. Figure 26.2
shows that the expected returns of these three portfolios are almost the same.
However, Fig. 26.3 shows that the MAD model has the lowest risk under a similar
expected return to that of the others. Especially, in January 24, 2009, the risk in the
MAD model was much lower than in the other two models.
Downside risk measurement is applied when investors only take the negative
return into consideration and focus on the loss of investment. In other words, their
concern is with the real loss, not with the positive deviation of the average return in
portfolios. Therefore, in Fig. 26.3, the risk generated by the DSR model is always
higher than other measurements.
In Fig. 26.4, the market value of the MV_S, MAD_S, and DSR_S models is
compared. Since these models take short selling into consideration, the portfolio
selection is more flexible, and the risk is much lower, as Fig. 26.6 shows. Under the
same expected return, the MAD_S model has the lowest risk among the three models,
using the mean absolute deviation risk measure. Figure 26.4 shows that it also has the
highest market value. Even though the market value of each model is increased after
short selling is allowed, the market value of the MAD_S model is always higher than
the other models and the benchmark. Evidently, the MAD model is suggested for use
as the best risk measurement tool with or without short sell.
26 A Comparison of Portfolios Using Different Risk Measurements 721
0.016
MV_S
0.014 MAD_S
DSR_S
0.012
0.01
Expected Return
0.008
0.006
0.004
0.002
0
1/24/07
3/24/07
5/24/07
7/24/07
9/24/07
11/24/07
1/24/08
3/24/08
5/24/08
7/24/08
9/24/08
11/24/08
1/24/09
3/24/09
5/24/09
7/24/09
9/24/09
−0.002
Fig. 26.5 The expected return of three risk measurements with short selling
0.035
MV_S
MAD_S
0.03 DSR_S
0.025
Risk
0.02
0.015
0.01
1/24/07
3/24/07
5/24/07
7/24/07
9/24/07
11/24/07
1/24/08
3/24/08
5/24/08
7/24/08
9/24/08
11/24/08
1/24/09
3/24/09
5/24/09
7/24/09
9/24/09
Fig. 26.6 The risk of three risk measurements with short selling
722 J.R. Yu et al.
26.5 Conclusion
This chapter compares six different rebalancing models, with/without short selling, in
order to determine which is more flexible for portfolio selection. One of the advan-
tages of the MAD and DSR models is that they can be linearized; thus, they are faster
and more efficient than the MV model, especially with large-scale problems.
The experimental results indicate that the MAD and MAD_S models are
efficient in handling data and show higher market value than the other models;
moreover, they have lower risks in situations with/without short selling.
However, there remain important risk measurements, such as VaR and CVaR, for
future research to investigate. Thus, future studies may focus on developing
rebalancing models with the measures of VaR and CVaR. Since the rebalancing period
is fixed, the dynamic rebalancing mechanism is required for the first change environ-
ment. In addition, a portfolio selection model able to predict future returns is required.
Appendix 1
Konno and Yamazaki (1991) assume that rit is the realization of random variable ri
X
n
during period t (t ¼ 1, . . ., T), and Ri ¼ T1 r it :
i¼1
The MAD model is as follows:
T X
1X
n
Max ðr it Ri Þwi
T t¼1 i¼1
X
n
s:t: Ri xi m,
i¼1
X
n
wi ¼ 1,
i¼1
wi 0,
for i ¼ 1, . . ., n.
X n
Let ðr it Ri Þwi ¼ dt ¼ dt þ þ d t ,
i¼1
dt þ and dt 0,
X
n
then ðr it Ri Þwi ¼ dt þ d t ,
i¼1
26 A Comparison of Portfolios Using Different Risk Measurements 723
X
n
dt d
t ¼ ðr it Ri Þwi þ d t ,
i¼1
!
1 Xn
dt ¼ dt ðr it Ri Þwi 0:
2 i¼1
X
n
2dt ¼ dt ðr it Ri Þwi :
i¼1
X
n
dt þ ¼ ðr it Ri Þwi þ dt ,
i¼1
X
n
dt dt þ ¼ ðr it Ri Þwi þ dt þ ,
i¼1
Xn
2dt þ ¼ ðr it Ri Þwi þ d t ,
i¼1
!
þ 1 Xn
dt ¼ ðr it Ri Þwi þ d t 0,
2 i¼1
dt þ , d
t 0:
T X 1X
1X
n
T
Min ðr it Ri Þwi ¼ d þ þ dt
T t¼1 i¼1 T t¼1 t
!
þ 1 Xn
dt ¼ ðr it Ri Þwi þ dt 0,
2 i¼1
X
n
) ðr it Ri Þwi þ dt 0:
i¼1
724 J.R. Yu et al.
!
1 X n
dt ¼ ðr it Ri Þwi þ d t 0,
2 i¼1
X
n
) ðr it Ri Þwi þ dt 0:
i¼1
1X T
Min dt
T t¼1
Xn
s:t: d t þ ðr it Ri Þwi 0, t ¼ 1, . . . , T,
i¼1
X
n
dt ðr it Ri Þwi 0, t ¼ 1, . . . , T,
i¼1
X
n
Ri xi m,
i¼1
X
n
wi ¼ 1,
i¼1
wi 0,
for i ¼ 1, . . . , n.
Appendix 2
The use of variance as a measure of risk makes no distinction between gains and
losses. The following mean semi-absolute deviation risk measurement proposed by
Speranza (1993) is used to find the portfolios with minimum semi-variance:
XT Xn X n
1
ðr it Ri Þwi ðr Ri Þwi
T
t¼1 i¼1
i¼1 it
Min
2
26 A Comparison of Portfolios Using Different Risk Measurements 725
X
n
s:t: Ri xi m,
i¼1
X
n
wi ¼ 1,
i¼1
wi 0,
for i ¼ 1, . . . n.
X n
Because of the absolute deviation, ðr it Ri Þwi of the DSR model can be
i¼1
linearized in the same manner as the MAD model:
X
n
ðr it Ri Þwi ¼ dt þ dt
i¼1
X n X n
ðr it Ri Þwi ðr Ri Þwi
i¼1 i¼1 it dt þ þ dt dt þ þ dt
¼ ¼ dt
2 2
!
1 Xn
dt þ ¼ ðr it Ri Þwi þ d t 0,
2 i¼1
d t 0:
1X T
Min d
T t¼1 t
X
n
s:t: Ri xi m,
i¼1
X
n
wi ¼ 1,
i¼1
wi 0,
X
n
d
t þ ðr it Ri Þwi 0, t ¼ 1, . . . , T,
i¼1
dt 0, t ¼ 1, . . . , T,
for i ¼ 1, . . ., n.
726 J.R. Yu et al.
Appendix 3
The aforementioned multiple objective models in Sect. 26.3 are solved by fuzzy
multiple objective programming (Zimmermann 1978; Lee and Li 1993) in order to
transform the multiple objective model into a single-objective model. Fuzzy mul-
tiple objective programming based on the concept of fuzzy set uses a min operator
to calculate the membership function value of the aspiration level, l, for all of the
objectives.
The following is a multiple objective programming problem (Lee and Li 1993):
s:t: Ax b,
x 0,
I þ ¼ Z 1 ; Z 2 ; . . . ; Z l ; W 1 ; W 2 ; . . . ; W l ,
I ¼ Z
1 ; Z2 ; . . . ; Zr ; W 1 ; W 2 ; . . . ; W r :
Z K ðxÞ Z
m k ðZ K Þ ¼ k
, k ¼ 1, 2, . . . , l,
Z k Z
k
W
s W s ðxÞ
ms ðZ s Þ ¼ , s ¼ 1, 2, . . . , r:
Ws Ws
Max l
s:t: l ðZK ðxÞ Z
k = Z k Z k , k ¼ 1, 2, . . . , l,
l W
s W s ðxÞÞ= W s W s , s ¼ 1, 2, . . . , r,
x 2 X,
References
Ang, A., Chen, J., & Xing, Y. (2006). Downside risk. The Review of Financial Studies, 19,
1191–1239.
Angel, J. J., Christophe, S. E., & Ferri, M. G. (2003). A close look at short selling on Nasdaq.
Financial Analysts Journal, 59, 66–74.
Chang, C. T. (2005). A modified goal programming approach for the mean-absolute deviation
portfolio optimization model. Applied Mathematics and Computation, 171, 567–572.
Deng, X. T., Wang, S. Y., & Xia, Y. S. (2000). Criteria models and strategies in portfolio selection.
Advanced Modeling and Optimization, 2, 79–104.
Deng, X. T., Li, Z. F., & Wang, S. Y. (2005). A minimax portfolio selection strategy with
equilibrium. European Journal of Operational Research, 166, 278–292.
Fusai, G., & Luciano, E. (2001). Dynamic value at risk under optimal and suboptimal portfolio
policies. European Journal of Operational Research, 135, 249–269.
Grootveld, H., & Hallerbach, W. (1999). Variance vs downside risk: Is there really that much
difference? European Journal of Operational Research, 114, 304–319.
Jorion, P. H. (1996). Value at risk: A new benchmark for measuring derivative risk. New York:
Irwin Professional Publishers.
Konno, H., & Yamazaki, H. (1991). Mean-absolute deviation portfolio optimization model and its
applications to Tokyo stock market. Management Science, 37, 519–531.
Kwan, C. C. Y. (1997). Portfolio selection under institutional procedures for short selling:
Normative and market equilibrium considerations. Journal of Banking & Finance, 21,
369–391.
Lee, E. S., & Li, R. J. (1993). Fuzzy multiple objective programming and compromise program-
ming with Pareto optimum. Fuzzy Sets and Systems, 53, 275–288.
Markowitz, H. M. (1952). Portfolio selection. The Journal of Finance, 7, 77–91.
Markowitz, H. M. (1959). Portfolio selection, efficient diversification of investments. New York:
Wiley.
Papahristoulou, C., & Dotzauer, E. (2004). Optimal portfolios using linear programming models.
Journal of the Operational Research Society, 55, 1169–1177.
Rockafellar, R. T., & Uryasev, S. (2002). Conditional value-at-risk for general loss distribution.
Journal of Banking and Finance, 26, 1443–1471.
Roy, A. D. (1952). Safety first and the holding of assets. Econometrica, 20, 431–449.
Simaan, Y. (1997). Estimation risk in portfolio selection: The mean variance model versus the
mean absolute deviation model. Management Science, 43, 1437–1446.
Speranza, M. G. (1993). Linear programming models for portfolio optimization. Finance, 14,
107–123.
Stevenson, S. (2001). Emerging markets, downside risk and the asset allocation decision. Emerg-
ing Markets Review, 2, 50–66.
728 J.R. Yu et al.
Topaloglou, N., Vladimirou, H., & Zenios, S. A. (2002). CVaR models with selective hedging for
international asset allocation. Journal of Banking and Finance, 26, 1535–1561.
Vercher, E., Bermudez, J. D., & Segura, J. V. (2007). Fuzzy portfolio optimization under downside
risk measures. Fuzzy Sets and Systems, 158, 769–782.
White, J. A. (1990). More institutional investors selling short: But tactic is part of a wider strategy.
Wall Street Journal, 7, 493–501.
Yu, J. R., & Lee, W. Y. (2011). Portfolio rebalancing model using multiple criteria. European
Journal of Operational Research, 209, 166–175.
Yu, M., Takahashi, S., Inoue, H., & Wang, S. (2010). Dynamic portfolio optimization with risk
control for absolute deviation model. Journal of Operational Research, 201, 349–364.
Zimmermann, H. J. (1978). Fuzzy programming and linear programming with several objective
functions. Fuzzy Sets and Systems, 1, 44–45.
Using Alternative Models and
a Combining Technique in Credit Rating 27
Forecasting: An Empirical Study
Contents
27.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 730
27.2 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 731
27.2.1 Ordered Probit Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 731
27.2.2 Ordered Logit Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 732
27.2.3 Combining Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 732
27.3 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 733
27.3.1 Model Estimates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 734
27.3.2 Credit Rating Forecasting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 738
27.3.3 Estimation Results Using the Combining Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . 738
27.3.4 Performance Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 741
27.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 741
Appendix 1: Ordered Probit Procedure for Credit Rating Forecasting . . . . . . . . . . . . . . . . . . . . . . . . 744
Appendix 2: Ordered Logit Procedure for Credit Rating Forecasting . . . . . . . . . . . . . . . . . . . . . . . . 747
Appendix 3: Procedure for Combining Probability Forecasts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 748
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 749
Abstract
Credit rating forecasting has long time been very important for bond classifica-
tion and loan analysis. In particular, under the Basel II environment, regulators
C.-F. Lee
Department of Finance and Economics, Rutgers Business School, Rutgers, The State University of
New Jersey, Piscataway, NJ, USA
Graduate Institute of Finance, National Chiao Tung University, Hsinchu, Taiwan
e-mail: cflee@business.rutgers.edu
K. Wang • Y. Yang (*)
Graduate Institute of Finance, National Chiao Tung University, Hsinchu, Taiwan
e-mail: lkwang@mail.nctu.edu.tw; yatingyang.iof98g@nctu.edu.tw
C.-C. Lien
Treasury Division, E.SUN Commercial Bank, Taipei, Taiwan
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 729
DOI 10.1007/978-1-4614-7750-1_27,
# Springer Science+Business Media New York 2015
730 C.-F. Lee et al.
in Taiwan have requested the banks to estimate the default probability of the
loan based on its credit classification. A proper forecasting procedure for credit
rating of the loan is crucially important in abiding the rule.
Credit rating is an ordinal scale from which the credit category of a firm can
be ranked from high to low, but the scale of the difference between them is
unknown. To model the ordinal outcomes, this study first constitutes an attempt
utilizing the ordered logit and the ordered probit models, respectively. Then, we
use ordered logit combining method to weigh different techniques’ probability
measures as described in Kamstra and Kennedy (International Journal of
Forecasting 14, 83–93, 1998) to form the combining model.
The samples consist of firms in the TSE and the OTC market and are divided
into three industries for analysis. We consider financial variables, market vari-
ables, as well as macroeconomic variables and estimate their parameters for
out-of-sample tests. By means of cumulative accuracy profile, the receiver
operating characteristics, and McFadden R2, we measure the goodness-of-fit
and the accuracy of each prediction model. The performance evaluations are
conducted to compare the forecasting results, and we find that combining
technique does improve the predictive power.
Keywords
Bankruptcy prediction • Combining forecast • Credit rating • Credit risk • Credit
risk index • Forecasting models • Logit regression • Ordered logit • Ordered
probit • Probability density function
27.1 Introduction
This study explores the credit rating forecasting techniques for firms in Taiwan. We
employ the ordered logit and the ordered probit models for rating classification and
then a combining procedure to integrate both. We then examine empirically the
performance of these alternative methods, in particular, whether the combining
forecasting performs better than any individual method.
Credit rating forecasting has long time been very important for bond classifica-
tion and loan analysis. In particular, under the Basel II environment, regulators in
Taiwan have requested the banks to estimate the default probability of the loan
based on its credit classification. A proper forecasting procedure for credit rating of
the loan is crucially important in abiding the rule.
Different forecasting models and estimation procedures have various underlying
assumptions and computational complexities. They have been used extensively by
researchers in the literature. Review papers like Hand and Henley (1997), Altman
and Sounders (1997), and Crouhy et al. (2000) have traced the developments of the
credit classification and bankruptcy prediction models over the last two decades.
Since Beaver’s (1966) pioneered work, there have been considerable researches
on the subject of the credit risk. Many of them (Altman 1968; Pinches and Mingo
1973; Altman and Katz 1976; Altman et al. 1977; Pompe and Bilderbeek 2005)
27 Using Alternative Models and a Combining Technique in Credit Rating Forecasting 731
use the multivariate discriminant analysis (MDA) which assumes normality for
the explanatory variables of the default class. Zmijewski (1984) utilizes the probit
model, and Ohlson (1980) applies the logit model in which discrete or continuous data
can be fitted.
Kaplan and Urwitz (1979), Ederington (1985), Lawrence and Arshadi (1995),
and Blume et al. (1998) show that it is a consistent structure considering credit
rating as ordinal scale instead of interval scale. That is, the different values of the
dependent variables as different classes represent an ordinal, but not necessarily
a linear scale. For instance, higher ratings are less risky than lower ratings, but we
don’t have a quantitative measure indicating how much less risky they are.
Kaplan and Urwitz (1979) conduct an extensive examination of alternative
prediction models including N-chotomous probit analysis which can explain the
ordinal nature of bond ratings. To test the prediction accuracy of various statistical
models, Ederington (1985) compares the linear regression, discriminant analysis,
ordered probit, and unordered logit under the same condition. He concludes that
the ordered probit can have the best prediction ability and the linear regression is
the worst.
In a survey paper on forecasting methods, Mahmoud (1984) concludes that
combining forecasts can improve accuracy. Granger (1989) summarizes the use-
fulness of combining forecasts. Clemen (1989) observes that combining forecasts
increase accuracy, whether the forecasts are subjective, statistical, econometric, or
by extrapolation. Kamstra and Kennedy (1998) integrate two approaches with logit-
based forecast-combining method which is applicable to dichotomous, polychoto-
mous, or ordered-polychotomous contexts.
In this paper, we apply the ordered logit and the ordered probit models in credit
rating classification for listed firms in Taiwan and then combine two rating models
with a logit regression technique. The performance of each model is then evaluated
and we find that combining technique does improve the predictive power.
27.2 Methodology
Credit rating is an ordinal scale from which the credit category of a firm can be
ranked from high to low, but the scale of the difference between them is unknown.
To model the ordinal outcomes, let the underlying response function be
Y ¼ Xb þ e (27.1)
2
1 e
fðeÞ ¼ pffiffiffiffiffiffi exp (27.3)
2p 2
ðe 2
1 t
F ðeÞ ¼ pffiffiffiffiffiffiexp dt: (27.4)
1 2p 2
For the ordered logit model, e has a logistic distribution with mean 0 and variance
p2/3. The probability density function is
expðeÞ
lð e Þ ¼ (27.5)
½1 þ expðeÞ2
expðeÞ
LðeÞ ¼ : (27.6)
1 þ expðeÞ
To combine the ordered logit and the ordered probit models for credit forecasting,
the logit regression method as described in Kamstra and Kennedy (1998) is applied.
We first assume that firm’s credit classification is determined by an index y.
Suppose there are J rating classes, ordered from 1 to J. If y exceeds the threshold
value tj, j ¼ 1, . . ., j 1, credit classification changes from j rating to j + 1 rating.
The probability of company i being in rating j is given by the integral of a standard
logit from tj 1 yi to tj yi.
Each forecasting method is considered as producing J 1 measures
oji ¼ tj yi, j ¼ 1, . . ., j 1 for each firm. These measures can be estimated as
P1i þ þ Pji
oji ¼ ln (27.7)
1 P1i Pji
27 Using Alternative Models and a Combining Technique in Credit Rating Forecasting 733
where Pji is a probability estimate for firm i in rating j. The combining method
proposed by Kamstra and Kennedy (1998) consists of finding, via MLE, an
appropriate weighted average of o’s in ordered logit and ordered probit
techniques.1
To validate the model, we use cumulative accuracy profile (CAP) and its
summary statistics, the accuracy ratio (AR). A concept similar to the CAP is the
receiver operating characteristic (ROC) and its summary statistics, the area under
the ROC curve (AUC). In addition, we also employ the McFadden’s R2 (pseudo R2)
to evaluate the performance of the credit rating model. McFadden’s R2 is defined as
1(unrestricted log-likelihood function/restricted log-likelihood function).
Data are collected from the Taiwan Economic Journal (TEJ) database for the period
between the first quarter in 2000 and the third quarter in 2005, with the last three
quarters used for out-of-sample tests. The sample consists of firms traded in the
Taiwan Security Exchange (TSE) and the OTC market.
The credit rating of the sample firms is determined by the Taiwan Corporate
Credit Risk Index (the TCRI). Among ten credit ratings, 1–4 represent the invest-
ment grade levels, 5–6 represent the low-risk levels, and 7–9 represent the high-risk
or speculative grade levels. The final rating, 10, represents the bankruptcy level.
Table 27.1 exhibits the descriptive statistics for the samples which are divided
into three industry categories. Panel A contains the in-sample observations, while
panel B shows the out-of-sample observations. There are 509 bankruptcy cases in
the traditional industry, 411 in the manufacturing sector, and 191 in the electronics
industry for in-sample data. For out-of-sample data, there are 63 in the traditional,
38 in the manufacturing, and 72 in the electronics industries, respectively.
Table 27.2 displays the frequency distributions of the credit ratings for
in-samples in these three industries.
1
See Kamstra and Kennedy (1998) for the detail description.
734 C.-F. Lee et al.
Bonfim (2009) finds that not only the firms’ financial situation has a central role
in explaining default probabilities, but also macroeconomic conditions are very
important when assessing default probabilities over time. Based on previous studies
in the literature, 62 explanatory variables including financial ratios, market condi-
tions, and macroeconomic factors are considered. We use the hybrid stepwise
method to find the best predictors in the ordered probit and ordered logit models.
The combining technique using logit regression is then applied.
Table 27.3 Regression results estimated by the ordered logit. This table represents the
regression results estimated by the ordered logit model. Panel A shows the 11 explanatory vari-
ables fitted in the traditional industry. Panel B shows the nine explanatory variables fitted in the
manufacturing industry. Panel C shows the ten explanatory variables fitted in the electronics
industry
Parameters
Standard
Explanatory variables Estimates errors
Panel A: Traditional
X7 Fixed Asset to Long Term Funds Ratio 0.968*** (0.098)
X12 Accounts Receivable Turnover Ratio 0.054*** (0.0087)
X19 Net Operating Profit Margin 5.135*** (0.546)
X27 Return on Total Assets (Ex-Tax, Interest Expense) 5.473*** (0.934)
X29 Depreciation to Sales Ratio 6.662*** (0.537)
X30 Interest Expense to Sales Ratio 22.057*** (1.862)
X35 Free Cash Flow to Total Debt Ratio 0.971*** (0.094)
X41 Capital Spending to Gross Fixed Assets Ratio 0.746*** (0.167)
X46 Debt Ratio 4.781*** (0.305)
X47 Retained Earning to Total Assets Ratio 5.872*** (0.336)
X50 Ln (Total Assets/GNP price-level index) 6.991*** (1.024)
Panel B: Manufacturing
X2 Equity to Total Asset Ratio 7.851*** (0.317)
X15 Total Assets Turnover Ratio 0.714*** (0.172)
X27 Return on Total Assets (Ex-Tax, Interest Expense) 7.520*** (0.932)
X40 Accumulative Depreciation to Gross Fixed Assets 1.804*** (0.199)
X47 Retained Earning to Total Assets Ratio 3.138*** (0.340)
X50 Ln (Total Assets/GNP price-level index) 5.663*** (1.213)
X52 1:If Net Income was Negative for the Last Two Years 1.274*** (0.108)
0: Otherwise
X54 Ln (Age of the firm) 0.486*** (0.102)
X60 Ln (Net Sales) 1.063*** (0.065)
Panel C: Electronics
X2 Equity to Total Asset Ratio 3.239*** (0.207)
X27 Return on Total Assets (Ex-Tax, Interest Expense) 6.929*** (0.349)
X30 Interest Expense to Sales Ratio 25.101*** (1.901)
X35 Free Cash Flow to Total Debt Ratio 0.464*** (0.031)
X40 Accumulative Depreciation to Gross Fixed Assets 0.781*** (0.127)
X44 Cash Reinvestment Ratio 0.880*** (0.167)
X45 Working Capital to Total Assets Ratio 3.906*** (0.179)
X47 Retained Earning to Total Assets Ratio 3.210*** (0.174)
X49 Market Value of Equity/Total Liability 0.048*** (0.004)
X50 Ln (Total Assets/GNP price-level index) 8.797*** (0.704)
***
Represents significantly different from zero at 1 % level
**
Represents significantly different from zero at 5 % level
*
Represents significantly different from zero at 10 % level
736 C.-F. Lee et al.
Table 27.4 Regression results estimated by the ordered probit. This table represents the
regression results estimated by the ordered probit model. Panel A shows the ten explanatory
variables fitted in the traditional industry. Panel B shows the ten explanatory variables fitted in the
manufacturing industry. Panel C shows the nine explanatory variables fitted in the electronics
industry
Parameters
Explanatory variables Estimates Standard errors
Panel A: Traditional
X7 Fixed Asset to Long Term Funds Ratio 0.442*** (0.058)
X12 Accounts Receivable Turnover Ratio 0.027*** (0.005)
X19 Net Operating Profit Margin 2.539*** (0.311)
X27 Return on Total Assets (Ex-Tax, Interest Expense) 3.335*** (0.534)
X29 Depreciation to Sales Ratio 3.204*** (0.287)
X30 Interest Expense to Sales Ratio 9.881*** (1.040)
X34 Operating Cash Flow to Total Liability Ratio 0.514*** (0.121)
X41 Capital Spending to Gross Fixed Assets Ratio 0.459*** (0.096)
X46 Debt Ratio 3.032*** (0.188)
X47 Retained Earning to Total Assets Ratio 3.317*** (0.192)
Panel B: Manufacturing
X2 Equity to Total Asset Ratio 4.648*** (0.183)
X10 Quick Ratio 0.036*** (0.009)
X11 Accounts Payable Turnover Ratio 0.021*** (0.004)
X15 Total Assets Turnover Ratio 0.742*** (0.072)
X27 Return on Total Assets (Ex-Tax, Interest Expense) 4.175*** (0.522)
X40 Accumulative Depreciation to Gross Fixed Assets 0.989*** (0.112)
X43 Cash Flow Ratio 0.190*** (0.045)
X47 Retained Earning to Total Assets Ratio 1.804*** (0.188)
X52 1:If Net Income was Negative for the Last Two Years 0.753*** (0.061)
0: Otherwise
X54 Ln (Age of the firm) 0.289*** (0.056)
Panel C: Electronics
X2 Equity to Total Asset Ratio 1.851*** (0.117)
X27 Return on Total Assets (Ex-Tax, Interest Expense) 3.826*** (0.197)
X30 Interest Expense to Sales Ratio 11.187*** (1.052)
X35 Free Cash Flow to Total Debt Ratio 0.233*** (0.017)
X45 Working Capital to Total Assets Ratio 2.101*** (0.101)
X47 Retained Earning to Total Assets Ratio 1.715*** (0.098)
X49 Market Value of Equity/Total Liability 0.027*** (0.002)
X50 Ln (Total Assets/GNP price-level index) 4.954*** (0.399)
X54 Ln (Age of the firm) 0.247*** (0.028)
***
Represents significantly different from zero at 1 % level
**
Represents significantly different from zero at 5 % level
*
Represents significantly different from zero at 10 % level
738 C.-F. Lee et al.
Table 27.5 Threshold values estimated by the ordinal analysis. This table shows the threshold
values estimated by the ordered logit and the ordered probit models. There are nine threshold
parameters given ten credit ratings
Tables 27.6 and 27.7 illustrate the prediction results of the ordered logit and the
ordered probit models. Following Blume et al. (1998), we define the most probable
rating as the actual rating or its immediate adjacent ratings. The ratio of the number
of the predicted ratings as the most probable ratings to the total number of the
ratings being predicted can assess the goodness-of-fit for the model. For out-of-
sample firms, the predictive power of the ordered logit model for each industry is
86.85 %, 81.06 %, and 86.37 %, respectively; and the predictive power of the
ordered probit model for each industry is 86.42 %, 80.21 %, and 84.87 %,
respectively. The results from two models are quite similar.
Table 27.8 depicts the regression results using the Kamstra-Kennedy combining
forecasting technique. The coefficients are the logit estimates on o’s. These o
values are all positive and strongly significant for each industry.
27 Using Alternative Models and a Combining Technique in Credit Rating Forecasting 739
Table 27.9 shows the prediction results of the combining forecasting model. For
out-of-sample test, the predictive power of the combining model for each industry
is 89.88 %, 82.77 %, and 88.02 %, respectively, which are higher than those of the
ordered logit or ordered probit models by 2–4 %.
To evaluate the performance of each model, Fig. 27.1 illustrates the ROC curves
estimated by the three models, respectively.
From these ROC curves we can distinguish the performance of each rating
model. Furthermore, we can compare the AUC and AR calculated from the ROC
and CAP (See Table 27.10).
For the traditional industry, the AUCs from the ordered logit, the ordered probit, and
the combining model are 95.32 %, 95.15 %, and 95.32 %, respectively. For the
manufacturing industry, they are 94.73 %, 93.66 %, and 95.51 %, respectively. And
for the electronics industry, they are 92.43 %, 92.30 %, and 94.07 %, respectively. These
results apparently show that the combining forecasting model performs better than any
individual one.
27.4 Conclusion
a 1.0 b 1.0
0.8 0.8
Hit Rate
Hit Rate
0.6 0.6
c 1.0
0.8
0.6
Hit Rate
0.4
Source of the Curve
Ordered Logit
0.2 Ordered Probit
Combine Forecast
Random Model
0.0
0.0 0.2 0.4 0.6 0.8 1.0
False Alarm Rate
Fig. 27.1 ROC curves, (a) Traditional, (b) Manufacturing, (c) Electronics
are divided into three industries, i.e., traditional, manufacturing, and electronics,
for analysis. Sixty-two explanatory variables consisting of financial, market, and
macroeconomics factors are considered. We utilize the ordered logit, the ordered
probit, and the combining forecasting model to estimate the parameters
and conduct the out-of-sample tests. The main result is that the combining
forecasting method leads to a more accurate rating prediction than that of any
single use of the ordered logit or ordered probit analysis. By means of cumula-
tive accuracy profile, the receiver operating characteristics, and McFadden R2,
we can measure the goodness-of-fit and the accuracy of each prediction model.
These performance evaluations depict consistent results that the combining
forecast performs better.
Credit rating is an ordinal scale from which the credit category of a firm can be
ranked from high to low, but the scale of the difference between them is unknown.
To model the ordinal outcomes, we follow Zavoina and McKelvey (1975) to begin
with a latent regression
Y ¼ Xb þ e (27.8)
where
2 3 2 3
Y 1 1 X11 XK1
Y ¼ 4 ⋮ 5, X ¼ 4 ⋮ ⋮ ⋮ 5
Y N 1 X1N XKN
2 3 2 3
b0 e0
b ¼ 4 ⋮ 5, e ¼ 4 ⋮ 5:
bK eN
2
There are detailed discussion about ordered data in Ananth and Kleinbaum (1997), McCullagh
(1980), Wooldridge (2010), and Greene (2011).
3
In our case, J ¼ 10 and K ¼ 62.
27 Using Alternative Models and a Combining Technique in Credit Rating Forecasting 745
where Y is the one we do observe, an ordinal version of Y*, and the t’s are
unknown parameters to be estimated with b. We assume
that 1 ¼ t0 t1 tJ ¼ + 1.
From Eqs. 27.8 and 27.10, we have
tj1 < Y i tj , tj1 < Xi b þ ei tj
tj1 Xi b ei tj Xi b
, < (27.11)
s s s
where F(.) represents the cumulative density function of standard normal distri-
bution. The model (27.12) is under-identified since any linear transformation of
the underlying scale variable Y*, if applied to the parameters and t0,. . .,tJ as well,
would lead to the same model. We will assume without loss of generality
that,t1 ¼ 0 and s ¼ 1 in order to identify the model. The model we will estimate
turns out to be
PrðY i ¼ jÞ ¼ F tj Xi b F tj1 Xi b : (27.13)
1 if Y i ¼ j
Y i, j ¼ :
0 otherwise
Then, for simple notation, we set Zi,j ¼ tj Xib and Fi,j ¼ F(Zi,j). The likelihood
function, L, is
Y i, j
n Y
Y J
L ¼ Lðb0 , . . . , bK , t2 , . . . , tJ1 jY Þ ¼ Fi, j Fi, j1 : (27.14)
i¼1 j¼1
∂Fi, j ∂Zi, j
¼ N i, j ¼ N i, j Xu, i for 0 u K
∂bu ∂bu (27.16)
∂Fi, j ∂Zi, j
¼ N i, j ¼ N i, j dl, j for 2 l J 1
∂tl ∂tl
and
∂N i, j
¼ Z i, j N i, j Xu, i for 0 u K
∂bu (27.17)
∂N i, j
¼ Zi, j N i, j dl, j for 2 l J 1
∂tl
By using Eqs. 27.16 and 27.17, we can calculate the J + K 1 partial derivatives
of Eq. 27.15 with respect to the unknown parameters, b and t, respectively:
∂ln L X n X J
N i, j1 N i, j Xu, i
¼ Y i, j for 0 u K
∂bu i¼1 j¼1
Fi, j Fi, j1
(27.18)
∂ln L Xn X J
N i, j dl, j N i, j1 dl, j1 Xu, i
¼ Y i, j for 2 l J 1:
∂tl i¼1 j¼1
Fi, j Fi, j1
∂2 ln L Xn X J
¼ Y i, j
∂bu ∂bV i¼1 j¼1
2
Fi, j Fi, j1 Z i, j1 N i, j1 Z i, j N i, j N i, j1 N i, j Xu, i Xv, i
2 ,
Fi, j Fi, j1
"
∂2 ln L ∂2 ln L X n X J
Fi, j Fi, j1 Zi, j N i, j dl, j Zi, j1 N i, j1 dl, j1
¼ ¼ Y i, j 2
∂bu ∂tl ∂tl ∂bu i¼1 j¼1 Fi, j Fi, j1
#
N i, j1 N i, j N i, j dl, j N i, j1 dl, j1
2 Xu, i ,
Fi, j Fi, j1
27 Using Alternative Models and a Combining Technique in Credit Rating Forecasting 747
"
∂2 ln L X n X J
Fi, j Fi, j1 Zi, j1 N i, j1 dm, j1 dl, j1 Zi, j N i, j dm, j dl, j
¼ Y i, j 2
∂tl ∂tm i¼1 j¼1 Fi, j Fi, j1
#
N i, j dl, j N i, j1 dl, j1 N i, j dm, j N i, j1 dm, j1
2 :
Fi, j Fi, j1
(27.19)
We then set the J + K 1 equations in Eq. 27.18 to zero to get the MLE
of the unknown parameters. The matrix of second partials should be negative
definite to insure that the solution is a maximum. The computer program
NPROBIT, which uses the Newton-Raphson method, can solve the nonlinear
equations in Eq. 27.18.
Y ¼ Xb þ e (27.20)
expðeÞ
lð e Þ ¼ (27.21)
½1 þ expðeÞ2
expðeÞ
LðeÞ ¼ : (27.22)
1 þ expðeÞ
Y*, the dependent variable of theoretical interests, is unobserved, but from which
we can classify the category j:
where Y is the one we do observe, an ordinal version of Y*, and the t’s are unknown
parameters satisfying t1 . . . tj and to be estimated with b.
748 C.-F. Lee et al.
From Eqs. 27.20 and 27.22, we form the proportional odds model:
exp tj Xi b
PrðY i jjXÞ ¼ (27.24)
1 þ exp tj Xi b
or equivalently
Pij
log it Pij ¼ log
1 Pij (27.25)
PrðY i jjXi Þ
log ¼ tj X i b
PrðY i > jjXi Þ
where Pij ¼ Pr(Yi j). Notice that in the proportional odds model, b is assumed to
be constant and not depend on j. The validity of this assumption can be checked
based on a w2 score test. The model that relaxes the proportional odds assumption
can be represented as
log it Pij ¼ tj Xi bj , (27.26)
where the regression parameter vector b is allowed to vary with j. Both models can
be fit through the procedure of maximum likelihood estimation.
To combine the ordered logit and the ordered probit models for credit forecasting,
the logit regression method as described in Kamstra and Kennedy (1998) is applied.
We first assume that firm’s credit classification is determined by an index y.
Suppose there are J rating classes, ordered from 1 to J. If y exceeds the threshold
value tj, j ¼ 1,. . . J 1, credit classification changes from j rating to j + 1 rating.
The probability of company i being in rating j is given by the integral of a standard
logit from tj 1 yi to tj yi.
Each forecasting method is considered as producing J 1 measures,
oji ¼ tj y, j ¼ 1, . . ., J 1 for each firm. These measures can be estimated as
P1i þ þ Pji
oji ¼ ln (27.27)
1 P1i Pji
where Pji is a probability estimate for firm i in rating j. For firm i we have
8 oji
>
>
e
for j ¼ 1
>
> oji
< 1 þoeji
> oj1i
e e
Pji ¼ for j ¼ 2 J 1 : (27.28)
>
> 1 þ e oji 1 þ eoj1i
>
>
>
:
1
for j ¼ J
1 þ eoj1i
27 Using Alternative Models and a Combining Technique in Credit Rating Forecasting 749
In our case, there are two forecasting techniques A and B. For firm i, the
combining probability estimate is
8 pj þpA ojiA þpB ojiB
>
>
e
for j ¼ 1
>
> pj þpA ojiA þpB ojiB
> þ
< pjeþpA ojiA þpB ojiB
1
e epj1 þpA oj1iA þpB oj1iB
Pji ¼ for j ¼ 2 J 1 :
>
> 1þe p þp o þp o 1 þ epj1 þpA oj1iA þpB oj1iB
>
j A jiA B jiB
>
> 1
: for j ¼ J
1 þ epj1 þpA oj1iA þpB oj1iB
(27.29)
This can be estimated using an ordered logit software package with the oA and
oB values as explanatory variables and the t1 and p2 parameters playing the role of
the unknown threshold values.
References
Altman, E. I. (1968). Financial ratios discriminant analysis and the prediction of corporate
bankruptcy. Journal of Finance, 23, 589–609.
Altman, E. I., & Katz, S. (1976). Statistical bond rating classification using financial and account-
ing data. In Proceedings of the conference on topical research in accounting, New York.
Altman, E. I., & Sounders, A. (1997). Credit risk measurement: Developments over the last
20 years. Journal of Banking and Finance, 21, 1721–1742.
Altman, E. I., Hadelman, R. G., & Narayanan, P. (1977). Zeta analysis, a new model to identify
bankruptcy risk of corporations. Journal of Banking and Finance, 10, 29–51.
Ananth, C. V., & Kleinbaum, D. G. (1997). Regression models for ordinal responses: A review of
methods and applications. International Journal of Epidemiology, 26, 1323–1333.
Beaver, W. (1966). Financial ratios as predictors of failure. Journal of Accounting Research, 4,
71–111.
Blume, M. E., Lim, F., & Mackinlay, A. C. (1998). The declining credit quality of U.S. corporate
debt: Myth or reality? Journal of Finance, 53, 1389–1413.
Bonfim, D. (2009). Credit risk drivers: Evaluating the contribution of firm level information and of
macroeconomic dynamics. Journal of Banking and Finance, 33, 281–299.
Clemen, R. T. (1989). Combining forecasts: A review and annotated bibliography. International
Journal of Forecasting, 5, 559–583.
Crouhy, M., Galai, D., & Mark, R. (2000). A comparative analysis of current credit risk models.
Journal of Banking and Finance, 24, 59–117.
Ederington, L. H. (1985). Classification models and bond ratings. The Financial Review, 20,
237–261.
Granger, C. W. (1989). Invited review: Combining forecasts-twenty years later. Journal of
Forecasting, 8, 167–173.
Greene, W. H. (2011). Econometric analysis. NJ: Prentice Hall.
Hand, D., & Henley, W. (1997). Statistical classification methods in consumer credit scoring:
A review. Journal of the Royal Statistical Society, Series A, 160, 523–541.
Kamstra, M., & Kennedy, P. (1998). Combining qualitative forecasts using logit. International
Journal of Forecasting, 14, 83–93.
Kaplan, R. S., & Urwitz, G. (1979). Statistical models of bond ratings: A methodological inquiry.
Journal of Business, 52, 231–261.
750 C.-F. Lee et al.
Lawrence, E. C., & Arshadi, N. (1995). A multinomial logit analysis of problem loan resolution
choices in banking. Journal of Money, Credit and Banking, 27, 202–216.
Mahmoud, E. (1984). Accuracy in forecasting: A survey. Journal of Forecasting, 3, 139–159.
McCullagh, P. (1980). Regression models for ordinal data. Journal of the Royal Statistical Society,
42, 109–142.
Ohlson, J. A. (1980). Financial ratios and the probabilistic prediction of bankruptcy. Journal of
Accounting Research, 18, 109–131.
Pinches, G. E., & Mingo, K. A. (1973). A multivariate analysis of industrial bond ratings. Journal
of Finance, 28, 1–18.
Pompe, P. M., & Bilderbeek, J. (2005). The prediction of bankruptcy of small- and medium-sized
industrial firms. Journal of Business Venturing, 20, 847–868.
Wooldridge, J. M. (2010). Econometric analysis of cross section and panel data (2nd ed.).
Cambridge, MA: MIT Press.
Zavoina, R., & McKelvey, W. (1975). A statistical model for the analysis of ordinal level
dependent variables. Journal of Mathematical Sociology, 4, 103–120.
Zmijewski, M. E. (1984). Methodological issues related to the estimation of financial distress
prediction models. Journal of Accounting Research, 22, 59–68.
Can We Use the CAPM as an Investment
Strategy?: An Intuitive CAPM and 28
Efficiency Test
Contents
28.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 752
28.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 756
28.3 Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 759
28.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 762
28.4.1 Strategy 1: The Investor Buys All the Undervalued Stocks . . . . . . . . . . . . . . . . . . 762
28.4.2 Strategy 2: The Investor Buys the Three Stock-Quartiles with the Highest
Degree of Undervaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 763
28.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 766
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 767
Strategy 1: Investor Buys All Undervalued Stocks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 768
Strategy 2: The Investor Buys the Most Undervalued Stocks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 774
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 788
Abstract
The aim of this chapter is to check whether certain playing rules, based on the
undervaluation concept arising from the CAPM, could be useful as investment
strategies, and can therefore be used to beat the Market. If such strategies work,
we will be provided with a useful tool for investors, and, otherwise, we will
The authors wish to thank Professor José Vicente Ugarte for his assistance in the statistical analysis
of the data presented in this chapter.
F. Gómez-Bezares
Universidad de Deusto, Bilbao, Spain
e-mail: f.gomez-bezares@deusto.es
L. Ferruz (*)
Facultad de Economı́a y Empresa, Departamento de Contabilidad y Finanzas, Universidad de
Zaragoza, Zaragoza, Spain
e-mail: lferruz@unizar.es
M. Vargas
Universidad de Zaragoza, Zaragoza, Spain
e-mail: mvargas@unizar.es
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 751
DOI 10.1007/978-1-4614-7750-1_28,
# Springer Science+Business Media New York 2015
752 F. Gómez-Bezares et al.
obtain a test whose results will be connected with the Efficient Market Hypoth-
esis (EMH) and with the CAPM.
The basic strategies were set out in Gómez-Bezares, Madariaga, and
Santibáñez (Análisis Financiero 68:72–96, 1996). Our purpose now is to recon-
sider them, to improve the statistical analysis, and to examine a more recent
period for our study.
The methodology used is both intuitive and rigorous: analyzing how many
times we beat the Market with different strategies, in order to check whether
beating the Market happens by chance. Furthermore, we set out to study,
statistically, when and by how much we beat it, and to analyze whether this is
significant.
Keywords
ANOVA • Approximately normal distribution • Binomial distribution • CAPM •
Contingency tables • Market efficiency • Nonparametric tests • Performance
measures
28.1 Introduction
1
Sharpe (1964), Lintner (1965).
2
One of its main defenders has been Fama (1970, 1991, 1998).
3
There are also other four key ideas.
4
A fundamental concept, predating the other two and clearly related to the CAPM and the EMH.
5
See Gómez-Bezares and Gómez-Bezares (2006). Also of interest could be Dimson and
Mussavian (1998, 1999).
6
The interesting work by Danielson, Heck and Shaffer (2008) is also worth looking at.
28 Can We Use the CAPM as an Investment Strategy? 753
the asset pricing model and efficiency are tested jointly (Fama 1970, 1991, 1998), in
such a way that, when they work, they both work,7 but when they fail we cannot
know which of them has failed.8
The CAPM tests fall far short of giving definitive results. Although some have
given it up for dead and buried because it cannot fit the data or because of
theoretical problems, one recent magnificent study by Levy (2010) rejects
the theoretical problems posed by psychologists and behavioral economists and,
based on ex ante data, says that there is experimental support for the CAPM.
Brav, Lehavy, and Michaely (2005), in an approach related to the above, set out
to test the model, not by looking at past returns (as a proxy of expected returns) but
instead at the expectations of Market analysts, on the basis that these may be
assumed to be unbiased estimates of the Market expectations. From their analysis,
they found a positive relationship between expected returns and betas.
Asset pricing models (such as the CAPM) look at the relationship between return
and risk. Recently, there has been more focus on liquidity. Liu (2006) builds an
enhanced CAPM based on two factors: the Market and liquidity – he obtains their
corresponding factorial weighting which he uses to explain the expected returns.
The model works correctly and allows him to account several anomalies; according
to Liu, his enhanced CAPM lends new support to the risk-return paradigm.
The methods most commonly used to contrast the CAPM are time-series and
cross-sectional studies, which each present different statistical problems. What we
set out to achieve in this chapter is to replicate a simple and intuitive but also
rigorous methodology that we believe is much less problematic from a statistical
point of view. The procedure was first proposed by Gómez-Bezares, Madariaga,
and Santibáñez (1996), and in this chapter we attempt to replicate it with a more
updated sample and greater statistical rigor. The basic idea is simple: an individual
who knows the CAPM calculates at the beginning of each month the return that
each stock has rendered in the past and compares it with the return it ought to have
given according to the CAPM. The stocks which gave higher returns than they
ought to have are cheap (undervalued) while those which gave lower return are
expensive (overvalued). If we assume that return levels are going to persist, the
investor should buy the cheap stocks and sell off the expensive ones in order to beat
the Market. We illustrate this line of reasoning in Fig. 28.1.
In Fig. 28.1, the betas appear on the horizontal axis and the expected return on
the vertical axis (Rf is the riskless rate; Rm is the Market return). We also show
the Security Market Line (SML), with its formula above it, which is the formula
for the CAPM. Our investor decides to buy the stocks which are cheap and
refrain from selling short the expensive stocks (due to the limitations on short
selling). Based on the data at his disposal, the stocks which have gained a value
more than that the SML indicates are undervalued, that is to say, they are above
7
We refer to the situation we have described previously which occurred at the beginning of
the 1970s.
8
See also Copeland, Weston, and Shastri (2005, p. 244).
28 Can We Use the CAPM as an Investment Strategy? 755
Rf
1 βi
the SML. Believing that this situation will continue, he will buy all of these
stocks, trusting that he will obtain an adjusted return higher than the Market
return. If this happens (which he can find out by using Jensen’s alpha), and this
strategy consistently proves to result in abnormal returns (positive Jensen alphas,
or to put it another way, the stocks stay above the SML in the next period), he
will have found a way to beat the Market; therefore it is not efficient. Moreover,
he shall be able to predict which stocks will, in the future, outperform the
SML and therefore do not comply with the CAPM, which goes against the
CAPM. There would then be one way left to save the EMH: risk should not
be measured using the beta (i.e., the CAPM is mistaken) and therefore a positive
value for alpha is not synonymous with beating the Market (which could not be
done consistently in an efficient Market); we could also save the CAPM as
follows: risk must be measured with the beta; however, the Market fails to value
stocks accurately and hence it can be beaten; what we cannot do is to save both
the EMH and the CAPM simultaneously.
On the other hand, if our investor cannot consistently beat the Market with his
strategy but can only achieve about a 50 % success rate, more or less, we must
conclude that the portfolio assembled with the previously undervalued stocks
sometimes performs better than the SML and other times not as well, purely by
chance, and then settles down to the value it ought to have according to the
CAPM. We will not have a Market-beating strategy, and the results are compatible
with the EMH; likewise, we will see that on a random basis the portfolios perform
better than the SML at times and other times worse, which is compatible with the
CAPM (every month there may be random oscillations around the SML); therefore
two key elements of the aforementioned paradigm would be rescued.
In our research, we use Jensen’s (1968, 1969) and Treynor’s (1965) indices,
since they consider systematic risk. These indices have been very widely used in the
literature.9
9
Recent studies that use these indices are, for example, Samarakoon and Hasan (2005), Abdel-
Kader and Qing (2007), Pasaribu (2009), Mazumder and Varela (2010), and Ferruz, Gómez-
Bezares, and Vargas (2010).
756 F. Gómez-Bezares et al.
28.2 Data
Our analysis is based on the 35 securities that comprise the IBEX 35 (the official
index of the Spanish Stock Exchange Market) in each month from practically the
beginning of trading in Continuous Market in Spain (1989) up to March 2007.10 We
chose this index because it includes the 35 most liquid companies in the Spanish
Market and, taking into account that the CAPM in Sharpe-Lintner’s version is valid
for liquid companies, it would have been difficult to find such companies in Spain
outside the IBEX 35.
10
We have not extended our analysis beyond March 2007 in order to exclude the period of the
global financial crisis which had a very damaging effect on major companies listed on the
IBEX 35, such as banks, as this could have distorted our results.
28 Can We Use the CAPM as an Investment Strategy? 757
11
The average Market capitalization of a stock in the Index is the arithmetic average of the result
we obtain when multiplying the stocks allowed to be traded in each session during the monitoring
period by the closing price of the stock in each of those sessions.
12
The monitoring period is the 6-month period ending prior to each ordinary meeting of the
Commission.
758 F. Gómez-Bezares et al.
The data on returns for all the stocks were obtained from Bloomberg.
In Table 28.1, we show the annual descriptive statistics from our database.
There are two panels in Table 28.1: in panel A we show the descriptive statistics
for the monthly returns of the previous period (December 1989 to November 1992).
In this period, the index did not exist as such, so we have looked at the return
of those stocks which, being included in the IBEX 35 at the beginning of
the contrasting period, provide data in the corresponding month as they then were
quoted in the Continuous Market. In panel B, we bring together the descriptive
Table 28.1 Annual descriptive statistics for the monthly returns on the stocks
PANEL A: PRECEDING PERIOD
Table 28.1 reports the annual descriptive statistics for our database. In Panel A, we show the
figures for the monthly returns for the previous period and in Panel B the figures for monthly
returns for the contrasting period
For Panel B, we consider only the returns on stocks which we have included in our study, that is
to say, those stocks which were part of the IBEX 35 at a given time and which had a track record of
monthly returns of at least 36 months. For panel A, as the index as such did not yet exist, we
include those stocks listed on the IBEX 35 at the beginning of the contrasting period, for which
there is data in the corresponding previous month; for example, the stocks we included for the
month of December 1989 are those which were quoted on the IBEX 35 in December 1992 but
which also were quoted in the Continuous Market back in December 1989. For the subsequent
months the number of stocks considered in the study rises consistently since the number of stocks
quoted continued to grow
28 Can We Use the CAPM as an Investment Strategy? 759
data for the monthly returns during the contrasting period, that is, it includes the
returns on the stocks which are included in our study (namely, the stocks which
comprised the IBEX 35 at each moment and for which there were at least 36 prior
quotations).
We built the Market portfolio,13 and as a risk-free asset we took the monthly
return on the 1-month Treasury Bills.
28.3 Methods
The aim of our study is to check whether it is possible to obtain abnormal returns
using the CAPM, that is to say, whether the returns derived from the use of the
model are greater than those which could be expected according to the degree of
systematic risk undertaken.
To this end, we analyzed the period December 1989 to March 2007. The study
focuses on the 35 securities comprising the IBEX 35 at any given moment during
the said period.
From this starting point, we propose two possible strategies for testing the efficient
Market hypothesis and the validity of the CAPM: the first strategy assumes an
individual who adjusts his portfolio at the end of each month, selling those stocks he
bought at the end of the previous month and buying those he considers to be
undervalued according to the CAPM; the second strategy is similar to the first, but
with the distinction that in this case the investor does not buy all of the undervalued
stocks but just 75 % of them, namely, those which are furthest removed from
the SML. We use two methods to calculate which securities are the most undervalued:
Jensen’s ratio and Treynor’s ratio. The reason for using a second method is in response
to the critique by Modigliani (1997) of Jensen’s alpha, in which she argues that
differences in return cannot be compared when the risks are significantly different.
To conduct our analysis, we begin by dividing the overall period analyzed
in our study (December 1989 to March 2007) into two subperiods: the first
(December 1989 to November 1992) allows us to calculate the betas of the
model at the beginning of the following (we carried out this calculation with the
36 previous month data). The second subperiod (December 199214 to March 2007)
allows us to carry out the contrast.
13
This has been constructed as a simple average of returns for the stocks comprising the
IBEX 35 each month, corrected for dividend distributions, splits, etc. Of course, in this case, it
is not necessary that there be at least 36 months of prior quotations for the stocks to be included in
the portfolio.
14
Initially, we intended to let the beginning of the second subperiod be the launch date of the
IBEX 35 (January 1992), however, given the requirement that we set ourselves of a prior 36-month
period to compute the betas of the model, we had to move this forward to December. Likewise, it
was originally our intention that the beginning of the first subperiod be the launch date of the
Continuous Market (April 1989) but most of the securities included in our study were listed for the
first time in December 1989, so data could only be obtained from that date onward.
760 F. Gómez-Bezares et al.
So our hypothetical investor will firstly, at the end of each month, observe the
stocks that comprise the IBEX 3515 and will buy those which are undervalued
(in our second analysis method he will buy just 75 % of the stocks which are the
most undervalued). A stock is undervalued if its return is higher than what is ought
to be based on the CAPM.
To obtain this we compute the monthly betas for each stock (bi) during the
contrasting period by regressing the monthly returns on each stock16 on the returns
on the Market portfolio17 in the 36 months immediately prior to this. Then, we
calculate the mean monthly returns on stock i (Ri ), on the Market portfolio (RM )
and on the risk-free asset (RF ) during the same period as a simple average of the
corresponding 36 monthly returns. From all of the foregoing we can compute the
mean monthly return (Ri ), which, according to the CAPM, the stock ought to have
gained during this 36-month period,
Ri ¼ RF þ RM RF bi (28.1)
We then compare it with the actual return gained (Ri ) to determine whether the
stock is undervalued.
Once we have determined which stocks are undervalued, according to our first
strategy the investor buys all of the undervalued stocks each month,18 while with
the second strategy, he disregards the first quartile and just buys 75 % of the most
undervalued stocks. The quartiles are assembled based on either Jensen’s alphas or
Treynor’s ratio.19
The next step consists in evaluating how well the CAPM functions; if the
undervalued stocks continue to be undervalued indefinitely, we would expect any
15
And for which there is a minimum of 36-monthly data prior to the month analyzed so that the
beta can be calculated. In the case of companies which have recently merged, the beta is calculated
by looking at the weighted average monthly returns on the companies prior to the merger. The
weighting is proportional to the relative importance of each company involved in the merger.
16
We look at the return obtained by our investor from capital gains, dividends, and the sale of
preemptive rights. The returns are also adjusted to take splits into account.
17
The return on the Market portfolio is calculated based on a simple average of the returns gained
by the stocks which comprise the IBEX 35 in each month. The reason we decided to build our own
portfolio rather than relying on a stock Market index such as the IBEX 35 itself is that we wanted
to obtain an index corrected by capital gains, dividends, splits, and preemptive rights. Moreover,
an equally weighted portfolio is theoretically superior.
18
Overvalued stocks are not included in our analysis as we decided to exclude short selling.
19
In this second variant, although the quartiles are built based on Treynor’s ratio, we continue to
use Jensen’s index to work out whether or not a stock is undervalued. This is because of the
problems we have encountered when determining which stocks are undervalued using Treynor’s
ratio. These are related to the return premiums and the negative betas which appear in some cases
in our database. Naturally, we take on board the problems of using Treynor’s ratio to construct the
quartiles, as there could be stocks with positive alphas and low Treynor ratios which would
exclude such stocks from the analysis. This is due to the form of Treynor’s ratio as a quotient.
28 Can We Use the CAPM as an Investment Strategy? 761
portfolio assembled with them to beat the Market in a given month, but this will not
occur if in that month the CAPM operates perfectly.
Jensen’s index allows us to determine whether we have been able to beat the
Market, in which case it must yield a positive result:
ap ¼ Rp RF bp ðRM RF Þ (28.2)
where,
bp is the beta for the portfolio, calculated as a simple average of the individual
betas20 of the stocks included in the portfolio (all of the undervalued stocks, or
alternatively the top 75 % thereof as we have outlined above).
Rp is the return on the portfolio and is calculated as the simple average of the
individual returns obtained each month for the stocks that comprise it.
RM and RF are the monthly returns on the Market portfolio and the risk-free asset
respectively, for the month in question.
ap is Jensen’s alpha for the portfolio p.
The Z-statistic, which follows an approximately normal distribution, (0.1),
allows us to determine whether the number of months in which our investor beats
the Market is due to chance:
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
Z ¼ ðY npÞ= npð1 pÞ (28.3)
20
Which have been calculated with the 36 previous months’ data.
21
If the CAPM and the efficient Market hypothesis are fulfilled exactly, the results of any portfolio
(adjusted for risk) should match, as an average, with those of the Market. Here we suppose that, by
pure chance, for any given month, they will end up 50 % above and 50 % below the Market
average.
762 F. Gómez-Bezares et al.
Pearson test together with the likelihood ratio. Finally, we conducted regression
analysis of the performance measurement as a dependent variable and the time
dimension (months) as an independent variable to see whether the performance
improved or worsened over time. These tests allow us to confirm our results, which
we shall see in the next section.
28.4 Results
In this first strategy, the investor observes, in a given month, the track record (for
the previous 36 months) of the 35 stocks which comprise the IBEX 35 at the
moment in question,22 and buys all those that are undervalued according to Jensen’s
index. The next step is to check whether he has managed to beat the Market with the
portfolio of undervalued stocks.
Table 28.2 shows the results of this strategy. We show for each of the 14 years in
our study and for the other 4 months the number of months (and the percentage) in
which the investor’s portfolio has beaten the Market. In addition, we show this
result for the whole period. Furthermore, and also with respect to the whole period,
we show the result for the Z-statistic.
As we can see in Table 28.2, only in 92 of the 172 months analyzed in our study
is the Market beaten, which is equal to 53.49 %23; moreover, the figure for
Z-statistic (which is below 1.96) confirms that the CAPM does not offer
a significantly better strategy than investing in the Market portfolio. In other
words, we could say that the success of the investor (in the months in which he
succeeds in beating the Market) could be due to chance. This conclusion is further
confirmed in Appendix by the mean tests which allow us to accept that the mean of
Jensen’s alphas is zero and by the binomial test which allows accepting a success
probability rate of 50 %.
If we now focus on the analysis of each year in our time frame, we can see that in
only two of the 14 years covered in our study, to be precise from December 2004 to
November 2006, the CAPM proves to be a useful tool for the investor allowing him
to use a strategy to beat the Market; in fact, in 75 % and 83.33 % of the months of
each of those 2 years our investor beats the Market, thus confirming the poor
performance of the model, or the inefficiency of the Market.
22
Which have a track record of at least 36 months in the Market. If, in a given month, there is
a stock which is listed on the IBEX35 but does not have at least 36 months of prior data, this will be
included in our study in the month in which it reaches this figure, unless by that stage it has already
been deleted from the IBEX35.
23
We have not included the transaction costs, which undoubtedly would be higher for the managed
portfolio and, therefore, would make that strategy less attractive.
28 Can We Use the CAPM as an Investment Strategy? 763
Table 28.2 Results of the Strategy 1 (buying all of the undervalued stocks)
In the first 2 years (from December 1992 to November 1994), the opposite
happens: in just 33.33 % of the months the investor’s strategy beats the Market.
In those months it seems, therefore, that he could use the opposite strategy to beat it.
The statistical tests carried out and included in the Appendix do not provide us
with clear conclusions as to the basis for the differences in performance of the
investor’s strategy between the different months and years, because the different
tests used gave us different results. However, the regression analysis does confirm
that the performance of the portfolio improves over time, as we obtain a positive
and significant beta.
With this strategy, which has two variants, what lets us find out whether a stock is or
is not undervalued is, just as with the previous strategy, the difference between the
mean monthly return actually achieved by the stock in a given month (using
information on the previous 36 months) and that which it ought to have achieved
according to the CAPM. However, with this second strategy, the investor disregards
the first quartile of undervalued stocks and buys the other 75 %, those that are the
most undervalued.
764 F. Gómez-Bezares et al.
In the 1st variant of this second strategy, the ranking of the stocks (in each
month) by quartiles is done using Jensen’s index and in the 2nd variant it is done
using Treynor’s ratio.
Table 28.3 Results of Strategy 2, 1st Variant (buying the top three quartiles of the most
undervalued stocks, constructed based on Jensen’s Index)
The opposite results are obtained for the first 2 years (from December 1992 to
November 1994), in which the percentages of success (25 % and 33.33 %, respec-
tively) are the lowest.
The ANOVA, the Kruskal–Wallis test, and the contingency table with Pearson’s
chi-square-statistic and the likelihood ratio all lead us to the conclusion24 that there
are no differences between the performance means achieved by the portfolio in the
months in our analysis, so the small differences could be due to mere chance.
However, these tests do not allow us to offer conclusions when we look at the years
that comprise our sample, as different conclusions can be drawn from the different
tests used.
Moreover, the regression analysis once again allows us to confirm that, as with
the previous strategy, the performance of the portfolio improves over time.
Table 28.4 Results of Strategy 2, 2nd Variant (buying the top three quartiles of the most
undervalued stocks, constructed based on Treynor’s Ratio)
24
See Appendix.
766 F. Gómez-Bezares et al.
Moreover, the value for Z-statistic once again leads us to accept the hypothesis that
any possible success or failure is due to chance. Meanwhile, the tests we include
in Appendix confirm the previous results: the mean tests support the
interpretation that the mean performance of the portfolios is zero, therefore we do
not beat the Market, and the binomial test suggests that the strategy’s success rate
is 50 %.
Focusing now on the analysis by years, there are 3 years (December 2002 to
November 2003 and December 2004 to November 2006) in which we can confirm
that the Market is not efficient or the model does not work well, since in those years
the CAPM seems to be a useful Market-beating tool. However, in the first 2 years of
our study as well as in the ninth year, the lowest success rates are delivered, which
does not allow us to confirm the efficient Market hypothesis, as by using the
opposite strategy we could have beaten it.
With regard to the statistical tests to analyze the robustness of the above-
mentioned results for the various years and months in our database and those
shown in Appendix, we find that there are no significant differences between the
performance means achieved by the portfolio in the different months, so any
difference can be due to chance. Nonetheless, we cannot draw conclusions on
a yearly basis as the tests produce different results.
Finally, once again the result of the previous strategies is confirmed: the perfor-
mance of the investor’s portfolio improves over time. Overall, simply looking at the
graphics for the three strategies where we can see the Jensen alphas achieved by our
investor over time, if the reader disregards the initial data, he will see that the
adjusted line is flat.
28.5 Conclusion
The aim of our study was to analyze whether the design of strategies based on
the CAPM can enable an investor to obtain abnormal returns. We also set out
to explore this using a methodology that was both intuitive and scientifically
rigorous.
We also set out to determine whether the efficient Market hypothesis
was fulfilled and whether we could accept the validity of the CAPM, since if
strategies that can beat the Market with a degree of ease exist, we cannot
confirm either the efficiency of the Market or the workability of the CAPM, as
this defends that the only way to obtain extra returns is to accept a higher degree
of systematic risk. Conversely, we would then be in possession of a tool enabling
the investor to achieve higher returns than the Market for a given level of
systematic risk.
Analyzing the behavior of an investor who can buy the stocks comprising the
IBEX 35 at any given moment, that is, the benchmark for the Spanish stock
28 Can We Use the CAPM as an Investment Strategy? 767
Market, and who reconfigures his portfolio on a monthly basis, we find that,
regardless of the strategy used (buying all of the undervalued stocks or buying
the 75 % most undervalued stocks, either measured with Jensen’s alpha or with
Treynor’s ratio), the investor manages to beat the Market about 50 % of the time.
We opted to exclude from our calculations the transaction costs, so we in fact
overvalued the investor’s results. Therefore, we can conclude that the CAPM is
not an attractive tool for an investor who wishes to achieve abnormal returns.
It seems that undervalued stocks rapidly fit the SML, and so from another
perspective, we can confirm the efficient Market hypothesis and the workability
of the CAPM. These conclusions are backed up by a series of statistical tests
included in Appendix.
A positive aspect of our study from the point of view of its applicability is that
the behavior we have envisaged for our virtual investor is perfectly replicable as
he acts only with the information available in any given month. Meanwhile, by
focusing on the stocks that make up the IBEX 35, our study’s conclusions are
applicable to the most representative and consolidated stocks on the Spanish
Market.
Furthermore, our system possesses interesting statistical advantages: it allows
for variation of the beta and of the risk premium over time, it avoids the risk of
illiquidity, and it reduces errors in measurement. We have also considered its
robustness.
Finally, it is important to point out that at all times we have accepted implicitly
the logic of the CAPM, when measuring performance with Jensen’s index (which
implies basing ourselves on the CAPM). For this reason our results enable us to
confirm that it is the efficiency of the Market which prevents abnormally high
returns compared to those the Market itself is able to achieve, and that the CAPM
is a model that works reasonably well; hence it cannot be used as a Market-
beating tool.
Appendix
In this section, we report a series of statistical tests done using the “Stata” program,
which support the results obtained in our study, thereby, we believe, making it more
robust.25 These statistics are drawn up for each of the strategies we tried out: buying
all of the undervalued stocks, buying the top 75 % most undervalued stocks using
Jensen’s Index to select them, and buying the 75 % most undervalued stocks
according to Treynor’s Ratio.
25
See Agresti (2007), Anderson et al. (2011), Conover (1999) and Newbold et al. (2009) for
a widening of the statistical processing used in this Appendix.
768 F. Gómez-Bezares et al.
1. Summary statistics
We work here, as for all of the strategies, with the Jensen values recorded for each
of the portfolios (172).
Percentiles Smallest
1% −0.1064671 −0.1087249
5% −0.0279054 −0.1064671
10% −0.0180511 −0.0708467 Obs. 172
25% −0.0077398 −0.0540319 Sum of Wgt. 172
50% 0.0012046 Mean 0.001246
Largest Std. Dev. 0.0213603
75% 0.0123422 0.0389192
90% 0.0237416 0.0444089 Variance 0.0004563
95% 0.0309298 0.0509601 Skewness −1.397642
99% 0.0509601 0.0773151 Kurtosis 10.85719
We would point out that in this table the Jensen values are far from normality, as
can be seen in the readings for asymmetry and kurtosis.
2. Mean test
Variable Obs. Mean Std. Err. Std. Dev. [95% Conf. Interval]
jensen 172 0.001246 0.0016287 0.0213603 −0.001969 0.004461
From this mean test, we can accept that the mean for the Jensen alphas is zero; in fact,
if we focus on the two-sided test we obtain a probability of 0.4453, higher than 5 %.
This allows us to conclude that we are not able to beat the Market with this strategy.
3. Nonparametric mean test
We can see again that Wilcoxon’s nonparametric test leads to the same conclu-
sion: we cannot beat the Market with this strategy.
4. Frequencies
BEAT
BEAT
NO YES Total
80 0 80
NO
46.51 0 46.51
0 92 92
YES
0 53.49 53.49
80 92 172
Total
46.51 53.49 100
From this contingency table, we see that the proportion of months in which we
beat the Market is very similar to the proportion in which we fail to do so.
5. Probabilities test
Variable N Observed K Expected K Assumed p Observed p
Jensen 172 92 86 0.50000 0.53488
We use this binomial test to determine whether the probability of beating the Market
is similar to the probability of failing to beat it (50–50). If we look at the two-sided test,
we obtain a probability of 0.401684, above 5 %, and so we accept the success rate is
around 50 %, which supports the results obtained with the contingency table.
6. Fit of Jensen by month
.1
.05
jensen
−.05
−.1
7. Linear fit
From the above table we can observe that the regression slope which
connects Jensen’s index (the dependent variable) and time measured in months
(the independent variable) gives a positive and significant coefficient for
beta which allows us to conclude that this strategy gives results that improve
over time.
8. One-way analysis of Jensen by month
.1
.05
−.05
−.1
y
ry
ch
il
ay
er
r
us
be
be
be
r
l
ar
Ju
Ap
ua
ob
M
ar
Ju
g
nu
em
em
em
Au
M
br
ct
Ja
ov
pt
ec
Fe
Se
D
28 Can We Use the CAPM as an Investment Strategy? 771
We carry out an ANOVA to determine whether the means for Jensen’s indices
obtained in the different months are uniform. We can see that F has a value of 1.23
and a related probability of 0.2729, above 5 %, which leads us to accept that the
means for Jensen’s indices are uniform throughout the months included in our study.
10. Kruskal-Wallis tests (Rank sums), by month
Month Obs. Rank Sum
January 15 949
February 15 1551
March 15 1310
April 14 1217
May 14 1093
June 14 1386
July 14 1270
August 14 798
September 14 1382
October 14 1025
November 14 1032
December 15 1865
chi-squared = 22.716 with 11 d.f.
probability = 0.0194
We use this contingency table to test the same point as with the ANOVA and the
Kruskal–Wallis test above, to determine whether there are any differences between
the months, and we find that both with Pearson’s chi-square test and with the
likelihood ratio the probability is less than 1 %, which leads us to conclude that
the chances of beating the Market are not the same month on month, so that there
will be some months in which it is easier to do so than others.
12. One-way analysis of Jensen by year
.1
.05
−.05
−.1
93
94
95
96
97
98
99
00
01
02
03
04
05
06
19
19
19
19
19
19
19
20
20
20
20
20
20
20
28 Can We Use the CAPM as an Investment Strategy? 773
We now use the Kruskal–Wallis test to find out the same point as we did with the
ANOVA. We see that we obtain a probability greater than 5 % which means we can
accept that the mean Jensen values are uniform from one year to another.
15. Contingency table (by year)
BEAT
Year
NO YES Total
1993 8 4 12
1994 9 3 12
1995 6 6 12
1996 5 7 12
1997 5 7 12
1998 6 6 12
1999 5 7 12
2000 6 6 12
2001 6 6 12
2002 7 5 12
2003 5 7 12
2004 5 7 12
2005 4 8 12
2006 1 11 12
Total 78 90 168
Pearson chi2(13) = 15.2205 Pr = 0.294
likelihood-ratio chi2(13) = 16.7608 Pr = 0.210
The contingency table confirms the result of the previous test, namely, that there
are no differences between the Jensen values for the different years so that any
small difference can be due to chance.
Percentiles Smallest
1% −0.1414719 −0.1486773
5% −0.0391241 −0.1414719
10% −0.0192669 −0.1117398 Obs. 172
25% −0.0103728 −0.0683232 Sum of Wgt. 172
50% 0.0017188 Mean 0.0011871
Largest Std. Dev. 0.0280576
75% 0.0154674 0.0545513
90% 0.0294085 0.0690987 Variance 0.0007872
95% 0.0390722 0.0696097 Skewness −1.842765
99% 0.0696097 0.0745625 Kurtosis 11.929
28 Can We Use the CAPM as an Investment Strategy? 775
We would point out that in this table the Jensen values are far from normality, as
can be seen in the readings for asymmetry and kurtosis.
2. Mean test
Variable Obs. Mean Std. Err. Std. Dev. [95% Conf. Interval]
jensen (Q) 172 0.001187 0.0021394 0.0280576 −0.0030359 0.0054101
From this mean test we assume that the mean for Jensen’s alphas is zero, in fact,
if we focus on the two-sided test we can detect a probability of 0.5797, higher than
5 %. This allows us to conclude that we are not able to beat the Market with this
strategy.
3. Nonparametric mean test
We can see again, that Wilcoxon’s nonparametric test leads to the same conclu-
sion: we cannot beat the Market with this strategy.
4. Frequencies
BEAT
BEAT
NO YES Total
78 0 78
NO
45.35 0 45.35
0 94 94
YES
0 54.65 54.65
78 94 172
Total
45.35 54.65 100
776 F. Gómez-Bezares et al.
From this contingency table, we see that the proportion of months in which we
beat the Market is very similar to the proportion of months in which we do not, with
a slightly higher probability of beating the Market.
5. Probabilities test
With this binomial test we set out to find whether the probability of beating the
Market is the same as the probability of not beating it; (50–50). If we look at the
two-sided test, we get a probability of 0.252662, above 5 %, hence we can accept
that the success rate is 50 %, which backs up the results obtained with the
contingency table.
6. Fit of Jensen (quartiles) by month
.1
.05
0
qjensen
−.05
−.1
−.15
1992m1 1994m1 1996m1 1998m1 2000m1 2002m1 2004m1 2006m1 2008m1
time
28 Can We Use the CAPM as an Investment Strategy? 777
7. Linear fit
From the above table we can observe that the regression slope, which connects
Jensen’s index (the dependent variable) and time measured in months (the inde-
pendent variable), gives a positive and significant coefficient for beta which allows
us to conclude that this strategy gives results that improve over time.
8. One-way analysis of Jensen (quartiles) by month
.1
.05
−.05
−.1
−.15
y
ry
ch
ril
ay
er
r
us
be
be
be
l
ar
Ju
Ap
ua
ob
M
ar
Ju
g
nu
em
em
Au
M
br
ct
te
Ja
ov
ec
Fe
p
Se
D
778 F. Gómez-Bezares et al.
We perform an ANOVA to see whether the means for Jensen values obtained for
the different months are uniform. We find that F-statistic has a value of 0.84 with
a related probability of 0.5997, above 5 %, which leads us to accept that the means
for Jensen values are uniform among the months included in our sample.
10. Kruskal-Wallis tests (Rank sums), by month
We now use the Kruskal–Wallis test to find out the same point as we did with the
ANOVA. We come to the same conclusion as we did with the ANOVA.
28 Can We Use the CAPM as an Investment Strategy? 779
The contingency table allows us to test the same point as above with the ANOVA
and the Kruskal–Wallis test, and leads to the same conclusions; hence, we can accept
that all months show a similar tendency in terms of beating the Market.
12. One-way analysis of Jensen (quartiles) by year
.1
.05
−.05
−.1
−.15
93
94
95
96
97
98
99
00
01
02
03
04
05
06
19
19
19
19
19
19
19
20
20
20
20
20
20
20
780 F. Gómez-Bezares et al.
We now perform an ANOVA by years to see whether the strategy gives similar
Jensen values among the years in our sample. We obtain a value for F-statistic of
2.69 with a related probability of 0.0019; hence we conclude that the means for
Jensen values are not uniform from one year to another.
14. Kruskal-Wallis tests (Rank sums), by year
Year Obs. Rank Sum
1993 12 735
1994 12 665
1995 12 980
1996 12 1042
1997 12 1002
1998 12 1064
1999 12 1045
2000 12 972
2001 12 846
2002 12 913
2003 12 1136
2004 12 1070
2005 12 1353
2006 12 1373
chi-squared = 17.864 with 13 d.f.
probability = 0.1627
28 Can We Use the CAPM as an Investment Strategy? 781
We now use the Kruskal–Wallis nonparametric test to find out the same point as we
did with the ANOVA. We see that we obtain a probability greater than 5 % which
means we can accept that the mean Jensen values are uniform from one year to another.
15. Contingency table (by year)
BEAT
Year
NO YES Total
1993 8 4 12
1994 9 3 12
1995 6 6 12
1996 4 8 12
1997 6 6 12
1998 5 7 12
1999 4 8 12
2000 6 6 12
2001 7 5 12
2002 7 5 12
2003 3 9 12
2004 6 6 12
2005 2 10 12
2006 2 10 12
Total 75 93 168
Pearson chi2(13) = 19.9673 Pr = 0.096
likelihood-ratio chi2(13) = 21.0731 Pr = 0.071
The contingency table confirms the result of the previous test, namely, that there
are no differences between the Jensen values for the different years so that any
small difference can be due to chance.
Percentiles Smallest
1% −0.0968895 −0.104714
5% −0.0382262 −0.0968895
10% −0.0231222 −0.0940315 Obs. 172
25% −0.0106754 −0.0702515 Sum of Wgt. 172
50% 0.0023977 Mean 0.0014829
Largest Std. Dev. 0.0256145
75% 0.0157561 0.0484799
90% 0.0285186 0.0629996 Variance 0.0006561
95% 0.04065 0.0648316 Skewness −0.9274022
99% 0.0648316 0.0745625 Kurtosis 6.503481
782 F. Gómez-Bezares et al.
We note that in this table the Jensen values are far from normality, as can be seen in
the readings for asymmetry and kurtosis, although this is less obvious than for the
previous strategies.
2. Mean test
Variable Obs. Mean Std. Err. Std. Dev. [95% Conf. Interval]
Jensen (Q) 172 0.001483 0.0019531 0.0256145 −0.0023724 0.0053381
mean = mean (jensen) t = 0.7592
Ho: mean = 0 degrees of freedom = 171
From this mean test, we can accept that the mean for the Jensen alphas is zero; in
fact, if we focus on the two-sided test, we obtain a probability of 0.4487, which is
higher than 5 %. This allows us to conclude that we cannot beat the Market with this
strategy.
3. Nonparametric mean test
sign Obs. sum ranks expected
positive 93 8460 7439
negative 79 6418 7439
zero 0 0 0
all 172 14878 14878
unadjusted variance 427742.50
adjustment for ties 0.00
adjustment for zeros 0.00
--------------
adjusted variance 427742.50
Ho: jensen = 0
z = 1.561
Prob > |z| = 0.1185
We can see again that Wilcoxon’s nonparametric test leads to the same conclu-
sion: we cannot beat the Market with this strategy.
4. Frequencies
BEAT
BEAT
NO YES Total
79 0 79
NO
45.93 0 45.93
0 93 93
YES
0 54.07 54.07
79 93 172
Total
45.93 54.07 100
28 Can We Use the CAPM as an Investment Strategy? 783
We use this binomial test to determine whether the probability of beating the
Market is similar to the probability of failing to beat it (50–50). If we look at the
two-sided test, we obtain a probability of 0.321574, above 5 %, and so we accept
the success rate is around 50 %, which supports the results obtained with the
contingency table.
6. Fit of Jensen (quartiles) by month
.1
.05
qtreynor
−.05
−.1
7. Linear fit
From the above table we can observe that the regression slope which
connects Jensen’s index (the dependent variable) and time measured in months
(the independent variable) gives a positive and significant coefficient for
beta which allows us to conclude that this strategy gives results that improve
over time.
8. One-way analysis of Jensen (quartiles) by month
.1
.05
−.05
−.1
ry
ch
il
ay
er
r
us
be
be
be
r
l
ar
Ju
Ap
a
ob
M
ar
Ju
g
nu
ru
em
em
em
Au
M
ct
b
Ja
ov
pt
ec
Fe
Se
D
28 Can We Use the CAPM as an Investment Strategy? 785
We perform an ANOVA to see whether the means for the Jensen values obtained
for the different months are uniform. We find that F-statistic has a value of 0.90
with a related probability of 0.5387, above 5 %, which leads us to accept that the
means for the Jensen values are uniform among the months included in our sample.
10. Kruskal-Wallis tests (Rank sums), by month
Month Obs. Rank Sum
January 15 1153
February 15 1442
March 15 1297
April 14 1107
May 14 1046
June 14 1430
July 14 1220
August 14 899
September 14 1352
October 14 1160
November 14 1044
December 15 1728
chi-squared = 12.840 with 11 d.f.
probability = 0.3039
786 F. Gómez-Bezares et al.
We now use the Kruskal–Wallis nonparametric test to find out the same point as we
did with the ANOVA. We come to the same conclusion as we did with the ANOVA.
11. Contingency table (by month)
BEAT
Month
NO YES Total
January 7 8 15
February 5 10 15
March 6 9 15
April 9 5 14
May 8 6 14
June 3 11 14
July 7 7 14
August 10 4 14
September 7 7 14
October 6 8 14
November 8 6 14
December 3 12 15
Total 79 93 172
Pearson chi2 (11) = 15.8416 Pr=0.147
likelihood-ratio chi2(11) = 16.5468 Pr = 0.122
The contingency table allows us to test the same point as above with the ANOVA
and the Kruskal–Wallis test, and leads to the same conclusions; hence we can accept
that all months show a similar tendency in terms of beating the Market.
12. One-way analysis of Jensen (quartiles) by year
.1
.05
−.05
−.1
93
94
95
96
97
98
99
00
01
02
03
04
05
06
19
19
19
19
19
19
19
20
20
20
20
20
20
20
28 Can We Use the CAPM as an Investment Strategy? 787
We now perform an ANOVA by years to see whether the strategy gives similar
Jensen values among the years in our sample. We obtain a value for F-statistic of
2.14 with a related probability of 0.0147; hence we conclude that for a significance level
of 1 % we can accept that the means for the Jensen values are the same from one year to
another, however for a significance level of 5 % we must reject this hypothesis and we
would conclude that there are differences between the Jensen values of different years.
14. Kruskal-Wallis tests (Rank sums), by year
Year Obs. Rank Sum
1993 12 740
1994 12 679
1995 12 1025
1996 12 1033
1997 12 1065
1998 12 1022
1999 12 877
2000 12 1026
2001 12 824
2002 12 920
2003 12 1166
2004 12 1110
2005 12 1322
2006 12 1387
chi-squared = 18.336 with 13 d.f.
probability = 0.1452
788 F. Gómez-Bezares et al.
We now use the Kruskal–Wallis test to find out the same point as we did with the
ANOVA. We see that we obtain a probability greater than 5 % which means we can
accept that the means for the Jensen values are uniform from one year to another.
15. Contingency table (by year)
BEAT
Year
NO YES Total
1993 8 4 12
1994 9 3 12
1995 4 8 12
1996 5 7 12
1997 6 6 12
1998 5 7 12
1999 6 6 12
2000 4 8 12
2001 8 4 12
2002 7 5 12
2003 4 8 12
2004 6 6 12
2005 2 10 12
2006 2 10 12
Total 76 92 168
Pearson chi2 (13) = 19.9908 Pr = 0.095
likelihood-ratio chi2(13) = 21.0581 Pr = 0.072
The contingency table confirms the result of the previous test, namely, that there
are no differences between the Jensen values for the different years and that any
small difference can be due to chance.
References
Abdel-Kader, M. G., & Qing, K. Y. (2007). Risk-adjusted performance, selectivity, timing ability
and performance persistence of Hong Kong mutual funds. Journal of Asia-Pacific Business,
8(2), 25–58.
Agresti, A. (2007). An introduction to categorical data analysis (2nd ed.). Hoboken: Wiley
Interscience.
Anderson, D. R., Sweeney, D. J., & Williams, T. A. (2011). Statistics for business and economics
(11th ed.). Mason: South-Western, Cengage Learning.
Black, F., Jensen, M. C., & Scholes, M. (1972). The capital asset pricing model: Some empirical tests.
In Jensen, M. C. (Ed.), Studies in the theory of capital Markets (pp. 79–121). New York: Praeger.
Brav, A., Lehavy, R., & Michaely, R. (2005). Using expectations to test asset pricing models.
Financial Management, 34(3), 31–64.
Brealey, R. A., Myers, S. C., & Allen, F. (2008). Principles of corporate finance (9th ed.). New
York: McGraw-Hill.
Conover, W. J. (1999). Practical nonparametric statistics (3rd ed.). New York: Wiley.
Copeland, T. E., Weston, J. F., & Shastri, K. (2005). Financial theory and corporate policy
(4th ed.). Boston: Pearson Addison Wesley.
28 Can We Use the CAPM as an Investment Strategy? 789
Danielson, M. G., Heck, J. L., & Shaffer, D. R. (2008). Shareholder theory – How opponents and
proponents both get it wrong. Journal of Applied Finance, 18(2), 62–66.
Dimson, E., & Mussavian, M. (1998). A brief history of Market efficiency. European Financial
Management, 4(1), 91–103.
Dimson, E., & Mussavian, M. (1999). Three centuries of asset pricing. Journal of Banking and
Finance, 23(12), 1745–1769.
Fama, E. F. (1970). Efficient capital markets: A review of theory and empirical work. Journal of
Finance, 25(2), 383–417.
Fama, E. F. (1991). Efficient capital markets: II. Journal of Finance, 46, 1575–1617.
Fama, E. F. (1998). Market efficiency, long-term returns, and behavioral finance. Journal of
Financial Economics, 49, 283–306.
Fama, E. F., & French, K. R. (1992). The cross-section of expected stock returns. Journal of
Finance, 47, 427–465.
Fama, E. F., & French, K. R. (2008). Dissecting anomalies. Journal of Finance, 63(4), 1653–1678.
Fama, E. F., & French, K. R. (2010). Luck versus skill in the cross-section of mutual fund returns.
Journal of Finance, 65(5), 1915–1947.
Fama, E. F., & MacBeth, J. D. (1973). Risk, return and equilibrium: empirical tests (pp. 607–636).
May-June: Journal of Political Economy.
Ferruz, L., Gómez-Bezares, F., & Vargas, M. (2010). Portfolio theory, CAPM, and performance
measures. In Lee, C. F. et al. (Eds.), Handbook of quantitative finance and risk management
(pp. 267–283). New York: Springer.
Gómez-Bezares, F., & Gómez-Bezares, F. R. (2006). On the use of hypothesis testing and other
problems in financial research. The Journal of Investing, 15(4), 64–67.
Gómez-Bezares, F., Madariaga, J. A., & Santibáñez, J. (1996). Valuation and efficiency models:
Does the CAPM beat the Market? Análisis Financiero, 68, 72–96.
Jensen, M. C. (1968). The performance of mutual funds in the period 1945–1964 (pp. 389–416).
May: Journal of Finance.
Jensen, M. C. (1969). Risk, the pricing of capital assets, and the evaluation of investment
portfolios. Journal of Business, 42, 167–247.
Levy, H. (2010). The CAPM is alive and well: A review and synthesis. European Financial
Management, 16(1), 43–71.
Lintner, J. (1965). The valuation of risk assets and the selection of risky investments in stock
portfolios and capital budgets. Review of Economics and Statistics, 47, 13–37.
Liu, W. (2006). A liquidity-augmented capital asset pricing model. Journal of Financial Econom-
ics, 82, 631–671.
Malkiel, B. G. (2003). The efficient Market hypothesis and its critics. Journal of Economic
Perspectives, 17(1), 59–82.
Mazumder, M. I., & Varela, O. (2010). Market timing, the trading of international mutual funds:
weekend, weekday and serial correlation strategies. Journal of Business, Finance & Account-
ing, 37(7–8), 979–1007.
Modigliani, L. (1997). Don’t pick your managers by their alphas. U.S. Investment Research,
U.S. Strategy, Morgan Stanley Dean Witter, June.
Newbold, P., Carlson, W. L., & Thorne, B. M. (2009). Statistics for business and economics
(7th ed.). New Jersey: Pearson Education Canada.
Pasaribu, R. B. (2009). Relevance of accrual anomaly information in stock portfolio
formation – case study: Indonesia stock exchange. Journal of Accounting and Business,
10(2), 1–45.
Samarakoon, L. P., & Hasan, T. (2005). Portfolio performance evaluation. In Lee, C. F. (Ed.), The
encyclopedia of finance (pp. 617–622). New York: Springer.
Sharpe, W. F. (1964). Capital asset prices: A theory of Market equilibrium under conditions of
risk. Journal of Finance, 19, 425–442.
Treynor, J. L. (1965). How to rate management of investment funds. Harvard Business Review, 43,
63–75.
Group Decision-Making Tools for
Managerial Accounting and Finance 29
Applications
Contents
29.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 793
29.2 Designing a Comprehensive Performance Evaluation System: Using the Analytic
Hierarchy Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 793
29.2.1 Hierarchical Schema for Performance Measurement . . . . . . . . . . . . . . . . . . . . . . . . . 795
29.2.2 Analytic Hierarchical Performance Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 796
29.2.3 An Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 798
29.2.4 Discussions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 800
29.2.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 801
29.3 Developing a Comprehensive Performance Measurement System in the Banking
Industry: An Analytic Hierarchy Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 801
29.3.1 Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 802
29.3.2 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 803
29.3.3 A Numerical Example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 804
29.3.4 Summary and Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 807
W. Kwak (*)
University of Nebraska at Omaha, Omaha, NE, USA
e-mail: wkwak@unomaha.edu
Y. Shi
University of Nebraska at Omaha, Omaha, NE, USA
Chinese Academy of Sciences, Beijing, China
e-mail: yshi@unomaha.edu
C.-F. Lee
Department of Finance and Economics, Rutgers Business School, Rutgers, The State University of
New Jersey, Piscataway, NJ, USA
Graduate Institute of Finance, National Chiao Tung University, Hsinchu, Taiwan
e-mail: cflee@business.rutgers.edu
H. Lee
Korea Advanced Institute of Science and Technology, Yuseong-gu, Daejeon, South Korea
e-mail: hsl@business.kaist.ac.kr
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 791
DOI 10.1007/978-1-4614-7750-1_29,
# Springer Science+Business Media New York 2015
792 W. Kwak et al.
Abstract
To deal with today’s uncertain and dynamic business environments with differ-
ent background of decision makers in computing trade-offs among multiple
organizational goals, our series of papers adopts an analytic hierarchy process
(AHP) approach to solve various accounting or finance problems such as devel-
oping a business performance evaluation system and developing a banking
performance evaluation system. AHP uses hierarchical schema to incorporate
nonfinancial and external performance measures. Our model has a broader set of
measures that can examine external and nonfinancial performance as well as
internal and financial performance. While AHP is one of the most popular
multiple goals decision-making tools, multiple-criteria and multiple-constraint
(MC2) linear programming approach also can be used to solve group decision-
making problems such as transfer pricing and capital budgeting problems. This
model is rooted by two facts. First, from the linear system structure’s point of
view, the criteria and constraints may be “interchangeable.” Thus, like multiple
criteria, multiple-constraint (resource availability) levels can be considered.
Second, from the application’s point of view, it is more realistic to consider
multiple resource availability levels (discrete right-hand sides) than a single
resource availability level in isolation. The philosophy behind this perspective
is that the availability of resources can fluctuate depending on the decision
situation forces, such as the desirability levels believed by the different
managers. A solution procedure is provided to show step-by-step procedure to
get possible solutions that can reach the best compromise value for the multiple
goals and multiple-constraint levels.
Keywords
Analytic hierarchy process • Multiple-criteria and multiple-constraint linear
programming • Business performance evaluation • Activity-based costing sys-
tem • Group decision making • Optimal trade-offs • Balanced scorecard •
Transfer pricing • Capital budgeting
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 793
29.1 Introduction
AHP is one of the most popular multiple goals decision-making tools (Ishizaka
et al. 2011). Designing a comprehensive performance measurement system has
frustrated many managers (Eccles 1991). The traditional performance measures
enterprises have used may not well fit in for the new business environment and
competitive realities. The figures that enterprises have traditionally used are not
very useful for the information-based society we are becoming. We suspect that
firms are much more productive than these out-of-date measures.
A broad range of firms is deeply engaged in redefining how to evaluate the
performance of their businesses. New measurements for quantification are needed
to perform business evaluation. Drucker (1993) put the ever-increasing measure-
ment dilemma this way:
Quantification has been the rage in business and economics these past 50 years. Accoun-
tants have proliferated as fast as lawyers. Yet we do not have the measurements we need.
Neither our concepts nor our tools are adequate for the control of operations, or for
managerial control. And, so far, there are neither the concepts nor the tools for business
control, i.e., for economic decision making. In the past few years, however, we have
become increasingly aware of the need for such measurements.
The AHP is a theory of measurement that has been extensively applied in modeling
the human judgment process (e.g., Lee 1993; Muralidhar et al. 1990). It decom-
poses a complex decision operation into a multilevel hierarchical structure. The
primary advantage of the AHP is its simplicity and the availability of the software.
Several desirable features of the AHP can help resolve issues in performance
evaluation. For example, nonfinancial and external effects can be identified and
integrated with financial and internal aspects of business performance through the
AHP. Furthermore, the AHP is a participation-oriented methodology that can aid
coordination and synthesis of multiple evaluators in the organizational hierarchy.
Participation makes a positive contribution to the quality of the performance
evaluation process. This point is further explored within the context of hierarchical
schema of performance evaluation as follows.
Performance measures have the relationship with management levels. They need
to be filtered at each superior/subordinate level in an organization; i.e., measures do
not need to be the same across management levels. Performance measures at each
level, however, should be linked to performance measures at the next level up.
Performance measurement information is tailored to match the responsibility of
each management level. For example, at the highest level, the CEO has responsi-
bility for performance of the total business. In contrast, the production manager’s
main interest may be in cost control because he or she is responsible for this. Taking
this idea into account leads to the notion that the performance measuring process
consists of different levels according to management levels.
Depending on organizational levels, therefore, a three management level model
will be suggested: top, middle, and operational. To recognize the ever-increasing
importance of nonfinancial measures, at the top management level, the hierarchy
consists of two criteria: nonfinancial and financial performance. One level lower
(middle-level management) includes performance criteria such as market share,
customer satisfaction, productivity, ROI, and profitability.
The lowest level (operational management level) includes measures that lead to
simplifications in the manufacturing process as related to high-level performance
measures. Examples are quality, delivery, cycle time, inventory turnover, asset turn-
over, and cost. Typically, the criteria at the lowest level may have several sub-criteria.
For example, quality measures may have four sub-criteria: voluntary, (appraisal and
prevention,) and failure, (internal and external,) costs. With relation to these criteria,
accounting information is controlled typically with respect to each product (or service)
or division (or department). As a result, product and division levels are added. One
should keep in mind that the above three-level hierarchical schema is dynamic over
time. As a company evolves, the hierarchy must be accordingly adjusted.
796 W. Kwak et al.
Another interesting aspect is that the hierarchy is not company invariant. The
hierarchy must be adjusted depending on unique situation faced by each individual
company or division. Basically, the hierarchal schema is designed to accommodate
any number of levels and alternatives. New level or alternative can be easily added or
deleted to the hierarchy once introduced. For example, another level regarding prod-
ucts may be added at the bottom in order to evaluate performance of different products.
Based on the hierarchical structure, the performance indices at each level of the
AHPM are derived by the analytic hierarchy process. The AHPM collects input
judgment in the form of matrix by pairwise comparisons of criteria. An eigenvalue
method is then used to scale weights of criteria at each level; i.e., the relative
importance of each criterion at each level is obtained. The relative importance is
defined as a performance index with respect to each alternative (i.e., criterion,
product, or division).
From now on, each step of the AHPM in obtaining the weights is explored. First,
a set of useful strategic criteria must be identified. Let nt be the total number
of criteria under consideration at the top management level. Typically,
nt ¼ 2 (nonfinancial and financial measures). The relative weight of each criteria
may be evaluated by pairwise comparison; i.e., two criteria are compared at one time
until all combinations of comparison are considered (only one pairwise comparison
is needed if nt ¼ 2). The experience of many users of this method and the experiments
reported (Saaty 1980) are likely to support that the 1–9 scale for pairwise compar-
isons captures human judgment fairly well while the scale can be altered to suit each
application. The result from all of pairwise comparisons is stored in an input matrix:
At ¼ atij ðan nt by nt matrixÞ:
weight for each criterion with relation to the ith criteria at one level higher. This
“local relative weight” is stored in a local weighing vector Wim. We need nt local
weighing vectors at this level. The “global relative weight” at this level is computed
and then stored in a vector Wm ¼ (wmi) that has the nm elements as
W m ¼ W t W 1m . . . W nt
m :
Here, DCmi is the amount of change of the ith criteria. For example, if the ROI is
increased from 5 % to 7 % and the market share is changed from 20 % to 25 % in the
automobile company discussed above,
2 5
em ¼ 0:25 þ 0:75 ¼ 0:2875:
5 20
We may conclude that the overall business performance has been increased by
$28.75 % with respect to middle management level. Similarly, the performance
change rates at any level can be obtained. Typically, they vary depending on levels
of the AHPM; i.e., they have different implications.
798 W. Kwak et al.
29.2.3 An Example
Wt ¼ ð0:4; 0:6Þ:
For the sake of convenience, ROI and profitability are not listed in this compar-
ison matrix.
The lower triangle is not listed because, in the eigenvalue method, the lower
triangle is simply the reciprocity of the upper triangle. As a result,
For each weighing computing, an inconsistency ratio was computed and checked
for acceptance; i.e., in this case, the ratio (g ¼ 0.017) was accepted because g 0.1.
For the financial criteria, ROI is estimated to be twice more important than
profitability, i.e.,
W 2m ¼ ð0:667, 0:333Þ:
Accordingly, the global relative weights of the managerial criteria (in the order
of market share, customer satisfaction, productivity, ROI, and profitability) are then
0:558 0:320 0:122 0 0
Wm ¼ ð0:4, 0:6Þ
0 0 0 0:667 0:333
¼ ð0:223, 0:128, 0:049, 0:400, 0:200Þ:
Here, the global relative weights of market share, customer satisfaction, produc-
tivity, ROI, and profitability are 22.3 %, 12.8 %, 4.9 %, 40 %, and 20 %, respec-
tively. Note that these percentages are elements of the above Wm.
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 799
Let us move down to the operational level. The following are local relative
weights of operational level criteria (in the order of quality, delivery, cycle time,
cost, inventory turnover, and asset turnover).
For the market share,
2 3
1 2 3 3 5 7
6 1=2 1 2 3 4 57
6 7
6 1=3 1=2 1 2 4 57
A0 ¼ 6
1 6 7
6 1=3 1=3 1=2 1 2 377
4 1=5 1=4 1=4 1=2 1 25
1=7 1=5 1=5 1=3 1=2 1
For convenience, the local weights are arranged in the order of quality, cycle
time, delivery, cost, inventory turnover, and asset turnover:
Similarly,
29.2.4 Discussions
It is too optimistic to argue that there can be one right way of measuring perfor-
mance. There are two factors to be considered. First, each business organization
requires its own unique set of measures depending on its environment. What is most
effective for a company depends upon its history, culture, and management style
(Eccles 1991). To be used for any business performance measurement, a well-
designed model must be flexible enough to incorporate a variety of measures while
retaining major aspects. Second, managers should change measures over time.
In the current competitive business environment, change is not incidental. It is
essential. Change must be managed. The AHPM is highly flexible so that managers
may adjust its structure to evolving business environments. This flexibility will
allow a company to improve its performance measurement system continuously.
Generally, a performance measurement system exists to monitor the implemen-
tation of planning of an organization and aid to motivate desirable individual
performance through a realistic communication of performance information in
related goals of business. This premise of performance measurement requires
a significant number of feedbacks and corrective actions in the practice of account-
ing information systems (Nanni et al. 1990), i.e., feedbacks between levels
and also within level. In the implementation of AHPM, these activities are
common procedures. A clean separation of levels in the hierarchy of the AHPM
is an idealization for simplifying the presentation. The flexibility of the AHPM,
however, accommodates as many feedbacks and corrective actions as possible.
Performance is monitored by group rather than by an individual because of
its ever-increasing importance of teamwork in the success of businesses. The
integrated structure of the AHPM facilitates group decision and thus increases the
chance that managers put trust in the resulting measures. This structure will
enhance more involvement of lower-level managers as well as workers
when a firm implements the AHPM. Furthermore, the hierarchy of the AHPM
corresponds to that of business organization. As a result, group decision at each
level is facilitated. For example, middle-level managers will be responsible for
determining weights for middle-level criteria while lower-level managers will be
responsible for operational level criteria.
The iterative process of weighing goals among managers, as the implementation
of AHPM progresses, will help them to understand which strategic factors are
important and how these factors are linked to other goals to be a successful
company as a group.
Information technology plays a critical role in designing a performance mea-
surement system to provide timely information to management. A computer-based
decision support system can be used to utilize this conceptual foundation in a real-
world situation. The AHPM can be easily stored in an enterprise database because
of the commercial software, Expert Choice. As a result, the AHPM can be readily
available to each management level via the network system of an enterprise.
The AHPM fits any corporate information architecture to pursue the company’s
long-term strategy.
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 801
29.2.5 Conclusions
(Liberatore et al. 1992), and real estate investment (Kamath and Khaksari 1991),
marketing applications (Dyer and Forman 1991), information system project
selection (Schniederjans and Wilson 1991), activity-based costing cost driver
selection (Schniederjans and Garvin 1997), developing lean performance measures
(DeWayne 2009), and customer’s choice analysis in retail banking (Natarajan
et al. 2010)). The AHP is relatively easy to use and its commercial software is
available. This study will be an analytical and comprehensive performance evaluation
model to cover a broader base of measures in the rapidly changing environment of
today’s banking industry. The following section presents background. The third
section discusses methodology. The fourth section presents a numerical example.
The last section summarizes and concludes this chapter.
29.3.1 Background
Recent research topic guide from Institute of Management Accountants lists “per-
formance measurement” as one of top priority research issues. Therefore, this
project will be interesting to bank administrators as well as managerial accountants.
With unprecedented competitive pressure from nonbanking institutions, dereg-
ulation, and the rapid acquisition of smaller banks by large national or regional
banks, the most successful banks in the new millennium will be the ones that adapt
strategically to a changing environment (Calvert 1990). Management accounting
and finance literature have emphasized using both financial and nonfinancial
measures as performance guidelines in the new environment (e.g., Chan and
Lynn 1991; Rotch 1990). However, most studies do not propose specifically how
we should incorporate these financial and nonfinancial factors into a formal model.
The performance measurement system proposed in this study is the formal
model applied using the AHP in the banking industry to cover a wide variety of
measures while providing operational control as well as strategic control. The AHP
can incorporate multiple-subjective goals into a formal model (Dyer and Forman
1991). Unless we design a systematic performance measurement system that
includes financial as well as nonfinancial control factors, there may be incorrect
behavior by employees because they misunderstand the organization’s goals and
how they relate to their individual performance.
Compared with previous evaluation methods, the model proposed in this study will
have advantages such as flexibility, continuous feedback, teamwork in goal setting,
and computational simplicity. To be used for any business performance measurement,
a well-designed model must be flexible enough to incorporate a variety of measures
while retaining major success factors. The AHP model is flexible enough for managers
to adjust its structure to a changing business environment through an iterative process
of weighing goals. This flexibility will allow a company to improve its performance
measurement system continuously. Through the iterative process of goal comparisons,
management could get continuous feedback for the priority of goals and work as
a team. The possible real-time control is of importance in the competitive business
environment we are facing today.
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 803
29.3.2 Methodology
The analytic hierarchy process (AHP) collects input judgments in the form of a matrix
by pairwise comparisons; i.e., two criteria are compared at one time. The experience of
many users of this method supports use of a 1–9 scale for pairwise comparisons to capture
human judgment while the scale can be altered to fit each application (Saaty 1980).
A simple example will be provided to explain how AHP operates. Consider the
situation where a senior executive has to decide on which of three managers to
promote to a senior position in the firm. The candidate’s profiles have been studied
and rated on three criteria: leadership, human relations skills, and financial man-
agement ability. First, the decision maker compares each of the criteria in pairs to
develop a ranking of the criteria. In this case, the comparisons would be:
1. Is leadership more important than human relations skills for this job?
2. Is leadership more important than financial management ability for this job?
3. Are human relations skills more important than financial management ability for
this job?
The response to these questions would provide an ordinal ranking of the three
criteria. By adding a ratio scale of 1–9 for rating the relative importance of one criterion
over another, a decision maker could make statements such as “leadership is four times
as important as human relations skills for this job,” “financial management ability is
three times as important as leadership,” and “financial management ability is seven
times as important as human relations skills.” These statements of pairwise compar-
isons can be summarized in a square matrix. The preference vectors are then computed
to determine the relative rankings of the three criteria in selecting the best candidate.
For example, the preference vectors of the three criteria are 0.658 for financial
management ability, 0.263 for leadership, and 0.079 for human relations skills.
Once the preference vector of the criteria is determined, each of the candidates
can be compared on the basis of the criteria in the following manner:
1. Is candidate A superior to candidate B in leadership skills?
2. Is candidate A superior to candidate C in leadership skills?
3. Is candidate B superior to candidate C in leadership skills?
Again, rather than using an ordinal ranking, the degree of superiority of one
candidate over another can be assessed. The same procedures can be applied to
human relations skills and financial management ability. The responses to these
questions can be summarized in matrices where the preference vectors are again
computed to determine the relative ranking of the three candidates for each criterion.
Accordingly, the best candidate should be the one who ranks “high” on the “more
important” criteria. The matrix multiplication of preference vectors of candidates on
evaluation criteria and the preference vector of evaluation criteria will provide the
final ranking of the candidates. In this example, the candidates are ranked A, B,
and C. This example provides the usefulness of the AHP for setting priorities for
both qualitative and quantitative measures (Chan and Lynn 1991).
We could apply the same procedures to a bank. Based on the hierarchical
structure of a banking institution, the relative weights of criteria at each level of
managers are derived by AHP. Here the relative importance of performance
804 W. Kwak et al.
This section presents a numerical example for a commercial bank. The Commercial
Omaha Bank (COB) is a local bank that specializes in commercial loans. Their head-
quarters are located in Omaha, Nebraska, and they have several branch offices through-
out rural areas of Nebraska. The top management of COB realized that the current
measurement system is not adequate for their strategic performance management and
identified the following measures based on the hierarchy of the organization for their
new performance measurement using AHP. These measures are shown in Table 29.1.
The OCB uses financial criteria such as return on assets and debt to equity and
nonfinancial criteria such as market share, productivity, and quality of service. At
the lowest management level, income to interest expense, service charges, interest
revenue, growth of deposits, default ratio, and customer satisfaction can be used.
Each computing step of the AHP is discussed as follows.
First, nonfinancial and financial criteria are computed and the result is entered in
a vector:
Wt ¼ ð0:5; 0:5Þ:
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 805
Here, the market share is estimated to be three times more important than produc-
tivity and four times more important than the quality of service. Productivity is
estimated to be three times more important than the quality of service. From this result,
For each weight computation, an inconsistency ratio (g) was computed and
checked for the acceptance level. If g 0.1, it is acceptable. For this example, it
is acceptable since g ¼ 0.065. If it is not acceptable, the input matrix should be
adjusted or recomputed.
For the financial criteria, ROA is estimated to be three times more important than
debt to equity ratio. Therefore,
W 2m ¼ ð0:75; 0:25Þ:
Here the global relative weights of market share, productivity, quality of service,
ROA, and debt to equity ratio are 30.4 %, 13.6 %, 6 %, 37.5 %, and 12.5 %,
respectively.
806 W. Kwak et al.
2 3
1 2 3 4 5 6
6 1=2 1 2 3 4 57
6 7
6 1=3 1=2 1 3 4 57
A0 ¼ 6
1
6 1=4
7
6 1=3 1=3 1 3 477
4 1=5 1=4 1=4 1=3 1 35
1=6 1=5 1=5 1=4 1=3 1
For simplicity of presentation, the local weights are arranged in the order of
income to interest expense, service charges, interest revenue, growth of deposits,
default ratio, and customer satisfaction:
Here the local weights with relation to market share are 36.7 %, 23.8 %, 18.3 %,
10.9 %, 6.5 %, and 3.8 %, respectively.
For other criteria at one level higher, the local weights can be calculated in the
same way. These are
For the next step, the global relative weights of lower-level management
criteria are
2 3
0:206 0:163 0:179 0:162 0:143 0:146
6 0:155 0:231 0:220 0:169 0:122 7
6 0:103 7
W0 ¼ ð0:304; 0:136; 0:060; 0:375; 0:125Þ 6 7
4 0:307 0:197 0:167 0:117 0:117 0:094 5
0:266 0:133 0:164 0:159 0:159 0:124
¼ ð0:297; 0:199; 0:176; 0:125; 0:112; 0:089Þ:
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 807
Ever since DuPont and General Motors Corporation of the USA initiated transfer
pricing systems for the interdivisional transfer of resources among their divisions,
many large organizations, with the creation of profit centers, have used a transfer
pricing system in one way or the other. Recently, transfer pricing problems have
become more important because most corporations increase transfer of goods or
services dramatically among their divisions as a result of restructuring or
downsizing their organizations (Tang 1992). Therefore, designing a good transfer
pricing strategy should be a major concern for both top management and divisional
managers (Curtis 2010).
Transfer pricing problems have been extensively studied by a number of scholars.
Many of them have recognized that a successful transfer pricing strategy should
consider multiple criteria (objectives), such as overall profit, total market share,
divisional autonomy, performance evaluation, and utilized production capacity
(Abdel-khalik and Lusk 1974; Bailey and Boe 1976; Merville and Petty 1978;
Yunker 1983; Lecraw 1985; Lin et al. 1993). However, few developed methods
808 W. Kwak et al.
have a capability of dealing with all possible optimal trade-offs of multiple criteria in
optimal solutions of the models with involvement of multiple decision makers.
In this chapter, we propose a multiple factor model to provide managers from
different background, who are involved in transfer price decision making of
a multidivisional corporation, with a systematic and comprehensive scenario
about all possible optimal transfer prices depending on both multiple-criteria and
multiple-constraint levels (in short, multiple factors). The trade-offs of the optimal
transfer prices, which have rarely been considered in the literature, can be used as
a basis for managers of corporations to make a high-quality decision in selecting
their transfer pricing systems for business competition.
This chapter proceeds as follows. First, existing transfer pricing models will be
reviewed. Then, the methodology of formulating and solving a transfer pricing
model with multiple factors will be described. A prototype of a transfer pricing
problem in a corporation will be illustrated to explain the implications of the
multiple factor transfer pricing model. Finally, conclusions and remaining research
problems will be presented.
get the benefit of efficiencies. Moreover, the manufacturing division manager in this
case should be evaluated based on cost, not on profit. Similarly, Gould (1964) and
Naert (1973) recommended economic models based on current entry and current exit
prices. Their models also focus on global profit maximization and have the same
problems as Hirshleifer’s. Note that when the transfer price is set based on marginal
costs, the division should be either controlled as a standard cost center or merged into
a larger profit center with a division that processes the bulk of its output.
Ronen and McKinney (1970) suggested dual prices in which a subsidy is given
to the manufacturing division by the central office in addition to marginal costs.
This subsidy would not be added to the price charged to the purchasing division.
They believed that autonomy is enhanced because the corporate office is only
a transmitter of information, not a price setter, and that the supplying division has
the same autonomy as an independent supplier. However, there might be a gaming
chance where all divisions are winners but the central office is a loser. There is also
a “marginal cost plus” approach that charges variable manufacturing costs (usually
standard variable costs) and is supplemented by periodic lump-sum charges to the
transferees to cover the transferor’s fixed costs, or their fixed costs plus profit, or
some kind of subsidy. It is difficult, however, to set a fixed fee that will satisfy both
the supplying division and the purchasing division. This method enables the
purchasing division to absorb all the uncertainties caused by fluctuations in the
marketplace. Moreover, the system begins to break down if the supplying division
is operating at or above the normal capacity since variable costs no longer represent
the opportunity costs of additional transfers of goods and services (Onsi 1970).
As an alternative method to identify marginal costs as the transfer prices,
a mathematical programming approach becomes more attractive for the transfer
pricing problem because it handles complex situations in a trade setting (Dopuch
and Drake 1964; Bapna et al. 2005). The application of linear programming to the
transfer pricing problem is based on the relationship between the primal and dual
solutions in the linear programming problem. Shadow prices, which reflect
the input values of scarce resources (or opportunity cost) implied in the primal
problem, can be used as the basis for a transfer price system. However, these
transfer prices have the following limitations for decision making: (i) those transfer
prices based on dual values of a solution tend to reward divisions with scarce
resources, (ii) the linear formulation requires a great deal of local information, and
(iii) transfer prices based on shadow prices do not provide a guide for performance
evaluation of divisional managers.
Based on the mathematical decomposition algorithm developed by Dantzig and
Wolfe (1960), Baumol and Fabian (1964) demonstrated how to reduce complex
optimization problems into sets of smaller problems solvable by divisions and the
central office (corporation). Although the final analysis of output decisions is made by
the central manager, the calculation process is sufficiently localized that central
management does not have to know anything about the internal technological arrange-
ments of the divisions. However, this approach does not permit divisional autonomy.
Ruefli (1971) proposed a generalized goal decomposition model for incorporating
multiple criteria (objectives) and some behavioral aspects within a three-level
810 W. Kwak et al.
29.4.2 Methodology
where C 2 Rqxn, A 2 Rmxn, and D 2 Rmxp are matrices of qxn, mxn, and mxp
dimensions, respectively; x 2 Rn are decision variables; l 2 Rq is called the criteria
parameter; and g 2 Rp is called the constraint level parameter. Both (g,l) are
assumed unknown.
The above MC2 problem has q criteria (objectives) and p constraint levels. If the
constraint level parameter g is known, then the MC2 problem reduces to an MC
linear programming problem (e.g., Yu and Zeleny 1975). In addition, if the criteria
parameter l is known, it reduces to a linear programming problem (e.g., Charnes
and Cooper 1961; Dantzig 1963).
Denote the index set of the basic variables {xji, . . ., xjm} for the MC2 problem by
J ¼ {j1, . . ., jm}. Note that the basic variables may contain some slack variables.
Without confusion, J is also called a basis for the MC2 problem. Since a basic
solution J depends on parameters (g, l), define that (i) a basic solution J is feasible
for the MC2 problem if and only if there exists a g0 > 0 such that J is a feasible
solution for the MC2 problem with respect to g0 and (ii) J is potentially optimal for
the MC2 problem if and only if there exist a g0 > 0 and a l0 > 0 such that J is an
optimal solution for the MC2 problem with respect to (g0, l0). Let G(J) be the
constraint level parameter set of all g such that the basis J is feasible and L(J) be the
criteria parameter set of all l that the basis J is dual feasible. Then, for a given basis
J of the MC2 problem, (i) J is a feasible solution if and only if the set G(J) is not
empty and (ii) J is potentially optimal if and only if both sets G(J) and L(J) are not
empty. For an MC2 problem, there may exist a number of potentially optimal
solutions {J} as parameters (g, l) vary depending on decision situations.
Seiford and Yu (1979) derived a simplex method to systematically locate the set
of all potentially optimal solutions {J}. The computer software of the simplex
method (called MC2 software) was developed by Chien et al. (1989). This software
consists of five subroutines in each iteration: (i) pivoting, (ii) determining primal
potential bases, (iii) determining dual potential bases, (iv) determining the effective
constraints for the primal weight set, and (v) determining the effective constraints
for the dual weight set (Chap. 8 of Yu 1985). It is written in PASCAL and operates
in a man-machine interactive fashion. The user cannot only view the tableau of each
iteration but also trace the past iterations. In the next section, the framework of
the MC2 problem, as well as its software, will be used to formulate and solve the
multiple factor transfer pricing problems.
812 W. Kwak et al.
be the budget availability level believed by the sth manager (or executive) for
producing the jth product by the ith division, s ¼ 1, . . . , h; dis be the production
capacity level believed by the sth manager for the ith division; dijs be the production
capacity level believed by the sth manager for the ith division to produce the
jth product; and eijs be the initial inventory level believed by the sth manager for
the ith division to hold the jth product. Then, the multiple factor transfer pricing
model is
Max Ski¼1 Slj¼1 pij xij
Max Ski¼1 Slj¼1 mij xij
Max Ski¼1 Slj¼1 sij xij
Max Ski¼1 Slj¼1 cij xij
Subject to Ski¼1 Slj¼1 bijxij b1 ij ; . . . ; bh ij (29.1)
Skj¼1 xij d1 ij , . . . dh i
xij d1 ij ; . . . ; dh ij
xij þ xiþ1, j e1 ij ; . . . ; eh ij
Xij 0, i ¼ 1, . . . , k, j ¼ 1, . . . , t:
In the next section, a prototype of the transfer pricing model in a corporation will
be illustrated to demonstrate the implications for decision makers.
All other products generate profits. We use a market price of each product to
maximize the market share goal. Note that the market prices of three products in
Table 29.2 are different.
In the company, besides the president, the vice presidents for production and for
finance are the main decision makers and may have different interpretations of the
same resource availability across the divisions. The vice president for production
views the constraint level based on the material, manpower, and equipment under
control, while the vice president for finance views the constraint level based on the
available cash flow. The president will make the final decision for the company’s
transfer price setting on the basis of the compromises of both vice presidents. All
divisional managers will carry out the president’s decision, although they have their
autonomy to provide the information about the divisional profits and costs for the
executives. The interpretations of the vice presidents for the resource constraint
levels are summarized in Table 29.3.
All executives and divisional managers agree on the units of resources consumed
to produce the products. The data is given in Table 29.4.
Let xij be the units of the jth product produced by the ith division, i ¼ 1, 2;
j ¼ 1, 2. Using the information in Tables 29.2, 29.3, and 29.4, the multiple factor
transfer pricing model is formulated as
Max 4x11 þ 8x12 þ 13x21 þ 5x22
Max 40x12 þ 46:2x21 þ 38x22
Max 4x11 þ 4x12 þ 3x21 þ 2x22
Subject to x11 þ x21 ð0; 100Þ
0:4x12 þ 0:4x21 þ 0:4x22 ð45; 000; 40; 000Þ (29.2)
X12 ð38; 000; 12; 000Þ
X21 ð45; 000; 50; 000Þ
X22 ð36; 000; 10; 000Þ
xij 0, i ¼ 1, 2; j ¼ 1, 2:
Since this multiple factor transfer pricing problem is a typical MC2 problem, the
MC2 software of Chien et al. (1989) can be used to solve the problem. Let l ¼ (l1,
l2, l3) be the weight parameter for the objectives, where l1 + l2 + l3 ¼ 1 and l1, l2,
l3 0. Let g ¼ (g1, g2) be the weight parameter for the constraint levels, where g1 +
g2 ¼ 1 and g1, g2 0. Because both weight parameters (g, l) are unknown before
design time, the solution procedure of MC2 linear programming must be used to
locate all possible potentially optimal solutions as (g, l) vary. The implications of
the potentially optimal solutions for accounting decision makers will be explained in
the next subsection. After putting (g, l) into the above model, it becomes
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 815
because there is the unused budget s2 for J2 and the parameter set G(J1) _ G(J2).
Because Product 1 in Division 1 is a by-product and its unit profit is $4, whenever
l1 < l3, both J1 and J2 are not optimal.
Let pq(Ji), q ¼ 1, . . . , 5; i ¼ 1, 2, be the shadow price of Ji for the qth constraint.
According to the marginal cost pricing approach, the optimal transfer prices of
{J1, J2} are designated as the shadow prices of {J1, J2}. These optimal transfer
prices are found in Table 29.6. Table 29.6 shows that (i) the relative transfer price
between x11 and x21 is p1(J1) ¼ 4 l1 4 l3; (ii) the transfer price for budget across
the divisions is p2(J1) ¼ 12.5 l1 + 95 l2 + 5 l3; (iii) the transfer price for x12 is
p3(J1) ¼ 3 l1 + 2 l2 + 2 l3; (iv) the transfer price for x21 is p4(J1) ¼ 4 l1 + 8.2 l2 +
5 l3; and (v) the transfer price for x22 is p5(J1) ¼ 0.
Optimal trade-offs related to transfer prices of the multiple factor model consist of
three components: (i) trade-offs among multiple objectives, (ii) trade-offs among
multiple-constraint levels, and (iii) trade-offs between multiple objectives and
multiple-constraint levels. The trade-offs among multiple objectives imply
that all possible optimal compromises of the multiple objectives are determined
by locating all possible weights of importance of these objectives. Similarly,
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 817
J1
J2
0 1 λ1
J2
0 1 λ1
prices associated with J2 are p1(J2) ¼ 4l1 4l3, p2(J2) ¼ 0, p3(J2) ¼ 8l1 + 4l3,
p4(J2) ¼ 9l1 + 7l3, and p5(J2) ¼ 5l1 + 2l3, respectively. The weighting values of
l1 and l3 in the transfer prices for J1 and J2 are given in Fig. 29.1, because
L(J1) ¼ L(J2).
(iii) If the utilized production capacity of the company is not considered (i.e., l3 ¼ 0),
then both J1 and J2 are optimal for which l1 + l2 ¼ 1 and l1, l2 _ 0, where the
graphical representation is shown in Fig. 29.2. Figure 29.2 implies that any
weighted combination of l1 and l2 taken from the indicated feasible segment
maximizes the utility values of both the overall company’s profit and the market
share of the company. The resulting optimal transfer prices of J1 are
p1(J1) ¼ 4l1, p2(J1) ¼ 12.5l1 + 95l2, p3(J1) ¼ 3l1 + 2l2, p4(J1) ¼ 4l1 +
8.2l2, and p5(J1) ¼ 0, while the resulting optimal transfer prices of J2 are
p1(J2) ¼ 4l1, p2(J2) ¼ 0, p3(J2) ¼ 8l1 + 40l2, p4(J2) ¼ 9l1 + 46.2l2, and
p5(J2) ¼ 5l1 + 38l2, respectively.
For the trade-offs among two constraint levels (the different opinions of two vice
presidents), the weight of the vice president for production is g1 while that of the vice
president for finance is g2 such that g1 + g2 ¼ 1, g1, g2 0. The range of g1 and g2 are
decomposed into two subsets: G(J1) ¼ {g1, g2 0 | 65g1 280g2 0 and g1 + g2 ¼ 1}
and G(J2) ¼ {g1, g2 0 | 65g1 + 280g2 0 and g1 + g2 ¼ 1} (see Table 29.5). The
graphical representation of G(J1) and G(J2) is shown in Fig. 29.3. Whenever the
weighting values of g1 and g2 are taken from G(J1), the corresponding compromise
of two vice presidents will result in the optimal transfer prices of J1 in Table 29.6.
Similarly, the decision situation of constraint levels for J2 can be explained.
Finally, there are many optimal trade-off situations between three objectives and
two constraint levels involved with the transfer pricing problem. For example, two
cases are illustrated as follows:
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 819
65γ1 = 280γ2
J1
0 1 γ1
λ1
J2 J1
(i) In the case that the utilized production capacity of the company is not considered
(i.e., l3 ¼ 0) in contrast to the weighting value of the market share and that of the
vice president for finance (i.e., 0 l2 1 and 0 g2 1), the optimal trade-offs
between the overall company’s profit and the constraint level believed by the vice
president for production for J1 and J2 are shown in Fig. 29.4. Here, if the values of
(g1, l1) are taken from 0 l1 1 to .81 g1 1, then the optimal transfer prices
820 W. Kwak et al.
.5
J2 J1
0 .81 1 γ1
associated with J1 will be chosen. They are p1(J1) ¼ 4l1, p2(J1) ¼ 12.5l1 + 95l2,
p3(J1) ¼ 3l1 + 2l2, p4(J1) ¼ 4l1 + 8.2l2, and p5(J1) ¼ 0. Otherwise, the values
of (g1, l1) are taken from 0 l1 1 to 0 g1 .81, and the optimal
transfer prices associated with J2, p1(J2) ¼ 4l1, p2(J2) ¼ 0, p3(J2) ¼ 8l1 + 40l2,
p4(J2) ¼ 9l1 + 46.2l2, and p5(J2) ¼ 5l1 + 38l2 will be chosen.
(ii) In the case that the weighting value of the overall company’s profit is fixed at .5
(i.e., l1 ¼ .5) and the weighting value of the market share and that of the vice
president for finance are any of 0 l2 1 and 0 g2 1, respectively, Fig. 29.5
shows the optimal trade-offs between the utilized production capacity of the
company and the constraint level believed by the vice president for production
for J1 and J2. In Fig. 29.5, if the values of (g1, l3) are taken from 0 l3 .5 to
.81 g1 1, then the optimal transfer prices associated with J1 are p1(J1) ¼ 2
4l3, p2(J1) ¼ 6.25 + 95l2 + 5l3, p3(J1) ¼ 1.5 + 2l2 + 2l3, p4(J1) ¼ 2 + 8.2l2 +
5l3, and p5(J1) ¼ 0, respectively. If the values of (g1, l3) are taken from 0 l3
.5 to 0 g1 .81, then the optimal transfer prices associated with J2 are
p1(J2) ¼ 2 4l3, p2(J2) ¼ 0, p3(J2) ¼ 4 + 40l2 + 4l3, p4(J2) ¼ 4.5 + 46.2l2 +
7l3, and p5(J2) ¼ 2.5 + 38l2 + 2l3, respectively. Note that when the weighting
values fall in the range of .5 l3 1 and 0 g1 1, there is not any optimal
trade-off because both J1 and J2 are not optimal solutions (recall that this is caused
by the by-product 1 in Division 1 that has -$4 as the unit profit).
It is worth noting some important implications for accounting decision makers
from the above trade-off analysis. First, the multiple factor transfer pricing model
has a capability of systematically locating all possible optimal transfer prices
through the optimal trade-offs of multiple objectives and multiple-constraint levels.
Since the set of all possible optimal transfer prices found by this model describes
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 821
every possible decision situation within the model framework, the optimal transfer
price obtained from either linear programming or goal programming models is
included in the subset as a special case.
Second, the multiple factor transfer pricing model can be applied to solve
transfer pricing problems not only with complex structure but also with some
organizational behavior contents, such as the organizational differentiation (see
Table 29.3 for different constraint levels). Consequently, this model can foster more
autonomous flexibility than any other mathematical programming model by
allowing central management or local managers to express their own preference
structures and weights.
Third, the proposed method facilitates decision makers’ participation that may
make a positive management achievement of organizational goals (Locke
et al. 1981). The model can aid coordination and synthesis of multiple conflicting
views. This may be quite effective in a transfer pricing situation in which many
objectives are contradictory to each other and these objectives are measured
differently by a number of decision participants.
Fourth, in most multiple-criteria solution techniques, including the goal pro-
gramming approach, if the decision makers are not satisfied with the optimal
solution obtained by using their preferred weights of importance for the multiple
objectives, then an iterative process has to be conducted for incorporating the
decision makers’ new preference on the weights and finding new solutions until
decision makers are satisfied. However, the proposed model eliminates such a time-
and cost-consuming iterative process since it already considers all possible optimal
solutions with respect to the changes of parameter (g, l). Whenever decision
makers want to change their preference on the weights, the corresponding optimal
transfer prices can be immediately identified from the results like in Table 29.6.
This model, in turn, allows the performance evaluation of optimal transfer prices.
Finally, the optimal transfer prices obtained by the multiple factor model have
a twofold significance in terms of decision characteristics. If the problem is viewed as
a deterministic decision problem, whenever the preferred weighting values of objec-
tives and constraint levels are known, the resulting optimal solutions can be identified
from the potentially optimal solutions of the model. Then, the corresponding optimal
transfer prices can be adopted to handle the business situation (recall the above trade-
off analysis). If the problem is viewed as a probabilistic decision problem, it involves
the assessment of the likelihood of (g, l) to occur at the various points of the range.
With proper assumptions, the uncertainty may be represented by random variables
with some known probability distribution. A number of known criteria such as
maximizing expected payoff, minimizing the variance of the payoff, maximin
payoff, maximizing the probability of achieving a targeted payoff, stochastic dom-
inance, probability dominance, and mean-variance dominance can be used to choose
the optimal transfer prices (see Shi 1991). In summary, the multiple factor transfer
pricing model fosters flexibility in designing the optimal transfer prices for the
corporation to cope with all possible changes of business competition. This model
is more likely to be a better aid for executives or managers to understand and deal
with their current or future transfer pricing problems.
822 W. Kwak et al.
29.4.5 Conclusions
A multiple factor transfer pricing model has been developed to solve the transfer
pricing problems in a multidivisional corporation. This model can provide
a systematic and comprehensive scenario about all possible optimal transfer prices
depending on multiple-criteria and multiple-constraint levels. The trade-offs of
optimal transfer prices offer a broad basis for managers of a corporation to flexibly
implement the optimal transfer pricing strategy and cope with various business
situations. Furthermore, this method also aids global optimization, division auton-
omy, and performance evaluation.
There are some research problems remaining to be explored. From a practical
point of view, the framework of this model can be applied to other accounting areas
such as capital budgeting, cost allocation, audit sampling objectives, and personnel
planning for an audit corporation if the decision variables and formulation are
expressed appropriately. From a theoretical point of view, the decomposition
algorithm of linear programming (Dantzig and Wolfe 1960) can be incorporated
into the MC2-simplex method (Seiford and Yu 1979) to sharpen the multiple factor
transfer pricing model’s capability of solving the large-scale transfer pricing prob-
lem. Thus, this incorporation may result in the development of more effective
solution procedures. How to incorporate other trade-off techniques, such as the
satisficing trade-off method (Nakayama 1994), into the MC2 framework for design-
ing optimal transfer pricing strategies is another interesting research problem.
present value (NPV) and internal rate of return (IRR), are the primary quantitative
methods in capital budgeting (Kim and Farragher 1981). The payback method
estimates how long it will take to recover the original investment. However,
this method incorporates neither the cash flows after the payback period nor the
variability of those cash flows (Boardman et al. 1982). The accounting rate of
return method measures a return on the original cost of the investment. Both of
the above methods ignore the time value of money. DCF methods may not
be adequate to evaluate new manufacturing or information systems, because of
a bias in favor of short-term investments with quantifiable benefits (Mensah and
Miranti 1989).
A current trend in capital budgeting methods utilizes mathematical program-
ming and higher discount rates to incorporate higher risk factors (Pike 1983;
Palliam 2005). Hillier (1963) and Huang (2008) suggested useful ways to eval-
uate risky investments by estimating expected values and standard deviations of
net cash flows for each alternative investment. They showed that the standard
deviation of cash flows is easily obtainable. With this information, a complete
description of the risky investment is possible via probability distribution of the
IRR, NPV, or annual cost of the proposed investment under the assumption of the
net cash flows from the investment, which are normally distributed. Similarly,
Turney (1990) suggested a stochastic dynamic adjustment model to incorporate
greater risk premiums when significant additional funds are required in multiple
time periods. Lin (1993) also proposed a multiple-criteria capital budgeting
model under risk. His model used chance constraints of uncertain cash flows
and accounting earnings as risk factors. Pike (1988) empirically tested the
correlation between sophisticated capital budgeting techniques and decision-
making effectiveness and found that management believed that sophisticated
investment techniques improve effectiveness in the evaluation and control of
large capital projects. Weingartner (1963) introduced a mathematical program-
ming approach in the capital budgeting problem. Other researchers have extended
Weingartner’s work with different directions (e.g., Baumol and Quandt 1965;
Bernard 1969; Howe and Patterson 1985). The development of chance-
constrained programming (CCP) by Charnes and Cooper (1961) also enriched
with applications of mathematical programming models in the capital budgeting
problem. They also developed an approximation solution method to the CCP with
zero-one variables using a linear constraint.
A typical mathematical capital budgeting approach maximizes DCFs that mea-
sure a project’s desirability on the basis of its expected net present value as
a primary goal. DCF analysis, however, ignores strategic factors such as future
growth opportunities (Cheng 1993). Furthermore, management can change their
plans if operating conditions change. For example, they can change input and
output mixes or abandon the project in a multi-period situation. The increasing
involvement of stakeholders, other than shareholders, in a business organization
supports a multiple-objective approach (Bhaskar 1979). Other empirical
studies also found that firms used multiple criteria in their capital budgeting
problems (e.g., Bhaskar and McNamee 1983; Thanassoulis 1985). The goal
824 W. Kwak et al.
premises. First, from the linear system structure’s perspective, the criteria and
constraint levels may be interchangeable. Thus, like multiple criteria, multiple-
constraint (resource availability) levels can be considered. Second, from the appli-
cation’s perspective, it is more realistic to consider multiple resource availability
levels (discrete right-hand sides) than a single resource availability level in isola-
tion. This recognizes that the availability of resources can fluctuate depending on
the decision situation forces, such as the preferences of the different managers. The
concept of multiple resource levels corresponds to the typical characteristic of
capital budgeting situations where the decision-making process should reflect
each manager’s preference on the new project. A theoretical connection between
MC linear programming and MC2 linear programming can be found in Gyetvan and
Shi (1992) and decision problems related to MC2 linear programming have been
extensively studied (see, e.g., Lee et al. (1990), Shi (1991), and Shi and Yu (1992)).
Key ideas of MC2 linear programming, as a primary theoretical foundation of this
article, are described as follows.
An MC2 linear programming problem can be formulated as
max lt Cx
s:t: Ax Dg
x0
where C 2 Rqxn, A 2 Rmxn, and D 2 Rmxp are matrices of qxn, mxn, and mxp
dimensions, respectively; x 2 Rn are decision variables; l 2 Rq is called the criteria
parameter; and g 2 Rp is called the constraint level parameter. Both (g, l) are
assumed unknown.
The above MC2 problem has q criteria (objectives) and p constraint levels. If the
constraint level parameter g is known, then the MC2 problem reduces to an MC
linear programming problem (e.g., Yu and Zeleny 1975). In addition, if the criteria
parameter l is known, it reduces to a linear programming problem (e.g., Charnes
and Cooper 1961; Dantzig 1963).
Denote the index set of the basic variables {xj1, . . . , xjm} for the MC2 problem by
J ¼ {j1, . . . , jm}. Note that the basic variables may contain some slack variables.
Without confusion, J is also called a basis for the MC2 problem. Since a basic
solution J depends on parameters (g, l), define that (1) a basic solution J is feasible
for the MC2 problem if and only if there exists a g0 > 0 such that J is a feasible
solution for the MC2 problem with respect to g0 and (2) J is potentially optimal for
the MC2 problem if and only if there exist a g0 > 0 and a l0 > 0 such that J is an
optimal solution for the MC2 problem with respect to (g0, l0). For an MC2 problem,
there may exist a number of potentially optimal solutions {J} as parameters (g, l)
vary depending on decision situations. Seiford and Yu (1979) derived a simplex
method to systematically locate the set of all potentially optimal solutions {J}.
In summary, a model within the framework of AHP and MC2 is proposed
to formulate and solve the multiple-objective capital budgeting problems with
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 827
equity ratio for the firm is assumed to be 1 and the inverse of debt to equity ratio is
used to measure leverage. For the coefficients of the constraints, let bij be the cash
outlay required by the jth project in the ith time period. For the coefficients of the
constraint levels, let Cki be the budget availability level believed by the kth manager
(or executive) for the firm in the ith time period, k ¼ 1, . . . , u.
In this article, for illustration, if we treat t ¼ 10, i ¼ 5, and u ¼ 5, then the
model is
X
10
max vj Xj
j¼1
X
10
max gij Xj 8i ¼ 1, . . . , 5
j¼1
X
10
max r ij Xj 8i ¼ 1, . . . , 5
j¼1
X
10
max lij Xj 8i ¼ 1, . . . , 5
j¼1
X
10
subject to bij Xj ðcli ; . . . ; c5i Þ 8i ¼ 1, . . . , 5
j¼1
and Xj ¼ f0; 1g
where k ¼ 1, . . . , 5, for cki, represents the five possible DMs (i.e., president,
controller, production manager, marketing manager, and engineering manager).
The weights of 16 total objectives can be expressed by (l1, . . . , l16) with
Slq ¼ 1, q ¼ 1, . . . , 16, 0 < lq < 1, and the weights of five DMs can be expressed
by (l1, . . . , l5) with Sgk ¼ 1, k ¼ 1, . . . , 5, 0 < gk < 1.
The above model is an integer MC2 problem with 16 objectives and five
constraint levels. Solving this problem by using a currently available solution
technique (Seiford and Yu 1979) is not a trivial task because of its substantial
computational complexity. To overcome this difficulty, we propose a two-phased
solution procedure as depicted in Fig. 29.6.
The first phase is referred to as the AHP phase in the sense that AHP is applied to
derive weights of l and g, which reduces the model’s complexity. The computation
of weights is effectively handled by Expert Choice (Forman et al. 1985), commer-
cial software for AHP. In this phase, we first induce preferences about objectives
from all five DMs involved. Note that each DM may have different preferences
about the four goals. The president may think maximization of ROI is
the most important goal, while the controller may think maximization of NPV is
the most important goal. AHP generates relative weights for each goal. Then,
multiple-constraint levels are incorporated using the MC2 framework. For instance,
DMs may estimate different budget availability levels for each time period. The
controller may believe that $40 million is available for the first year, while the
production manager may believe that the company can spend $50 million for this
time period. This scenario is more realistic if we consider the characteristics of
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 829
AHP
Weights of
λ and γ
HC2
Framework IP
Capital
Budgeting
No
Accept ?
Yes
Final Solution
a capital budgeting problem from the perspective of the different utility functions or
risk factors of each DM. An example using AHP to generate relative weights for the
five DMs’ constraint levels (preferences) can be found in Appendix 1.
The second phase is called the integer programming (IP) phase. After applying
the AHP phase to the model, the model is reduced to a linear IP problem, in
which the potentially optimal projects must be selected. To solve this problem,
we employ the ZOOM software (Singhal et al. 1989) that was originally introduced
for solving zero-one integer linear programming problems.
Table 29.7 Net present value data for the Lorie-Savage Corporation (in millions)
Cash outlays for each period
Project Net present value Period 1 Period 2 Period 3 Period 4 Period 5
1 $20 $24 $16 $6 $6 $8
2 16 12 10 2 5 4
3 11 8 6 6 6 4
4 4 6 4 7 5 3
5 4 1 6 9 2 3
6 18 18 18 20 15 15
7 7 13 8 10 8 8
8 19 14 8 12 10 8
9 24 16 20 24 16 16
10 4 4 6 8 6 4
2. Maximize the growth of the firm due to each project. Growth can be measured by net
sales increase from each project. This can measure a strategic success factor of a firm.
3. Maximize the profitability of the company. Profitability can be measured by ROI
for each time period generated by each project.
4. Maximize the flexibility of financing. The flexibility of financing (leverage) can
be measured by debt to equity ratio. If a firm’s debt to equity ratio is higher than
the industry average or the optimum level, the firm’s cost of debt financing will
become more expensive. In this example, we use the inverse of the debt to equity
ratio and assume 1 as the optimum debt to equity ratio.
The data related to these objectives is given in Table 29.7. In this example, all
projects are assumed to be independent. In Table 29.7, the firm’s cost of capital is
assumed to be known a priori and to be independent of the investment decisions.
Based on these assumptions, the net present value of each project can be defined as
the sum of the cash flows discounted by the cost of capital. Cash outlay is the
amount of expenditure required for project j, j ¼ 1, 2, 3, . . . , 10, in each time period.
To measure growth, the net sales increase for each time period for each project is
estimated. These data are provided in Table 29.8.
To measure the profitability of each project, ROI is estimated after reflecting
additional income and capital expenditures from each investment for each time
period. These data are provided in Table 29.9.
To measure leverage, the inverse of debt to equity ratio is used after adopting
each project. Here, the optimum debt to equity ratio is assumed to be one for the
Lorie-Savage Corporation. These data are provided in Table 29.10.
The five key DMs in this company (president, controller, production manager,
marketing manager, and engineering manager) have different beliefs regarding
resource availability. For example, for budget availability levels, each DM may
have a different opinion. Of course, the president will make the final decision based
on the opinions of other managers. However, the DMs’ preferences of collection
process should improve the quality of the final decision. Budget availability level
data are provided in Table 29.11.
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 831
Table 29.8 Net sales increase data for the Lorie-Savage Corporation (in millions)
Net sales increase for each project
Project Period 1 Period 2 Period 3 Period 4 Period 5
1 $120 $130 $145 $150 $170
2 100 120 140 150 160
3 80 90 95 95 100
4 40 50 50 55 60
5 40 45 50 55 60
6 110 120 140 150 165
7 60 70 65 70 80
8 110 120 100 110 120
9 150 170 180 190 200
10 35 40 40 50 50
Table 29.9 Return on investment data for the Lorie-Savage Corporation (in percentage)
Return on investment for each project
Project Period 1 Period 2 Period 3 Period 4 Period 5
1 10 12 14 15 17
2 10 12 18 16 17
3 12 15 15 15 18
4 8 15 10 8 12
5 15 10 8 20 18
6 12 12 10 15 15
7 8 12 10 12 12
8 14 16 13 15 16
9 12 10 9 12 12
10 10 8 9 8 12
Table 29.11 Budget availability level data for the Lorie-Savage Corporation (in millions)
Estimate budget availability
Decision maker Period 1 Period 2 Period 3 Period 4 Period 5
President $50 $30 $30 $35 $30
Controller 40 45 30 30 20
Production manager 55 40 20 30 35
Marketing manager 45 30 40 45 30
Engineering 50 40 45 30 35
The parameters for budget availability levels are derived by using AHP. Here,
one level of an AHP is used. The aggregated objective function also can be obtained
by using AHP. The difference here is that two levels of an AHP must be used to
provide preferences for the objective functions. The first level of this AHP corre-
sponds to the four groups of objectives. The second level corresponds to time
periods within each objective group. Note that the second level is not required for
the objective of maximizing the net present value. Hence, four pairwise comparison
matrices are required for eigenvalue computation. More detailed information about
the formulation of this problem is presented in the appendix. In sum, the objective
function aggregated from the 16 objectives is written as
29.5.3 Conclusions
29.6 Conclusions
Appendix 1
For illustrative purposes, we show how to use AHP to induce the five DMs’
preferences of budget availability and to compute the relative weights. Generally,
AHP collects input judgments of DMs in the form of a matrix by pairwise compar-
isons of criteria (i.e., their budget availability levels). An eigenvalue method is then
used to scale weights of such criteria. That is, the relative importance of each
criteria is computed. The result from all of pairwise comparison is stored in an input
matrix as follows:
President Controller Production Marketing Engineering
Manager Manager Manager
2 3
1 3 4 5 6
6 7
6 1 2 5 57
6 7
6 1 3 47
6 7
6 7
4 1 25
1
References
Abdel-khalik, A. R., & Lusk, E. J. (1974). Transfer pricing – A synthesis. The Accounting Review, 8–24.
Arrington, E. E., Hillison, W., & Jensen, R. E. (1984). An application of analytical hierarchy
process to model expert judgment on analytic review procedures. Journal of Accounting
Research, 22, 298–312.
Ayvaz, E., & Pehlivanli, D. (2011). The use of time driven activity based costing and analytic
hierarchy process method in the balanced scorecard implementation. International Journal of
Business and Management, 6, 146–158.
Bailey, A. D., & Boe, W. J. (1976). Goal and resource transfers in the multigoal organization. The
Accounting Review, 559–573.
Bapna, R., Goes, P., & Gupta, A. (2005). Pricing and allocation for quality-differentiated online
services. Management Science, 51, 1141–1150.
Baumol, W. J., & Fabian, T. (1964). Decomposition, pricing for decentralization and external
economies. Management Science, 11, 1–32.
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 837
Baumol, W. J., & Quandt, R. E. (1965). Investment and discount rate under capital rational –
A programming approach. Economic Journal, 75(298), 173–329.
Bernard, R. H. (1969). Mathematical programming models for capital budgeting – A survey
generalization and critique. Journal of Financial and Quantitative Analysis, 4, 111–158.
Bhaskar, K. A. (1979). Multiple objective approach to capital budgeting. Accounting and Business
Research, 10, 25–46.
Bhaskar, I. C., & McNamee, P. (1983). Multiple objectives accounting and finance. Journal of
Business Finance and Accounting, 10, 595–621.
Boardman, W., Reinhart, I., & Celec, S. E. (1982). The role of payback period in the theory and
application of duration to capital budgeting. Journal of Business Finance and Accounting, 9,
511–522.
Borkowski, S. C. (1990). Environmental and organizational factors affecting transfer pricing:
A survey. Journal of Management Accounting Research, 2, 78–99.
Boucher, T. O., & MacStravic, E. L. (1991). Multi-attribute evaluation within a present value
framework and its relation to the analytic hierarchy process. The Engineering Economist, 37, 1–32.
Calvert, J. (1990). Quality customer service: A strategic marketing weapon for high performance
banking. Industrial Engineering, 54–57.
Campi, J. P. (1992). It’s not as easy as ABC. Journal of Cost Management Summer, 6, 5–11.
Chan, Y. L. (2004). Use of capital budgeting techniques and an analytic approach to capital
investment decisions in Canadian municipal governments. Public Budgeting & Finance, 24,
40–58.
Chan, Y. L., & Lynn, B. E. (1991). Performance evaluation and the analytic hierarchy process.
Journal of Management Accounting Research, 3, 57–87.
Charnes, A., & Cooper, W. W. (1961). Management models and industrial applications of linear
programming (Vols. I & II). New York: Wiley.
Cheng, J. K. (1993). Managerial flexibility in capital investment decision: Insights from the real-
options. Journal of Accounting Literature, 12, 29–66.
Chien, I. S., Shi, Y., & Yu, P. L. (1989). MC2 program: A Pascal program run on PC or Vax
(revised version). Lawrence: School of Business, University of Kansas.
Choi, T. S., & Levary, R. R. (1989). Multi-national capital budgeting using chance-constrained
goal programming. International Journal of Systems Science, 20, 395–414.
Corner, I. L., Decicen, R. F., & Spahr, R. W. (1993). Multiple-objective linear programming in
capital budgeting. Advances in Mathematical Programming and Financial Planning, 3,
241–264.
Curtis, S. L. (2010). Intercompany licensing rates: Implications from principal-agent theory.
International Tax Journal, 36, 25–46.
Dantzig, G. B. (1963). Linear programming and extensions. New Jersey: Princeton University
Press.
Dantzig, G. B., & Wolfe, P. (1960). Decomposition principles for linear programs. Operations
Research, 8, 101–111.
Deckro, R. F., Spahr, R. W., & Herbert, J. E. (1985). Preference trade-offs in capital budgeting
decisions. IIE Transactions, 17, 332–337.
DeWayne, L. S. (2009). Developing a lean performance score. Strategic Finance, 91, 34–39.
Dopuch, N., & Drake, D. F. (1964). Accounting implications of a mathematical programming
approach to the transfer price problem. Journal of Accounting Research, 2, 10–24.
Drucker, P. F. (1993). We need to measure, not count. Wall Street Journal, April 13. Also Expert
Choice Voice, July.
Dyer, R. F., & Forman, E. H. (1991). An analytic approach to marketing decisions. Englewood,
NJ: Prentice Hall.
Eccles, R. G. (1983). Control with fairness in transfer pricing. Harvard Business Review, 61,
149–161.
Eccles, R. G. (1991). The performance measurement manifesto. Harvard Business Review, 69,
131–137.
838 W. Kwak et al.
Eccles, R. G., & Pyburn, P. J. (1992). Creating a comprehensive system to measure performance.
Management Accounting, 74, 41–44.
Eddie, W., Cheng, L., & Li, H. (2001). Analytic hierarchy process: An approach to determine
measures for business performance. Measuring Business Excellence, 5, 30–36.
Fisher, J. (1992). Use of nonfinancial performance measures. Journal of Cost Management, 6,
31–38.
Forman, E. H., Saaty, T. L., Selly, M. A., & Waldron, R. (1985). Expert choice. Pittsburgh:
Decision Support Software.
Goicoechea, A., Hansen, D. R., & Duckstein, L. (1982). Multi-objective decision analysis appli-
cations. NewYork: Wiley.
Gonzalez, G., Reeves, R., & Franz, L. S. (1987). Capital budgeting decision making: An interac-
tive multiple objective linear integer programming search procedure. Advances in Mathemat-
ical Programming and Financial Planning, 1, 21–44.
Gould, J. R. (1964). Internal pricing in corporations when there are costs of using an outside
market. The Journal of Business, 37, 61–67.
Gyetvan, F., & Shi, Y. (1992). Weak duality theorem and complementary slackness theorem for
linear matrix programming problems. Operations Research Letters, 11, 244–252.
Hardy, C., & Reeve, R. (2000). A study of the internal control structure for electronic data interchange
systems using the analytic hierarchy process. Accounting and Finance, 40, 191–210.
Harker, P. T., & Vargas, L. G. (1987). The theory of ratio scale estimation: Saaty’s analytic
hierarchy process. Management Science, 33, 1383–1403.
Hillier, F. S. (1963). The derivation of probabilities information for the evaluation of risky
investments. Management Science, 443–457.
Hirshleifer, J. (1956). On the economies of transfer pricing. The Journal of Business, 172–184.
Howe, K. M., & Patterson, I. H. (1985). Capital investment decisions under economies of scale in
flotation costs. Financial Management, 14, 61–69.
Howell, R. A., & Schwartz, W. A. (1994). Asset deployment and investment justification, In
Handbook of cost management. Boston: Warren Gorham Lamont.
Huang, X. (2008). Mean-variance model for fuzzy capital budgeting. Computers & Industrial
Engineering, 55, 34.
Huang, S., Hung, W., Yen, D. C., Chang, I., & Jiang, D. (2011). Building the evaluation model of the
IT general control for CPAs under enterprise risk management. Decision Support Systems, 50, 692.
Ignizio, J. P. (1976). An approach to the capital budgeting problem with multiple objectives. The
Engineering Economist, 21, 259–272.
Ishizaka, A., Balkenborg, D., & Kaplan, T. (2011). Influence of aggregation and measurement
scale on ranking a compromise alternative in AHP. The Journal of the Operational Research
Society, 62, 700–711.
Kamath, R. R., & Khaksari, S. Z. (1991). Real estate investment: An application of the analytic
hierarchy process. The Journal of Financial and Strategic Decisions, 4, 73–100.
Kaplan, R. S. (1986). Accounting lag: The obsolescence of cost accounting systems. California
Management Review, 28, 175–199.
Kaplan, R. S., & Norton, D. P. (1992). The balanced scorecard-measures that drive performance.
Harvard Business Review, 69, 71–79.
Karanovic, G., Baresa, S., & Bogdan, S. (2010). Techniques for managing projects risk in capital
budgeting process. UTMS Journal of Economics, 1, 55–66.
Kim, S. H., & Farragher, E. I. (1981). Current capital budgeting practices. Management Account-
ing, 6, 26–30.
Lecraw, D. J. (1985). Some evidence on transfer pricing by multinational corporations. In
A. M. Rugman & L. Eden (Eds.), Multinationals and transfer pricing. New York:
St. Martin’s Press.
Lee, C. F. (1993). Statistics for business and financial economics. Lexington: Heath.
Lee, H. (1993). A structured methodology for software development effort prediction using the
analytic hierarchy process. Journal of Systems and Software, 21, 179–186.
29 Group Decision-Making Tools for Managerial Accounting and Finance Applications 839
Lee, S. M., & Lerro, A. (1974). Capital budgeting for multiple objectives. Financial Management,
3, 51–53.
Lee, Y. R., Shi, Y., & Yu, P. L. (1990). Linear optimal designs and optimal contingency plans.
Management Science, 36, 1106–1119.
Lee, H., Nazem, S. M., & Shi, Y. (1995). Designing rural area telecommunication networks via
hub cities. Omega: International Journal of Management Science, 22(3), 305–314.
Li, Y., Huang, M., Chin, K., Luo, X., & Han, Y. (2011). Integrating preference analysis and
balanced scorecard to product planning house of quality. Computers & Industrial Engineering,
60, 256–268.
Liberatore, M. J., Mohnhan, T. F., & Stout, D. E. (1992). A framework for integrating capital
budgeting analysis with strategy. The Engineering Economist, 38, 31–43.
Lin, T. W. (1993). Multiple-criteria capital budgeting under risk. Advances in Mathematical
Programming and Financial Planning, 3, 231–239.
Lin, L., Lefebvre, C., & Kantor, J. (1993). Economic determinants of international transfer pricing
and the related accounting issues, with particular reference to Asian Pacific countries. The
International Journal of Accounting, 49–70.
Locke, E. A., Shaw, K. N., & Lathom, G. P. (1981). Goal setting and task performance:
1969–1980. Psychological Bulletin, 90, 125–152.
Lorie, J. H., & Savage, L. I. (1955). Three problems in rationing capital. Journal of Business, 28,
229–239.
Mensah, Y. M., & Miranti, P. J., Jr. (1989). Capital expenditure analysis and automated
manufacturing systems: A review and synthesis. Journal of Accounting Literature, 8, 181–207.
Merville, L. J., & Petty, W. J. (1978). Transfer pricing for the multinational corporation. The
Accounting Review, 53, 935–951.
Muralidhar, K., Santhanam, R., & Wilson, R. L. (1990). Using the analytic hierarchy process for
information system project selection. Information and Management, 18, 87–95.
Naert, P. A. (1973). Measuring performance in a decentralized corporation with interrelated
divisions: Profit center versus cost center. The Engineering Economist, 99–114.
Nakayama, H. (1994). Some remarks on trade-off analysis in multi-objective programming. The XIth
International conference on multiple criteria decision making, Coimbra, Portugal, Aug. 1–6.
Nanni, A. J., Dixon, J. R., & Vollmann, T. E. (1990). Strategic control and performance measure-
ment. Journal of Cost Management, 4, 33–42.
Natarajan, T., Balasubramanian, S. A., & Manickavasagam, S. (2010). Customer’s choice amongst
self-service technology (SST) channels in retail banking: A study using analytical hierarchy
process (AHP). Journal of Internet Banking and Commerce, 15, 1–15.
Onsi, M. (1970). A transfer pricing system based on opportunity cost. The Accounting Review, 45,
535–543.
Palliam, R. (2005). Application of a multi-criteria model for determining risk premium. The
Journal of Risk Finance, 6, 341–348.
Pike, R. (1983). The capital budgeting behavior and corporate characteristics of capital-
constrained firms. Journal of Business Finance and Accounting, 10, 663–671.
Pike, R. (1988). An empirical study of the adoption of sophisticated capital budgeting practices
and decision-making effectiveness. Accounting and Business Research, 18, 341–351.
Reeves, G. R., & Franz, L. S. (1985). A simplified interactive multiple objective linear program-
ming procedure. Computers and Operations Research, 12, 589–601.
Reeves, G. A., & Hedin, S. R. (1993). A generalized interactive goal programming procedure.
Computers and Operations Research, 20, 747–753.
Reeves, G. F., Lawrence, S. M., & Gonzalez, J. J. (1988). A multiple criteria approach to aggregate
industrial capacity expansion. Computers and Operations Research, 15, 333–339.
Ronen, J., & McKinney, G. (1970). Transfer pricing for divisional autonomy. Journal of Account-
ing Research, 8, 99–112.
Rotch, W. (1990). Activity-based costing in service industries. Journal of Cost Management,
4–14.
840 W. Kwak et al.
Contents
30.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 842
30.2 Preliminary Knowledge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 844
30.2.1 Change of Measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 844
30.2.2 Hierarchical Clustering with K-Means Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 845
30.2.3 Standard Errors in Finance Panel Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 846
30.3 Employee Stock Options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 847
30.3.1 Model-Based Approach to Subjective Value . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 847
30.3.2 Compensation-Based Approach to Subjective Value . . . . . . . . . . . . . . . . . . . . . . . . . 851
30.4 Simulation Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 852
30.4.1 Exercise Behavior . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 852
30.4.2 Factors’ Effects on ESOs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 853
30.4.3 Offsetting Effect of Sentiment and Risk on ESO . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 857
30.5 Empirical Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 857
30.5.1 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 857
30.5.2 Preliminary Findings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 859
30.5.3 Implications of Regression Results and Variable Sensitivities . . . . . . . . . . . . . . . 861
30.5.4 Subjective Value and Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 864
30.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 866
Appendix 1: Derivation of Risk-Neutral Probability by Change of Measure . . . . . . . . . . . . . . . . . 867
Appendix 2: Valuation of European ESOs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 869
Appendix 3: Hierarchical Clustering with a K-Means Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 870
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 871
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 841
DOI 10.1007/978-1-4614-7750-1_30,
# Springer Science+Business Media New York 2015
842 L.-j. Chen and C.-d. Fuh
Abstract
This study presents model-based and compensation-based approaches to
determining the price-subjective value of employee stock options (ESOs).
In the model-based approach, we consider a utility-maximizing model in
which the employees allocate their wealth among company stock, a market
portfolio, and risk-free bonds, and then we derive the ESO formulas, which
take into account illiquidity and sentiment effects. By using the method of
change of measure, the derived formulas are simply like those of the market
value with altered parameters. To calculate the compensation-based subjec-
tive value, we group employees by hierarchical clustering with a K-means
approach and back out the option value in an equilibrium competitive
employment market.
Further, we test illiquidity and sentiment effects on ESO values by running
regressions that consider the problem of standard errors in the finance panel data.
Using executive stock options and compensation data paid between 1992 and
2004 for firms covered by the Compustat Executive Compensation Database, we
find that subjective value is positively related to sentiment and negatively related
to illiquidity in all specifications, consistent with the offsetting roles of sentiment
and risk aversion. Moreover, executives value ESOs at a 48 % premium to the
Black-Scholes value and ESO premiums are explained by a sentiment level of
12 % in risk-adjusted, annualized excess return, suggesting a high level of
executive overconfidence.
Keywords
Employee stock option • Sentiment • Subjective value • Illiquidity • Change of
measure • Hierarchical clustering with K-means approach • Standard errors in
finance panel data • Exercise boundary • Jump diffusion model
30.1 Introduction
1
The option values are estimated by the calculation in ExecuCompustat database.
30 Statistics Methods Applied in Employee Stock Options 843
financial constraints (Core and Guay 2001). In addition, ESOs attract highly
motivated and able employees (Core and Guay 2001; Oyer and Schaefer 2005).
All of these factors contribute to the importance of ESOs for corporate gover-
nance and finance research.
The illiquidity problem of ESOs cannot be neglected. ESOs usually have
a vesting period during which they cannot be exercised, and hence they cannot
be redeemed for a fixed period of time. Furthermore, ESOs are generally not
publicly traded and their sale is not permitted. Standard methods for valuing
options are difficult to apply to ESOs. Because of the illiquidity of ESOs, many
employees hold undiversified portfolios that include large stock options of their
own firms. Because of the impossibility of full diversification, the value per-
ceived by employees (subjective value) may be quite different from the traded
option. We consider the subjective value to be what a constrained agent would
pay for the ESOs and the market value to be the value perceived by an
unconstrained agent. Many papers address the differences between subjective
and market value, each concluding that the subjective value of ESOs should be
less than the market value (Lambert et al. 1991; Hall and Murphy 2002;
Ingersoll 2006).
If ESOs are generally worth less than market value, why do employees continue
to accept and, indeed, sometimes prefer ESO compensation? One reasonable
explanation is employee sentiment. Sentiment means positive private information
or behavioral overconfidence regarding the future risk-adjusted returns of the firm.
Simply, the employee believes that he possesses private information and can benefit
from it. Or the employee overestimates the return of the firm and believes ESOs are
valuable. Some empirical evidence supports this conjecture. Oyer and Schaefer
(2005) and Bergman and Jenter (2007) posit that employees attach a sentiment
premium to their stock options; firms exploit this sentiment premium to attract and
retain optimistic employees. Hodge et al. (2009) provide survey evidence and find
that managers subjectively value the stock option greater than its Black-Scholes
value.
Three statistics methods are applied in our ESO study (Chen and Fuh 2011;
Chang et al. 2013), including change of measure, hierarchical clustering with a
K-means approach, and estimation of standard errors in finance panel data.
We derive a solution for ESO value that is a function of both illiquidity
and sentiment in a world where employees balance their wealth between the
company’s stock, the market portfolio, and a risk-free asset. By using the method
of change of measure, we find a probability measure and then the ESO formulas are
derived easily. In addition, from the ESO pricing formulas, we are able to not only
estimate the subjective values but also study the exercise policies. Early exercise is
a pervasive phenomenon and, importantly, the early exercise effect is critical in
valuation of ESOs, especially for employees who are more risk averse and when
there are more restrictions on stock holding.
Applying a comprehensive set of executive options and compensation data,
this study empirically determines subjective value, grouping employees by
hierarchical clustering with a K-means approach and backing out the option
844 L.-j. Chen and C.-d. Fuh
In our ESO study, we assume stock price follows a jump-diffusion process. Here,
we briefly introduce change of measure for a compound Poisson and Brownian
motion. Suppose that we have a probability space (O, F , ℙ) on which is defined
a Brownian motion Wt. Suppose that on this same probability space there is defined
a compound Poisson process
X
Nt
Qt ¼ Yi
i¼0
with intensity l and jumps having density function f(y). Assume that there is
a single filtration F t, t0, for both the Brownian motion and the compound Poisson
process. In this case, the Brownian motion and compound Poisson process must be
independent (see Shreve 2008).
30 Statistics Methods Applied in Employee Stock Options 845
ðt
e t ¼ Wt þ
W YðsÞds
0
commonly used metrics of similarity for numeric data. For non-numeric data,
metrics such as the Hamming distance are used. It is important to note that the
actual observations (instances) are not needed for hierarchical clustering, because
only the matrix of distances is sufficient. The user can obtain different clustering
depending on the level that is cut.
A partitional clustering algorithm obtains a single partition of the data instead
of a clustering structure. A problem accompanying the use of a partitional
algorithm is the choice of the number of desired output clusters (usually
called k). One of the most commonly used partitional clustering algorithms is
the K-means clustering algorithm. It starts with a random initial partition and
keeps reestimating cluster centers and reassigning the patterns to clusters based
on the similarity between the pattern and the cluster centers. These two steps are
repeated until a certain intra-cluster similarity objective function and inter-cluster
dissimilarity objective function are optimized. Therefore, sensible initialization
of centers is an important factor in obtaining quality results from partitional
clustering algorithms.
Hierarchical and partitional approaches each have their advantages and disad-
vantages. Therefore, we apply a hybrid approach combining hierarchical and
partitional approaches in this study.
In finance panel data, the residuals may be correlated across firms or across time,
and OLS standard errors can be biased. Petersen (2009) compares the different
methods used in the literature and provides guidance to researchers as to which
method should be used.
The standard regression for a panel data is
Y it ¼ Xit b þ Eit i ¼ 1, . . . , N; t ¼ 1, . . . , T
where there are observations on firms i across years t. X and ϵ are assumed to be
independent of each other and to have finite variance. OLS standard errors are
unbiased when the residuals are independent and identically distributed. However, it
may result in incorrect inference when the residuals are correlated across observations.
In finance study, there are two general forms of dependence: time-series depen-
dence (firm effect), in which the residuals of a given firm may be correlated across
years for a given firm, and cross-sectional dependence (time effect), in which the
residuals of a given year may be correlated across different firms. Considering the
firm effect, the residuals and independent variable are specified as
Both the independent variable and the residual are correlated across observations
of the same firm but are independent across firms. Petersen (2009) shows that OLS,
30 Statistics Methods Applied in Employee Stock Options 847
Fama-MacBeth, and Newey-West standard errors are biased, and only clustered
standard errors are unbiased as they account for the residual dependence created by
the firm effect.
Considering the time effect, the residuals and independent variable are
specified as
OLS standard errors still underestimate the true standard errors. The clustered
standard errors are much more accurate, but unlike the results with the firm effect, they
underestimate the true standard error. However, the bias in the clustered standard error
estimates declines with the number of clusters. Because the Fama-MacBeth procedure
is designed to address a time effect, the Fama-MacBeth standard errors are unbiased.
In both firm and time effect, the residuals and independent variable are
specified as
This section introduces two methods to price the subjective value of ESOs. We call
them the model-based approach and the compensation-based approach. To derive
the model-based subjective value, we use the technique of change of measure and
find a probability measure P* such that the option value can be generated simply. To
calculate compensation-based subjective value, we group employees by hierarchi-
cal clustering with a K-means approach and back out the option value in an
equilibrium-competitive employment market.
From Chen and Fuh (2011), we have a three asset economy in which the
employee allocates their wealth among three assets: company stock S, market
portfolio M, and risk-free bond B. Because of the illiquidity of ESO, the
employee is constrained to allocate a fixed fraction a of their wealth to company
stock (via some form of ESO). Define the jump-diffusion processes for the three
assets as follows:
848 L.-j. Chen and C.-d. Fuh
8
>
> dS XNt
>
> ¼ m dt þ s dW þ ndW þ d ðY i 1Þ,
>
> s s m s
< S i¼0
dM (30.1)
>
> ¼ ðmm dm Þdt þ sm dW m ,
> M
>
>
>
: dB ¼ rdt,
B
where ms ¼ m d lk. m, mm, r are instantaneous expected rates of return for the
stock, market portfolio, and risk-free bond, respectively. d and dm are dividends for the
stock and market portfolio, respectively. The Brownian motion process Wm represents
the normal systematic risk of the market portfolio. The Brownian motion process Ws
and jump process Nt are the idiosyncratic risk of the company stock, where Nt captures
the jump risk of company stock and follows a Poisson distribution with average
frequency l. Yi 1 represents the percentage of stock variation when the ith jump
occurs. Denote E(Yi 1) ¼ k and E(Yi 1)2 ¼ k2 for all i. ss and sm are the normal
systematic portions of total volatility for the stock and the market portfolio, respec-
tively, whereas n is the normal unsystematic volatility of the stock. The two Brownian
motions and the jump process are presumed independent. For simplicity, we assume
that CAPM holds so that the efficient portfolio is the market.
Following the idea of Ingersoll (2006), we solve the constrained optimization
problem and determine the employee’s indirect utility function. The marginal
utility is then used as a subjective state price density to value compensation. We
assume that the employee’s utility function U(·) is set as U(C) ¼ Cg/g with
a coefficient of relative risk aversion, 1 g. The process of the employee’s
marginal utility or the pricing kernel can be derived as2:
dJ W
^ dW m ð1 gÞandW s
¼ ^r dt s
JW
XNt n o (30.2)
þd ½aðY i 1Þ þ 1g1 1 ,
i¼0
∂J ½W ðtÞ, t
where J W ¼ ∂W ðtÞ
is the marginal utility, J[W(t), t] and W(t) are the
employee’s
2 2 1 2 total utility and wealth at time t, ^r ¼ r ð1 gÞ
mm r
a n þ 2 a g l þ a lk , and s ^ ¼ sm .
The rational equilibrium value of the ESO at time t, F(St, t), satisfies the Euler
equation
Et fJ W ½W ðT Þ, T FðST ; T Þg
FðSt ; tÞ ¼ , (30.3)
J W ½W ðtÞ, t
where F(ST, T) is the payoff at the maturity T. To easily calculate the ESO value, we
find a probability measure P* by using change of measure method and then the
second equality in the following Eq. (30.4) is satisfied.
2
Because the process can be similarly derived from Chang et al. (2008), it is omitted.
30 Statistics Methods Applied in Employee Stock Options 849
Et fJ W ½W ðT Þ, T FðST ; T Þg
F ð St ; t Þ ¼
J W ½W ðtÞ, t (30.4)
¼ erðTtÞ Et ½FðST ; T Þ,
dP ¼ ZðtÞ , Z(t) ¼ e
Z ðT Þ r*t
where dP JW[W(t), t], r* is subjective bond yield, and E*t is the
expectation under P and information at time t. Under P*, the stock process can be
*
expressed as
dS XNt
¼ ðr d Þdt þ sN dW t þ d ðY i 1Þ,
S i¼0
where
1
r ¼ r ð1 gÞða lk þ g lk2 a2 þ a2 n2
2
lðx 1Þ,
1
d ¼ d ð1 gÞ alk þ g lk2 a2 ð1 a a n2
2
lðx 1Þ þ lk,
s2N ¼ s2s þ n2 , l ¼ l x, x ¼ E½aðY i 1Þ þ 1g1 ,
W*t is the standard Brownian motion and Nt is a Poisson process with rate l*.
A detailed explanation is given in Appendix 1.
Ker t E F d2
Yj
ln St i¼0 Y i =K þ r d þ 12 s2N t
d1 ¼ pffiffiffi ,
sN t
pffiffiffi
d2 ¼ d1 sN t:
CA ðS0 ; T Þ
X
n1
r‘Dt
¼ CE ðS0 ; T Þ þ e E0 ½S‘ 1 edDt
‘¼1
K 1 erDt I fS‘ S g
‘
X
n
rjDt
e E0 ½CA ðSj , ðn jÞDtÞ Sj K I fSj1 S g I fSj <S g : (30.6)
j1 j
j¼2
The critical price S*i at time ti for i ¼ 1, , n is defined as the solution to the
following equation:
Si K
¼ CE Si , ðn iÞDt
X
ni1
r ‘Dt
þ e Ei Siþ‘ 1 ed Dt
‘¼1
r Dt
K 1e I fSiþ‘ S g
iþ‘
X
ni
er jDt
Ei CA Siþj , ðn i jÞDt
j¼2
Siþj K I fSiþj1 S g I fSiþj <S g ,
iþj1 iþj
believes that they possess private information and can benefit from it. Or
they have behavioral overconfidence regarding future risk-adjusted return
of their firm and believe ESOs are valuable. We consider the impact of sentiment
on ESO values and the exercise decision. Now, the drift term of stock
process in Eq. 30.1 becomes ms ¼ m + s d lk, where sentiment level is
denoted by s. In other words, the employee overestimates or rationally adjusts the
risk-adjusted return of the company owing to inside information by s, then
the same analysis in Theorem 30.1 and 30.2 are valid with a simple adjustment
in parameters. The adjusted interest rate and dividend yield used in pricing are
1
r ¼ r þ as ð1 gÞða lk þ g lk2 a2 þ a2 n2 lðx 1Þ,
2
1
d ¼ d ð1 aÞs ð1 g alk þ glk2 a2 ð1 aÞa n2 lðx 1Þ þ lk:
2
3
Gabaix and Landier (2008) finds that total market value as a proxy for firm size has the strongest
predictive power on compensation. We, however, redo all tests using number of employees as the
size proxy and find qualitatively identical results.
4
The details of hierarchical clustering with a K-means approach and its performance are presented
in Appendix 2.
852 L.-j. Chen and C.-d. Fuh
million in nonoption compensation must then value their options at two million in
order to agree to continued employment. Importantly, it may be the case that the
market value of these options is only $100,000, but the CEO subjectively values them
at $500,000 because they believe the market to have undervalued the options.
Although this method of calculation is clearly not perfectly precise, we tried numerous
robustness checks by using different grouping criteria, all of which arrived at quali-
tatively identical results. Some intangible sources of value, such as training, learning
opportunities, and advantageous work environments, are not controlled here but may
be relatively unimportant given that this is an executive database of listed firms.
Section 30.3 provided a pricing formula for ESOs that includes illiquidity of the
options and sentiment of the employees. Moreover, from this ESO pricing formula,
we can not only estimate the subjective values but also study the exercise policies.
The exercise boundary is endogenously derived by finding the minimum stock price
such that the option value equals its intrinsic value for each time. In other words, the
employee exercises the option when the stock price is above the exercise boundary.
To illustrate our model, this section discusses factors that affect ESO values and
exercise decisions including stock holding constraint, sentiment, level of risk
aversion, moneyness, dividend, time to maturity, total volatility, and normal
unsystematic volatility. We also empirically test these effects in next section.
According to the collected data from Compustat, the model parameters stock
price S, strike price K, total volatility s, dividend yield d, interest free rate r, and
time to maturity t are set to 25, 25, 0.3, 2 %, 5 %, 10, respectively. Normal
unsystematic volatility n is two-thirds of the total volatility following calibrations
applied by Compustat and the majority of papers in the area.5 We employ the
common parameterization for the coefficient of relative risk aversion 1 g ¼ 2 and
three jump size models: double exponential, bivariate jump, and Y ¼ 0 (no residual
value).6 Additionally, default intensity l ¼ 0.01, following Duffee (1999) and
Fruhwirth and Sogner (2006), which use US and German bond data, respectively.
5
See Bettis et al. (2005), Aggarwal and Samwick (2003), Ingersoll (2006), and Bryan et al. (2000).
6
The parameters of double exponential are estimated by daily return data from 1992 to 2004.
A jump occurs if return goes beyond 10 %, which relates to an approximately three-standard
deviation daily return during this period.
30 Statistics Methods Applied in Employee Stock Options 853
Understanding the factors that affect ESO values and the exercise decision is
important for firms designing stock option programs. As we mentioned before,
854 L.-j. Chen and C.-d. Fuh
Fig. 30.1 Exercise boundaries. This figure presents the exercise boundaries according to stock
holding constraint a, sentiment s, level of risk aversion 1 g, moneyness In: K ¼ 20, At: K ¼ 25,
Out: K ¼ 30, where K is exercise price, dividend yield d, time to maturity t, total volatility s, and
idiosyncratic risk n, respectively
ESO values and exercise decisions are closely related. Factors affecting the
employees exercise policies will directly influence the valuation of ESOs. This
section discusses the impact of factors on ESOs. The results are shown in
Table 30.1, which presents the studied factors effect on ESO value, discount
ratio, and early exercise premium, where ESO value is calculated by formula
(30.6), discount ratio is defined as one minus the ratio of subjective to market
value, and early exercise premium is the difference between American and Euro-
pean ESO value.
30 Statistics Methods Applied in Employee Stock Options 855
Unlike traditional options, ESOs usually have a vesting period during which they
cannot be exercised and employees are not permitted to sell their ESOs. In this
situation, employees receive the ESOs in a very illiquid market. Table 30.1 shows
that subjective values (a 6¼ 0) are uniformly smaller than the market values (a ¼ 0).
These results are consistent with Lambert et al. (1991) and Hall and Murphy (2002),
in which the subjective value is lower than market value due to the constrained
fixed holding in the underlying stock. The more risk averse the employee (more
positive 1 g) and more restrictions on the stock holding (larger a), the more they
lean to depreciate the option values and incur the higher early exercise premium.
Note that early exercise effect on ESO values cannot be ignored in these situations.
Because of the restriction of ESOs, many employees have undiversified portfolios
with large stock options for their own firms. Therefore, a risk-averse employee
discounts the ESO values. Discount ratios increase with stock holding constraint
and the degree of risk aversion. In other words, employees who are more risk averse
and more restricted on the stock holdings need to compensate more risk premium.
856 L.-j. Chen and C.-d. Fuh
Employee sentiment enhances the option value and reduces the early exercise
premium. One would expect for options with high sentiment having higher discount
that the option value declines sharply when employees face an undiversification
problem. The moneyness of each option Sok is the stock price at issuance divided
by strike price. If the option is in (out of) the money, Sok is greater (less) than 1. In the
money options have higher values, lower discount and higher early exercise pre-
miums. Interestingly, even in the money options having less discount than out of the
money, employees still tend more to early exercise in the money options to diversify
their wealth portfolio risk. Larger dividends depreciate the option values and induce
employees to exercise their options sooner, even they have lower discount ratios.
More interestingly, the early exercise premium is not zero when no dividends are paid.
This is a departure from traditional option theory, although it is consistent with the
phenomenon that ESOs are exercised substantially before maturity date, even ESOs
not paying dividends, because of the lack of diversification. Options with longer
lifetime have more value; at the same time, they have higher discount ratios and
early exercise premiums. Although not reported in the table, the lifetime of option
may be negatively related to European ESO value. This is due to the longer one must
wait and the greater risk caused by undiversification affecting the ESO value.
Whereas options may provide incentives for employees to work harder, they can
also induce suboptimal risk-taking behavior. General option pricing results show
that value should increase with risk while employees need to compensate more risk
premium at the same time. It is not necessarily the case that subjective value is
positive related to risk, as is the traditional result.7 We have the usual finding that
total volatility increases the option value; however, on the contrary, with respect to
normal unsystematic volatility, the subjective value decreases with it. In the Black-
Scholes framework, this risk is eliminated under a risk-neutral measure. However,
in our model, the employee has an illiquid holding and full diversification is
impossible. Hence, a risk-averse employee depreciates the ESO values. The dis-
count and early exercise premium increasing with the volatility risk also can be
found in Table 30.1. This is intuitive, because the more volatile the stock price, the
higher the opportunity cost of not being able to exercise. Therefore, employees have
more incentives to early exercise volatile options.8
Further we examine the effect of vesting on subjective value. Panel B compares
the ESOs that vest immediately, after 2, and 4 years, respectively. Vesting obvi-
ously reduces the ESO values since it restricts the exercise timing. Discount ratio
increasing with vesting period implies that market value is affected less than the
subjective value. Because the constrained ESOs are usually exercised much earlier
than unconstrained ESOs and more tend to fall afoul of the vesting rule. While
7
Nohel and Todd (2005), Ryan and Wiggins (2001), and others show that option values increase
with risk, however, they do not study the impact of increased idiosyncratic risk. Carpenter (2000)
presents examples where convex incentive structures do not imply that the manager is more
willing to take risks. The model used in Chang et al. (2008) is able to capture this result.
8
While Panel A shows the results of options that vest immediately, we can also consider the
vesting effect and there is no significant qualitative difference.
30 Statistics Methods Applied in Employee Stock Options 857
vesting has negative effect on the American ESOs, it has no effect on the European
ESOs, therefore, early exercise premium decreases with vesting.
The level of sentiment is estimated from two perspectives. First, we consider the
sentiment effect on ESO value (SenV), and then the estimated sentiment level can be
calculated whereby subjective value with sentiment is equal to market value.
Secondly, we estimate sentiment level from the early exercise perspective (SenE),
that is, what value of sentiment such that employees exercise their options at the time
that unconstrained investors do. The sentiment level of European ESOs (SenVE) is
also calculated. Due to the limitation of European options (they are not allowed to
early exercise), the sentiment level can only be estimated from value perspective.
Estimation of sentiment is shown in Table 30.2. Here, we only list the estimated
sentiment level at the money option because there is no obvious relationship
between sentiment level and moneyness. Table 30.2 shows that sentiment estimated
from the American and European ESO formulas have similar patterns. The more
risk averse the employee and more restricted on the stock holding, the higher the
sentiment level is needed. SenVE is slight higher than SenV because of more
restrictions in the European contract.
30.5.1 Data
Data for this study are collected from the Compustat Executive Compensation
(Execucomp) database. From this database, all executive stock options issued
between 1992 and 2004 are collected with stock price at issuance S, strike price
K, maturity date T, implied volatility Vol, and dividend yield Div. While the median
option is issued at the money, the mean is in the money (Sok ¼ 1.012). Note that
virtually all options are issued at the money (Sok ¼ 1). Indeed, this is true for about
90 % of our dataset. Average values of time to maturity, implied volatility, and
dividend yield are 9.3 years,9 0.43 % and 1.37 %, respectively.
9
For some issues for which there is no time stamp, we assume an issuance date of July 1, inasmuch
asthis would be the middle of the fiscal year for the vast majority of firms.
858 L.-j. Chen and C.-d. Fuh
10
For cash paid for purchasing additional stock, where direct data is unavailable, we use the change
in stock holdings times the year-end stock price to calculate this value.
30 Statistics Methods Applied in Employee Stock Options 859
In other words, assume that the executive has no wealth before entering the firm; all
NCasht are realized at the end of the fiscal year and invested at the risk-free rate rt+1
f
during the next fiscal year. Then, a is the sum of all illiquid firm-specific holdings,
including unvested restricted stocks and options, divided by total cumulative wealth.
Average a is about 35 %, implying that the illiquid firm-specific holdings account for
more than one-third of executive total wealth.11 We also calculate a by an iterated
approach that synchronizes a and subjective value simultaneously. Qualitative findings
with respect to sentiment are identical. Several proxies of sentiment are used in an
empirical study including previous year risk-adjusted CAPM alpha and Fortune mag-
azine’s list of top 100 firms to work for, and estimated from our ESO pricing formula.
Substituting the subjective value implied by compensation data into our model
along with the options variables given in our dataset, we are able to back out
sentiment levels, Sen. Results are presented in Table 30.4. There are about 105,000
options issued by each firm (AvgIss) over the test period, with a total of nearly
2,700 firms and 82,000 total observations accounted for. Industry breakdowns,
while exhibiting some fluctuations in point estimates, show that results across
industries are qualitatively similar. While the mean Black-Scholes value of options,
BSOPM, is about $13.09, with some variation across industries, the mean subjective
value, Sub, is more than $19.38, reflecting a 48 % premium. That is, although
virtually all of the theoretical literature implies a subjective value discount, empir-
ical data show that executives generally value ESOs more highly than their Black-
Scholes values. Though not reported in the table, t-tests show that subjective values
11
Holland and Elder (2006) find that rank-and-file employees exhibit an a close to 10 % and concur
that subjective value is decreasing in a because of risk aversion and under-diversification.
860 L.-j. Chen and C.-d. Fuh
Table 30.4 Summary statistics for subjective value and sentiment by industry
Sector BSOPM Sub Sen AvgCom AvgIss Obs
10 Energy 12.589 16.239 0.089 1,774.08 78.46 4,307
15 Materials 10.822 17.613 0.076 1,399.90 61.89 6,412
20 Industrials 12.569 20.305 0.115 1,591.53 70.64 12,134
25 Con. Dis. 12.177 19.799 0.106 2,030.63 98.64 15,925
30 Con. Sta. 12.053 18.121 0.076 2,405.29 111.40 4,347
35 Health care 16.290 21.098 0.115 2,338.68 105.61 8,883
40 Financials 13.440 21.770 0.066 2,892.28 99.37 10,441
45 Inf. Tec. 15.853 18.499 0.229 2,748.47 164.53 14,614
50 Tel. Ser. 12.768 22.941 0.162 5,310.29 272.27 1,324
55 Utilities 5.475 14.633 0.063 1,370.11 67.66 3,931
Others 9.487 9.583 0.208 1,360.34 111.24 56
Total 13.088 19.385 0.120 2,213.73 105.13 82,374
This table presents, by industry: Black-Scholes value BSOPM, subjective value Sub, sentiment
level Sen, average total compensation AvgCom, number of options issued AvgIss, and number of
observations by individual Obs. AvgCom and AvgIss are reported in thousands. Sen is calculated
using the European ESO formula (30.5), where the distribution of jump size follows y ¼ 0. Con.
Dis., Con. Sta., Inf. Tec., and Tel. Ser. refer to Consumer Discretionary, Consumer Staples,
Information Technology and Telecommunication Services, respectively
are statistically significantly higher than Black-Scholes values at the 1 % level for
almost all industries and in aggregate. The only exception is the others industry,
where Sen is still significantly positive but Sub is about equal to BSOPM owing to
a particularly high a in this industry.
Given the large proportion of executive income that is attributed to illiquid, firm-
specific options holdings, this finding suggests substantial overconfidence or pos-
itive inside information regarding their firm’s future prospects. Indeed, the data
show that the average executive prices ESOs such that the firm should outperform
the market’s expectations by an average of 12 % per annum (Sen). T-tests show that
these values are significantly different from zero at the 1 % level in all industries
and in aggregate.
Table 30.5 shows the mean and median values of Rt and Sub in each subsample,
where Rt is the CAPM alpha. Top is a dummy variable taking value 1 if the
executive works for a firm listed in Fortune magazine’s top 100 companies to
work for. Results show that firms with higher previous-year return tend to have
significantly higher subjective values. This is true of both the mean and median
value. Interestingly, subsequent return momentum is not consistently present in this
data, at least as regards mean values. Firms listed in the top 100 in fact make
significantly lower risk-adjusted returns in the year in which they are so listed.
However, they enjoy substantially higher subjective value. This indicates that
sentiment may generally be independent of performance but does significantly
affect subjective value.
Figure 30.2 shows that relative subjective values are greater than one but
relatively stable over time. In contrast, the number of issuances generally increases.
30 Statistics Methods Applied in Employee Stock Options 861
AvgIss Sub/BSOPM
200 4.00
150 3.00
100 2.00
50 1.00
0 0.00
1992 1994 1996 1998 2000 2002 2004
Year
Fig. 30.2 Summary AvgIss and Sub/BSOPM by year. AvgIss and Sub/BSOPM for each year are
graphed in this figure. AvgIss, BSOPM, and Sub are number of options issued, Black-Scholes
value, and subjective value, respectively. The y-axis of the histogram is on the left and that of the
line chart is on the right
The industry with the second highest subjective values (Financials) has a below-
average number of issuances. These observations highlight the importance of
looking at pricing, rather than issuance alone, as high subjective values do not
imply that ESOs will be a more popular financing tool.
We now shift our attention to the testable implications of our model, namely
confirming the relations between key options’ variables and subjective value.
Specifically, we apply the following regression equation:
where Int is the intercept term and all variables are defined as before. Note that for
all results presented here, the calculation of significance is via clustered standard
errors by firm, though OLS results are nearly identical.
862 L.-j. Chen and C.-d. Fuh
First, we apply gross subjective value Sub as the dependent variable. The first
three tests in Table 30.6 use CAPM risk-adjusted alpha from the year prior to option
issuance Rt1 as a proxy for sentiment under the conjecture that those stocks that
performed better in the previous year generate more positive sentiment prior to
options being issued. Note that our model implies that only the risk-adjusted excess
return should be priced because the market portion of the firm’s return is eliminated
via the risk-neutral measure. Bergman and Jenter (2007), in contrast, test the gross
prior year return. Because a year’s worth of data is required to calculate these
alphas, the dataset is reduced to about 57,000 observations. We find that a is
significantly negatively related to subjective value. This matches our intuition
30 Statistics Methods Applied in Employee Stock Options 863
that the larger the proportion of one’s portfolio held in options, the less diversified
the portfolio, and the less valuable the ESO. Sen, on the other hand, is positively
related, and significantly so. In other words, positive sentiment is associated with
higher subjective value. Note that these results control for the usual options’ pricing
factors. Whereas Div is significantly negative related as expected, Sok and t are not
consistently significantly related, and Vol is negatively related. As explored more
fully later, this last negative relation is quite telling and is consistent with our model
as the sensitivity of subjective value to idiosyncratic risk is negative.
Further, the data are split into two groups according to the sign of the product of
Rt and Sen. A positive (negative) sign implies that the positive sentiment measure is
(not) accompanied by strong performance. The positive case (insider), then, can be
explained by nonsentiment-related factors. The executive may have private inside
information and hence be able to forecast future returns. They also have the ability
to affect future returns so that optimism may be self-fulfilling. The negative case
(true sentiment), on the other hand, has not such a concern inasmuch as it would
imply that positive (negative) sentiment is followed by poor (good) performance.
As it turns out, similar results are obtained in both cases: sentiment is positively
related to subjective value while a is negatively related, both significantly
so. As a result, it is not likely that insider information explains whole sentiment
effect on subjective value.
Next, the Top dummy is selected as a proxy for sentiment. Once again, sentiment
is significantly positively related to subjective value while a is significantly nega-
tively related. All other relations are as above.
We also back Sen out of the European ESO formula (30.5) under the aforemen-
tioned three different jump size assumptions.12 Because our model itself determines
the relation between subjective value and Sen, the purpose of these tests is simply to
observe the other variable relations as well as the stability of the model to the
specification of the jump process. Results are quite consistent across the three
processes tested here. All other coefficients remain qualitatively as before with
the coefficient of a, importantly, remaining significantly negative in all cases.
Finally, in order to more clearly test the difference in impact of sentiment for
insider versus true sentiment events, we interact the event identification dummy
with our sentiment proxy as follows:
All variables are defined as before and Sen is the previous-period CAPM alpha,
also as before. DIn is a dummy variable that takes value 1 if the event is insider and
12
Here, sentiment is estimated from the European option formula. It can also be calculated from
the American option formula but with more exhaustive computations. As we mentioned before,
sentiment estimated from the European and American ESO formulas have similar patterns. It may
not affect the regression results much.
864 L.-j. Chen and C.-d. Fuh
Table 30.7 Regression results for insider versus true sentiment events
Int a DInSen DTSen Sok t Vol Div
Sen ¼ Rt1
Coefficient 1.9070 0.2214 0.0372 0.0044 0.0343 0.2866 0.3872 0.0876
(p-value) (<.0001) (<.0001) (<.0001) (<.0001) (0.3769) (0.0876) (<.0001) (<.0001)
This table presents regression results for insider versus true sentiment events, where DIn is
a dummy variable taking value 1 if the event is insider and 0 otherwise, DT takes value 1 if it is
true sentiment and 0 otherwise, and Sen is again defined as the CAPM alpha
We now turn our attention to the sensitivity of subjective value to risk. While we
note that our model implies a positive relation between total risk and subjective
value, it further dictates that the sensitivity of subjective value to idiosyncratic risk
is negative, a notion supported by our empirical findings. This indicates that
increased levels of risk may negatively affect subjective value owing to the inability
of executives to fully diversify their holdings. In contrast, the Black-Scholes as well
as the majority of options pricing models prescribe no role to idiosyncratic risk, that
is, the sensitivity should be zero, and are generally not be able to capture the
empirical finding that subjective value is negatively related to risk.
In applying the empirical data to the formulae for the sensitivities of subjective
value to various forms of risk, our model does indeed generate a negative relation
between firm-specific risk and subjective value, a finding that is consistent also with
the empirical observations of Meulbroek (2001). This finding is particularly impor-
tant as managers can easily affect the firm’s idiosyncratic risk level through various
moral hazard-related activities.
In Table 30.8, risk sensitivities are calculated, vegas, for all options issues in our
dataset assuming there are no illiquid holdings (UV), that is, a ¼ 0, and using our default
value for a (V), with and without consideration of sentiment. The first two columns find
as expected that the sensitivity with respect to total risk is positive, for both UV and V,
regardless of whether sentiment is considered or not. This is true of all jump specifica-
tions. In every case, the sensitivity is higher when sentiment is not considered. Looking
at the vegas with respect to jump frequency, UV(freq) can be either positive or negative
depending on the jump specification, while V(freq) is always negative. Interestingly,
30 Statistics Methods Applied in Employee Stock Options 865
UV is positive for the constant jump model but negative for the other two models,
pointing out the importance of jump specification when liquidity is not also considered.
The magnitude of UV is always smaller than that of V.
Perhaps the most interesting factor affecting our subjective value in our model is
idiosyncratic risk, for which the estimate is always negative and is significantly
larger in magnitude than the other vegas. While the jump size vega also plays a role
and is likewise always negative, the magnitude of this effect is much smaller. This
finding highlights the role of idiosyncratic risk in our model and explains why the
empirical sensitivity of subjective value to volatility is found to be negative,
contrary to generally accepted moral hazard models that dictate that option com-
pensation encourages risk taking. If agents are sufficiently under-diversified, the
risk premium from taking on excess idiosyncratic risk offsets gains from convexity
and discourages risk-taking behavior. The corresponding UVs for idiosyncratic and
jump size risk are both zero as these do not play a role in determining market value
866 L.-j. Chen and C.-d. Fuh
when there are no under-diversified holdings. Also, the Vs are substantially more
negative when sentiment is introduced, pointing out the sharply offsetting effects of
positive sentiment and risk aversion in this model. Which piece dominates then
depends on the risk aversion parameter and a of the employee.
30.6 Conclusion
This chapter applies three statistics methods in ESO study, including change of
measure, hierarchical clustering with a K-means approach, and estimation
of standard errors in finance panel data. We use a model for employee stock options
that incorporates illiquidity of the options, a jump diffusion for the stock price
evolution that includes various jump processes, and the potential roles of employee
sentiment and insider information in a world where employees balance their wealth
among the company’s stock, the market portfolio, and a risk-free asset. Our option
contract is American type and the optimal exercise boundary is derived endoge-
nously. From the ESO pricing formula, we are able to not only estimate the
subjective values but also study the exercise policies.
The subjective value placed on ESOs implied by compensation data is calculated
by applying empirical data. Specifically, using data provided by Compustat, execu-
tives are grouped by a hierarchical clustering with a K-means approach based on
a number of firm and individual criteria. By assuming that all executives in the same
cluster receive the same total compensation, a notion that relies on the existence of
competitive labor markets, we then back out the valuation placed by each executive on
their respective ESO. These groups include consideration of nonoption compensation,
rank, industry, year, firm size, and immediate exercise value. Though the extant
literature predicts that employees should discount the value of their options, we find
that executives in fact value their options more highly than implied by Black-Scholes,
applying an average premium of 48 %. As such, the cost of issuance for the firm is
vastly lower than the benefit perceived by employees, suggesting that ESO compen-
sation should be an even larger part of executive compensation. We then relate
subjective value to sentiment levels and generate the novel finding that executives
must expect their firm’s risk-adjusted returns to outpace that predicted by the market
by 12 % in order to justify the subjective value placed on ESOs. This expectation may
be the result of private information regarding the growth prospects of the firm.
Testing subjective value and its relation to pertinent variables, subjective value is
negatively related to the proportion of wealth held in illiquid firm-specific holdings and
positively related sentiment. In other words, the larger the illiquid ESO position is, the
larger the discount risk aversion prescribed and the lower the subjective value implied in
the compensation package. On the other hand, the more positive the employee’s view of
future risk-adjusted returns, the more valuable the ESO. Interestingly, subjective value
may be negatively related to risk as the inability of executives to fully diversify their
holdings may lead to risk premia that outweigh the value placed on risk by the convexity
of options payouts. Note that this relation is particularly negative with regard to
idiosyncratic risk and is empirically also negative for risk associated with both jump
30 Statistics Methods Applied in Employee Stock Options 867
frequency and size. Because these aspects of return are precisely those that may be most
directly controlled by executives, traditional moral hazard arguments relating solely to
the convexity of the options payout may not hold.
Firms increasingly grant nontraditional employee stock options to link stock price
performance and managerial wealth and provide greater incentives to employees.
While this study focuses on the traditional employee stock option, the main intuition
can be involved in nontraditional ESOs. Premium stock option, performance-vested
stock option, repriceable stock option, purchased stock option, reload stock option,
and index stock option are the objects of future study. We can derive the option
formulas and compare the value, incentive effect, and cost per unit of subjective
incentive across the nontraditional ESOs and the traditional ones. This future study
provides a firm a proper compensation vehicle according to its characteristics.
Here we introduce the idea for deriving the ESO formulas and define using
probability measure P* by the technique of change of measure.
The process of an employee’s marginal utility or the pricing kernel is:
dJ W
^ dW m þ ^n dW s þ Y^ 1 dN t ,
¼ ^r dt s
JW
where
1
^r ¼ r ð1 gÞ a2 n2 þ a2 glk2 þ alk ,
2
mm r
^¼
s ,
sm
^n ¼ ð1 gÞa n,
^ ¼ ½aðY 1Þ þ 1g1 ,
Y
then
YNt
1 2 1 2 ^ i:
J W ½W ðtÞ, t ¼ J W ½W ð0Þ, 0exp ^r s ^ W m ðtÞ þ ^n W s ðtÞ
^ ^n t þ; s Y
2 2 i¼0
Let B(t, T) be the price of a zero coupon bond with maturity T, then
J W ½W ðT Þ, T
Bðt; T Þ ¼ E BðT; T ÞjF t g
J W ½W ðtÞ, t
( Y )
Nt
1 2 1 2 ^i
¼ E exp ^r s ^ ^n t þ s ^ W m ðtÞ þ ^n W s ðtÞ Y
2 2 i¼0
where
n o
^ ¼ E ½aðY 1Þ þ 1g1 :
t ¼ T t, x ¼ E Y
1
r lnBðt; T Þ
Tt
1
¼ r ð1 gÞðalk þ g lk2 a2 þ a2 n2 lðx 1Þ:
2
Let
YNt
1 2 1 2 ^i ,
ZðtÞ e J W ¼ exp
rt
s ^ W m ðtÞ þ ^n W s ðtÞ
^ ^n lðx 1Þt þ s Y
2 2 i¼0
J W ½W ðT Þ, T Z ðT Þ
¼ erðTtÞ :
J W ½W ðtÞ, t Z ðtÞ
The rational equilibrium value of the ESO at time t, F(St, t), satisfies the Euler
equation,
J W ½W ðT Þ, T
Fð S t ; t Þ ¼ Et Fð ST ; T Þ
J W ½W ðtÞ, t
¼ erðTtÞ Et ½FðST ; T Þ,
dP ¼ Zð0Þ, and Et is the expectation under P and information at time t. Under
Z ðt Þ * *
where dP
the probability measure P , the processes W m ¼ W m s
*
^ t and W s ¼ W s ^n t are
Brownian motions, Nt is a Poisson process with intensity l* ¼ lx, and the jump
sizes follow density f*Y(y),
1
f Y ðyÞ ¼ ½aðy 1Þ þ 1g1 f Y ðyÞ:
x
Therefore,
dS
¼ ms dt þ ss dW m þ ndW s þ ðY 1ÞdN t
S
¼ r d ð1 gÞan2 lk dt
þ ss dW m þ ndW s þ ðY 1ÞdN t
ðr d Þdt þ sN dW t þ ðY 1 dN t ,
30 Statistics Methods Applied in Employee Stock Options 869
where
1 2
d ¼ d ð1 gÞ a lk þ g lk2 a ð1 a an
2
2
lðx 1Þ þ lk,
s2N ¼ s2s þ n2 ,
sN W t ¼ ss W m þ nW s :
In other words,
YNt
1 2
St ¼ S0 exp r d sN t þ sN W t Yi:
2 i¼0
Fig. 30.3 Natural log of total compensation. The box plots show the natural log of total
compensation for the two largest industries in our sample. Executives are grouped according to
position, the firm’s total market value, nonoption compensation, and the immediate exercise
value of the options for each industry by using a hierarchical clustering with K-means
approach
where
h Yj i
ln St i¼0 Y i =K þ r d þ 12 s2N t
d 1 ¼ pffiffiffi ,
sN t
pffiffiffi pffiffiffi
d 2 ¼ d 1 sN t, a1 ¼ d2 t:
References
Aggarwal, R. K., & Samwick, A. A. (2003). Performance incentives within firms: The effect of
managerial responsibility. Journal of Finance, 58, 1613–1649.
Agrawal, A., & Mandelker, G. N. (1987). Managerial incentives and corporate investment and
financing decisions. Journal of Finance, 42, 823–837.
Bergman, N. K., & Jenter, D. (2007). Employee sentiment and stock option compensation. Journal
of Financial Economics, 84, 667–712.
Bettis, C. J., Bizjak, J. M., & Lemmon, M. L. (2005). Exercise behavior, valuation, and the
incentive effects of employee stock options. Journal of Financial Economics, 76, 445–470.
Bryan, S., Hwang, L., & Lilien, S. (2000). CEO stock-based compensation: An empirical analysis
of incentive-intensity, relative mix and economic determinants. Journal of Business, 73,
661–693.
Carpenter, J. N. (2000). Does option compensation increase managerial risk appetite? Journal of
Finance, 55, 2311–2331.
Chang, C., Fuh, C. D., & Hsu, Y. H. (2008). ESO compensation: The role of default risk, employee
sentiment, and insider information. Journal of Corporate Finance, 14, 630–641.
Chang, C., Chen, L. J., & Fuh, C. D. (2013). The pricing of risk and sentiment: A study of
executive stock options. Financial Management, 42, 79–99.
Chen, L. J., & Fuh, C. D. (2011). A study of employee stock options and the exercise decision.
International Research Journal of Finance and Economics, 67, 142–159.
Core, J. E., & Guay, W. (2001). Stock option plans for non-executive employees. Journal of
Financial Economics, 61, 253–287.
Dittmann, I., & Maug, E. (2007). Lower salaries and no options? On the optimal structure of
executive pay. Journal of Finance, 62, 303–343.
Duffee, G. R. (1999). Estimating the price of default risk. The Review of Financial Studies, 12,
197–226.
Fruhwirth, M., & Sogner, L. (2006). The Jarrow/Turnbull default risk model-evidence from the
German market. European Journal of Finance, 12, 107–135.
Gabaix, X., & Landier, A. (2008). Why has CEO pay increased so much? Quarterly Journal of
Economics, 123, 49–100.
Gukhal, C. R. (2001). Analytical valuation of American options on jump-diffusion processes.
Mathematical Finance, 11, 97–115.
Hall, B. J., & Murphy, K. J. (2002). Stock options for undiversified executives. Journal of
Accounting and Economics, 33, 3–42.
Hemmer, T., Matsunaga, S., & Shevlin, T. (1996). The influence of risk diversification on the early
exercise of employee stock options by executive officers. Journal of Accounting and Econom-
ics, 21, 45–68.
Hemmer, T., Kim, O., & Verrechia, R. K. (2000). Introducing convexity into optimal compensa-
tion contracts. Journal of Accounting and Economics, 28, 307–327.
Hodge, F. D., Rajgopal, S., & Shevlin, T. (2009). Do managers value stock options and
restricted stock consistent with economic theory? Contemporary Accounting Research, 26,
899–932.
Holland, L. C., & Elder, E. M. (2006). Employee stock options in compensation agreements:
A financing explanation. Journal of Corporate Finance, 12, 367–379.
872 L.-j. Chen and C.-d. Fuh
Huddart, S., & Lang, M. (1996). Employee stock option exercises: An empirical analysis. Journal
of Accounting and Economics, 21, 5–43.
Hull, J., & White, A. (2004). How to value employee stock options. Financial Analysts Journal,
60, 114–119.
Ingersoll, J. E. (2006). The subjective and objective evaluation of incentive stock options. The
Journal of Business, 79, 453–487.
Jain, A. K., Murty, M. N., & Flynn, P. J. (1999). Data clustering: A review. ACM Computing
Surveys, 31, 264–323.
Jensen, M. C., & Meckling, W. H. (1976). Theory of the firm: Managerial behavior, agency costs,
and ownership structure. Journal of Financial Economics, 3, 305–360.
Lambert, R. A., Larcker, D. F., & Verrecchia, R. E. (1991). Portfolio considerations in valuing
executive compensation. Journal of Accounting Research, 29, 129–149.
Liao, F. Y., & Lyuu, Y. D. (2009). An expanded model for the valuation of employee stock
options. Journal of Futures Markets, 29, 713–735.
Meulbroek, L. K. (2001). The efficiency of equity-linked compensation: Understanding the full
cost of awarding executive stock options. Financial Management, 30, 5–44.
Nohel, T., & Todd, S. (2005). Compensation for managers with career concerns: The role of stock
options in optimal contracts. Journal of Corporate Finance, 11, 229–251.
Oyer, P., & Schaefer, S. (2005). Why do some firms give stock options to all employees?
An empirical analysis of alternative theories. Journal of Financial Economics, 76, 99–133.
Petersen, M. (2009). Estimating standard errors in finance panel data sets: Comparing approaches.
Review of Financial Studies, 22, 435–480.
Ryan, H. E., & Wiggins, R. A. (2001). The influence of firm- and manager-specific
characteristics on the structure of executive compensation. Journal of Corporate Finance, 7,
101–123.
Shreve, S. E. (2008). Stochastic calculus for finance II: Continuous-time models. New York:
Springer.
Structural Change and Monitoring Tests
31
Cindy Shin-Huei Wang and Yi Meng Xie
Contents
31.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 874
31.2 Structural Breaks in Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 875
31.2.1 Stationary Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 875
31.2.2 Multiple Breaks Case . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 876
31.2.3 Prediction Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 878
31.2.4 Structural Changes in Long-Memory Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 879
31.3 Persistent Change in Time Series . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 884
31.3.1 Tests for a Change in Persistence for I(0) or I(1) Process . . . . . . . . . . . . . . . . . . . . 884
31.3.2 Unit Root with Persistent Change . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 887
31.4 Tests for Special Structural Breaks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 888
31.4.1 Bubble Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 888
31.4.2 Cointegration Breakdown Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 893
31.5 Monitoring Structural Breaks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 896
31.5.1 Monitoring in Comparably Ideal Conditions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 896
31.6 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 900
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 900
Abstract
This chapter focuses on various structural change and monitoring tests for a class
of widely used time series models in economics and finance, including I(0), I(1),
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 873
DOI 10.1007/978-1-4614-7750-1_31,
# Springer Science+Business Media New York 2015
874 C.S.-H. Wang and Y.M. Xie
Keywords
Cointegration breakdown test • Structural break • Long-memory process •
Monitoring fluctuation test • Boundary function • CUSUM squared test •
Prediction test • Bubble test • Unit root time series • Persistent change
31.1 Introduction
The structural break refers to the phenomenon illustrating the time series comes
across unanticipated and significantly influential shifts. Generally, most macroeco-
nomic and financial time series are subject to occasional structural breaks.
Ignoring the presence of structural breaks can lead to seriously biased estimates
and forecasts and the unreliability of models (see Clements and Hendry 1998,
1999). Since economic activities often experience rather long period, there exist
plausible possibilities that structural breaks could occur. Hence, it is very important
to monitor and modify the changes of the model characteristics. If those model
characteristics were not corrected in real time, it may lead to wrong decisions on the
policy making and investment plans for policy makers and practitioners. Conse-
quently, those wrong decisions could result in the fluctuations or crises in economy
and finance.
Examples in the real world are a large sum of financial crises ever since before
the Industrial Revolution. For instance, ten enormous bubbles took place ahead of
the twenty-first century documented by Kindleberger (2005), and the latest
subprime crisis occurred in 2008. All crises crushed the wealth of the whole
world again and again. After the first reported financial crisis – Dutch tulip bulb
bubble in 1636 – all meltdowns in the economy systems were obviously character-
ized by the deviation from steadily development.
It implies that structural breaks could happen prior to the bubbles or explosions.
Accordingly, an accurate approach to detecting structural breaks correctly can be
treated as a measurement to notice the possible incoming financial crisis or to
monitor market activities. On the other hand, developing a suitable structural
31 Structural Change and Monitoring Tests 875
Structural break has been an intriguing issue since Chow (1960) test resolves how to
detect a single structural change with known break date, and more ensuing litera-
tures expand that in different dimensions. Break numbers can be multiple while
break dates are no longer required as necessitates. Moreover, econometricians
apply research closely to reality, adjusting these models from classical stationary
process to nonstationary and long memory, which are found more likely to reflect
economic and financial data. Now, we will offer systematical analysis of these
marked developments.
It is the Chow (1960) test that firstly detect a potential structural break for a given
break date. Suppose the model is
0 0
yt ¼ bt Xt1 þ et , (31.1)
where our null hypothesis is bt ¼ b for all t, et is a martingale difference sequence
with respect to the s fields, and Xt is a k 1 vector of regressor, which are here
0
assumed to be constant and/or I(0) with EXtXt ¼ SX and, possibly, a nonzero mean.
Particularly, Xt1 can include lagged dependent variables as long as they are
I(0) under the null. When the break date is unknown, a natural solution proposed
by Quandt (1960) calculates break date, and it is extended by Davies (1977) to
general models with parameters unidentified under the null. And Andrews (1993)
876 C.S.-H. Wang and Y.M. Xie
reports asymptotic critical values of this QLR statistics. in the form of Quandt
likelihood ratio (QLR) statistic.
To the contrast of the null, the alternative hypothesis of a single break in some or
all of the coefficients is
bt ¼ b, tr and bt ¼ b þ g, t > r, (31.2)
where r, k + 1 < r < T, is the “break date” (or “change point”) and g 6¼ 0.
When the potential break date is known, a natural test for change in b is the
Chow (1960) test, which can be implemented in asymptotically equivalent Wald,
Lagrange multiplier (LM), and LR forms. In the Wald form, the test for a break at
a fraction r/T through the sample is
SSR1, T SSR1, r þ SSRrþ1, T
FT ðr=T Þ ¼ , (31.3)
SSR1, r þ SSRrþ1, T =ðT 2kÞ
where SSR1, r is the sum of squared residuals from the estimation of Eq. 31.1 on
observation 1,. . ., r, etc. For fixed r/T, FT(r/T) has an asymptotic w2k distribution
under the null. However, when break date is unknown, the situation can be more
complicated. An expectable idea is to estimate the break date and then compute
Eq. 31.3 for that break. But since the change point is selected by virtue of an
apparent break at the point, the null distribution of the resulting test is not the same
as if the break date were chosen without regard to the data. Thus means of
determining r/T must be further specified before the distribution of the resulting
test can be obtained. Quandt (1960) and Davies (1977) come up with another FT
statistics called Quandt likelihood ratio, thus QLR statistic:
QLR ¼ MAX FT ðr Þ (31.4)
r¼r 0 , ..., r1
Intuitively, it has power against a change in b even without a known break date.
Although it confuses econometricians for years to calculate its null asymptotic
distribution, Andrews (1993, Table I) reports computable critical values for it. For
more details, please see Stock (1994).
This issue of detecting structural breaks becomes even richer after the possible
multiple breaks are considered no matter the numbers and dates of breaks are
known or not. Bai and Perron (1998) proposed a comprehensive treatment for
estimating linear models with multiple structural breaks. They set the multiple
linear regression with m breaks (m + 1 regimes) as follows:
yt ¼ x0t b þ z0 dj þ ut , (31.5)
where
Li, ¼ t; T^i1 þ T^i T^i1 t T^i T^i T^i1 (31.8)
and s^ 2 is a consistent estimate of s2 under the null hypothesis. Note that for i ¼ 1,
^
ST T 1, . .., T^ði 1Þ,t, T^i , . .., T^l is understood as ST t; T^1 ;.. .; T^l and for i ¼ l + 1
as ST T^1 ;. ..; T^l ;t .
As well, Bai and Perron (1996) also provide a valid proof for consistency of the
estimated break fractions along with its limiting distributions. They also apply their
model to the situations where autocorrelation could happen and compare it with
sequential estimation.
878 C.S.-H. Wang and Y.M. Xie
Prediction tests which are often applied to forecast stock and exchange rate also
lose power when potential structural change is incorporated without reasonable
adjustment. While numerical literatures explore suitable models to fit the reality,
their efforts still fail for rather long period or large forecasting sample as proven in
Welch and Goyal (2008). This means tests for structural stability are equivalently
vital in forecasting. For time series satisfying characters of stationarily I(0),
prediction tests are operated on the difference between observed variables and
predictions and only when it is before the suspected structural change could such
tests be valid. Lutkepohl (1989) proposes a prediction tests for structural stability of
univariate times series and compares its efficacy to the univariate time series
investigated by Lutkepohl (1988).
Lutkepohl (1989) assumes a multiple time series generated by a stationary vector
stochastic process yt ¼ (y1t,. . ., ykt) in autoregressive (AR) representation
X
1
yt ¼ v þ Ai yti þ ui (31.9)
i¼0
X
1
yt ¼ m þ Fi uti , F0 ¼ I k , (31.10)
i¼0
X
1
y t ð hÞ ¼ m ¼ Fi utþhi ,
i¼h
X
h1
SðhÞ ¼ Fi Su F0 :
i¼0
for h ¼ 1, 2, . . ., can be used for testing whether a structural change has occurred
after period t. As usual w2(K) denotes the chi-squared distribution with Kdf.
Besides the univariate description of this system, another test relevant for
multivariate systems is also designed by Lutkepohl (1989). Since the vector of
one-step forecast errors to h-step forecast errors is also normally distributed,
Here
SðhÞ ¼ Fh ðI h Su Þ Fh 60 , (31.13)
_ 1 I k 0F
Fh ¼ I k 0F _ h1 Fh2 I_ (31.14)
k
Hence,
for h ¼ 1, 2, . . ., is another sequence of statistics that can be used for stability tests.
This L(h) statistics are particularly easy to compute and according to Lutkepohl
(1989), while univariate time series tests can be relevant for multivariate systems,
a structural change in a multivariate system will not be necessarily be reflected in
the univariate models for the individual variables of the system. And Lutkepohl
(1989) points that since the coefficients Fi, Ai, and Su are usually unknown so that
the statistics cannot be computed, the replacement by estimators is needed. Thus the
estimated processes T^ðhÞ ! w2 ðK Þ and L^ðhÞ ! w2 ðhK Þ can match the null hypoth-
esis of structural change. Generally, Lutkepohl (1989) proposes a prediction test for
checking the structural stability of a system of time series. Despite the fact it lacks
power in modelling the form of the change, it serves as a useful tool without
imposing severe restrictions on the data. Another merit is that very few data are
required after the time period or the time point at which a structural change is
suspected.
So far long memory has been recognized as one of most common features shared by
macroeconomic and financial time series. Let Xt be a stationary process with auto-
covariances r(k) ¼ cov(Xt, Xt+k), then Xt is said to have long memory if as jkj ! 1,
880 C.S.-H. Wang and Y.M. Xie
1
r ðk Þ L 1 ðk Þj k j 2H2
, H2 ;1
2
1 X 1
f ðx Þ ¼ r ðkÞeikx L2 ðxÞjxj1=2H
2p k¼1
is jxj ! 0 for some L2(·) slowly varying the origin. The explanation for this
phenomenon is not elusive since these economic data always persist for a rather
long period and dependence between time thus has a lasting influence. Yet tradi-
tional wisdoms usually appear stuck in analysis of long memory, and thus need for
structural break detection requires econometricians to develop singular new esti-
mations for use. Among these manufactory efforts attempt to deal with long
dependence either invariant to time or matching the time future. Hereafter more
methods will be discussed.
f t ¼ o þ gt1 þ vt , (31.16)
By definition, t is the expected mean squared error of the best linear prediction
of Xt given Xs, s t 1. The long-memory behavior is characterized by H, the
additional parameters 2, . . ., m allow for flexible modelling of short-term features.
Define
ðp
ak ð Þ ¼ eikx f 1 fx; ð1; Þgdx:
p
X
N
Q ð Þ ¼ aij ðÞ Xi X Xj X :
i, j¼1
For Gaussian processes, the asymptotic covariance matrix is the same as for the
exact maximum likelihood estimator. As proved by Beran and Terrin (1996), for
0 < t < 1,
882 C.S.-H. Wang and Y.M. Xie
X
bNtc
Q1 ðt; Þ ¼ aij ðÞ Xi X Xj X ,
i, j¼1
X N
Q2 ðt; Þ ¼ aij ðÞ Xi X Xj X :
bNtcþ1
k2 ¼ 2D
11 (), where D is the k k matrix with elements
1
ðp
∂
Dij ¼ ð2pÞ1 log f fx; ð1; Þgdx:
p ∂i
where B1 and B2 are two independent standard Brownian motions. Based on this
distribution, the test statistics is suggested as
T N ¼ supd<t<1d
Ze N ðt Þ
(31.18)
bðsÞ ¼ bA , 0 s t
¼ bB , t < s < 1,
for a known number t between zero and one, where the unobservable error ut has
auto-covariances featuring long memory. Thus the null hypothesis is H0 : bA ¼ bB
against the alternative H1 : bA 6¼ bB.
For the case of non-stochastic regressors, a type of Wald testing procedure is
applied. For a given t, define h ¼ [tn], where [·] indicated integer part. Correspond-
ingly, let X1 ¼ (x1, . . ., xh)0 , X2 ¼ (xh+ 1, . . ., xn)0 , Y1 ¼ (y1, . . ., yh)0 , Y2 ¼ (yh+ 1, . . ., yn)0 ,
and then estimate bA and B by
0 1 0
b^A ¼ X1 X1 X1 Y 1 ,
0 1 0
b^B ¼ X2 X2 X2 Y 2
Put
0 1 0 0 1 0 0
W ¼ X1 X1 X1 , X2 X2 X2 ,
0
u ¼ ð u1 ; . . . ; un Þ ,
b^A b^B ¼ W u:
0
0 1=2
W GW b^A b^B N ð0, I k Þ:
Hidalgo and Robinson (1996) also allowed for a degree of generality in the
specification of the error structure, and the test is mainly designed for liner
regression model by applying a semiparametric but a parametric model. Besides,
the change point is assumed to be already known. Yet Kuan and Hsu (1998)
point that such a test would incorporate size distortion that could lead to
a change point while there is none. This is not peculiar since according to
Kuan and Hsu (1998) when the memory parameter is in (0, 0.5), many well-
known structural changes are supposed to suggest a nonexistent change. They
also indicate a spurious change which might arise from stationary data with long
884 C.S.-H. Wang and Y.M. Xie
Permanent property shifts of the time series are characters of many key macroeco-
nomic and financial variables in developed economies. Hence, correctly character-
izing the time series whether stationary or nonstationary would be very helpful to
build the accurate models. This issue has been investigated in terms of the data
property change from I(0) to I(1) or the vice versa. The change might be responsible
for portfolio adjustment especially for financial crises. Furthermore, recent studies
indicate that other than those classical structural break types, there could also exist
persistent changes in processes characterized by long memory, for example, the
long-memory parameter changes from d1 to d2 or the vice versa, where d is the
fractional differencing parameter.
A normal question in testing for the persistent change is about what the process used
to be and what direction it changes to. To solve that systematically, we could set the
data-generating process (DGP) as following:
yt ¼ d t þ vt , (31.21)
vt ¼ rt vt1 þ et , t ¼ 1, . . . T: (31.22)
Eq. 31.21 consists of the deterministic kernel dt which is a constant either plus
linear time trend or nothing. The error term et is stationary.
Within the model (31.21), there exist four possibilities. The first is that yt is all
the time I(0) as rt ¼ 1 for all t, which can be defined as H1, and the second is
denoted as H01, reflecting Yt changing from I(0) to I(1) at time bt*c, where b·c
represents the integer part. In this situation, rt ¼ r, jrj < 1 t bt*c and rt ¼ 1 for
t > bt*c in the context of (i). Noted that the change-point proportion t* is assumed
to be unknown while set in L ¼ [LL, LU], which is an interval symmetric around 0.5
and between zero and one. The third denoted as H10 is that yt is form I(1) to I(0) at
time t*.
31 Structural Change and Monitoring Tests 885
Extant tests to detect the above structural changes are CUSUM of squares-based
tests and regression-based tests. While the former has been improved by Leybourne
et al. (2006) as displaying no tendency to spurious over-rejection, modified tests
from regression angle by LKSN still lack the power for special structural change,
taking large sample, for instance, where tests reject H1 with probability one even
though there is no change. Owing to these reasons, we here provide the CUSUM
test according to Leybourne et al. (2006) as the ideal way in detecting this kind of
persistent change.
In order to test the null hypothesis of constant I(0) behavior H1 against a change
in persistence from I(0) to I(1), H01, a standardized CUSUM of squared subsample
OLS residuals is needed:
2
X½tT
½ tT v^2
t¼1 t, t
K f ð tÞ ¼ (31.23)
w^ 2f ðtÞ
where
X
½tT
v^t, t ¼ yt yðtÞ with yðtÞ ¼ ½tT 1 yt :
t¼1
And similarly H(10) equivalent to change from I(0) to I(1) in the reversed series,
let xt
y(T t + 1), occurring at time (T [t*T]) and the reversed series can be
obtained as
XT½tT
ðT ½tT Þ2 e
v 2t, t
K r ð tÞ ¼ t¼1
(31.24)
^ 2r ðtÞ
o
where
X
T btT
e
v t, t ¼ xt xð1 tÞ with xð1 tÞ ¼ ðT ½tT Þ1 xt :
t¼1
Notice for both forward and reversed series, the long-run variance o is replaced
by the estimator
X
m
^ 2f ðtÞ ¼ ^g 0 þ 2
o ws, m ^g s , (31.25)
s¼1
X
btT c
^g s ¼ btT c1 D^
v t, t D^
v ts, t , (31.26)
t¼1
Therefore, a test against persistent change could be based on the ratio R of the
forward and reverse CUSUMs of squared statistics, i.e.,
886 C.S.-H. Wang and Y.M. Xie
inf t2L K f ðt Þ N
Rð t Þ ¼ r ¼:
inf t2L K ðt Þ D
þ
X
½tT X
T ½tT
K ðtÞ ¼ ½tT 2 K ðtÞ ¼ ðT ½tT Þ2
r
v^2t, t and e
v 2t, :
t¼1 t¼1
And the large sample behavior of R, N, and D under both H01 and H10 will be
provided in their Theorem 2 along with consistent estimators of t*. Additionally,
the results of R in their Theorem 2 imply a consistent test of H1 against H01 or H10,
respectively, by the left-tail and the right-tail distribution of R, while the
unstandardized form yields a consistent estimate of break date t*. What’s more,
even the direction of the change is unknown, such a test with two tails can be
appropriate against either H01 and H10. As for the limiting distribution of R under
H0, it remains available for both the demeaned and de-trend cases. Thus, we could
further know that under H0, R !p 1.
Although Leybourne et al. (2006) rule out the size distortion when applied to
process displaying constant persistence, they neglect such a test for long-memory
case could cause over-rejection, which is often very serious. Sibbertsen and Kruse
(2009) provide a solution based on simulation, by which they find despite that
critical values may need modified, estimators for break dates still hold consistently.
In this sense, the yt generated from DGP is assumed to follow an ARFIMA(p,d,q)
process, which means only the memory parameter d determine the degree of
integration of yt. Then the hypothesis should be written as
H 0 : d ¼ d0 for all t,
H1 : d ¼ d1 for t ¼ 1, . . . , ½tT
d ¼ d2 for t ¼ ½tT þ 1, . . . , T
nonstationary long memory at some unknown breakpoint with the vice versa also
valid. Similarly, parts constituting R statistics are thus defined by
½tT
X X
T ½tT
K f ðtÞ ¼ ½tT 2d0 v^t, t , r ðtÞ ¼ ðT ½tT Þ2d0 v^t, t :
t¼1 t¼1
However, Sibbertsen and Kruse (2009) find that asymptotic property has
changed since the limiting distribution depends strongly on the memory parameter,
which still leads to heavy size distortion. To deal with this problem, they simulate
the adjusted critical values which are dependent on d. Although in this way
practitioners benefit as such sample size will not influence critical values as
significantly much as memory parameter, they have to decide the exact d for this
simulation.
Conventional unit root tests usually seem problematic when they are applied to
nonstationary volatility, although they are proved to perform well in other situa-
tions. Breaks following a stationary Markov switching process and time-varying
conditional variances are proved not to cause significant size distortion and impact
on unit root and cointegration tests. Similarly, stationary time-varying conditional
variances including (ARCH) are also known to have no impact on unit root and
cointegration tests. Conversely, it can be different for permanent changes in
volatility (so that volatility is nonstationary), since it is easy to greatly affect unit
root inference. Hamori and Tokihisa (1997) show that a single abrupt increase in the
innovation variance increases the size of augmented Dickey and Fuller (1979,
1981) (ADF) tests when no deterministic components are present. The similar
situation happens to Kim et al. (2002) who report severe oversizing in the
constant-corrected ADF tests in the presence of an early single abrupt decrease in
the innovation variance. Cavaliere (2004) proposes a more general framework for
investigating the effects of permanent changes in volatility by showing that the
limiting null distributions of the Phillips and Perron (1988) (PP) test statistics
depend on a particular function of the underlying volatility process which can
lead to either oversizing or undersizing in the tests.
Cavaliere and Taylor (2006) consider both smooth volatility changes and mul-
tiple volatility shifts and construct a numerical solution to obtain approximate
quantiles from the asymptotic null distributions of the standard unit root statistics.
They focus on M unit root tests by Perron and Ng (1996) which is still robust when
applied to autocorrelation. Set the following model to generate the time series
process
0
0
Xt gt ¼ a Xt1 g Zt1 þ ut t ¼ 1, 2, . . . , T (31.29)
888 C.S.-H. Wang and Y.M. Xie
where zt is a vector of deterministic components, thus set as zt ¼ (1, t, . . .,t p)0 , while
et ¼ st et, et iid (0, 1) which is responsible for the heterogeneity. And besides X0 is
of Op(1). Furthermore, the M statistics for a given sample {Xt}T0 is defined as
where X^ are the OLS residuals from the regression of Xt on zt, t ¼ 0,. . ., T and s2AR(k)
t
There also exist several types of structural breaks that have exclusive features
distinctly from common economic phenomena. For example, a financial bubble
being originated or bursting indicates a significant and catastrophical structural
break, which is capable of wiping out the great wealth of people as much as
possible. Additionally, the cointegration, recognized as a common relationship
between economic individuals especially in international finance and macroeco-
nomics, could be with breaks. Thus, testing a break in cointegrating relationship
correctly could be a solution of financial crises.
attentions from the specific financial asset to another and points out these behaviors
take no consideration of deviations in fundamentals. More importantly, most
historical events coincided with Shiller’s findings; thus those opinions become
more convincing. A famous example is the remark given by Greenspan, the
ex-chairman of the Federal Reserve Board, which illustrates that it would be fairly
difficult to know when irrational exuberance has unduly escalated asset values and
thus the USA experienced the bubble collapse.
Furthermore, those claims have already been proved. Kirman and Teyssiere
(2005) rebuked that the deviations from fundamental assets are not the outcome
of irrational herding but the significant shifting compositions of expectations, which
result from opinion diffusion processes. This view is backed up by an analysis of
two types of agents in financial market, which are fundamentalists and chartists.
The former denotes agents who believe asset price Pt is related to underlying
fundamental Pt , i.e., a constant P. The latter depends on the Pt which is cumulated
by historical prices. Moreover, there also exists heterogeneity between these agents.
Thus for the former,
X
Mf
Ef ðPtþ1 jI t Þ ¼ Pt þ vj Ptjþ1 Ptj (31.30)
j¼1
where vj, j ¼ 1, . . ., Mf are positive constants and Mf is the memory of the funda-
mentalists. And in return for the latter
X
Mc
Ec ðPtþ1 jI t Þ ¼ hj Ptj (31.31)
j¼0
where hj, j ¼ 1, . . ., M are constants, M is the memory of the chartists. And based
c c
on the two opinions, the market view of prices are synthetic by the weighted
average of these forecast
Em Ptþ1 jI t Þ ¼ wt Ef ðPtþ1 jI t Þ þ ð1 wt ÞEc ðPtþ1 jI t Þ (31.32)
And it can be easily understood that wt functioned in Eq. 31.32 originates from
the following mechanism,
XN
1
wt ¼ N 1 ♯ i : qi; t :
i¼1
2
890 C.S.-H. Wang and Y.M. Xie
where xt stands for log stock or log dividend and the certain order of J is suggested
by Campbell and Perron (1991), while NID denotes independent and normal
distribution. Thus the unit root null hypothesis is H0 : d ¼ 1, and the right-tailed
alternative is H1 : d > 1.
31 Structural Change and Monitoring Tests 891
After this, a series of forward recursive regressions are applied to Eq. 31.34
repeatedly with sequential subsets of the sample data incremented by one observa-
tion at each pass. Set number of observations in the first regression as t0 ¼ [nr0] and
subsequent regressions employ the originating data set supplied by successive
observations giving a sample of size t ¼ [nr] for r0 r 1. And the corresponding
t-statistic by ADFr and ADF1 corresponds to the full sample. Under the null,
ðr
WdW
ADFr ) ð r 0
1=2 ,
W2
0
and ðr
WdW
supr2½r0 ;1 ADF ) supr2½r0 ;1 ð 0 1=2
r
2
W
0
where W is the standard Brownian motion. Comparing supr2½r0 ;1 ADF with the
ð r ð r 1=2 !
2
right-tailed critical values with supr2½r0 ;1 WdW W will lead to
0 0
a unit root test for against explosiveness. And to pinpoint the structural break, the
recursive test statistics ADFr are needed against the right-tailed critical values of
the asymptotic of the standard Dickey-Fuller t-statistic. In particular, if re denotes
the origination date and rf is the collapse date of explosive behavior in the data,
estimates of these dates are as follows:
n o
r^e ¼ INF s : ADFs > cvadf
bn , ^
r f ¼ inf s^
re inf s : ADF s < cv adf
bn (31.35)
sr0
where cvadfbn ðsÞ denotes the right-sided critical value of ADFs corresponding to
a significant level of bn. Notice the consistent estimation of the bubble period r^e, r^f
requires the significance level bn approaching zero asymptotically and correspond-
ingly cvadf
bn ðsÞ diverging to infinity in order to eliminate type I error as n ! 1. This
can be implemented in a convenient way employing cvadf bn ðsÞ ¼ logðlogðnsÞÞ=100.
Particularly, it replaces the explicit setting for bn which is more complicated. This
test referred by Phillips et al. (2009) is advantageous for the reason that it doesn’t
allow for the possibility of periodically collapsing bubbles, which are often
observed in practical economic and financial applications.
What’s more, there are more other attempts that could avoid potential setbacks
in performance of testing for bubbles. Among them, large sums of progressing
improvement are from unit root tests for explosive characters. For example,
bootstrapping is applied to improve the power of tests, of which Paparoditis and
Politis (2003) make use for the heavy-tail problem. For DGP in Eq. 31.34, it is
assumed that
892 C.S.-H. Wang and Y.M. Xie
For a test against jyj < 1, the residual block bootstrap method is carried as
pffiffiffi k, k ¼ 1,. . ., B,
follows: for each block bootstrap series satisfying H0, set the sample
and the integer b ¼ b(T) < T is employed such that 1=b þ b= T ! 0 as T ! 1.
Define k ¼ [T( 1)/b] and draw with replacement k integers i0,. . ., ik1 from the set
1,. . ., T b, and build the bootstrap samples ej(k) as follows:
The purpose of the unit root and subsampling tests here is to approximate the
distribution of the unit root statistics x without knowledge of the tail index alpha
which is difficult to estimate.
Another method employing subsampling is advocated by Jach and Kokoszka
(2004) consists in constructing T b processes which satisfies H0,
where b is the size of the subsampling blocks. Let e y b ðkÞ be the LS estimator of
the
regression of e j(k) on e j1 (k). Then the distribution of x would be estimated by
b ey b ðÞ 1 .
Moreover Wu and Xiao (2008) have proposed a procedure similar to
cointegration tests by Xiao and Phillips (2002) to help detect collapsible bubbles
against which Evans (1991) pointed out that unit root tests would lose the power.
31 Structural Change and Monitoring Tests 893
Their procedure is based on the magnitude of variation of the partial sum processes
Xk
Sk ¼ ^e of the residuals of regression (31.36). If there is no bubble, the
t¼1
magnitude of fluctuation of the process Sk is proportional to k1/2, while the presence
of a bubble makes the process Sk diverging to 1. Their statistic denoted by R enjoys
the advantage of no influence from serial correlation and the correlation between
the residuals and the fundamentals under the null hypothesis. As showed below, R
converges to the supremum of a functional of Brownian motions, i.e.,
k
e
R :¼ max1kT pffiffiffi
k1 Sþ
k T 1 þ
d
S T ! dsup 0t1
V ðtÞ
^ e, d T
o
ð 1 ð 1 1 ð
where e ðtÞ ¼ W ðtÞ tW ð1Þ ,
V W d ¼ W 1 ð tÞ dW 1 S
0
SS
0
,
d d
0 0 0
S(t) 0 ¼ (1, W2(t)), W1(t) and W2(t) are Brownian motions that are independent
^ t, d is a nonparametric long-run variance estimator.
of each other; o
In addition to detecting the structural breaks in a time series, tests for a breakdown
in the cointegration relationship between two nonstationary I(1) process are also of
interest. For example, correlation of global financial market becomes more obvious
as a result of rapid international capital flows, which feature I(1) processes and are
generally the linear combinations of nonstationary time series. As shown by most
empirical evidences, the existing cointegration relationship often comes to the end
prior to the finite crisis. Such a structural break is vital since it is an indicator for
finance crash. Traditional wisdom formulates the cointegration breakdown tests
always from the assumption that the post-breakdown period is relatively long, and
this is often strong and unrealistic especially for practitioners, see Hansen (1992)
and Quintos and Phillips (1993). Andrews and Kim (2003) on the contrary abandon
this assumption and introduce tests for cointegration breakdown with fixed post-
breakdown time length m, under the condition that the sample T + m goes to infinity.
Clearly, their implementation concentrating on end-of-sample could be conve-
niently extended to breakdown tests occurring at the beginning or in the middle
of the sample.
The data-generating process is as follows:
0
x b0 þ ut for t ¼ 1, . . . , T
yt ¼ 0
x bt þ ut for t ¼ T þ 1, . . . , T þ m
where yt, xt 2 R and xt, b0, bt 2 Rk and the regressors for all time periods are
I(1) processes with potential deterministic and stochastic trend or other stationary
random variables. Thus the null and the alternative hypotheses are
894 C.S.-H. Wang and Y.M. Xie
bt ¼ b0 for all t ¼ T þ 1, . . . , T þ m and
H0 :
8 t : t ¼ 1, . . . , T þ m are stationary and ergodic
u
< bt 6¼ b0 for some t ¼ T þ 1, . . . , T þ m and=or :
H 1 : the distribution of uTþ1 , . . . , uTþm differs from
:
the distribution of u1 , . . . , um
and for j ¼ 1,. . ., T + 1. For the P tests in Andrews and Kim (2003), o is some
nonsingular m matrix and Im denotes the m dimensional identity matrix. Naturally,
an estimator of b0 denoting as b^ is based on the least squares with observations
t ¼ r,. . ., s for 1 + m as following:
0 1 0
b^rs ¼ Xrs Xrs Xrs Y rs
Although other estimators like the fully modified estimator of Phillips and
Hansen (1990) and the ML estimator of Johansen can also be applied, the priority
for explanation is given to LS estimator.
Then the first test statistics, Pa is defined as
X
T þm 2
Pa ¼ PTþ1 b^1T ¼ yt xt b^1T :
0
t¼Tþ1
Clearly, b^ðjÞ by the estimators mentioned above is consistent for b0 under suitable
assumptions. Define
Pa, j ¼ Pj b^ðjÞ for j ¼ 1, . . . , T m þ 1:
X
Tmþ1
F^PðaÞ , T ðxÞ ¼ 1 m þ 1 l Pa, j :
t¼1
Then define the test statistic Pa to be the 1 a sample quantile, q^pa , 1 a and
The Pb test supersedes the Pa for it is less variable as the estimator b^1ðTþm=2eÞ
depends on the observation indexed by t ¼ T + 1,. . ., T + 2e.
Moreover, a naturally less variable statistic Pc is dependent on the complete
sample estimator b^1ðTþmÞ :
Pc ¼ PTþ1 b^1ðTþmÞ Pc, j ¼ P b^2ðjÞ for j ¼ 1, . . . , T m þ 1:
where
estimator of b using observations indexed by t ¼ 1, . . . , T with
b^2ðjÞ ¼
t 6¼ j, . . . , j þ 2e1
for j ¼ 1,. . ., T m + 1. The P tests are suitable for models where errors are
uncorrelated, and including weights in the statistics based on an estimator of the
error covariance matrix will be advantageous if the errors are correlated. Tests with
the weights are the same as Pa Pc except that o ¼ Im is replaced by
!1
X
T þ1 0
^ 1ðTþmÞ ¼
o 1þ1 U^ j, jþm1 U^ j, jþm1
j¼1
896 C.S.-H. Wang and Y.M. Xie
where
U^ j, jþm1 ¼ Y j, jþm1 Xj, jþm1 b^1ðTþmÞ ,
Foregoing tests almost deal with in-sample data, and it may be tempting to apply
them to real-time data. Under this distinct condition, however, traditional method
for detecting structural breaks mainly within a historical data set normally generates
comparatively large chance of mistaken instability, which leads to variant over-
rejection. Typically, such a test will signal a nonexistent break with probability one.
Thus due to the law of the iterated logarithm, no matter how perfectly these
methods perform as one-shot type detection, their sound performance could not
be translated to monitor out-of-sample stability. Conversely, monitoring test could
detect the break on time without these concerns. Focusing on monitoring structural
break, this section will explain how this relatively advantageous test works better
timely. Plus, it is also to provide evidence to support that the good performance still
holds even in long memory, as articulated before to some degree an ideal descrip-
tion of economy issues.
Let {Sn} be the partial sum process constructed from ei and assume the process ei
follows the FCLT (functional central limiting theorem) as l ! g1 0 m
1/2
S[ml] )
1
l(l), l 2 [0, 1), where Sn ¼ ∑ i ¼ 1ei and s0 ! n E(Sn) < 1. Chu et al. (1996)
n 2 2
pffiffiffiffi
limm!1 P jSn j mgððn=mÞ, for some n m ¼ PfjW ðtÞjðtÞ, t 1g
(31.37)
Robinson and Siegmund (1970) give their proof for an i.i.d. ei in Eq. 31.38 could
hold for a certain class of continuous functions g(t) that satisfy t1/2 g(t) is
ultimately nondecreasing as t ! 1 and other assumptions. Though this is useful,
yet such assumption is not applicable all the time for research. Segen and Sanderson
(1980) provide a partial resolution for they show that under the boundary function
t1/2 g(t) is nondecreasing, stochastic sequence Sn that satisfies the FCLT continues
to hold. Yet such condition turns out more restrictive than that of Robbins and
Siegmund. Chu et al. (1996) propose a new limiting similar relation without
imposing these restrictions. Typically, the boundary function g(t) considered to
be pertinent to economy is given as
n
1=2 o
P jW ðtÞj t a2 þ ln tÞ , for some t 1 ¼ 2½1 FðaÞ þ afðaÞ
(31.38)
n
1=2 o
P jW ðtÞj ðt þ 1Þ1=2 a2 þ lnðt þ 1Þ , some t > 0 ¼ exp a2 (31.39)
where F and f stand for the cdf and pdf, respectively, of a standard normal random
variable.
A monitoring is a stopping time, determined by a detecting statistic (detector) Gn
and a threshold g(m,n), according to tg(Gn)n, Gn > g(m, n). Firstly consider the
situation where the detector is in form of CUSUM. Let
Xn Xn
0 1
^
b¼ XX X Y be the OLS estimator at time n. Also define
i¼1 i i i¼1 i i
0 0
1
recursive residuals as wk ¼ 0 and wn ¼ ^e =nn1=2 , nn ¼ 1 + Xn(∑n1 i¼1 XiXi) Xn,
^e n ¼ Y n Xn b^n1 , n ¼ k + 1,. . .m,. . .. Thus the nth-cumulated sum of recursive
0
Xn Xkþ em t
1 1
residuals is Qm t ¼ ^
s o
i¼k i
¼ ^
s i¼k
oi, for ðn kÞ=em < ðn k þ 1Þ=em,
where s ^ is a consistent if s, m
e ¼ ðm kÞ is the integer of ðm e Þt. Thus under H0,
898 C.S.-H. Wang and Y.M. Xie
e 1=2 Qm
t!m t , t 2 ½0; 1Þ ) tðtÞ, t 2 ð0; 1Þ (31.40)
the model suitable for historical period is no longer reliable. Moreover, an extension
of this can be applied to FL monitoring procedure as
n
pffiffiffiffiffiffiffiffiffiffiffiffi n o
limm!1 P
Qm
m kg , for some n m
n
mk
(31.41)
¼ P jW ðtÞj gðtÞ, for some t 1
It needs notice that CUSUM algorithms, except those recursive residuals, also
hold for such mechanism, e.g., LR statistics for dependent sequences. Yet LR
detector sometimes could be superfluously complex in general and leads to
a complicated computation. According to the above statement, a CUSUM moni-
toring can be implemented as follows:
0
Suppose (i) Yt ¼ Xtb0 + et, t ¼ 1, . . ., m + 1, . . ., where Xt is a k 1 random vector
0
such that m ∑ t¼1Xt and m1 ∑m
1 m
t¼1XtXt converge in probability to b, a
non-stochastic k 1 vector and M, a k matrix of full rank, respectively; (ii) et is
a martingale difference sequence with respect to a sequence of s-algebra Ft such
that E(e2) < 1 and E(e2t jFt1) ¼ s20 for all t, where Ft is generated by. . ., (Yt2,
0 0
Xt1), (Y1, Xt); (iii) the sequence Xtet obeys the functional central limit theorem,
then ( )
m
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
n þ m k 1=2
e
limm!1 P
Q n
n þ m k a þ ln 2
, for some, n 1
mk
¼ expða2 =2Þ;
(31.42)
n 1=2 h n i1=2
e m
pffiffiffiffiffiffiffiffiffiffiffiffi
limm!1 P
Q n
m k a þ ln
2
, some n
mk mk
¼ 2½1 F½a þ afðaÞ
(31.43)
31 Structural Change and Monitoring Tests 899
Additionally, by the right-handed side of Eqs. 31.42 and 31.43, the asymptotic
size control of the CUSUM monitoring is fairly effortless. For CUSUM monitoring
based on Q e m , the 10 % and 5 % asymptotic size correspond to a2 ¼ 4.6 and
n
6, respectively. Equally handily, we obtain the 10 % and 5 % asymptotical size of
the CUSUM monitoring based on Qm by setting a2 ¼ 6.25 and 7.78 in Eq. 31.43,
respectively.
Furthermore, from Eq. 31.36 we also can construct the fluctuation monitoring of
sequential parameter estimates. The key condition is that Xtet obeys the multivariate
FCLT as
1=2
X
½ml
l1=2 V 0 Xt et ) l W ðlÞ, l 2 ½0; 1Þ
t¼1
0
where V0 ¼ limm!1m1E(SmSm) with Sm ¼ ∑m t¼1Xtet, and W(l) is a k-dimensional
Wiener process. Following this, the fluctuation detector can be defined as
Z^n ¼ nD1=2 ^ b^ , n
b (31.44)
m n m
where Dm ¼ M1 1
m V 0 MmP is Op(1) and uniformly positive definite such that
!
X
m
0
Xt Xt =m Mm ! 0:
t¼1
h n h n ii1=2
i
1=2 n m
ðbÞ : limm!1 P
^z ½ml
m a þ In
2
,
m nm nm
ðfor some m and some iÞ
¼ 1 ½1 2½1 FðaÞ þ afðaÞk ,
In this chapter, we discuss two classes of structural breaks – the retrospective tests
and the monitoring tests. From the investor’s points of view, it is crucial to use
suitable structural break tests to detect turning points of the financial and macro-
economics data accurately and further adjust the portfolios immediately. With the
increase of the data property caused by highly frequent market fluctuations, the
currently existing structural break tests could not fully detect locations of breaks
without the issue of size distortion. More precisely, more and more sophisticated
structural breaks tests with respect to any possible market performances would be
expected to be created in the near future.
References
Andrews, D. W. K. (1993). Tests for parameter instability and structural change with unknown
change point. Econometrica, 61, 821–856.
Andrews, D. W. K., & Kim, J. (2003). End-of-sample cointegration breakdown tests. Cowles
Foundation Discussion Paper No. 1404.
Bai, J., & Perron, P. (1996). Computation and analysis of multiple structural change models.
Manuscript in Preparation, Universite de Montreal.
Bai, J., & Perron, P. (1998). Estimation and testing linear models with multiple structural changes.
Econometrica, 66, 47–78.
Beran, J. (1991). M-estimators of location for data with slowly decaying serial correlations.
Biometrika, 76, 261–269.
Beran, J., & Terrin, N. (1996). Testing for a change of the long-memory parameter. Biometrika,
83, 627–638.
Campbell, J. Y., & Perron, P. (1991). Pitfalls and opportunities: What macroeconomists should
know about unit roots. National Bureau of Economic Research Macroeconomics Annual, 6,
201–210.
Cavaliere, G. (2004). Unit root tests under time-varying variances. Econometric Reviews, 23,
259–292.
Cavaliere, G., & Taylor, A. M. R. (2006). Testing for unit roots in time series models with
non-stationary volatility. Journal of Econometrics, 140, 919–947.
Chan, N. H., & Tran, L. T. (1989). On the first order autoregressive process with infinite variance.
Econometric Theory, 5, 354–362.
Chow, G. C. (1960). Tests of equality between sets of coefficients in two linear regressions.
Econometrica, 28, 591–605.
Chu, C. J., Stinchcombe, M., & White, H. (1996). Monitoring structural change. Econometrica, 64,
1045–1065.
Clements, M. P., & Hendry, D. F. (1998). Forecasting economic processes. International Journal
of Forecasting, 14, 111–131.
Clements, M. P., & Hendry, D. F. (1999). On winning forecasting competitions in economics.
Spanish Economic Review, 1, 123–160.
Davies, R. B. (1977). Hypothesis testing when a nuisance parameter is present only under the
alternative. Biometrika, 64, 247–254.
Diba, B. T., & Grossman, H. I. (1988). Explosive rational bubbles in stock prices? The American
Economic Review, 78, 520–530.
Dickey, D. A., & Fuller, W. A. (1979). Distribution of the estimators for autoregressive time
series with a unit root. Journal of the American Statistical Association, 74, 427–431.
31 Structural Change and Monitoring Tests 901
Dickey, D. A., & Fuller, W. A. (1981). Likelihood ratio statistics for autoregressive time series
with a unit root. Econometrica, 49, 1057–1072.
Evans, G. W. (1991). Pitfalls in testing for explosive bubbles in asset prices. American Economic
Review, 81, 922–930.
Granger, C. (1980). Long memory relationships and aggregation of dynamics models. Journal of
Econometrics, 14, 227–238.
Granger, C. W. J., & Joyeux, R. (1980). An introduction to long-memory time series and fractional
differencing. Journal of Time Series, 1, 15–29.
Hamori, S., & Tokihisa, A. (1997). Testing for a unit root in the presence of a variance shift.
Economics Letters, 57, 245–253.
Hansen, B. E. (1992). Tests for parameter instability in regressions with I(1) processes. Journal of
Business and Economic Statistics, 10, 321–336.
Hidalgo, J., & Robinson, P. M. (1996). Testing for structural change in a long-memory environ-
ment. Journal of Econometrics, 70, 159–174.
Horvath, L., & Kokoszka, P. S. (2003). A bootstrap approximation to a unit root test statistic for
heavy tailed observations. Statistics and Probability Letters, 62, 163–173.
Hsiao, C., & Robinson, P. M. (1978). Efficient estimation for a dynamic error-shock model.
International Economic Review, 19, 467–480.
Jach, A., & Kokoszka, P. S. (2004). Subsampling unit root tests for heavy-tailed observations.
Methodology and Computing in Applied Probability, 6, 73–97.
Kazakevicius, V., Leipus, R., & Viano, M. (2004). Stability of random coefficient ARCH models
and aggregation schemes. Journal of Econometrics, 120, 139–158.
Kim, T. H., Leybourne, S., & Newbold, P. (2002). Unit root tests with a break in innovation
variance. Journal of Econometrics, 109, 365–387.
Kindleberger, C. P. (2005). Manias, panics, and crashes: A history of financial crises. New Jersey:
Wiley.
Kirman, A., & Teyssiere, G. (2005). Testing for bubbles and change-points. Journal of Economic
Dynamic Control, 29, 765–799.
Kokoszka, P. S., & Parfionovas, A. (2004). Bootstrap unit root tests for heavy tailed time series. In
S. T. Rachev (Ed.), Handbook of computational and numerical methods in finance
(pp. 175–197). Boston: Birkhäuser.
Kuan, C., & Hsu, C. (1998). Change-point estimation of fractionally integrated processes. Journal
of Time Series Analysis, 19, 693–708.
Leybourne, S., Taylor, R., & Kim, T. (2006). Cusum of squares-based tests for a change in
persistence. Journal of Time Series Analysis, 28, 408–433.
Lutkepohl, H. (1988). Prediction tests for structural stability. Journal of Econometrics, 39,
267–296.
Lutkepohl, H. (1989). Prediction tests for structural stability of multiple time series. Journal of
Business Economic Statistics, 7, 129–135.
Mandelbrot, B. B., & Vanness, J. W. (1968). Fractional Brownian motions, fractional noises and
applications. SIAM Review, 10, 422–437.
Paparoditis, E., & Politis, D. (2003). Residual-based block bootstrap for unit root testing.
Econometrica, 71, 813–855.
Perron, P., & Ng, S. (1996). Useful modifications to some unit root tests with dependent errors and
their local asymptotic properties. Review of Economic Studies, 63, 435–463.
Phillips, P. C. B., & Hansen, B. E. (1990). Statistical inference in instrumental variables regression
with I(1) processes. Review of Economic Studies, 57, 99–125.
Phillips, P. C. B., & Perron, P. (1988). Testing for a unit root in time series regression. Biometrika,
75, 335–346.
Phillips, P. C. B., Wu, Y., & Yu, J. (2009). Explosive behavior in the 1990s Nasdaq: When did
exuberance escalate asset values?. Cowles Foundation Discussion Paper No.1699
Phillips P. C. B., Shi, S. & Yu, J. (2011). Testing for multiple bubbles. Singapore Management
University working paper.
902 C.S.-H. Wang and Y.M. Xie
Quandt, R. E. (1960). Tests of the hypothesis that a linear regression system obeys two separate
regimes. Journal of the American Statistical Association, 55, 324–330.
Quintos, C. E., & Phillips, P. C. B. (1993). Parameter constancy in cointegrating regressions.
Empirical Economics, 18, 675–706.
Robinson, H., & Siegmund, D. (1970). Boundary crossing probabilities for the Wiener process and
sample sums. The Annals of Mathematical Statistics, 41, 1410–1429.
Robinson, P. M., & Zaffaroni, P. (1998). Nonlinear time series with long memory: A model for
stochastic volatility. Journal of Statistical Planning and Inference, 68, 359–371.
Segen, J., & Sanderson, A. (1980). Detection of changes in a time series. IEEE Transaction in
Information Theory, 26, 249–257.
Shiller, R. J. (2000). Irrational exuberance. Princeton: Princeton University Press.
Sibbertsen, P., & Kruse, R. (2009). Testing for a break in persistence under long-range dependen-
cies. Journal of Time Series Analysis, 30, 268–285.
Stock, J. H. (1994). Unit roots, structural breaks and trends, Chapter 46. In Handbook of
econometrics (pp. 2740–2841). Amsterdam: Elsevier
Taylor, S. (1986). Modeling financial time series. New Jersey: Wiley.
Welch, I., & Goyal, A. (2008). A comprehensive look at the empirical performance of equity
premium prediction. Review of Financial Studies, 21, 1455–1508.
Wu, G. & Xiao, Z. (2008). Are there speculative bubbles in stock markets? Evidence from an
alternative approach. Statistics and its Interface, 1, 307–320.
Xiao, Z., & Phillips, P. C. B. (2002). A CUSUM test for cointegration using regression residuals.
Journal of Econometrics, 108, 43–61.
Yajima, Y. (1988). On estimation of a regression model with long-memory stationary errors.
Annals of Statistics, 16, 791–807.
Zaffaroni, P. (2000). Contemporaneous aggregation of GARCH processes. Journal of Time Series
Analysis, 28, 521–544.
Zaffaroni, P. (2006). Aggregation and memory of models of changing volatility. Journal of
Econometrics, 136, 237–249.
Consequences for Option Pricing
of a Long Memory in Volatility 32
Stephen J. Taylor
Contents
32.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 904
32.2 Long Memory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 906
32.2.1 Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 906
32.2.2 Fractionally Integrated White Noise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 907
32.2.3 Evidence for Long Memory in Volatility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 908
32.2.4 Explanations of Long Memory in Volatility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 909
32.3 Long Memory Volatility Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 910
32.3.1 ARCH Specifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 910
32.3.2 Estimates for the S & P 100 Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 911
32.3.3 Stochastic Volatility Specifications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 914
32.4 Option Pricing Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 914
32.4.1 A Review of SV and ARCH Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 914
32.4.2 The ARCH Pricing Framework . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 915
32.4.3 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 916
32.5 Illustrative Long Memory Option Prices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 917
32.5.1 Inputs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 917
32.5.2 Parameter Selections . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 918
32.5.3 Comparisons of Implied Volatility Term Structures . . . . . . . . . . . . . . . . . . . . . . . . . . 919
32.5.4 Comparisons of Smile Effects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 923
32.5.5 Sensitivity Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 925
32.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 926
Appendix 1: Series Expansions and Truncation Approximations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 929
Appendix 2: Simulation Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 930
Appendix 3: Impact of a Volatility Risk Premium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 931
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 932
S.J. Taylor
Lancaster University Management School, Lancaster, UK
e-mail: s.taylor@lancaster.ac.uk
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 903
DOI 10.1007/978-1-4614-7750-1_32,
# Springer Science+Business Media New York 2015
904 S.J. Taylor
Abstract
Conditionally heteroscedastic time series models are used to describe the
volatility of stock index returns. Volatility has a long memory property in the
most general models and then the autocorrelations of volatility decay at
a hyperbolic rate; contrasts are made with popular, short memory specifications
whose autocorrelations decay more rapidly at a geometric rate.
Options are valued for ARCH volatility models by calculating the discounted
expectations of option payoffs for an appropriate risk-neutral measure. Monte
Carlo methods provide the expectations. The speed and accuracy of the calcu-
lations is enhanced by two variance reduction methods, which use antithetic and
control variables. The economic consequences of a long memory assumption
about volatility are documented, by comparing implied volatilities for option
prices obtained from short and long memory volatility processes.
Results are given for options on the S & P 100-share index, with lives up to
2 years. The long memory assumption is found to have a significant impact upon
the term structure of implied volatilities and a relatively minor impact upon
smile shapes. These conclusions are important because evidence for long mem-
ory in volatility has been found in the prices of many assets.
Keywords
ARCH models • Implied volatility • Index options • Likelihood maximization •
Long memory • Monte Carlo • Option prices • Risk-neutral pricing • Smile
shapes • Term structure • Variance reduction methods
32.1 Introduction
Long memory effects in a stochastic process are effects that decay too slowly to be
explained by stationary processes defined by a finite number of autoregressive and
moving-average terms. Long memory is often represented by fractional integration
of shocks to the process, which produces autocorrelations which decrease at
a hyperbolic rate compared with the faster geometric rate of stationary ARMA
processes.
Long memory in volatility occurs when the effects of volatility shocks decay
slowly. This phenomenon can be identified from the autocorrelations of measures
of realized volatility. Two influential examples are the study of absolute
daily returns from stock indices by Ding et al. (1993) and the investigation of
daily sums of squared 5-min returns from exchange rates by Andersen
et al. (2001b).
Stochastic volatility causes option prices to display both smile and term structure
effects. An implied volatility obtained from the Black-Scholes formula then
depends on both the exercise price and the time until the option expires. Exact
calculation of smile and term effects is only possible for special volatility processes,
32 Consequences for Option Pricing of a Long Memory in Volatility 905
with the results of Heston (1993) being a notable example. Monte Carlo methods
are usually necessary when the volatility process has a long memory and these were
first applied by Bollerslev and Mikkelsen (1996, 1999).
This chapter documents the economic consequences of a long memory
assumption about volatility. This is achieved by comparing implied volatilities
for option prices obtained from short and long memory specifications.
It is necessary to use a long history of asset prices when applying a long memory
model and this chapter uses levels of the S & P 100 index from 1984 to 1998. For
this data it is found that a long memory assumption has a significant economic
impact upon the term structure of implied volatilities and a relatively minor impact
upon smile effects. Some related empirical results are provided by Ohanissian
et al. (2004).
Option traders have to make assumptions about the volatility process.
The effects of some assumptions are revealed by the prices of options with long
lives. Bollerslev and Mikkelsen (1999) find that the market prices of exchange
traded options on the S & P 500 index, with lives between 9 months and 3 years,
are described more accurately by a long memory pricing model than by the
short memory alternatives. Thus these option prices reflect the long memory
phenomenon in volatility, although it is found that significant biases remain
unexplained.
Three explanatory sections precede the illustrative option pricing results in
Sect. 32.5. Section 32.2 defines and characterizes long memory and then reviews
the empirical evidence for these characteristics in volatility. The empirical evidence
for the world’s major markets appears compelling and explanations for the source
of long memory effects in volatility are summarized.
Section 32.3 describes parsimonious volatility models which incorporate long
memory, either within an ARCH or a stochastic volatility framework. The former
framework is easier to use and we focus on applying the fractionally integrated
extension of the exponential GARCH model, which is known by the acronym
FIEGARCH. An important feature of applications is the unavoidable truncation
of an autoregressive component of infinite order. Empirical results are provided for
10 years of S & P 100 returns.
Section 32.4 provides the option pricing methodology. Contingent claim prices
are obtained by simulating terminal payoffs using an appropriate risk-neutral
measure. Numerical methods enhance the accuracy of the simulations and these
are described in Appendix 2.
Section 32.5 compares implied volatilities for European option prices obtained
from short and long memory volatility specifications, for hypothetical S & P
100 options whose lives range from 1 month to 2 years. Options are valued on
10 dates, one per annum from 1989 to 1998. The major impact of the long memory
assumption is seen to be the very slow convergence of implied volatilities to a limit
as the option life increases. This convergence is so slow that the limit cannot be
estimated precisely. Section 32.6 contains conclusions.
906 S.J. Taylor
32.2.1 Definitions
s2 X 1
f ðoÞ ¼ r cos ðtoÞ, o > 0, (32.2)
2p t¼1 t
varðxtþ1 þ . . . þ xtþn Þ X
n1
nt
Vn ¼ ¼ 1 þ 2 r (32.3)
ns2 t¼1
n t
X
n
rt ! C1 , f ðoÞ ! C2 , V n ! C3 , as n ! 1, o ! 0, (32.4)
t¼1
for constants C1, C2, C3. Examples are provided by stationary ARMA processes.
These processes have geometrically bounded autocorrelations, so that jrtj Cft
for some C > 0 and 1 > f > 0, and hence Eq. 32.4 is applicable.
In contrast to the above results, all the limits given by Eq. 32.4 do not exist for
a typical covariance stationary long memory process. Instead, it is typical that the
autocorrelations have a hyperbolic decay, the spectral density is unbounded for low
frequencies, and the variance ratio increases without limit. Appropriate limits are
then provided for some positive d < 12 by
rt f ðoÞ Vn
! D1 , ! D2 , 2d ! D3 , as n ! 1, o ! 0, (32.5)
t2d1 o2d n
for positive constants D1, D2, D3. The limits given by Eq. 32.5 characterize
the stationary long memory processes that are commonly used to represent
32 Consequences for Option Pricing of a Long Memory in Volatility 907
d ð d 1Þ 2 d ð d 1Þ ð d 2Þ 3
ð1 LÞd ¼ 1 dL þ L L þ ... (32.6)
2! 3!
where L is the usual lag operator, so that Lyt ¼ yt1. Then a fractionally integrated
white noise (FIWN) process {yt} is defined by
ð1 LÞd yt ¼ et (32.7)
2
with the et assumed to have zero mean and variance se. Throughout this chapter it is
assumed that the differencing parameter d is constrained by 0 d < 1.
The mathematical properties of FIWN are summarized in Baillie (1996). The
process is covariance stationary if d < 12 and then the following results apply. First,
the autocorrelations are given by
d d ð d þ 1Þ d ð d þ 1Þ ð d þ 2Þ
r1 ¼ , r2 ¼ , r3 ¼ , :::::: (32.8)
1d ð1 d Þð2 d Þ ð1 d Þð2 d Þð3 d Þ
Gð1 dÞGðt þ dÞ
rt ¼ , (32.9)
GðdÞGðt þ 1 dÞ
with
rt G ð1 d Þ
! as t ! 1: (32.10)
t 2d1 Gðd Þ
so that
s2e 2d
f ðoÞ ffi o for o near 0: (32.12)
2p
Also,
Vn Gð1 dÞ
! as n ! 1: (32.13)
n2d ð1 þ 2d ÞGð1 þ dÞ
When d 12, the FIWN process has infinite variance and thus the
autocorrelations are not defined, although the process has some stationarity
properties for 12 d < 1.
X
N
^ 2t ¼
s r 2t, j : (32.14)
j¼1
and other microstructure effects introduce bias into the estimate. Using 5-min
returns provides conclusive evidence for long memory effects in the estimates s ^ 2t
in four studies: Andersen et al. (2001b) for 10 years of DM/$ and Yen/$ rates;
Andersen et al. (2001b) for 5 years of stock prices for the 30 components of the
Dow-Jones index; Ebens (1999) for 15 years of the same index; and Areal and
Taylor (2002) for 8 years of FTSE-100 stock index futures prices. These papers
provide striking evidence that time series of estimates s ^ 2t display all three
properties of a long memory process: hyperbolic decay in the autocorrelations,
spectral densities at low frequencies that are proportional to o2d, and variance
ratios whose logarithms are very close to linear functions of the aggregation
period n. It is also found that estimates of d are between 0.3 and 0.5, with most
estimates close to 0.4.
A general set of long memory stochastic processes can be defined by first applying
the filter (1 L)d and then assuming that the filtered process is a stationary ARMA
(p, q) process. This defines the ARFIMA (p, d, q) models of Granger (1980), Granger
and Joyeux (1980), and Hosking (1981). This approach can be used to obtain long
memory models for volatility, by extending various specifications of short memory
volatility processes. We consider both ARCH and stochastic volatility specifications.
The conditional distributions of returns rt are defined for ARCH models using
information sets It1 which are here assumed to be previous returns {rt i, i 1},
conditional mean functions mt(It1), conditional variance functions ht(It1), and
a probability distribution D for standardized returns zt. Then the terms
rt m
zt ¼ pffiffiffiffi t (32.15)
ht
are independently and identically distributed with distribution D and have zero
mean and unit variance.
Baillie (1996) and Bollerslev and Mikkelsen (1996) both show how to define
a long memory process for ht by extending either the GARCH models of Bollerslev
(1986) or the exponential ARCH models of Nelson (1991). The GARCH extension
cannot be recommended because the returns process then has infinite variance for
all positive values of d, which is incompatible with the stylized facts for asset
returns. For the exponential extension, however, ln(ht) is covariance stationary for
d < 12 ; it may then be conjectured that the returns process has finite variance for
particular specifications of ht.
Like Bollerslev and Mikkelsen (1996, 1999), this chapter applies the
FIEGARCH(1, d, 1) specification:
Calculations using Eq. 32.16 require truncation of a series expansion in the lag
operator L. The relevant formulae are listed in Appendix 1. We apply the following
ARMA(N, 1) approximation:
X
N
lnðht Þ ¼ a þ bj ln htj a þ gðzt1 Þ þ cgðzt2 Þ, (32.18)
j¼1
Representative parameters are used in Sect. 32.5 to illustrate the option pricing
consequences of long memory in volatility. These parameters are estimated from
daily returns rt for the S & P 100 index, excluding dividends, calculated from index
levels pt as rt ¼ ln (pt/pt1).
The conditional variances are evaluated for t 1 by setting N ¼ 1,000 in
Eq. 32.18, with ln (htj) replaced by a and g(ztj) replaced by zero whenever
t j 0. The log-likelihood function is calculated for the 2,528 trading days
during the 10-year estimation period from 3 January 1989 to 31 December 1998,
which corresponds to the times 1, 221 t 3, 748 for our dataset; thus the first
1,220 returns are reserved for the calculation of conditional variances before 1989
which are needed to evaluate the subsequent conditional variances.
Results are first discussed when returns have a constant conditional mean which
is estimated by the sample mean. The conditional variances are obtained recur-
sively from Eqs. 32.15, 32.17, 32.18, and 32.32. The conditional distributions are
assumed to be normal when defining the likelihood function. This assumption is
known to be false but it is made to obtain consistent parameter estimates (Bollerslev
and Wooldridge 1992). Preliminary maximizations of the p likelihood
ffiffiffiffiffiffiffiffi showed that
a suitable value for C ¼ E[|zt|] is 0.737, compared with 2=p ffi 0:798 for the
standard normal distribution. They also showed that an appropriate value of the
location parameter a of ln (ht) is 9.56; the log-likelihood is not sensitive to minor
X N
deviations from this level because a is multiplied by a term 1 bj in Eq. 32.18
j¼1
which is small for large N. Consequently, the results summarized in Table 32.1 are
given by maximizing the log-likelihood function over some or all of the parameters
y, g, f, c, d.
The estimates of y and g provide the usual result for a series of US stock index
returns that changes in volatility are far more sensitive to the values of negative
returns than those of positive returns, as first reported by Nelson (1991). When zt is
negative, g(zt) ¼ (g y)( zt) gC, otherwise g(zt) ¼ (g + y)zt gC. The ratio
(g y)/(g + y) is at least 4 and hence is substantial for the estimates presented in
Table 32.1.
912 S.J. Taylor
Table 32.1 Parameter estimates for short and long memory ARCH models
Model Constraints y g f c d RLL
AR(1) c¼d¼0 0.06 0.10 0.982 0.0
ARMA(1,1) d¼0 0.09 0.15 0.988 0.44 3.0
FI(d) f¼c¼0 0.12 0.19 0.66 13.9
ARFI(1, d) c¼0 0.11 0.17 0.30 0.59 17.2
ARFIMA(1, d, 1) None 0.10 0.15 0.16 0.98 0.57 21.7
ARFIMA(1, d, 1) f+c+d1 0.12 0.18 0.27 0.68 0.59 19.7
ARFIMA(1, d, 1) d ¼ 0.4, f + c 0.6 0.11 0.18 0.64 0.04 0.4 11.4
Parameters are estimated by maximizing the log-likelihood of daily returns from the S & P
100 index, from 3 January 1989 to 31 December 1998. Returns are modelled as
pffiffiffiffi
r t ¼ m þ ht z t ,
lnðht Þ ¼ a þ ð1 fLÞ1 ð1 LÞd ð1 þ cLÞgðzt1 Þ,
gðzt Þ ¼ yzt þ gðjzt j CÞ
The zt are assumed to be i.i.d., standard normal variables when defining the likelihood function.
The values C ¼ 0.737 and a ¼ 9.56 are used for all likelihood calculations. The relative
log-likelihood (RLL) for a model equals the maximum log-likelihood (MLL) for that model minus
the MLL for the AR(1) model. The MLL for the AR(1) model is 8561.6
The first two rows of Table 32.1 report estimates for short memory
specifications of the conditional variance. The AR(1) specification has
a persistence of 0.982 which is typical for this volatility model. The ARMA
(1,1) specification has an additional parameter and increases the log-likelihood
by 3.0. The third row shows that the fractional differencing filter alone
(d > 0, f ¼ c ¼ 0) provides a better description of the volatility process than
the ARMA(1,1) specification; with d ¼ 0.66 the log-likelihood increases by
10.9. A further increase of 7.8 is then possible by optimizing over all three
volatility parameters, d, f, and c, to give the parameter estimates1 in the fifth
row of Table 32.1.
The estimates for the most general specification identify two issues of concern.
First, d equals 0.57 for our daily data which is more than the typical estimate of 0.4
produced by the studies of higher frequency data mentioned in Sect. 32.2.3. The
same issue arises in Bollerslev and Mikkelsen (1996) with d estimated as 0.63
(standard error 0.06) from 9,559 daily returns of the S & P 500 index, from 1953 to
1990. Second, the sum d + f + c equals 1.39. As this sum equals c1 in Eqs. 32.35
and 32.36, more weight is then given to the volatility shock at time t 2 than to the
shock at time t 1 when calculating ln(ht). This is counterintuitive. To avoid this
outcome, the constraint d + f + c 1 is applied and the results given in the
penultimate row of Table 32.1 are obtained. The log-likelihood is then reduced by
2.0. Finally, if d is constrained to be 0.4, then the log-likelihood is reduced by an
additional 8.3.
1
A conservative robust standard error for our estimate of d is 0.12, using information provided by
Bollerslev and Mikkelsen (1996).
32 Consequences for Option Pricing of a Long Memory in Volatility 913
0.9
0.8
0.7
Coefficient
0.2
0.1
0
0 50 100 150 200 250 300 350 400 450 500
Lag
The estimates obtained here for f and c, namely, 0.27 and 0.68 for the most
general specification, are rather different to the 0.78 and 0.68 given by
Bollerslev and Mikkelsen (1999, Table 1), although the estimates of d are similar,
namely, 0.59 and 0.65. However, the moving-average representations obtained
from these sets of parameter estimates are qualitatively similar. This is shown on
Fig. 32.1 which compares the moving-average coefficients cj defined by (32.36).
The coefficients are positive and monotonic decreasing for the four sets of
parameter values used to produce Fig. 32.1. They show the expected hyperbolic
decay when d > 0 and a geometric decay when d ¼ 0. The values of bj in
Eqs. 32.32 and 32.38 that are used to calculate the conditional variances
decay much faster. For each curve on Fig. 32.1, c10 > 0.33 and c100 > 0.07 while
0 < b10 < 0.02 and 0 < b100 < 0.0003.
The results reported in Table 32.1 are for a constant conditional mean, mt ¼ pm.
ffiffiffiffi
Alternative specifications such as mt ¼ m + brt1, mt ¼ m 12ht, and mt ¼ m þ l ht
give similar values of the log-likelihood when the volatility parameters are set to
the values in the final row of Table 32.1. First, including the lagged return rt1 is
not necessary because the first-lag autocorrelation of the S & P 100 returns
equals 0.022 and is statistically insignificant. Second, including the
adjustment 12ht makes the conditional expectation of (pt pt1)/pt1 constant
when the conditional distribution is normal. The adjustment reduces the
log-likelihood by an unimportant 0.3. Third, incorporating the ARCH-M
parameter l gives an optimal value of 0.10 and an increase in the
log-likelihood of 1.5. This increase is not significant using a non-robust
likelihood-ratio test at the 5 % level.
914 S.J. Taylor
Two shocks per unit time characterize stochastic volatility (SV) models, in contrast
to the single shock zt that appears in ARCH models. A general framework for long
memory stochastic volatility models is given for returns rt by
r t ¼ m þ s t ut (32.19)
This framework has been investigated by Breidt et al. (1998), Harvey (1998),
and Bollerslev and Wright (2000), all of whom provide results for the simplifying
assumption that the two i.i.d. processes {ut} and {vt} are independent. This
assumption can be relaxed and has been for short memory applications (Taylor
1994; Shephard 1996).
Parameter estimation is difficult for SV models, compared with ARCH models,
because SV models have twice as many random innovations as observable vari-
ables. Breidt et al. (1998) describe a spectral-likelihood estimator and provide
results for a CRSP index from 1962 to 1989. For the ARFIMA(1, d, 0) specification,
they estimate d ¼ 0.44 and f ¼ 0.93. Bollerslev and Wright (2000) provide detailed
simulation evidence about semiparametric estimates of d, related to the frequency
of the observations.
It is apparent that the ARCH specification (Eqs. 32.15–32.17) has a similar
structure to the SV specification (Eqs. 32.19–32.20). Short memory special cases of
these specifications, given by d ¼ q ¼ 0, have similar moments (Taylor 1994). This
is a consequence of the special cases having the same bivariate diffusion limit when
appropriate parameter values are defined for increasingly frequent observations
(Nelson 1990; Duan 1997). It seems reasonable to conjecture that the multivariate
distributions for returns defined by (32.15–17) and (32.19–20) are similar, with the
special case of independent shocks {ut} and {vt} corresponding to the symmetric
ARCH model that has y ¼ 0 in Eq. 32.17.
The pricing of options when volatility is stochastic and has a short memory has
been studied using a variety of methods. The most popular methods commence
with separate diffusion specifications for the asset price and its volatility. These
are called stochastic volatility (SV) methods. Option prices then depend on
several parameters including a volatility risk premium and the correlation
between the differentials of the Wiener processes in the separate diffusions.
32 Consequences for Option Pricing of a Long Memory in Volatility 915
When pricing options it will be assumed that returns are calculated from prices
(or index levels) as rt ¼ ln (pt/pt1) and hence exclude dividends. A constant risk-
free interest rate and a constant dividend yield will also be assumed and, to simplify
the notation and calculations, it will be assumed that interest and dividends are paid
once per trading period. Conditional expectations are defined with respect to current
and prior price information represented by It ¼ {pti, i 0}.
To obtain fair option prices in an ARCH framework, it is necessary to make
additional assumptions in order to obtain a risk-neutral measure Q. Duan (1995) and
Bollerslev and Mikkelsen (1999) provide sufficient conditions to apply a risk-
neutral valuation methodology. For example, it is sufficient that a representative
agent has constant relative risk aversion and that returns and aggregate growth rates
in consumption have conditional normal distributions. Kallsen and Taqqu (1998)
derive the same solution as Duan (1995) without making assumptions about utility
functions and consumption. Instead, they assume that intraday prices are determined by
geometric Brownian motion with volatility determined once a day from a discrete-time
ARCH model.
916 S.J. Taylor
32.4.3 Implementation
with r the risk-free interest rate for one trading period. We now specify an
appropriate way to simulate pt0 +n under a risk-neutral measure Q. Monte Carlo
methods are then used to estimate the conditional expectation in Eq. 32.21.
Following Duan (1995), it is assumed that the distribution of observed returns is
defined by some probability measure P, for which
rt m
zt ¼ pffiffiffiffi t P i:i:d:N ð0; 1Þ (32.23)
ht
r t ðr d 12ht Þ Q
zt ¼ pffiffiffiffi i:i:d:N ð0; 1Þ: (32.25)
ht
1 pffiffiffiffi
mt ¼ r d h t þ l ht (32.26)
2
Option prices are evaluated when the conditional variances are given by an
ARMA(N, 1) approximation to the FIEGARCH(1, d, 1) specification. From
Eqs. 32.18 and 32.27,
!
X
N
1 j
bj L ðlnðht Þ aÞ ¼ ð1 þ cLÞgðzt1 Þ ¼ ð1 þ cLÞg zt1 l , (32.28)
j¼1
and
32.5.1 Inputs
Many parameters and additional inputs are required to implement the FIEGARCH
option pricing methodology. To apply that methodology to value European options,
we specify 18 numbers, a price history, and a random number generator, as follows:
• Contractual parameters – time until exercise T measured in years, the exercise
price X, and whether a call or a put option.
• The current asset price S ¼ pt0 and a set of previous prices {pt, 1 t < t 0 }.
• Trading periods per annum M, such that consecutive observed prices are sepa-
rated by 1/M years and likewise for simulated prices {pt, t 0 < t t 0 + n} with
n ¼ MT.
• Risk-free annual interest rate R, from which the trading period rate r ¼ R/M is
obtained.
• Annual dividend yield D giving a constant trading period payout rate of d ¼ D/M;
both R and D are continuously compounded and applicable for the life of the
option contract.
• The risk premium l for investment in the asset during the life of the option,
pffiffiffiffi such
that one-period conditional expected returns are mt ¼ r d 12ht þ l ht, for the
real-world measure P.
• Parameters m and l0 that define conditional expected pffiffiffiffi returns during the time
period of the observed prices by mt ¼ m 12ht þ l0 ht , again for measure P.
• Eight parameters that define the one-period conditional variances ht. The inte-
gration level d, the autoregressive parameter f, and the truncation level
N determine the terms bj in the AR(N) filter in Eq. 32.28. The mean a and the
moving-average parameter c complete the ARMA(N, 1) specification for ln(ht)
In Eq. 32.28. The values of the shocks to the ARMA(N, 1) process depend on g
918 S.J. Taylor
and y, which, respectively, appear in the symmetric function g(jzt j C) and the
asymmetric function yzt whose total determines the shock
pffiffiffiffiffiffiffiterm
ffi g(zt); the con-
stant C is a parameter C0 for observed prices but is 2=p when returns are
simulated.
• K, the number of independent simulations of the terminal asset price ST ¼ pt0 +n.
• A set of Kn pseudorandom numbers distributed uniformly between 0 and 1, from
which pseudorandom standard normal variates can be obtained. These numbers
typically depend on a seed value and a deterministic algorithm.
Option values are tabulated for hypothetical European options on the S & P
100 index. Options are valued for 10 dates defined by the last trading days of the
10 years from 1989 to 1998 inclusive. For valuation dates from 1992 onwards, the
size of the price history is set at t0 ¼ 2,000; for previous years the price
history commences on 6 March 1984 and t0 < 2,000. It is assumed that there are
M ¼ 252 trading days in 1 year and hence exactly 21 trading days in one simulated
month. Option values are tabulated when T is 1, 2, 3, 6, 12, 18, and 24 months.
Table 32.2 lists the parameter values used to obtain the main results. The
annualized risk-free rate and dividend yield are set at 5 % and 2 %, respectively.
The risk parameter l is set at 0.028 to give2 an annual equity risk premium of 6 %.
The mean return parameter m is set to the historic mean of the complete set of S & P
100 returns from March 1984 to December 1998 and l0 is set to zero.
There are two sets of values for the conditional variance process because the
primary objective here is to compare option values when volatility is assumed to
have either a short or a long memory. The long memory parameter set takes the
integration level to be d ¼ 0.4, because this is an appropriate level based upon the
high frequency reviewed in Sect. 32.2.3. The remaining variance parameters are
then based on Table 32.1; as the moving-average parameter is small, it is set to zero
and the autoregressive parameter is adjusted to retain the unit total, d + f +
c ¼ 1. The AR filter3 is truncated at lag 1,000, although the results obtained will
nevertheless be referred to as long memory results. The short memory parameters
are similar to those for the AR(1) estimates provided in Table 32.1. The parameters
g and y are both 6 % less in Table 32.2 than in Table 32.1 to ensure that selected
moments are matched for the short and long memory specifications; the uncondi-
tional mean and variance of ln(ht) are then matched for the historic measure P,
although the unconditional means differ by approximately 0.10 for the risk-neutral
measure Q as noted in Appendix 3.
pffiffiffiffiffi
2
The conditional
pffiffiffiffi expectations of rt for measures P and Q differ by l ht and a typical average
value of ht is 0.00858. Assuming 253 trading days in 1 year gives the stated value of l.
3
The filter coefficients sum to b1 + . . . + b1,000 ¼ 0.983. After b1 ¼ 1 and b2 ¼ 0.12, all of the
coefficients are near zero, with b100 ¼ 0.00017 and b1,000 ¼ 7 106.
32 Consequences for Option Pricing of a Long Memory in Volatility 919
Option prices are estimated from K ¼ 10, 000 independent simulations of prices
{pt, t 0 < t t 0 + n} with n ¼ 504. Applying the antithetic and control variate
methods described in Appendix 2 then produces results for a long memory process
in about 12 min, when the processor speed is 2 GHz. Most of the time is spent
evaluating the high-order AR filter; the computation time is about 1 min for the
short memory process.
The values of all options are reported using annualized implied volatilities rather
than prices. Each implied volatility (IV) is calculated from the Black-Scholes
formula, adjusted for continuous dividends. The complete set of IV outputs for
one set of inputs forms a matrix with rows labelled by the exercise prices X and
columns labelled by the times to expiry T; examples are given in Tables 32.5 and
32.6 and are discussed later.
Initially we only consider at-the-money options, for which the exercise price equals
the forward price F ¼ Se(RD)T, with IV values obtained by interpolation across two
adjacent values of X. As T varies, the IV values represent the term structure of implied
volatility. Tables 32.3 and 32.4, respectively, summarize these term structures for the
short and long memory specifications. The same information is plotted on Figs. 32.2
920 S.J. Taylor
Table 32.3 At-the-money implied volatilities for a short memory volatility process
Implied volatilities for options that expire after 1, 2, 3, 6, 12, 18, and 24 months
0 1 2 3 6 12 18 24
Year
1989 0.1211 0.1238 0.1267 0.1291 0.1340 0.1384 0.1404 0.1414
1990 0.1378 0.1380 0.1388 0.1395 0.1409 0.1421 0.1427 0.1431
1991 0.1213 0.1240 0.1271 0.1296 0.1339 0.1383 0.1403 0.1413
1992 0.1085 0.1129 0.1175 0.1210 0.1288 0.1356 0.1385 0.1401
1993 0.0945 0.1008 0.1069 0.1119 0.1226 0.1321 0.1363 0.1382
1994 0.1165 0.1201 0.1236 0.1263 0.1321 0.1375 0.1398 0.1411
1995 0.1118 0.1158 0.1200 0.1234 0.1304 0.1367 0.1391 0.1405
1996 0.1462 0.1449 0.1446 0.1445 0.1442 0.1438 0.1440 0.1441
1997 0.1883 0.1791 0.1730 0.1683 0.1599 0.1526 0.1500 0.1486
1998 0.1694 0.1640 0.1607 0.1581 0.1531 0.1488 0.1476 0.1469
Mean 0.1315 0.1323 0.1339 0.1352 0.1380 0.1406 0.1419 0.1425
St. dev. 0.0291 0.0243 0.0205 0.0175 0.0116 0.0063 0.0043 0.0032
The parameters of the EGARCH price process are listed in Table 32.2. The half-life of a volatility
shock is 1.8 months
European option prices are estimated from 10,000 simulations of the asset prices on the dates that
the options expire. The implied volatilities are for at-the-money options whose exercise prices
equal the forward rates for the expiry dates. The standard errors of the implied volatilities are
between 0.0001 and 0.0003
Options are valued on the final trading day of ten consecutive years. The returns history is drawn
from the set of daily returns from the S & P 100 index from 6 March 1984–31 December 1998.
A set of t0 ¼ 2,000 historical returns is used from 1992 onwards and as many as possible before
then. The column for T ¼ 0 provides the annualized volatilities on the valuation dates
and 32.3, respectively. The IV values for T ¼ 0 are obtained from the conditional
variances on the valuation dates. The standard errors of the tabulated implied volatil-
ities increase with T. The maximum standard errors for at-the-money options are,
respectively, 0.0003 and 0.0004 for the short and long memory specifications.
The ten IV term structures for the short memory specification commence
between 9.5 % (1993) and 18.8 % (1997) and converge towards the limiting
value of 14.3 %. The initial IV values are near the median level from 1989 to
1991, are low from 1992 to 1995, and are high from 1996 to 1998. Six of the term
structures slope upwards, two are almost flat, and two slope downwards. The shapes
of these term structures are completely determined by the initial IV values because
the volatility process is Markovian.
There are three clear differences between the term structures for the short and
long memory specifications that can be seen by comparing Figs. 32.2 and 32.3.
First, the long memory term structures can and do intersect because the volatility
process is not Markovian. Second, some of the term structures have sharp kinks
for the first month. This is particularly noteworthy for 1990 and 1996 when the
term structures are not monotonic. For 1990, the initial value of 14.1 % is
followed by 15.6 % at 1 month and a gradual rise to 16.2 % at 6 months and
a subsequent slow decline. For 1996, the term structure commences at 15.6 %,
32 Consequences for Option Pricing of a Long Memory in Volatility 921
Table 32.4 At-the-money implied volatilities for a long memory volatility process
Implied volatilities for options that expire after 1, 2, 3, 6, 12, 18, and 24 months
0 1 2 3 6 12 18 24
Year
1989 0.1194 0.1356 0.1403 0.1429 0.1467 0.1496 0.1507 0.1515
1990 0.1413 0.1556 0.1592 0.1609 0.1624 0.1624 0.1614 0.1607
1991 0.1215 0.1368 0.1416 0.1441 0.1478 0.1502 0.1510 0.1516
1992 0.1239 0.1261 0.1283 0.1301 0.1338 0.1375 0.1395 0.1409
1993 0.1080 0.1101 0.1128 0.1150 0.1195 0.1245 0.1272 0.1292
1994 0.1114 0.1189 0.1213 0.1228 0.1256 0.1284 0.1300 0.1314
1995 0.0965 0.1041 0.1067 0.1085 0.1119 0.1160 0.1187 0.1210
1996 0.1564 0.1357 0.1311 0.1295 0.1283 0.1288 0.1297 0.1308
1997 0.1697 0.1650 0.1626 0.1609 0.1574 0.1540 0.1523 0.1515
1998 0.1734 0.1693 0.1682 0.1672 0.1650 0.1623 0.1610 0.1602
Mean 0.1322 0.1357 0.1372 0.1382 0.1399 0.1414 0.1422 0.1429
St. dev. 0.0267 0.0221 0.0211 0.0204 0.0186 0.0165 0.0151 0.0141
The parameters of the FIEGARCH price process are listed in Table 32.2 and include an integration
level of d ¼ 0.4.
European option prices are estimated from 10,000 simulations of the asset prices on the dates that
the options expire. The implied volatilities are for at-the-money options whose exercise prices
equal the forward rates for the expiry dates. The standard errors of the implied volatilities are all
less than 0.0004
Options are valued on the final trading day of ten consecutive years. The returns history is drawn
from the set of daily returns from the S & P 100 index from 6 March 1984–31 December 1998.
A set of t0 ¼ 2,000 historical returns is used from 1992 onwards, and as many as possible before
then. The column for T ¼ 0 provides the annualized volatilities on the valuation dates
falls to 13.6 % after 1 month, and reaches a minimum of 12.8 % after 6 months
followed by a slow incline. The eight other term structures are monotonic and
only those for 1997 and 1998 slope downwards. Third, the term structures
approach their limiting value very slowly.4 The 2-year IVs range from 12.1 %
to 16.1 %, and it is not possible to deduce the limiting value, although
15.0–16.0 % is a plausible range.5 It is notable that the dispersion between the
ten IV values for each T decreases slowly as T increases, from 2.2 % for 1-month
options to 1.4 % for 2-year options.
There are substantial differences between the two IV values that are
calculated for each valuation date and each option lifetime. Figure 32.4
shows the differences between the at-the-money IVs for the long memory
4
The results support the conjecture that IV(T) ffi a1 + a2T2d1 for large T with a2 determined by the
history of observed returns.
5
An estimate of the constant a1 (defined in the previous footnote) is 16.0 %. An estimate of 15.0 %
follows by supposing the long memory limit is 105 % of the short memory limit, based on the limit
of ln(ht) being higher by 0.1 for the long memory process as noted in Appendix 3. The difference in
the limits is a consequence of the risk premium obtained by owning the asset; its magnitude is
mainly determined by the pronounced asymmetry in the volatility shock function g(zt).
922 S.J. Taylor
0.20
At-the-money implied volatility
0.18
0.16
0.14
0.12
0.10
0.08
0 3 6 9 12 15 18 21 24
Option life in months
Fig. 32.2 Ten volatility term structures for a short memory process
0.20
At-the-money implied volatility
0.18
0.16
0.14
0.12
0.10
0.08
0 3 6 9 12 15 18 21 24
Option life in months
Fig. 32.3 Ten volatility term structures for a long memory process with d = 0.4
specification minus the number for the short memory specification. When
T ¼ 0, these differences range from 1.9 % (1997) to 1.5 % (1992), for
3 month options from 1.5 % (1995, 1996) to 2.1 % (1990), and for 2-year
options from 1.9 % (1995) to 1.7 % (1990). The standard deviation of the ten
differences is between 1.1 % and 1.4 % for all values of T considered so it is
common for the short and long memory option prices to have IVs that differ by
more than 1 %.
32 Consequences for Option Pricing of a Long Memory in Volatility 923
0.025
ATM implied when d=0.4 minus value when d=0
0.020
0.015
0.010
0.005
0.000
0 3 6 9 12 15 18 21 24
−0.005
−0.010
−0.015
−0.020
−0.025
Option life in months
The columns of the IV matrix provide information about the strength of the
so-called smile effect for options prices. These effects seem to be remarkably
robust to the choice of valuation date and they are not very sensitive to the choice
between the short and long memory specifications. This can be seen by considering
the ten values of DIV ¼ IV(T,X1) IV(T,X2) obtained for the ten valuation dates, for
various values of T, various pairs of exercise prices X1, X2, and a choice of volatility
process. First, for 1-month options with S ¼ 100, X1 ¼ 92, and X2 ¼ 108, the values
of DIV range from 3.0 % to 3.3 % for the short memory specification and from
3.7 % to 4.0 % for the long memory specification. Second, for 2-year options with
X1 ¼ 80 and X2 ¼ 120, the values of DIV range from 1.8 % to 2.0 % and from 1.8 %
to 1.9 %, respectively, for the short and long memory specifications.
Figure 32.5 shows the smiles for 3-month options valued using the short memory
model, separately for the ten valuation dates. As may be expected from the above
remarks, the ten curves are approximately parallel to each other. They are almost all
monotonic decreasing for the range of exercise prices considered, so that a U-shaped
function (from which the idea of a smile is derived) cannot be seen. The near
monotonic decline is a standard theoretical result when volatility shocks are nega-
tively correlated with price shocks (Hull 2000). It is also a stylized empirical fact for
US equity index options; see, for example, Rubinstein (1994) and Dumas et al. (1998).
Figure 32.6 shows the 3-month smiles for the long memory specification. The
shapes on Figs. 32.5 and 32.6 are similar, as all the curves are for the same expiry
time, but they are more dispersed on Fig. 32.6 because the long memory effect
induces more dispersion in at-the-money IVs. The minima of the smiles are
generally near an exercise price of 116. Figure 32.7 shows further long memory
smiles, for 2-year options when the forward price is 106.2. The parallel shapes are
924 S.J. Taylor
0.22
0.20
0.18
Implied volatility
0.16
0.14
0.12
0.10
0.08
84 88 92 96 100 104 108 112 116 120
Exercise price
Fig. 32.5 Ten smile shapes for three-month options and a short memory process
0.22
0.20
0.18
Implied volatility
0.16
0.14
0.12
0.10
0.08
84 88 92 96 100 104 108 112 116 120
Exercise price
Fig. 32.6 Ten smile shapes for three-month options and a long memory process with d = 0.4
clear; the two highest curves are almost identical, and the third, fourth, and fifth
highest curves are almost the same.
Tables 32.5 and 32.6 provide matrices of implied volatilities for options valued
on 31 December 1998. When either the call or the put option is deep out-of-the-
money, it is difficult to estimate the option price accurately because the risk-
neutral probability q(X) of the out-of-the-money option expiring in-the-money is
small. Consequently, the IV information has not been presented when the
corresponding standard errors exceed 0.002; estimates of q(X) are less than 3 %.
32 Consequences for Option Pricing of a Long Memory in Volatility 925
0.18
0.17
0.16
Implied volatility
0.15
0.14
0.13
0.12
0.11
0.10
80 84 88 92 96 100 104 108 112 116 120
Exercise price
Fig. 32.7 Ten smile shapes for two-year options and a long memory process with d = 0.4
The standard errors of the IVs are least for options that are near to at-the-money
and most of them are less than 0.0005 for the IVs listed in Tables 32.5 and 32.6.
All the sections of the smiles summarized by Tables 32.5 and 32.6 are monotonic
decreasing functions of the exercise price. The IV decreases by approximately
4–5 % for each tabulated section.
The sensitivity of the IV matrices to three of the inputs has been assessed for
options valued on 31 December 1998. First, consider a change to the risk parameter
l which corresponds to an annual risk premium of 6 % for the tabulated results.
From Sect. 32.4.3, option prices should be lower for large T when l is reduced to
zero. Changing l to zero reduces the at-the-money IV for 2-year options from
16.0 % to 15.4 % for the long memory inputs, with a similar reduction for the short
memory inputs. Second, consider reducing the truncation level N in the AR(N) filter
from 1,000 to 100. Although this has the advantage of a substantial reduction in the
computational time, it changes the IV numbers by appreciable amounts and cannot
be recommended; for example, the 2-year at-the-money IV then changes from
16.0 % to 14.7 %.
The smile shapes on Figs. 32.5, 32.6 and 32.7 are heavily influenced by the
negative asymmetric shock parameter y, which is substantial relative to the sym-
metric shock parameter g. The asymmetry in the smile shapes can be expected to
disappear when y is zero, which is realistic for some assets including exchange
rates. Figures 32.8 and 32.9 compare smile shapes when y is changed from the
926 S.J. Taylor
Table 32.5 A matrix of implied volatilities for a short memory volatility process
Implied volatilities for options that expire after 1, 2, 3, 6, 12, 18, and 24 months
1 2 3 6 12 18 24
X
72 0.1705 0.1651
76 0.1744 0.1670 0.1624
80 0.1835 0.1696 0.1637 0.1600
84 0.1994 0.1911 0.1771 0.1655 0.1607 0.1577
88 0.1959 0.1876 0.1818 0.1711 0.1617 0.1579 0.1556
92 0.1845 0.1780 0.1734 0.1653 0.1579 0.1552 0.1535
96 0.1735 0.1693 0.1660 0.1599 0.1546 0.1526 0.1515
100 0.1644 0.1615 0.1593 0.1549 0.1514 0.1502 0.1496
104 0.1577 0.1550 0.1533 0.1502 0.1481 0.1479 0.1478
108 0.1537 0.1498 0.1480 0.1461 0.1452 0.1457 0.1461
112 0.1510 0.1461 0.1437 0.1422 0.1425 0.1436 0.1445
116 0.1525 0.1435 0.1407 0.1388 0.1400 0.1417 0.1429
120 0.1421 0.1384 0.1359 0.1376 0.1399 0.1415
126 0.1325 0.1345 0.1373 0.1394
132 0.1304 0.1317 0.1348 0.1374
138 0.1293 0.1328 0.1353
144 0.1272 0.1305 0.1336
150 0.1255 0.1286 0.1322
156 0.1269 0.1306
162 0.1246 0.1294
168 0.1220 0.1282
174 0.1277
180 0.1274
The parameters of the EGARCH price process are listed in Table 32.2. The half-life of a volatility
shock is 1.8 months
Options are valued on 31 December 1998. The returns history is the set of 2,000 daily returns from
the S & P 100 index from 4 February 1991 to 31 December 1998. European option prices are
estimated from 10,000 simulations of the asset prices on the dates that the options expire
Implied volatilities shown in Roman font have standard errors (s.e.) that are at most 0.0005 and
those shown in italic font have s.e. between 0.0005 and 0.0020; results are not shown when the
s.e. exceeds 0.0020, when options are either deep in- or out-of-the-money
values used previously to zero, with g scaled to ensure the variance of ln(ht) is
unchanged for measure P. Figure 32.8 shows that the 1-month smile shapes become
U-shaped when y is zero, while Fig. 32.9 shows that the IV are then almost constant
for 1-year options.
32.6 Conclusions
The empirical evidence for long memory in volatility is strong, for both equity and
foreign exchange markets. This evidence may more precisely be interpreted as
32 Consequences for Option Pricing of a Long Memory in Volatility 927
Table 32.6 A matrix of implied volatilities for a long memory volatility process
Implied volatilities for options that expire after 1, 2, 3, 6, 12, 18, and 24 months
1 2 3 6 12 18 24
X
72 0.1811 0.1765
76 0.1829 0.1783 0.1743
80 0.1895 0.1792 0.1754 0.1723
84 0.1970 0.1847 0.1760 0.1728 0.1703
88 0.1954 0.1889 0.1798 0.1729 0.1703 0.1683
92 0.1953 0.1860 0.1815 0.1750 0.1699 0.1680 0.1664
96 0.1814 0.1771 0.1746 0.1707 0.1672 0.1657 0.1645
100 0.1699 0.1691 0.1685 0.1666 0.1645 0.1635 0.1627
104 0.1611 0.1623 0.1631 0.1627 0.1617 0.1615 0.1610
108 0.1557 0.1569 0.1581 0.1593 0.1592 0.1596 0.1595
112 0.1525 0.1526 0.1538 0.1559 0.1570 0.1577 0.1579
116 0.1499 0.1505 0.1529 0.1549 0.1559 0.1564
120 0.1480 0.1480 0.1504 0.1528 0.1543 0.1550
126 0.1439 0.1469 0.1502 0.1520 0.1530
132 0.1440 0.1478 0.1497 0.1512
138 0.1421 0.1456 0.1477 0.1492
144 0.1436 0.1456 0.1473
150 0.1417 0.1438 0.1457
156 0.1402 0.1421 0.1444
162 0.1409 0.1428
168 0.1388 0.1416
174 0.1370 0.1405
180 0.1399
The parameters of the FIEGARCH price process are listed in Table 32.2 and include an integration
level of d ¼ 0.4
Options are valued on 31 December 1998. The returns history is the set of 2,000 daily returns from
the S & P 100 index from 4 February 1991 to 31 December 1998. European option prices are
estimated from 10,000 simulations of the asset prices on the dates that the options expire
Implied volatilities shown in Roman font have standard errors (s.e.) that are at most 0.0005, and
those shown in italic font have s.e. between 0.0005 and 0.0020; results are not shown when the
s.e. exceeds 0.0020, when options are either deep in- or out-of-the-money
evidence for long memory effects, because there are short memory processes that
have similar autocorrelations and spectral densities, except at very low frequencies.
The theory of option pricing when volatility follows a discrete-time ARCH
process relies on weak assumptions about the continuous-time process followed
by prices and the numerical implementation of the theory is straightforward.
Application of the theory when the volatility process is fractionally integrated
does, however, require pragmatic approximations because the fundamental filter
(1 L)d is an infinite order polynomial that must be truncated at some power N.
Option prices are sensitive to the truncation point N, so that large values and long
price histories from an assumed stationary process are required.
928 S.J. Taylor
0.22
0.20
0.18
Implied volatility
0.16
0.14
0.20
d=0.4, theta<0
0.19
d=0, theta<0
d=0.4, theta=0
0.18
d=0, theta=0
Implied volatility
0.17
0.16
0.15
0.14
0.13
0.12
84 88 92 96 100 104 108 112 116 120
Exercise price
they may not have a monotonic shape. Also, term structures calculated on
different valuation dates sometimes intersect each other. None of these
possibilities occurs for a Markovian short memory specification. Long
memory term structures do not converge rapidly to a limit as the
lifetime of options increases. It is difficult to estimate the limit for the typical
value d ¼ 0.4.
Implied volatilities as functions of exercise prices have similar shapes for short
and long memory specifications. The differences in these shapes are minor in
comparison to the differences in the term structure shapes.
It is common for the short and long memory implied volatilities to differ by
more than 1 % for options on the S & P 100 index, regardless of the option
lifetime and the exercise price; if the short memory implied is at a typical level
of 14 %, then the long memory implied is often below 13 % or above 15 %.
Consequently, the economic consequences of a long memory assumption are
important.
Acknowledgments The support of Inquire Europe and the helpful advice of Bernard Dumas, Ton
Vorst, and Granville Tunnicliffe-Wilson are gratefully acknowledged.
Series expansions in the lag operator L are required to evaluate the following
conditional variance:
X
1
jd1
ð1 LÞd ¼ 1 aj Lj , a1 ¼ d, aj ¼ aj1 , j 2, (32.31)
j¼1
j
X
1
ð1 fLÞð1 LÞd ¼ 1 bj Lj , b1 ¼ d þ f, bj ¼ aj faj1 , j 2, (32.32)
j¼1
X
1
ð1 fLÞð1 LÞd ð1 þ cLÞ1 ¼ 1 fj Lj ,
j¼1
X
j1
f1 ¼ d þ f þ c, fj ¼ bj ðcÞj þ ðcÞjk bk , j 2: (32.33)
k¼1
930 S.J. Taylor
X
N
lnðht Þ ¼ a þ fj ln htj a þ gðzt1 Þ, (32.37)
j¼1
X
N
lnðht Þ ¼ a þ bj ln htj a þ gðzt1 Þ þ cgðzt2 Þ: (32.38)
j¼1
As j ! 1, the coefficients bj and fj converge much more rapidly to zero than the
coefficients cj. Consequently it is best to use either the AR or the ARMA
approximation.
Suppose there are returns observed at times 1 t t0 , whose distributions are given
by measure P, and that we then want to simulate returns for times t > t0 using
measure Q. Then ln (ht) is calculated for 1 t t0 + 1 using the observed returns,
with ln (ht) ¼ a and g(zt) ¼ 0 for t < 1, followed by simulating z*t Q N(0,1) and
hence obtaining rt and ln (ht+1) for t > t0 .
Care is required when calculating the constant C in Eq. 32.29 because the
observed conditional distributions of the terms zt are not normal while the simula-
tions assume that they are. Consequently, we define
C ¼ C0 , t t0 ,
pffiffiffiffiffiffiffiffi (32.39)
¼ 2=p, t > t0 ,
i.i.d. N(0, 1) sequence {z*t } to define the further i.i.d. N(0, 1) sequences, { zt*},
{z•t }, and { •
zt }, with
•
the terms zt chosen so that there
isnegative
correlation
between zt and zt ; this is achieved by defining F z t þ F zt ¼ 1 þ 12sign zt .
The four sequences provide claim prices whose average, y say, is much less variable
than the claim price from a single sequence. An overall average y^ is then obtained
from a set of K values fyk , 1 k K g.
The control variate method makes use of an unbiased estimate y^CV of a known
parameter yCV, such that y^ is positively correlated with y^CV. A suitable parameter,
when pricing a call option in an ARCH framework, is the price of a call option when
volatility is deterministic. The deterministic volatility process is defined by
replacing all terms ln(ht), t > t 0 + 1, by their expectations under P conditional on
the history It0 . Then yCV is given by the obvious modification of the Black-Scholes
formula, while y^CV is obtained by using the same 4K sequences of i.i.d. variables
that define y^. Finally, a more accurate estimate of the option price is then given by
y~ ¼ y^ b(^ y CV yCV) with b chosen to minimize the variance of y~.
On average the term structure of implied volatilities will slope upwards for the
FIEGARCH option pricing model. This occurs because the expectation of ln(ht)
depends on the measure6 when l 6¼ 0. The unconditional expectation equals a for
measure P. It is different for measure Q because
pffiffiffiffiffiffiffiffi l2 g
EQ g zt l ¼ ly þ g EQ zt l 2=p ffi ly þ pffiffiffiffiffiffi (32.40)
2p
when l is small, and this expectation is in general not zero.7 For a fixed t0 , as t ! 1,
!1
X
N
E ½lnðht Þ jI t0
! a þ
Q
1 bj ð1 þ cÞEQ g zt l (32.41)
j¼1
6
The dependence of moments of ht on the measure is shown by Duan (1995, p. 19) for the GARCH
(1,1) model.
pffiffiffiffiffiffiffiffi
7
When z N(0,1), E½jz lj
¼ 2=pexp 12l2 þ lð2FðlÞ 1Þ with F the cumulative distri-
bution function of z.
932 S.J. Taylor
The magnitude of the volatility risk premium can be important and, indeed, the
quantity defined by the limit in Eq. 32.41 becomes infinite8 as N ! 1 when d is
positive. A plausible value of l for the S & P 100 index is 0.028, obtained by assuming
that the equity risk premium is 6 % per annum. For the short memory parameter values
in the first row of Table 32.1, when d ¼ 0 and N ¼ 1,000, the limit of EQ[ln(ht) |It0 ] a
equals 0.10. This limit increases to 0.20 for the parameter values in the final row of
Table 32.1, when d ¼ 0.4 andpNffiffiffiffi¼ 1,000. The typical effect of adding 0.2 to ln(ht) is to
multiply standard deviations ht by 1.1 so that far-horizon expected volatilities, under
Q, are slightly higher than might be expected from historical standard deviations.
References
Andersen, T. G., & Bollerslev, T. (1997). Heterogeneous information arrivals and return volatility
dynamics: Uncovering the long-run in high frequency returns. Journal of Finance, 52, 975–1005.
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Ebens, H. (2001a). The distribution of stock
return volatility. Journal of Financial Economics, 61, 43–76.
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Labys, P. (2001b). The distribution of realized
exchange rate volatility. Journal of the American Statistical Association, 96, 42–55.
Areal, N. M. P. C., & Taylor, S. J. (2002). The realized volatility of FTSE-100 futures prices.
Journal of Futures Markets, 22, 627–648.
Baillie, R. T. (1996). Long memory processes and fractional integration in econometrics. Journal
of Econometrics, 73, 5–59.
Bakshi, G., Cao, C., & Chen, Z. (1997). Empirical performance of alternative option pricing
models. Journal of Finance, 52, 2003–2049.
Barndorff-Nielsen, O. E., & Shephard, N. (2001). Non-Gaussian Ornstein-Uhlenbeck based
models and some of their uses in financial economics. Journal of the Royal Statistical
Society B, 63, 167–241.
Bollerslev, T. (1986). Generalized autoregressive conditional heteroscedasticity. Journal of
Econometrics, 31, 307–327.
Bollerslev, T., & Mikkelsen, H. O. (1996). Modeling and pricing long memory in stock market
volatility. Journal of Econometrics, 73, 151–184.
Bollerslev, T., & Mikkelsen, H. O. (1999). Long-term equity anticipation securities and stock
market volatility dynamics. Journal of Econometrics, 92, 75–99.
Bollerslev, T., & Wooldridge, J. M. (1992). Quasi-maximum likelihood estimation and inference
in dynamic models with time varying covariances. Econometric Reviews, 11, 143–172.
Bollerslev, T., & Wright, J. H. (2000). Semiparametric estimation of long-memory volatility
dependencies: The role of high-frequency data. Journal of Econometrics, 98, 81–106.
Breidt, F. J., Crato, N., & de Lima, P. (1998). The detection and estimation of long memory in
stochastic volatility. Journal of Econometrics, 83, 325–348.
Brockwell, P. J., & Davis, R. A. (1991). Time series: Theory and methods (2nd ed.). New York:
Springer.
Christoffersen, P., Jacobs, K., Ornthanalai, C., & Wang, Y. (2008). Option valuation with long-run
and short-run volatility components. Journal of Financial Economics, 90, 272–297.
Christoffersen, P., Dorion, C., Jacobs, K., & Wang, Y. (2010). Volatility components: Affine restric-
tions and non-normal innovations. Journal of Business and Economic Statistics, 28, 483–502.
Comte, F., & Renault, E. (1998). Long memory in continuous-time stochastic volatility models.
Mathematical Finance, 8, 291–323.
X
1 X1
As (1 L)d1 ¼ 0 for d > 0, it follows from Eqs. 32.31 and 32.32 that
8
aj ¼ bj ¼ 1:
j¼1 j¼1
32 Consequences for Option Pricing of a Long Memory in Volatility 933
Dacorogna, M. M., Muller, U. A., Nagler, R. J., Olsen, R. B., & Pictet, O. V. (1993).
A geographical model for the daily and weekly seasonal volatility in the foreign exchange
market. Journal of International Money and Finance, 12, 413–438.
Ding, Z., Granger, C. W. J., & Engle, R. F. (1993). A long memory property of stock market
returns and a new model. Journal of Empirical Finance, 1, 83–106.
Duan, J.-C. (1995). The GARCH option pricing model. Mathematical Finance, 5, 13–32.
Duan, J.-C. (1997). Augmented GARCH(p, q) process and its diffusion limit. Journal of Econo-
metrics, 79, 97–127.
Duffie, D., Pan, J., & Singleton, K. J. (2000). Transform analysis and asset pricing for affine jump-
diffusions. Econometrica, 68, 1343–1376.
Dumas, B., Fleming, J., & Whaley, R. E. (1998). Implied volatility functions: Empirical tests.
Journal of Finance, 53, 2059–2106.
Ebens, H. (1999). Realized stock volatility. Working paper, John Hopkins University.
Ederington, L. H., & Lee, J. H. (1993). How markets process information: News releases and
volatility. Journal of Finance, 49, 1161–1191.
Gallant, A. R., Hsu, C. T., & Tauchen, G. E. (1999). Using daily range data to calibrate volatility
diffusions and extract the forward integrated variance. Review of Economics and Statistics, 81,
617–631.
Granger, C. W. J. (1980). Long memory relationships and the aggregation of dynamic models.
Journal of Econometrics, 14, 227–238.
Granger, C. W. J., & Ding, Z. (1996). Varieties of long memory models. Journal of Econometrics,
73, 61–77.
Granger, C. W. J., & Joyeux, R. (1980). An introduction to long-memory time series models and
fractional differencing. Journal of Time Series Analysis, 1, 15–29.
Harvey, A. C. (1998). Long-memory in stochastic volatility. In J. Knight & S. Satchell (Eds.),
Forecasting volatility in financial markets. London: Butterworth-Heinemann.
Heston, S. (1993). A closed-form solution for options with stochastic volatility with applications to
bond and currency options. Review of Financial Studies, 6, 327–343.
Hosking, J. (1981). Fractional differencing. Biometrika, 68, 165–176.
Hull, J. (2000). Options, futures and other derivative securities (4th ed.). New Jersey: Prentice-Hall.
Kallsen, J., & Taqqu, M. (1998). Option pricing in ARCH-type models. Mathematical Finance, 8,
13–26.
Nelson, D. B. (1990). ARCH models as diffusion approximations. Journal of Econometrics, 45, 7–38.
Nelson, D. B. (1991). Conditional heteroskedasticity in asset returns: A new approach.
Econometrica, 59, 347–370.
Ohanissian, A., Russell, J. R., Tsay, R. S. (2004). True or spurious long memory in volatility: Does
it matter for pricing options? Working paper, University of Chicago.
Ohanissian, A., Russell, J. R., & Tsay, R. S. (2008). True or spurious long memory? A new test.
Journal of Business and Economic Statistics, 26, 161–175.
Pong, S., Shackleton, M. B., & Taylor, S. J. (2008). Distinguishing short and long memory
volatility specifications. Econometrics Journal, 11, 617–637.
Ritchken, P., & Trevor, R. (1999). Pricing options under generalized GARCH and stochastic
volatility processes. Journal of Finance, 54, 377–402.
Rubinstein, M. (1994). Implied binomial trees. Journal of Finance, 49, 771–818.
Shephard, N. (1996). Statistical aspects of ARCH and stochastic volatility. In D. R. Cox, D. V.
Hinkley, & O. E. Barndorff-Nielsen (Eds.), Likelihood, time series with econometric and other
applications (pp. 1–67). London: Chapman and Hall.
Taylor, S. J. (1994). Modeling stochastic volatility: A review and comparative study. Mathemat-
ical Finance, 4, 183–204.
Taylor, S. J. (2005). Asset price dynamics, volatility, and prediction. Princeton: Princeton Uni-
versity Press.
Taylor, S. J. (2008). Modelling financial time series (2nd ed.). Singapore: World Scientific
Publishing.
Seasonal Aspects of Australian Electricity
Market 33
Vikash Ramiah, Stuart Thomas, Richard Heaney, and
Heather Mitchell
Contents
33.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 936
33.2 Australian Electricity Spot Price . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 938
33.3 Australian Demand for Electricity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 945
33.3.1 Demand Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 945
33.3.2 Demand Returns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 947
33.3.3 Modelling Seasonality in Demand for Electricity . . . . . . . . . . . . . . . . . . . . . . . . . . . . 952
33.4 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 952
33.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 955
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 956
Abstract
Australian electricity spot prices differ considerably from equity spot prices in
that they contain an extremely rapid mean-reversion process. The electricity spot
price could increase to a market cap price of AU$12,500 per megawatt hour
(MWh) and revert back to a mean level (AUD$30) within a half-hour interval.
This has implications for derivative pricing and risk management. For example,
while the Black and Scholes option pricing model works reasonably well for
equity market-based securities, it performs poorly for commodities like electric-
ity. Understanding the dynamics of electricity spot prices and demand is also
V. Ramiah (*)
School of Economics, Finance and Marketing, RMIT University, Melbourne, Australia
e-mail: vikash.ramiah@rmit.edu.au
S. Thomas • H. Mitchell
RMIT University, Melbourne, VIC, Australia
e-mail: stuart.thomas@rmit.edu.au; heather.mitchell@rmit.edu.au
R. Heaney
Accounting and Finance, The University of Western Australia, Perth, Australia
e-mail: richard.heaney@uwa.edu.au
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 935
DOI 10.1007/978-1-4614-7750-1_33,
# Springer Science+Business Media New York 2015
936 V. Ramiah et al.
Keywords
Electricity • Spot price • Seasonality • Outlier • Demand • Econometric modelling
33.1 Introduction
1
See Montero et al. (2011) for further evidence in the Spanish market.
938 V. Ramiah et al.
Western Australia and the Northern Territory are geographically remote and to date
have not been integrated into the “national” market. Physical transmission of power
between regions is achieved via interconnectors that physically link the five different
states. The spot electricity market in the AEMO is where all market generators and
market customers settle their electricity sales and purchases based on a spot price.
The market participants are generally large consumers of electricity, generators, and
speculators. Large consumers of electricity like car manufacturers and supermarkets
enter this market to enjoy a lower cost of production and to hedge their positions.
Speculators, on the other hand, trade future contracts, forwards, options, caps, floors,
and swaps for profit motives.
The half-hourly pool price data is sourced directly from AEMO (previously known
as National Electricity Market Management Company, NEMMCO) for the period
from 1 January 1999 to 31 January 2006. Prices are expressed in Australian dollars
per megawatt hour ($/MWh). The sample size is 124,224 observations for each
region except for TAS1 that joined the pool at a later stage. Figure 33.1 shows the
data series for NSW1 with evidence of considerable spiked behavior. Descriptive
statistics for the price and return series which are shown in Table 33.1 tend to
confirm these outliers.
10000
NSW
8000
6000
4000
2000
−2000
25000 50000 75000 100000 125000
Fig. 33.1 Electricity price series for NSW1 ($/MWh) for the period 01.01.1999–31.01.2006
33
Table 33.1 The descriptive statistics of spot prices and returns for the different Australian regions
Range Minimum Maximum Mean Std. deviation Skewness Kurtosis JB stats
Spot prices
NSW1 9,912.13 3.10 9,909.03 34.02 178.11 35.37 1468.54 89701
QLD1 9,098.74 156.14 8,942.60 36.65 158.45 28.56 1058.17 46527
SA1 9,822.43 822.45 8,999.98 41.91 155.18 25.40 800.35 26598
VIC1 7,746.07 329.91 7,416.16 30.19 103.26 35.38 1502.93 93950
TAS1 10,332.00 332.00 10,000.00 87.30 367.26 21.46 505.16 10584
Seasonal Aspects of Australian Electricity Market
Returns
RNSW1 453.00 222.00 231.00 0.0301 1.36 41.46 15,480.38 9981513
RQLD1 380.49 11.25 369.24 0.0611 1.74 111.49 19,122.19 15233050
RSA1 423.74 32.94 390.80 0.0603 1.68 123.00 24,820.63 25665638
RVIC1 351.93 209.50 142.43 0.0262 1.05 15.41 17,959.48 13434834
RTAS1 2416.00 83.50 2,332.50 0.6681 31.42 59.14 3,754.10 586865
939
940 V. Ramiah et al.
production to meet demand at that interval. The spot price is then calculated as an
arithmetic average of the six dispatch prices in a half hour. All generators who are
called into production during a given half-hour trading interval receive this spot
price for the quantity of electricity delivered during the trading interval. A generator
may bid a negative price into the pool for its self-dispatch quantity as a tactical
move to ensure that they are among the first to be called in to generate as closing
down their operations due to not being called into production may be more costly
than producing at negative prices. The implication of a negative price is that the
standard return calculation for prices may contain this bias, and to that end, the
following return formula is used:
ðPt Pt1 Þ
RET t ¼ : (33.1)
jPt1 j
X
5 X6 X12 X
2006
RET R, t ¼ a0 þ b1, i RET R, t1 þ b2, j DAY j þ b3, k MTH k þ b4, l YRl
i¼1 j1 k¼1, 6¼9 l¼1999, 6¼2001
X
48 X
N R, S
X
N R, N
þ b5, m HH m þ b6, o SPIKER, o þ b7, p NEGR, p þ et
m¼1, 6¼24 o¼1 p¼1
(33.2)
where RETR,t represents the discrete return for region R at time t, a0 represents the
constant term, and DAYj represents the dummy variable for each day of the week
942 V. Ramiah et al.
(j ¼ 1 for Monday, 2 for Tuesday, . . . , 6 for Saturday). The dummy variable for
Sunday is dropped on the basis that the returns are relatively lower on that day. We
use similar justification when it comes to discarding other periods. MTHk represents
the dummy variable for each month (k ¼ 1 for January, 2 for February, . . . , 12 for
December). The dummy variable for September is dropped; YRl represents the
dummy variable for each year included in the sample period (l ¼ 1999, . . . , 2006).
The dummy variable for 2001 is dropped; HHm represents the dummy variable for
each half-hourly trading interval (m ¼ 1 for 00:00 h, 2 for 00:30 h, . . . , 48 for
23:30 h). The dummy variable for 11:30 h is dropped; SPIKER,o represents a set of
NR,S dummy variables, one for each extreme spike as previously defined, with NR,S
representing the number of extreme returns observed in region R for the period of the
study; NEGR,N represents the dummy variable for the return associated with an
occurrence of a negative price (p ¼ 1, . . . , NR,N), with NR,N representing the number
of occurrences of a negative price for region R during for the period of the study.
The values for the trading interval at 11:30 h, Sunday, September, and the year
2001 were dropped to avoid exact collinearity and to allow comparison of these
values with the remaining seasonal coefficients for HHm, DAYj, MTHk, and YRl. The
equation was initially estimated for each region with 20 lagged returns (RETRt1, . . . ,
RETRt20). F-tests for redundant variables were performed for all regions, and AIC
and SBC values support the finding that lags 1 through 5 were significant. Lags
6 onwards were not found to be significant and were discarded. Standard tests and
residual diagnostics revealed no misspecification in the above model.
Results of the regression analysis are presented in Table 33.2. Coefficients and
t-statistics (t-stats) are presented for each seasonal dummy variable. The results
generated from this model are consistent with Kaminski (1997), Clewlow and
Strickland (2000) and De Jong and Huismann (2002) with regard to the seasonal
aspects and spike behavior of the price series. Seasonal effects vary between
regions, and time-of-day effects are generally more significant than other
seasonalities. Positive returns are observed at times of peak population activity in
the morning and early evening, and negative returns observed at most other times.
In general, significant negative returns are found for the small hours of the morning
between 12:30 a.m. and approximately 4:00 a.m. in all regions. NSW1 exhibits an
unexplained positive return at 1:30 a.m., reverting to negative returns for the
remainder of the early morning. The hours between 5:00 a.m. and 9:30 a.m.
inclusively exhibit significant positive effect in all regions, reverting to generally
negative returns in the late morning. Negative returns are found in SA1 during
midafternoon. Significant positive effects are observed for all regions in the early
evening, generally between the hours of 5:00 and 7:00 p.m., reverting to significant
negative effects for the remainder of the evening, until positive effects emerge in
the late evening at 10:30 p.m. and at midnight. The periods of positive return in the
morning and early evening are consistent with peaks in activity in the population.
The positive returns observed around 11:00 p.m. are consistent with increased
demand for electricity arising from off-peak hot water systems generally switching
on at 11:00 p.m.
33 Seasonal Aspects of Australian Electricity Market 943
The demand data used in this study are half-hourly observations of total demand.
The data is obtained from the same source as the price data and covers the same
period and regions. The basic quantity of interest in demand modelling and
forecasting is typically the periodic “total system demand” or “total demand.”
The total demand value reported by AEMO is a derived value, somewhat different
from demand (as may be represented by traded volume), as it may be understood in
conventional financial markets. Suppliers and distributors lodge schedules and
bids for the sale and purchase of electricity with AEMO at 12:30 p.m. on the day
prior to actual dispatch of electricity for each interval. AEMO compiles this data
and mates it with a short-term forecast of system demand and grid capacity
to determine an expected dispatch quantity and dispatch order of generators
(Smith 2000).
This study uses AEMO’s reported “total demand” values for each region,
expressed in megawatts (MW) by half-hour trading interval for the sample period.
Total demand is defined by AEMO as the total forecast regional demand against
which a dispatch solution is performed. For any particular interval and region, this
is determined as described by Eq. 33.3:
X
n X
n X
n
DT ¼ Gi Li þ NI i þ AIL þ FðDÞ þ ADE (33.3)
i¼1 i¼1 i¼1
946 V. Ramiah et al.
7500
7000
6500
6000
MW
5500
5000
4500
VIC1 Demand - 4/6/00 to 11/6/00
4000
Fig. 33.2 Plot of VIC1 demand for the week commencing Monday 4/6/00, illustrating the regular
intraday and daily seasonal patterns in the demand series
where:
DT is total demand.
Gi is “generator initial MW (SCADA),” the sum of initial MW values for all
scheduled generation units within the region, measured at their generator termi-
nals and reported by SCADA.
Li is “load initial MW (SCADA),” the scheduled base-load generation level for the
interval.
NIi is “net interconnector initial MW into region,” the net of all interconnector flows
into and out of the region.
AIL is “total allocated interconnector losses” represented by ∑(MW losses
X regional loss allocation). “MW losses” represent actual power losses due to
physical leakage from the transmission system. Regional loss allocation is an
NEMMCO predetermined static loss factor for each interconnector.
F(D) is demand forecast, a per-interval demand adjustment that relates the demand
at the beginning of the interval to the target at the end of the interval.
ADE is “aggregate dispatch error,” an adjustment value used by the NEM to
account for disparities between scheduled and actual dispatch for all scheduled
generation units in the region.
Figure 33.2 is provided to illustrate the presence of seasonal patterns in the
intraday behavior of electricity demand in Victoria over a 10-day period in 2000.
Descriptive statistics for the demand series are shown in Table 33.3.
We report the mean, standard deviation, minimum, maximum, range, skewness,
kurtosis, Jarque-Bera statistic, and Augmented Dickey-Fuller2 statistics for each
2
Additional statistics are provided for the demand series but not for the price series. This is because
the original paper on the price series reports these values.
33 Seasonal Aspects of Australian Electricity Market 947
Table 33.3 Descriptive statistics for demand by region, January 1999 to January 2006
Demand NSW1 QLD1 SA1 VIC1
Meana 8123.89 5197.74 1451.56 5411.02
S.D.a 1306.02 864.33 264.45 757.26
Maximuma 12884.15 8231.95 2873.03 8545.39
Minimuma 4624.03 2945.96 778.00 2726.88
Skewness 0.03 0.12 0.74 0.04
Kurtosis 2.64 2.67 4.44 2.83
JB stat 700.17 874.81 22026.19 175.72
ADFb 22.83 17.50 28.30 24.77
N 124224 124224 124224 124224
a
Mean, standard deviation, maximum, and minimum are expressed in megawatts (MW)
b
Augmented Dickey-Fuller (ADF) statistic rejects the hypothesis of a unit root at the 1 % level of
confidence
region’s demand series. NSW1 has the highest mean, median, and maximum demand
observations of the five regions for the period. New South Wales is Australia’s most
populous state so we would expect that demand for electric power to be highest in the
NSW1 region. The other regions follow generally in order of state population, with
VIC1 next highest, followed by QLD1, and SA1. The standard deviation and range of
demand levels are generally high, widely dispersed across the regions, and broadly
consistent with the pattern of means, ranging from 264 MW for SA1 to 1,306 MW for
NSW1. The distributions of demand observations are slightly positively skewed for
all regions. NSW1, QLD1, and VIC1 are slightly platykurtic, while SA1 is
leptokurtic. Jarque-Bera (JB) statistics reject the hypothesis of normal distribution
at the 1 % level of significance for all four regions, and Augmented Dickey-Fuller
statistics reject the hypothesis of a unit root at the 1 % level of significance.
Percent
8000 0
−1
7000
−2
MW
6000
5000
Fig. 33.3 Plot of VIC1 demand and returns on demand for the week commencing 4/6/00.
Demand is in MW and returns are percentage returns
In light of the fact that AEMO’s total demand is reported at half-hourly intervals
in discrete time, the demand return series used in this study were generated as half-
hourly discrete returns rather than log returns, according to Eq. 33.4 as follows:
ðDt Dt1 Þ
RDt ¼ (33.4)
Dt1
where RDt is discrete demand return at time t, Dt is half-hourly demand at time t, and
Dt1 is the previous half-hourly total demand, i.e., at time t1. The results of tests for
the presence of a unit root give us confidence that the demand and return series are
stationary, and we prefer this discrete return specification over log returns, as a log
return specification will dampen the spike effects we are attempting to capture.
We define a spike in demand returns as any observed demand return greater than
four standard deviations larger than the mean. Tables 33.4 and 33.5 collates the
occurrences of spikes as defined. Panel (a) shows the occurrence of spikes by region
and in aggregate for weekday, month, and year. Panels (b) and (c) show the
occurrence of spikes by half-hourly trading interval.
Table 33.4 Panel (a) shows that in aggregate there are 208 spikes in demand
returns observed across all regions during the sample period. VIC1 shows the
highest incidence of demand spikes with 92 (44 %) of the 208 observed, followed
by NSW1 with 81 (38 %), and SA1 with 22 (10 %) spikes during the sample period.
By day of the week, Monday shows the highest incidence with 67 (32 %)
tapering gradually to Sunday with 15 occurrences (7 %). August shows the highest
incidence by month with 42 (20 %), and of these 37 occur in NSW1. The next
33 Seasonal Aspects of Australian Electricity Market 949
Table 33.4 Panel (a) Summary of occurrences of extreme demand spikes by region, day of week,
month, and year
Interval NSW1 QLD1 SA1 VIC1 Total
Sun 1 4 4 6 15
Mon 39 3 8 17 67
Tue 12 2 1 21 36
Wed 12 3 4 13 32
Thu 13 0 3 15 31
Fri 4 0 0 15 19
Sat 0 1 2 5 8
Jan 0 0 0 23 23
Feb 0 0 0 17 17
Mar 1 1 5 31 38
Apr 0 0 0 8 8
May 13 0 2 0 15
Jun 20 2 3 0 25
Jul 3 4 4 0 11
Aug 37 1 1 3 42
Sep 4 1 0 1 6
Oct 2 0 1 1 4
Nov 1 4 4 3 12
Dec 0 0 2 5 7
1999 15 4 4 78 101
2000 13 0 2 9 24
2001 17 0 2 4 23
2002 16 3 0 1 20
2003 10 1 0 0 11
2004 7 4 3 0 14
2005 3 1 11 0 15
2006 0 0 0 0 0
Total 81 13 22 92 208
highest incidences by month are March (38) and June (25), with spikes predomi-
nantly in VIC1 for March and in NSW1 in June. The highest incidence by year
occurs in 1999 with 101 spikes (49 %), dropping markedly in 2000 (24) and 2001
(23). The incidence of spikes appears to have settled somewhat from 2003 onwards
at around 15–16 spikes per year.3
Table 33.5 Panel (b) shows the incidence of extreme spikes in demand returns by
half-hourly trading interval. There are concentrations of spikes occurring at the 06:30
(91 spikes, of which 78 occur in NSW1) and 23:30 (80 spikes, all of which occur in
VIC1). A subperiod analysis of demand returns suggests that sharp peaks in demand
3
Sample data for 2006 only includes the month of January and is unlikely to be representative of
the full year.
950 V. Ramiah et al.
Table 33.5 Panel (b) Occurrence of extreme demand spikes by half-hourly trading interval
Interval NSW1 QLD1 SA1 VIC1 Total
H0000 0 0 1 0 1
H0030 0 0 0 0 0
H0100 0 0 0 0 0
H0130 0 0 0 0 0
H0200 0 0 0 0 0
H0230 0 0 0 0 0
H0300 0 0 0 0 0
H0330 0 0 0 0 0
H0400 0 0 0 0 0
H0430 0 0 0 0 0
H0500 0 1 0 0 1
H0530 1 0 0 2 3
H0600 0 0 0 0 0
H0630 78 7 0 6 91
H0700 0 0 5 0 5
H0730 0 1 2 0 3
H0800 0 1 1 1 3
H0830 0 0 1 0 1
H0900 1 1 2 1 5
H0930 0 0 0 0 0
H1000 0 0 0 0 0
H1030 0 0 0 0 0
H1100 0 0 0 0 0
H1130 0 0 0 0 0
H1200 0 0 0 1 1
H1230 0 0 0 0 0
H1300 0 0 1 0 1
H1330 0 0 2 0 2
H1400 0 0 0 0 0
H1430 0 0 1 0 1
H1500 0 0 0 0 0
H1530 0 0 0 0 0
H1600 0 0 0 0 0
H1630 0 0 0 0 0
H1700 0 0 0 0 0
H1730 0 1 0 0 1
H1800 1 1 1 0 3
H1830 0 0 5 1 6
H1900 0 0 0 0 0
H1930 0 0 0 0 0
H2000 0 0 0 0 0
H2030 0 0 0 0 0
(continued)
33 Seasonal Aspects of Australian Electricity Market 951
0.08
0.06
Changes in demand
0.04
0.02
0
0
0
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
:3
00
01
02
03
04
05
06
07
08
09
10
11
12
13
14
15
16
17
18
19
20
21
22
23
−0.02
−0.04
−0.06
Time
Fig. 33.4 Half-hourly demand returns for VIC1 for the period 1999–2005
returns are persistent throughout the sample period, as illustrated by Fig. 33.4.
Figure 33.4 shows the pattern of demand returns for VIC1 for the period
1999–2005 and is illustrative of the pattern in NSW1. The 06:30 peak in demand
appears consistent with the commencement of the morning peak in activity in the
population. We believe that the 2,330 peak coincides with the activation of off-peak
hot water systems set to take advantage of overnight off-peak retail electricity tariffs.
Descriptive statistics for the half-hourly demand return series are shown in
Table 33.6. We report the mean, standard deviation, minimum, maximum, range,
skewness, kurtosis, and Augmented Dickey-Fuller statistics for each region’s
demand return series.
Mean, standard deviation, maximum, and minimum are expressed in terms of
half-hourly percentage return and are broadly consistent across NSW1, QLD1, SA1,
and VIC1. The distributions of demand returns for all four regions demonstrate
positive skewness and high positive kurtosis. Jarque-Bera (JB) statistics reject the
null hypothesis of normal distribution at the 1 % level of significance for all four
regions. This fat-tailed character is consistent with studies on price behavior (see
Huisman and Huurman 2003; Higgs and Worthington 2005; Wolack 2000) and
appears driven by the presence of spikes in demand returns. Augmented Dickey-
Fuller (ADF) statistics robustly reject the hypothesis of a unit root at the 1 % level of
significance for all five regions, again consistent with the findings of the earlier studies.
952 V. Ramiah et al.
Table 33.6 Descriptive statistics for half-hourly demand returns, by region, January 1999 to
January 2006
Returns NSW1 QLD1 SA1 VIC1
Meana 0.05 0.04 0.05 0.04
S.D.a 3.15 2.87 3.31 2.98
Maximuma 44.50 52.64 38.41 49.52
Minimuma 26.55 31.94 38.97 33.81
Skewness 1.13 0.94 0.39 1.11
Kurtosis 5.00 5.43 4.45 5.01
JB stat 47,301.83 49,025.43 14,003.32 46,573.67
ADFb 42.57 43.29 41.51 41.51
N 1,24,223 1,24,223 1,24,223 1,24,223
a
Mean, standard deviation, maximum, and minimum are expressed as percentage values
b
Augmented Dickey-Fuller (ADF) statistic rejects the hypothesis of a unit root at the 1 % level of
confidence
In this section, we adapt and adjust Eq. 33.2 to model seasonality in demand return.
The equation is as follows:
X
6 X
11 X
2006
RDit ¼ f0 þ b1 LRt þ b2 DAY i þ b3 MTH i þ b4 YRi
n¼1, i6¼Sun n¼1, i6¼Sep n¼1999, i6¼2001
(33.5)
X
47 XNS
þ b5 HH i þ b6 DDSPIKEi þ et
n¼1, i6¼1, 130 hrs n¼1
where RDit represents the discrete demand return for region i at time t, f0 represents
the constant term, LRit represents the lagged demand return for region i at time t,
DDSPIKEi represents the dummy variable set for each occurrence of extreme return
as previously defined, and the remaining variables are defined as in Eq. 33.2. The
trading interval at 11:30 h, Sunday, September, and the year 2001 were incorpo-
rated into the constant term a in the model as the base case for each dummy series.
These base cases were selected as the trading interval, day, month, and year, in
which demand return activity was consistently lowest in all four regions.
Results of the regression analysis are presented in Tables 33.7 and 33.8. Coeffi-
cients and t-statistics are presented for each seasonal dummy variable and for
lagged returns. In view of the very large number of individual spikes in demand
returns (208 spikes identified for the sample period across all regions), coefficients
for individual spikes are not explicitly reported though results are discussed.
33 Seasonal Aspects of Australian Electricity Market 953
Table 33.7 Panel (a) – Results Of regression analysis for half-hourly demand return against
seasonal dummy variables, by region for day, month, and year
NSW1 QLD1 SA1 VIC1
Coeff t-stats Coeff t-stats Coeff t-stats Coeff t-stats
C .00603 8.05 .00170 3.66 .00356 4.02 .00228 7.43
Mon 0.00099 7.88 0.00084 3.24 0.00091 4.57 0.00058 4.24
Tue 0.00014 1.41 0.00020 0.80 0.00011 0.70 0.00004 0.30
Wed 0.00010 0.96 0.00014 0.57 0.00005 0.34 0.00006 0.42
Thu 0.00004 0.30 0.00020 0.79 0.00047 2.36 0.00002 0.17
Fri 0.00099 7.27 0.00105 4.05 0.00092 4.28 0.00062 4.58
Sat 0.00025 1.87 0.00033 1.28 0.00018 0.85 0.00021 1.58
Jan 0.00018 0.38 0.00001 0.03 0.00009 0.12 0.00008 0.46
Feb 0.00005 0.10 0.00005 0.13 0.00016 0.23 0.00004 0.20
Mar 0.00003 0.05 0.00008 0.22 0.00006 0.09 0.00018 0.98
Apr 0.00009 0.19 0.00025 0.68 0.00014 0.20 0.00011 0.61
May 0.00009 0.20 0.00009 0.24 0.00022 0.31 0.00004 0.24
Jun 0.00003 0.07 0.00010 0.27 0.00038 0.54 0.00003 0.18
Jul 0.00011 0.23 0.00012 0.34 0.00016 0.23 0.00002 0.12
Aug 0.00003 0.06 0.00005 0.13 0.00040 0.61 0.00000 0.00
Oct 0.00000 0.01 0.00001 0.02 0.00009 0.14 0.00001 0.03
Nov 0.00008 0.18 0.00001 0.02 0.00013 0.19 0.00004 0.22
Dec 0.00012 0.26 0.00010 0.28 0.00041 0.59 0.00000 0.03
1999 0.00009 0.20 0.00006 0.24 0.00019 0.31 0.00031 2.26
2000 0.00003 0.06 0.00007 0.25 0.00002 0.03 0.00000 0.01
2002 0.00002 0.05 0.00001 0.03 0.00004 0.06 0.00001 0.10
2003 0.00001 0.02 0.00001 0.03 0.00003 0.05 0.00002 0.14
2004 0.00005 0.12 0.00001 0.04 0.00004 0.07 0.00002 0.15
2005 0.00002 0.05 0.00001 0.03 0.00009 0.15 0.00003 0.18
2006 0.00013 0.13 0.00008 0.11 0.00129 0.86 0.00005 0.15
R2 0.87 0.83 0.82 0.82 0.87
Adj R2 0.87 0.83 0.82 0.82 0.87
We find that coefficients are relatively small for NSW1, QLD1, SA1, and VIC1.
Day-of-week effects are positive and significant for Monday in NSW1, QLD1, SA1,
and VIC1. Thursday shows significant negative effect in demand returns in SA1 only.
There is significant negative Friday effect in all regions. There appears to be no clear
pattern evident for monthly effect across regions. There is no clear pattern evident in
yearly effect, although significant negative effect is observed for 1999 in VIC1. Half-
hourly time-of-day effects offer more interesting results and are broadly more
consistent across regions than the seasonal effects previously discussed. In general,
significant negative demand returns are found for the small hours of the morning
from 12:30 a.m. until approximately 4:00–5:00 a.m. in all regions. VIC1 exhibits an
unexplained highly significant positive return at 1:30 a.m., reverting to negative
returns until for the remainder of the early morning. There is wide variation in the
pattern of demand returns during the waking day. NSW1 and VIC1 show broadly
954 V. Ramiah et al.
Table 33.8 Panel (b): Results of regression analysis for demand return against seasonal dummy
variables, by halfhourly trading interval by region, 0000–2,330 h
NSW1 QLD1 SA1 VIC1
Coeff t-stats Coeff t-stats Coeff t-stats Coeff t-stats
H0000 0.013 14.30 0.040 91.44 0.044 49.76 0.055 139.42
H0030 0.008 8.91 0.057 114.74 0.013 14.63 0.017 48.79
H0100 0.014 15.48 0.047 89.84 0.004 4.89 0.010 26.90
H0130 0.014 15.96 0.043 87.43 0.062 70.67 0.086 238.13
H0200 0.020 22.39 0.036 74.58 0.064 68.82 0.066 169.17
H0230 0.019 21.03 0.029 61.46 0.053 55.84 0.023 62.27
H0300 0.009 10.63 0.019 41.36 0.040 42.74 0.012 33.13
H0330 0.005 5.91 0.010 23.47 0.031 33.90 0.006 17.25
H0400 0.005 5.71 0.004 9.95 0.023 25.48 0.002 5.04
H0430 0.015 17.58 0.005 12.12 0.013 15.02 0.007 20.52
H0500 0.020 22.54 0.013 29.54 0.004 4.49 0.009 26.51
H0530 0.048 54.16 0.034 76.97 0.005 5.53 0.027 77.09
H0600 0.039 43.14 0.045 93.50 0.022 24.88 0.021 59.57
H0630 0.049 53.66 0.079 157.08 0.028 31.42 0.045 123.52
H0700 0.045 48.87 0.084 133.88 0.052 58.70 0.017 44.80
H0730 0.011 12.11 0.082 133.22 0.041 44.66 0.022 57.89
H0800 0.036 40.06 0.058 94.63 0.023 25.87 0.028 78.38
H0830 0.020 22.05 0.035 67.65 0.031 35.51 0.009 24.48
H0900 0.003 3.91 0.018 38.06 0.019 21.19 0.002 6.33
H0930 0.015 17.25 0.014 30.12 0.007 8.13 0.007 20.14
H1000 0.001 1.24 0.004 8.62 0.005 5.81 0.001 3.97
H1030 0.002 2.45 0.001 1.36 0.002 3.36 0.002 5.85
H1100 0.002 3.03 0.002 6.49 0.001 0.98 0.003 7.98
H1200 0.004 5.80 0.001 5.05 0.001 2.53 0.002 6.15
H1230 0.003 3.82 0.004 10.98 0.002 2.08 0.001 2.39
H1300 0.002 2.03 0.005 10.64 0.000 0.23 0.001 3.86
H1330 0.004 5.03 0.001 3.23 0.002 2.92 0.010 28.51
H1400 0.002 2.14 0.001 2.88 0.006 6.97 0.005 12.89
H1430 0.005 5.82 0.000 0.89 0.010 11.42 0.000 0.61
H1500 0.005 5.44 0.003 6.23 0.006 6.46 0.002 4.52
H1530 0.007 7.57 0.001 1.98 0.010 11.66 0.003 7.70
H1600 0.011 12.34 0.004 9.22 0.009 9.88 0.005 14.29
H1630 0.008 8.69 0.008 17.99 0.004 4.93 0.002 6.33
H1700 0.019 22.22 0.016 36.28 0.003 3.59 0.009 23.96
H1730 0.014 15.77 0.021 46.66 0.005 5.65 0.005 13.52
H1800 0.027 30.31 0.027 58.50 0.010 11.86 0.010 28.01
H1830 0.001 1.18 0.013 28.47 0.015 16.97 0.006 16.80
H1900 0.001 1.01 0.006 14.25 0.002 2.06 0.004 10.01
H1930 0.004 5.07 0.010 21.73 0.008 8.85 0.005 13.97
H2000 0.002 2.52 0.014 32.78 0.012 14.03 0.002 5.10
(continued)
33 Seasonal Aspects of Australian Electricity Market 955
33.5 Conclusions
The current literature establishes that electricity time series differ from traditional
financial data having greater incidence of spikes than is generally observed in
financial data and this results in extreme volatility. The work done overseas is of
limited practical application to the Australian market as our market structure is
different. The lesson learned from our study is that even within the same country,
variation exists as different states reflect variation in geographic differences. The
implication is that specific models must be developed for each geographic region. Our
work is innovative in that we develop an econometric time series seasonal model that
can be applied in areas where seasonality is suspected. For instance, the model was
initially applied to price and return series and then later on to demand return series.
This model has the capability to be applied in other financial time series models.
Developing models to explain and to predict electricity prices is a significant
task. This is important both for the market participants who operate in the physical
market and for those trading in electricity derivatives. Modelling electricity prices
and trading in this market is challenging, so much so that investment banks avoid
trading electricity derivatives because of the incidence of spikes. This provides
a strong incentive for this research and future research into the electricity market.
A better understanding of the time series nature of the electricity prices may help in
the development of more efficient forecasting models which will help to lower risk
956 V. Ramiah et al.
management costs and thus reduce the cost of managing electricity price exposures.
Electricity generators, electricity retailers, transmission and distribution network
suppliers, electricity retailer companies, and electricity consumers will all benefit
from development of more accurate models in this area.
Acknowledgment The authors wish to acknowledge the financial support of the Australian
Centre for Financial Studies, via grant 18/2005. The research on the seasonality within the spot
price has been published in 2011, “Seasonal factors and outlier effects in rate of return in
Australia’s National Electricity Market,” Applied Economics, Vol. 43, Issue 3; pp. 355–369.
Binesh Seetanah provided invaluable research assistance. Any remaining errors are the responsi-
bility of the authors.
References
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of
Political Economy, 81(3), 637.
Bunn, D., & Karakatsani, N. (2003). Forecasting electricity prices. Working paper, London
Business School.
Clewlow, L., & Strickland, C. (2000). Valuing energy options in a one-factor model fitted to
forward prices. Working paper, University of Sydney.
Eichler, M., Grothe, O., Manner, H., & Tuerk, D. (2012). Modeling spike occurrences in electricity
spot prices for forecasting. Working paper, RM/12/029, Maastricht University.
Engle, R., Granger, C., & Hallman, J. (1989). Merging short and long-run forecasts: An applica-
tion of seasonal cointegration to monthly electricity sales forecasting. Journal of Economet-
rics, 50, 297–327.
Escribano, A. J., Peaea J., & Villaplana, P. (2002). Modelling electricity prices: International
evidence. Working paper, Universidad Carlos III de Madrid.
Goto, M., & Karolyi, G. A. (2004). Understanding electricity price volatility within and across
markets. Dice Centre working paper 2004–12, pp. 1–41.
Harvey, A., & Koopman, S. J. (2002). Forecasting hourly electricity demand using time-varying
splines. Journal of the American Statistical Association, 88(424), 1228–1236.
Huisman, R., & Huurman, C. (2003). Fat tails in power prices. ERIM report series reference
no. ERS-2003-059-F&A. SSRN: http://ssrn.com/abstract¼450992
Johnson, B., & Barz, G. (1999). Selecting stochastic processes for modelling electricity prices. In
Energy modelling and the management of uncertainty. Risk Publications.
Kaminski, V. (1997). The challenge of pricing and risk managing electricity derivatives. In The
U.S. Power Market (pp. 149–71).
Knittel, C. R., & Roberts, M. (2001). An empirical examination of deregulated electricity prices.
Working paper, PWP-087, University of California Energy Institute.
Montero, J. M., Garcia, M. C., & Fernandez-Aviles, G. (2011). Modelling the volatility of the
Spanish wholesale electricity spot market. Asymmetric GARCH models vs. Threshold ARSV
model. Revista de Estudios de Economı́a Aplicada, 29(2), 1–20.
Pardo, A., Meneu, V., & Valor, E. (2002). Temperature and seasonality influences on Spanish
electricity load. Energy Economics, 24, 55–70.
Smith, M. (2000). Modelling and short-term forecasting of New South Wales electricity system
load. Journal of Business & Economic Statistics, 18(4), 465–478.
Thomas, S., Ramiah, V., Heaney, R., & Mitchell, H. (2011). Seasonal factors and outlier effects in
rate of return in Australia’s national electricity market. Applied Economics, 43(3), 355–369.
Pricing Commercial Timberland Returns
in the United States 34
Bin Mei and Michael L. Clutter
Contents
34.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 958
34.2 Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 960
34.2.1 CAPM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 960
34.2.2 Fama-French Three-Factor Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 961
34.2.3 CAPM and Fama-French Three-Factor Model Under the State Space
Framework . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 961
34.2.4 Stochastic Discount Factor Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 962
34.3 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 963
34.3.1 Timberland Returns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 963
34.3.2 Basis Assets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 964
34.3.3 Other Indices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 964
34.4 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 965
34.4.1 Estimation of the CAPM and the Fama-French Three-Factor Model . . . . . . . . 965
34.4.2 State Space Estimation of the CAPM and the Fama-French
Three-Factor Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 965
34.4.3 Abnormal Performance Measured by the SDF Approach . . . . . . . . . . . . . . . . . . . . 966
34.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 967
Appendix 1: State Space Model with the Kalman Filter . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 970
Appendix 2: Heuristic Proof of Equation 34.6 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 971
Appendix 3: NCREIF Timberland Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 972
Appendix 4: EViews Code for Estimating the CAPM and the Fama-French
Three-Factor Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 972
Appendix 5: Steps for the SDF Approach Using Excel Solver . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 974
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 975
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 957
DOI 10.1007/978-1-4614-7750-1_34,
# Springer Science+Business Media New York 2015
958 B. Mei and M.L. Clutter
Abstract
Commercial timberland assets have attracted more attention in recent decades.
One unique feature of this asset class roots in the biological growth, which is
independent of traditional financial markets. Using both parametric and non-
parametric approaches, we evaluate private- and public-equity timberland
investments in the United States. Private-equity timberland returns are proxied
by the NCREIF Timberland Index, whereas public-equity timberland returns are
proxied by the value-weighted returns on a dynamic portfolio of the US publicly
traded forestry firms that had or have been managing timberlands. The results
from parametric analysis reveal that private-equity timberland investments
outperform the market and have low systematic risk, whereas public-equity
timberland investments fare similarly as the market. The nonparametric
stochastic discount factor analyses reveal that both private- and public-equity
timberland assets have higher excess returns.
Static estimations of the capital asset pricing model and Fama-French
three-factor model are obtained by ordinary least squares, whereas dynamic
estimations are by state space specifications with the Kalman filter. In estimating
the stochastic discount factors, linear programming is used.
Keywords
Alternative asset class • Asset pricing • Evaluation • Fama-French three-factor
model • Nonparametric analysis • State space model • Stochastic discount factor •
Timberland investments • Time series • Time-varying parameter
34.1 Introduction
34.2 Methods
For the parametric method, an explicit model is needed. Two candidate models
prevalent in the finance literature are the CAPM and the Fama-French three-factor
model. The parametric method is often criticized for the “joint hypothesis tests”
problem, i.e., testing the asset pricing model and the abnormal performance (market
efficiency) simultaneously. The nonparametric method does not require such an
explicit model specification and is therefore not subject to these critiques. The SDF
approach is a general, nonparametric asset pricing approach and is complement to
the parametric approaches.
34.2.1 CAPM
where ex post realized returns Ri and Rm rather than ex ante expected returns E[Ri]
and E[Rm] are used. The intercept ai is called Jensen’s (1968) alpha. A positive a
suggests that the individual asset outperforms the market and earns a higher than
risk-adjusted return, whereas a negative a suggests that the individual asset
underperforms the market and earns a lower than risk-adjusted return. Therefore,
Jensen’s alpha has become a commonly used measure of abnormal performance,
and testing whether it is zero has been widely used in the empirical asset pricing
literature.
34 Pricing Commercial Timberland Returns in the United States 961
Given the empirical evidence that small size stocks outperform large size stocks,
and value (high book-to-market) stocks outperform growth (low book-to-market)
stocks on average, Fama and French (1993) develop a model that includes these
extra two factors to adjust for risk:
where RRMRF ¼ Rm Rf is the same market factor as in the CAPM, representing the
market risk premium; RSMB ¼ Rsmall Rbig is the size factor, representing the return
difference between a portfolio of small stocks and a portfolio of large stocks (SMB
stands for “small minus big”); RHML ¼ RhighBM RlowBM is the book-to-market
factor, representing the return difference between a portfolio of high book-to-
market stocks and a portfolio of low book-to-market stocks (HML stands for
“high minus low”); and b’s are called factor loadings, representing each asset’s
sensitivity to these factors. When estimating Fama-French three-factor model, ex
post realized returns are used, as in the case of the CAPM, and an intercept is added
to capture the abnormal performance,
The CAPM (Eq. 34.2) and the Fama-French three-factor model (Eq. 34.4) are
usually estimated by OLS, possibly with some correction for the autocorrelations
in the errors. One restrictive nature of the OLS method is that the coefficients in the
regression are imposed to be constant. This condition may be unrealistic in real
asset pricing modeling. For instance, one would suspect that both a’s and b’s should
be time varying. To solve this problem, we can estimate the CAPM and the Fama-
French three-factor model in the state space framework with the Kalman filter
(Appendix 1). Using the CAPM as an example, in the state space framework, the
system of equations is specified as
Ri, t Rf , t ¼ ai, t þ bi, t Rm, t Rf , t þ mi, t
ai, tþ1 ¼ ai, t þ xt (34.5)
bi, tþ1 ¼ bi, t þ tt
where mi,t, xt, and tt are normally and independently distributed mean-zero
error terms. In the state space model, the first equation in (34.5) is called
the observation or measurement equation, and the second and third equations are
called the state equations. In this particular case, each state variable follows
a random walk.
962 B. Mei and M.L. Clutter
One advantage of the state space approach with time-varying parameters is that
it can incorporate external shocks, such as policy and regime shifts, economic
reforms, and political uncertainties, into the system, especially when the shocks
are diffuse in nature (Sun 2007). This approach has been applied to a variety of
issues, including demand systems (e.g., Doran and Rambaldi 1997), aggregate
consumptions (e.g., Song et al. 1996), policy analysis (e.g., Sun 2007), and price
modeling and forecasting (e.g., Malaty et al. 2007).
The single-period asset pricing models ignore the consumption decisions. In effect,
investors make their consumption and portfolio choices simultaneously in an
intertemporal setting. In the framework of an exchange economy in which an
investor maximizes the expectation of a time-separable utility function (Lucas
1978), it can be proved that (Appendix 2)
Et 1 þ Ri, tþ1 Mtþ1 ¼ 1, (34.6)
where Ri,t+1 is the return on asset i in the economy and Mt+1 is known as the
stochastic discount factor, or intertemporal marginal rate of substitution, or pricing
kernel (e.g., Campbell et al. 1997).
Hansen and Jagannathan (1991) demonstrated how to identify the SDF from
a set of basis assets, i.e., the derivation of the volatility bounds. These bounds
are recognized as regions of admissible mean-standard deviation pairs of the
SDF. Their major assumptions are the law of one price and the absence of arbitrage
opportunities. Accordingly, there are two particular solutions for the SDF: the law
of one price SDF and the no-arbitrage SDF. The process of retrieving the reverse-
engineered law of one price SDF is equivalent to the following constrained
optimization problem:
" #1=2
1 X T
Min sMt ¼ ðM t v Þ
Mt T 1 t¼1
1X T
s:t: Mt ¼ v (34.7)
T t¼1
1X T
Mt 1 þ Ri, t ¼ 1
T t¼1
for a range of selected v (mean of Mt) and for all assets i ¼ 1,2, ,N. Under the
stronger condition of no arbitrage, another positivity constraint on Mt is needed.
Therefore, the only difference between the law of one price SDF and the
no-arbitrage SDF is whether Mt is allowed to be negative. In this study,
no-arbitrage SDF is used. Following Hansen and Jagannathan (1991),
34 Pricing Commercial Timberland Returns in the United States 963
This equation presents the basis for testing the risk-adjusted performance of
a portfolio (Chen and Knez 1996). Namely, one can test whether
ai ¼ Et ai, t ¼ Et Ri, tþ1 Rf Mtþ1 ¼ 0: (34.9)
Ahn et al. (2003) pointed out that this measure generalizes Jensen’s alpha
and does not count on a specific asset pricing model. Based on this method,
they reassess the profitability of momentum strategies and found that their non-
parametric risk adjustment explains almost half of the anomalies.
34.3 Data
Returns for both private- and public-equity timberland investments are analyzed.
Although TIMOs have become the major timberland investment management
entities for institutional investors as well as high-net-wealth families and individ-
uals, their financial data are rarely publicly available. To provide a performance
benchmark, several TIMOs, together with National Council of Real Estate
Investment Fiduciaries (NCREIF) and the Frank Russell Company, initiated the
NCREIF Timberland Index in early 1992 (Binkley et al. 2003) (Appendix 3).
NCREIF members can be divided into data contribution members, professional
members, and academic members. Data contribution members include investment
managers and plan sponsors who own or manage real estate in a fiduciary setting.
Professional members include providers of accounting, appraisal, legal, consulting,
or other services to the data contribution members. Academic members include
full-time professors of real estate. Data contribution members submit their data on
a quarterly basis for computation of the NCREIF Property Index. Regarding
the NCREIF Timberland Index, it is some TIMOs that are the major data
contribution members. The quarterly NCREIF Timberland Index is reported at
both regional (the South, the Northeast, and the Pacific Northwest) and national
levels, and extends back to 1987. In this study, the national-level NCREIF
Timberland Index (1987Q1–2010Q4) is used as a return proxy for the US
private-equity timberland investments.
964 B. Mei and M.L. Clutter
Panel A of Table 34.1 presents the OLS estimation of the CAPM and the Fama-
French three-factor model using the quarterly NCREIF Timberland Index after
adjustment for the seasonality. A significant positive a from the CAPM suggests
that private-equity timberland investments have a risk-adjusted excess return of
about 8.20 % (2.05 % 4) per year. This excess return is slightly larger
after accounting for Fama-French factors. Market b’s from both models are insig-
nificantly different from zero, but significantly less than one. This means that
private-equity timberland investments are not only weakly correlated with the
market but also less risky than the market. The small magnitudes with high
p-values of the coefficients for SMB and HML signify that these two extra factors
add limited explanatory power to the CAPM in pricing private-equity timberland
returns.
In contrast, the CAPM and the Fama-French three-factor model fit the returns
on the dynamic portfolio of forestry firms much better, as implied by the higher R2
values (Panel A of Table 34.2). This is within our expectation since these forestry
firms are publicly traded and are more exposed to the market. However, a’s
are insignificant albeit positive, indicating no abnormal performance. Market b’s
are significantly different from zero, but not from one. In addition, b’s for SMB
and HML in Fama-French three-factor model are significant at the 5 % level or
better, meaning these factors capture some variations in the portfolio returns that
are not explained by the market premium. As a result, the abnormal performance (a)
has dropped by 53 %. The magnitudes of b’s indicate that the dynamic portfolio is
dominated by mid-large firms with middle book-to-market ratios.
Panel B of Table 34.1 presents the state space estimation of the CAPM and the
Fama-French three-factor model using the NCREIF Timberland Index. Those OLS
coefficient estimates are used as the starting values. Only a is specified as a state
variable (stochastic level) in that little time variation is observed in b, and both AIC
and SBC favor the deterministic-b model. Back to the model specification in system
(Eq. 34.5), this is equivalent to restrict tt ¼ 0. The AIC and SBC are marginal-
ly larger than those for the OLS estimation because of the relatively small sample
size. Figure 34.1 depicts the evolution of the risk-adjusted excess returns of the
NCREIF Timberland Index estimated from the CAPM. For most time in the last
22 years, the NCREIF Timberland Index has achieved positive abnormal returns
with an average of 10.20 % per year (calculated from the estimated a series).
Nevertheless, in certain years (2001–2003), the a is low and even negative,
966 B. Mei and M.L. Clutter
Table 34.1 Estimation of the CAPM and the Fama-French three-factor model using the NCREIF
Timberland Index (1987Q1–2010Q4)
CAPM FF3
Coefficient Estimate p-value Coefficient Estimate p-value
Panel A: OLS estimation
a 2.05 0.001 a 2.11 0.001
b 0.01 0.773 bRMRF 0.02 0.675
bSMB 0.04 0. 634
bHML 0.05 0.318
H0: b ¼ 1 0.000 H0: bRMRF ¼ 1 0.000
R2 0.14 R2 0.15
Log likelihood 251.24 Log likelihood 250.59
S.E. of regression 3.78 S.E. of regression 3.82
Durbin-Watson stat. 1.94 Durbin-Watson stat. 1.96
Akaike info. criterion 5.53 Akaike info criterion 5.56
Schwarz criterion 5.61 Schwarz criterion 5.69
F-stat. 7.09 F-stat. 3.82
Panel B: State space estimation
a 0.54 0.692 a 0.79 0.549
b 0.02 0.919 bRMRF 0.01 0.856
bSMB 0.11 0.263
bHML 0.05 0.581
H0: b ¼ 1 0.000 H0: bRMRF ¼ 1 0.000
Log likelihood 275.00 Log likelihood 273.59
Akaike info. criterion 5.86 Akaike info criterion 5.80
Schwarz criterion 5.94 Schwarz criterion 5.94
(1) OLS estimates after correction for the fourth-order autocorrelation in the residuals. (2) Only a
is specified stochastic under the state space framework, while b is specified deterministic due to its
lack of variation and AIC criterion
The mean of the no-arbitrage SDF Mt is specified in the selected range of [0.9750, 1]
with an increment step of 0.0025. When the five-industry portfolios plus
the long-term treasury bonds are used as the basis assets, the global minimum
variance of Mt is identified at v ¼ 0.9800; when the ten-industry portfolios plus
34 Pricing Commercial Timberland Returns in the United States 967
Table 34.2 Estimation of the CAPM and the Fama-French three-factor model using returns on
a dynamic portfolio of publicly traded forestry firms in the United States (1987Q1–2010Q4)
CAPM FF3
Coefficient Estimate p-value Coefficient Estimate p-value
Panel A: OLS estimation
a 0.59 0.521 a 0.28 0.750
b 0.95 0.000 bRMRF 0.87 0.000
bSMB 0.39 0.031
bHML 0.31 0.001
H0: b ¼ 1 0.330 H0: bRMRF ¼ 1 0.249
R2 0.48 R2 0.54
Log likelihood 344.00 Log likelihood 337.27
S.E. of regression 8.80 S.E. of regression 8.29
Durbin-Watson stat. 2.19 Durbin-Watson stat. 2.24
Akaike info. criterion 7.21 Akaike info criterion 7.11
Schwarz criterion 7.26 Schwarz criterion 7.22
F-stat. 85.16 F-stat. 36.58
Panel B: State space estimation
a 0.59 0.500 a 0.24 0.786
b 0.95 0.000 bRMRF 0.84 0.000
bSMB 0.39 0.021
bHML 0.32 0.001
H0: b ¼ 1 0.390 H0: bRMRF ¼ 1 0.446
Log likelihood 355.38 Log likelihood 348.61
Akaike info. criterion 7.47 Akaike info criterion 7.37
Schwarz criterion 7.55 Schwarz criterion 7.50
Only b is specified stochastic under the stochastic framework, while a is specified deterministic
due to its lack of variation and AIC criterion
the long-term treasury bonds are used instead, the global minimum variance of Mt is
identified at v ¼ 0.9750.
The SDF performance measures for both the NCREIF Timberland Index, and the
returns on the dynamic portfolio of publicly traded timber firms are reported in
Table 34.3. The a values for both return indices have increased, and the latter has
become marginally significant. This indeed implies that intertemporal consumption
decisions play a key role in pricing timberland assets. In a word, there is clear
evidence of statistically as well as economically significant excess returns for the
NCREIF Timberland Index, but only some evidence of economically significant
excess returns for the portfolio of publicly traded timber firms.
34.5 Conclusion
10 % Alpha
+/− 2 RMSE
8
−2
−4
−6
1988 1990 1992 1994 1996 1998 2000 2002 2004 2006 2008 2010
Fig. 34.1 Evolution of a over time from the state space estimation of the CAPM using the
NCREIF Timberland Index (1987Q1–2010Q4). Note: The time-varying a estimated from Fama-
French three-factor model exhibits similar patterns, thus is not shown separately. The graph is
available from the authors upon request
3.0
Beta
+/− 2 RMSE
2.5
2.0
1.5
1.0
0.5
1988 1990 1992 1994 1996 1998 2000 2002 2004 2006 2008 2010
Fig. 34.2 Evolution of b over time from the state space estimation of the CAPM using returns on
a dynamic portfolio of publicly traded forestry firms in the United States (1987Q1–2010Q4). Note:
The time-varying b estimated from Fama-French three-factor model exhibits similar patterns, thus
is not shown separately. The graph is available from the authors upon request
34 Pricing Commercial Timberland Returns in the United States 969
Table 34.3 Performance measures of timberland returns by the nonparametric SDF approach
(1987Q1–2010Q4)
Performance measure (a) p-value (one tail)
Mean of Mt (v) S.D. of Mt ðsMt Þ (1) (2) (1) (2)
Panel A: Five-industry portfolios plus long-term T-bonds
0.9775 0.199 2.63 1.59 0.000 0.119
0.9800 0.176 2.60 1.36 0.000 0.156
0.9825 0.217 2.57 1.13 0.000 0.202
Panel B: Ten-industry portfolios plus long-term T-bonds
0.9725 0.244 2.76 2.03 0.000 0.056
0.9750 0.237 2.75 1.82 0.000 0.082
0.9775 0.255 2.77 1.60 0.000 0.116
Column (1) is for the NCREIF Timberland Index, and Column (2) is for the returns on a dynamic
portfolio of the US publicly traded forestry firms that had or have been managing timberlands
the NCREIF Timberland Index fell by 6.5 %, the largest drop it ever had, which was
primarily caused by the capital loss from the shrinking timberland values. In the
same period, the S&P 500 index went up by 7.8 %. The overall decreasing trend in
b for the dynamic portfolio of forestry firms may be related to the massive
restructurings of these firms. For instance, Plum Creek, Potlatch, and Rayonier
have converted themselves into timber REITs in recent years. With improved tax
efficiency and increased concentration on timberland management, these timber
REITs are expected to be less risky.
Another interesting fact noted in this study is that, despite the current economic
downturn triggered by the subprime residential mortgage blowup, private-equity
timberland returns remain relatively strong. While the CRSP market index went
down 39 % in 2008, the NCREIF Timberland Index achieved a 9 % return, or on the
risk-adjusted basis, an excess return of 10 % (calculated using the estimated a series
in 2008). In contrast, the portfolio value of publicly traded timber firms fell 39 %
just like the market. However, it should be noted that most of those forestry firms do
have non-timberland business, such as paper and lumber mills, which may be more
sensitive to the overall economic conditions. A close examination of the three
publicly traded timber REITs reveals that they were less affected by the gloomy
market. Looking ahead, global economic crisis will last for some time, multiple
factors will affect timberland returns, and the net effect on timberland properties
has yet to be observed (Washburn 2008).
It should be noted that there have been some concerns about the data and method
consistency of the NCREIF Timberland Index. As pointed out by Binkley
et al. (1996), there are no standardized appraisal and valuation practice in forestry,
so heterogeneity may exist in the data. In addition, due to lack of quarterly
appraisals for many properties in the NCREIF Timberland Index, quarterly return
series may be less useful than the annual ones. Finally, the NCREIF Timberland
Index is a composite performance measure of a very large pool of commercial
forestland properties acquired in the private market for investment purposes.
Hence, individual investors should use cautions when interpreting the NCREIF
Timberland Index.
The multivariate time-series model can be represented by the following state space
form:
yt ¼ Z t at þ et , et NIDð0; H t Þ (34.10)
called the observation or measurement equation, and the second is called state
equation. The parameters at, Ht, and Qt in the system of equations can be estimated
jointly by the maximum likelihood method with the recursive algorithm Kalman
filter. The intention of filtering is to update the information of the system each time
a new observation yt is available, and the filtering equations are
v t ¼ y t Z t at ,
0
Ft ¼ Z t Pt Z t þ H t ,
0
K t ¼ T t Pt Z t F1
t , (34.12)
Lt ¼ T t K t Z t ,
atþ1 ¼ T t at þ K t vt ,
0 0
Ptþ1 ¼ T t Pt Lt þ Rt Qt Rt ,
For t ¼ 1, , N. The mean vector a1 and the variance matrix P1 are known for
the initial state vector a1 (Durbin and Koopman 2001; Harvey 1989).
where x jt is the amount of security j purchased at time t, pjt is the price of security j at
time t, Wt is the individual’s endowed wealth at time t, Ct is the individual’s
consumption at time t, djt is the dividend paid by security j at time t, and b is time
discount. Express Ct in terms of xjt , and differentiate the objective function with
respect to xjt , then we can get the following first-order condition:
h i h i
Et U 0 ðCt Þpjt ¼ Et bU0 ðCtþ1 Þ pjtþ1 þ d jtþ1 (34.14)
for all j. After rearranging the terms, we can reach Eq. 34.6, where
bU 0 ðCtþ1 Þ
Mt ¼
U 0 ðC t Þ
(34.15)
pj þ d j
Rtþ1 ¼ tþ1 j tþ1 1:
pt
972 B. Mei and M.L. Clutter
The NCREIF Timberland Index has two components, the income return and the
capital return. The income return is also known as EBITDDA return, which
represents earnings before interest expenses, income taxes, depreciation, depletion,
and amortization. The capital return is derived from land appreciation. The formu-
las to calculate these returns are
EBITDDAt
IRt ¼ (34.16)
MV t1 þ 0:5ðCI t PSt þ PPt EBITDDAt Þ
where IRt and CRt are the income return and capital return, respectively; EBITDDAt
equals the net operating revenue obtained from the tree farm (primarily from timber
sales); CIt equals the capitalized expenditures on the tree farm (e.g., forest regen-
eration and road construction); PSt equals the net proceeds from sales of land from
the tree farm; PPt equals the gross costs of adding land to the tree farm; and MVt
equals the market value of the tree farm (Binkley et al. 2003).
Show KFcapm2.output
sspace KFff1
’ Time-varying alpha in the Fama-French three-factor
model
KFff1.append @signal (NCREIF-RF) ¼ sv1 + c(1)*MktRf + c(2)
*SMB + c(3)*HML + [var ¼ exp(c(4))]
KFff1.append @state sv1 ¼ sv1(-1) + [var ¼ exp(c(5))]
KFff1.append @param c(1) 0.02 c(2) -0.04 c(3) -0.05 c(4) 0
c(5) 0
KFff1.ml(showopts, m ¼ 500, c ¼ 0.0001, m)
Show KFff1.output
sspace KFff2
’ Time-varying beta in the Fama-French three-factor model
KFff2.append @signal (port-RF) ¼ c(1) + sv1*MktRf + c(2)
*SMB + c(3)*HML + [var ¼ exp(c(4))]
KFff2.append @state sv1 ¼ sv1(-1) + [var ¼ exp(c(5))]
KFff2.append @param c(1) 0 c(2) 0 c(3) 3 c(4) 3 c(5) 0
KFff2.ml(showopts, m ¼ 500, c ¼ 0.0001, m)
Show KFff2.output
’ Save the workfile
wfsave Timberland_Jul2011
First, choose minimizing the standard deviation of the SDFs as the objective
function.
Second, set the mean of the SDFs equal to a predetermined value. This is
constraint No.1.
Third, for each basis asset (industry group) in the industry portfolio, add
1X T
one constraint according to Mt 1 þ Ri, t ¼ 1 . That is, add five more
T t¼1
constraints when using the five-industry portfolio, whereas add ten more constraints
34 Pricing Commercial Timberland Returns in the United States 975
when using the ten-industry portfolio. Fourth, specify the solutions to be nonneg-
ative and solve for the SDFs. Fifth, use the SDFs to price timberland returns
according to Eq. 34.9. Repeat steps 1–5 with a different value as the given mean of
the SDFs.
References
Ahn, D.-H., Conrad, J., & Dittmar, R. F. (2003). Risk adjustment and trading strategies. Review of
Financial Studies, 16, 459–485.
Binkley, C. S., Raper, C. F., & Washburn, C. L. (1996). Institutional ownership of U.S. Timberland:
History, rationale, and implications for forest management. Journal of Forestry, 94, 21–28.
Binkley, C. S., Washbrun, C. L., & Aronow, M. E. (2003). The NCREIF timberland property
index: Research notes 2003. Boston: Hancock Timber Resource Group.
Campbell, J. Y., Lo, A. W., & MacKinlay, A. C. (1997). The econometrics of financial markets.
Princeton: Princeton University Press.
Cascio, A. J., & Clutter, M. L. (2008). Risk and required return assessments of equity Timberland
investments in the United States. Forest Products Journal, 58, 61–70.
Caulfield, J. P. (1998). Timberland return drivers and investing styles for an asset that has come of
age. Real Estate Finance, 14, 65–78.
Chen, Z., & Knez, P. J. (1996). Portfolio performance measurement: Theory and applications.
The Review of Financial Studies, 9, 511–555.
Conroy, R., & Miles, M. (1989). Commercial forestland in the pension portfolio: The biological
beta. Financial Analysts Journal, 45, 46–54.
Doran, H. E., & Rambaldi, A. N. (1997). Applying linear time-varying constraints to
econometric models: With an application to demand systems. Journal of Econometrics, 79,
83–95.
Durbin, J., & Koopman, S. J. (2001). Time series analysis by state space methods. New York:
Oxford University Press.
Fama, E. F., & French, K. R. (1993). Common risk factors in the returns on stocks and bonds.
Journal of Financial Economics, 33, 3–56.
Fortson, J. C. (1986). Factors affecting the discount rate for forestry investments. Forest Products
Journal, 36, 67–72.
Haight, R. G., & Holmes, T. P. (1991). Stochastic price models and optimal tree cutting: Results
for Loblolly pine. Natural Resource Modeling, 5, 424–443.
Hansen, L. P., & Jagannathan, R. (1991). Implications of security market data for models of
dynamic economies. Journal of Political Economy, 99, 225–262.
Harris, T., Baldwin, S., & Siry, J. (2010). United States Timberland markets 2000 to mid-2010:
Transactions, values & market research. Athens: Timber Mart-South.
Harvey, A. C. (1989). Forecasting, structural time series models and the Kalman filter.
Cambridge, UK: Cambridge University Press.
Heikkinen, V. P. (2002). Co-integration of timber and financial markets-implications for portfolio
selection. Forest Science, 48, 118–128.
Jensen, M. C. (1968). The performance of mutual funds in the period 1945–1964. The Journal of
Finance, 23, 389–416.
Kenneth, R. French’s website. http://mba.tuck.dartmouth.edu/pages/faculty/ken.french/data_library.
html.
King, B. F. (1966). Market and industry factors in stock price behavior. The Journal of Business,
39, 139–190.
Liao, X., Zhang, Y., & Sun, C. (2009). Investments in timberland and softwood timber as parts of
portfolio selection in the United States: A cointegration analysis and capital asset pricing
model. Forest Science, 55, 471–479.
976 B. Mei and M.L. Clutter
Lintner, J. (1965). The valuation of risk assets and the selection of risky investments in stock
portfolios and capital budgets. The Review of Economics and Statistics, 47, 13–37.
Lucas, R. E., Jr. (1978). Asset prices in an exchange economy. Econometrica, 46, 1429–1445.
Malaty, R., Toppinen, A., & Viitanen, J. (2007). Modeling and forecasting Finnish pine sawlog
stumpage prices using alternative time-series methods. Canadian Journal of Forest Research,
37, 178–187.
Markowitz, H. (1952). Portfolio selection. Journal of Finance, 7, 77–91.
Redmond, C. H., & Cubbage, F. W. (1988). Portfolio risk and returns from timber asset invest-
ments. Land Economics, 64, 325–337.
Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk.
Journal of Finance, 19, 425–442.
Song, H., Liu, X., & Romilly, P. (1996). A time varying parameter approach to the Chinese
aggregate consumption function. Economics of Planning, 29, 185–203.
Sun, C. (2007). Variation of federal cost-share programs in the United States and the inducement
effects on tree planting. Journal of Forest Economics, 12, 279–296.
Sun, C., & Zhang, D. (2001). Assessing the financial performance of forestry-related investment
vehicles: Capital Asset Pricing Model vs. Arbitrage Pricing Theory. American Journal of
Agricultural Economics, 83, 617–628.
Washburn, C. (2008). Forestry investments in a slowing global economy. London/Athens: Inter-
national Forest Investment/Center for Forest Business.
Washburn, C. L., & Binkley, C. S. (1990). On the use of period-average stumpage prices to
estimate forest asset pricing models. Land Economics, 66, 379–393.
Washburn, C. L., & Binkley, C. S. (1993). Do forest assets hedge inflation? Land Economics, 69,
215–224.
Zinkhan, F. C. (1988). Forestry projects, modern portfolio theory, and discount rate selection.
Southern Journal of Applied Forestry, 12, 132–135.
Zinkhan, F. C. (2008). History of modern timberland investments. In International forest invest-
ment: An executive education conference (Munich, Germany), Center for Forest Business,
Athens.
Zinkhan, F. C., & Cubbage, F. W. (2003). Financial analysis of timber investments. In E. O. Sills
& K. L. Abt (Eds.), Forests in a market economy. Dordrecht: Kluwer.
Optimal Orthogonal Portfolios with
Conditioning Information 35
Wayne E. Ferson and Andrew F. Siegel
Contents
35.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 978
35.2 Optimal Orthogonal Portfolios: The Classical Case . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 979
35.3 The Conditional Setting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 982
35.4 The Main Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 983
35.5 Empirical Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 987
35.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 992
Appendix 1: Theorems and Proofs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 993
Efficient Portfolio Solutions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 993
Proofs of Propositions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 997
Proof of Propositions 2 and 3 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 997
Proof of Proposition 4 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 998
Appendix 2: Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 998
The Parametric Bootstrap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 999
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1001
Abstract
Optimal orthogonal portfolios are a central feature of tests of asset pricing models
and are important in active portfolio management problems. The portfolios com-
bine with a benchmark portfolio to form ex ante mean variance efficient portfolios.
This paper derives and characterizes optimal orthogonal portfolios in the presence
of conditioning information in the form of a set of lagged instruments. In this
setting, studied by Hansen and Richard (1987), the conditioning information is
used to optimize with respect to the unconditional moments. We present an
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 977
DOI 10.1007/978-1-4614-7750-1_35,
# Springer Science+Business Media New York 2015
978 W.E. Ferson and A.F. Siegel
Keywords
Asset pricing tests • Conditioning information • Minimum-variance efficiency •
Optimal portfolios • Predicting returns • Portfolio management • Stochastic
discount factors • Generalized method of moments • Maximum likelihood •
Parametric bootstrap • Sharpe ratios
35.1 Introduction
The optimal orthogonal portfolio, also known as the most mispriced portfolio or the
active portfolio, is a central concept in asset pricing tests and in modern portfolio
management. In asset pricing problems, it represents the difference between the
performance of a benchmark portfolio and the maximum potential performance in
a sample of assets (Jobson and Korkie 1982). In modern portfolio management, it
shows how to actively tilt away from a given benchmark portfolio to achieve
portfolio efficiency (Gibbons et al. 1989; Grinold and Kahn 1992).
Optimal orthogonal portfolios are studied by Roll (1980), MacKinlay (1995),
Campbell et al. (1987), and others. However, these studies restrict the analysis to
a setting where the portfolio weights are fixed over time. In contrast, studies in asset
pricing use predetermined variables to model conditional expected returns, corre-
lations, and volatility. Portfolio weights may be functions of the predetermined
variables, and they will generally vary over time. Quantitative portfolio managers
routinely use conditioning information in optimized portfolio strategies. Therefore,
it is important to understand optimal orthogonal portfolios in a conditional setting.
This paper derives, characterizes, and illustrates optimal orthogonal portfolios in
a conditional setting. The setting is one where the conditional means and variances
of returns are time varying and optimal time-varying portfolio weights achieve
unconditional mean variance efficiency with respect to the information, as
described by Hansen and Richard (1987) and Ferson and Siegel (2001).1 Ferson
1
An alternative is to study conditional efficiency, where the weights minimize the conditional
variance. This may be handled by simply reinterpreting the classical analysis.
35 Optimal Orthogonal Portfolios with Conditioning Information 979
and Siegel (2001) argue that this setting is interesting from the perspective of active
portfolio management, where the client cannot observe the information that
a portfolio manager may have. Ferson and Siegel (2003, 2009) show that this
setting is also interesting from the perspective of testing asset pricing models, but
they do not develop the optimal orthogonal portfolio.
We show that the optimal orthogonal portfolio has time-varying weights, and
we derive the weights in a closed form. The portfolio weights for unconditionally
efficient portfolios in the presence of conditioning information are derived by
Ferson and Siegel (2001). They consider the case with no risk-free asset and the
case with a fixed risk-free asset whose return is constant over time. We generalize
these solutions to the case with a “conditionally” risk-free asset whose return is
known at the beginning of the period and is thus included in the lagged condi-
tioning information and may vary over time. We derive solutions for the optimal
orthogonal portfolios with conditioning information, including cases where there
is no risk-free asset, a constant risk-free rate, or a time-varying conditionally risk-
free rate. We show that a “law of conservation of squared Sharpe ratios” holds,
implying that the optimal orthogonal portfolio’s squared unconditional Sharpe
ratio is the difference between that of the benchmark portfolio and the maximum
unconditional squared Sharpe ratio that is possible using the assets and condi-
tioning information. Empirical examples illustrate the performance of the optimal
orthogonal portfolios with conditioning information and the behavior of the
portfolio weights.
Section 35.2 briefly reviews optimal orthogonal portfolios in the classical case
with no conditioning information. Section 35.3 describes the setting for our analysis
with conditioning information. Section 35.4 presents the main theoretical results,
and Sect. 35.5 presents our empirical examples. Section 35.6 concludes. Appendix 1
includes the proofs of the main results, and Appendix 2 describes the methodology
of our empirical examples in detail, including a general description of the
parametric bootstrap.
In the classical case, portfolio weights are fixed constants over time, and there is no
conditioning information. Optimal orthogonal portfolios are tied to mean variance
efficiency. Mean variance efficient portfolios maximize the expected return, given
the variance of the return. Since Markowitz (1952) and Sharpe (1964), such
portfolios have been at the core of financial economics.
The mean variance efficiency of a given portfolio can be described using a system
of time-series regressions. If rt ¼ Rt g0 is the vector of N excess returns at time t,
measured in excess of a given risk-free or zero-beta return, g0, and rpt ¼ Rpt g0 is
the excess return on a benchmark portfolio, the regression system is
r t ¼ a þ br pt þ ut ; t ¼ 1, . . . , T, (35.1)
980 W.E. Ferson and A.F. Siegel
0 2
xa
xc ¼ Arg Max : (35.2)
x Varðx0 r Þ
It is clear from the definition in Eq. 35.2 why the portfolio is referred to as the
most mispriced. The vector a captures the “mispricing” of the tested asset returns in
Eq. 35.1 when evaluated using the benchmark rp, and the portfolio xc has the largest
squared alpha relative to its variance. This interpretation also reveals why the
portfolio is of central interest in active portfolio management. Given a benchmark
portfolio rp, an active portfolio manager places bets by deviating from the portfolio
weights that define the benchmark. The manager is rewarded for bets that deliver
higher returns and penalized for increasing the volatility. The portfolio in
(2) describes the active bets that achieve the largest amount of extra return for the
variance. Thus, the solution is also referred to as the active portfolio by Gibbons
et al. (1989). (See Grinold and Kahn (1992) for an in-depth treatment of modern
portfolio management.)
In the classical case of fixed portfolio weights, the solution to Eq. 35.2 is given
0 1
by xc ¼ 1 V 1 a V 1 a , where 1 is an N-vector of ones and V ¼ Cov(r), the
covariance matrix of the returns. Using this solution, several well-known properties
2
Note the distinction between minimum variance efficient portfolios, which minimize the variance
for the given mean return, and mean variance efficient, which maximize the mean return given its
variance. The latter set of portfolios is a subset of the former, typically depicted as the positively
sloped portion of the minimum variance efficient boundary when graphed with mean return on the
y-axis and standard deviation or variance of return on the x-axis. The portfolio rp is mean variance
efficient when a ¼ 0 and E(rp) exceeds the expected excess return of the global minimum variance
portfolio.
35 Optimal Orthogonal Portfolios with Conditioning Information 981
!
0 1
^S 2 ðRÞ ^S 2 Rp
W ¼ T^a ½Covð^a Þ ^a ¼ T 2
_ 2 ðN Þ
w (35.3)
^
1 þ S Rp
( 0 2 )
E xr
S ðRÞ ¼ max
2
: (35.4)
x Varðx0 r Þ
3
See Roll (1980), Gibbons et al. (1989), MacKinlay (1995), and Campbell et al. (1987) for
analyses of optimal orthogonal portfolios in the classical case with no conditioning information.
982 W.E. Ferson and A.F. Siegel
h 0 i h 0 i
Var Rp, tþ1 Var x ðZ t ÞRtþ1 if E Rp, tþ1 ¼ E x ðZ t ÞRtþ1 (35.5)
h 0 i h 0 i
E x ðZt ÞRtþ1 ¼ g0 þ g1 Cov x ðZ t ÞRtþ1 , Rp, tþ1 : (35.6)
Equation 35.5 is the definition of efficiency with respect to Z. It states that Rp, t+1
is on the minimum-variance boundary formed by all possible portfolios that use the
assets in R and the conditioning information. Equation 35.6 states that the familiar
expected return-covariance relation from Fama (1973) and Roll (1977) must hold
with respect to the efficient portfolio. In Eq. 35.6, the coefficients g0 and g1 are fixed
scalars that do not depend on the functions x() or the realizations of Zt.
We consider portfolios formed from the risky assets using weights xq ¼ xq(Z)
where the weights must sum to 1 (for all Z) when there is no risk-free asset. This
constraint is relaxed when there is a conditional or unconditional risk-free asset
(where the implicit weight in the risk-free asset is then set at 1 minus the sum of the
weights in the risky assets). When there is no risk-free asset, the portfolio return
0
is Rq,t+1 ¼ xq(Zt)Rt+1, and when there is a risk-free asset, the portfolio return is
0
Rq, tþ1 ¼ Rf ðZ t Þ þ xq ðZt Þ Rtþ1 Rf ðZt Þ1 whether or not Rf (Z) is constant. In either
case, we denote the (unconditional) portfolio mean mq ¼ E(Rq) and variance
s2q ¼ Var(Rq). When there exists a conditional time-varying risk-free asset, the
portfolio takes advantage of the ability to adapt both the percentage invested in
risky assets and their portfolio weights in response to the information Z. This may
be interpreted as “market timing” and security selection, respectively. We define
the optimal orthogonal portfolio with respect to a given benchmark portfolio
P formed from the risky assets using (possibly) time-varying weights xp ¼ xp(Z).
Definition The most mispriced (or optimal orthogonal) portfolio, Rc, with respect
to the benchmark portfolio Rp and conditioning information Z, with portfolio
weight denoted xc(Z), uses the conditioning information to maximize a2c /s2c where
the unconditional variance of Rc is s2c , the unconditional mean is mc ¼ E(Rc), the
unconditional alpha of Rc with respect to Rp is ac ¼ mc [g0 + (mp g0)scp/s2p], the
zero-beta rate is g0, and the unconditional covariance is scp ¼ Cov(Rc, Rp).
Proposition 2 The unique most mispriced (or optimal orthogonal) portfolio Rc with
respect to a given benchmark portfolio Rp (with weights xp and expected return mp),
conditioning information Z and given zero-beta rate, g0, has the following portfolio
weight in each of the following cases. If there is no risk-free rate, then the weights
conditionally sum to 1 as defined by
( 0
! )
L1 L1 1 L
xc ðZ Þ ¼ A 0 þ ½ðc þ 1Þms þ b L 0 mðZÞ þ Bxp , (35.7)
1 L1 1 L1
Rf (Z) (which may be either constant or time varying, but if constant,4 then we must
choose g0 ¼ Rf along with any ms 6¼ Rf), then the solution is
xc ðZÞ ¼ A ðc þ 1Þms þ b Rf Q mðZ Þ Rf 1 þ Bxp , (35.8)
where
ðms g0 Þ=s2s
A¼ , (35.9)
ðms g0 Þ=s2s mp g0 =s2p
4
Equation 35.11 cannot be used to determine ms when Rf (Z) is almost surely constant due to
division by zero, and, in this case, every choice ms 6¼ Rf uses the same (rescaled) portfolio of risky
assets Q mðZ Þ Rf 1 in the formation of an efficient portfolio xs(Z).
35 Optimal Orthogonal Portfolios with Conditioning Information 985
mp g0 =s2p
B¼ , (35.10)
ðms g0 Þ=s2s mp g0 =s2p
1
0
Q ¼ Q ðZ Þ E R Rf 1 R Rf 1 Z
1 (35.13)
0
¼ mðZÞ Rf 1 mðZÞ Rf 1 þ S
= e ðZ Þ ,
n h 0 io1 h i1
L ¼ LðZ Þ E RR Z
0
¼ mðZÞm ðZÞ þ S
= e ðZ Þ : (35.14)
h i
ðms g0 Þ=s2s Rs mp g0 =s2p Rp
Rc ¼ h i (35.15)
ðms g0 Þ=s2s mp g0 =s2p
Or
s2p Rs sps Rp
Rc ¼ ¼ ARs þ BRp with A þ B ¼ 1, (35.16)
s2p sps
where we assume that Rs and Rp are not perfectly correlated and that s2p 6¼ ssp.
Proof: See Appendix 1.
Propositions 2 and 3 extend the concept of the active or optimal orthogonal
portfolio to the setting of efficiency with respect to conditioning information. Given
a benchmark portfolio Rp, its optimal orthogonal portfolio with respect to Z shows
how to tilt away from the benchmark weights to obtain efficiency with respect to Z.
The portfolio Rc has weights that depend on Z. Thus, the optimal tilt away from
a benchmark uses the manager’s information Z in a dynamic way.
Equation 35.16 shows how the optimal orthogonal portfolio Rc can be formed by
combining an efficient-with-respect-to-Z portfolio Rs with Rp. The portfolio Rc is
the regression error of Rs projected on Rp, normalized so that the weights sum to
1, as can be seen by solving Eq. 35.16 for Rs. Thus, the portfolio Rc is uncorrelated
with Rp.
We defined the most mispriced portfolio with conditioning information as
maximizing the squared alpha relative to the unconditional variance of Rc. Since
Rc is orthogonal to Rp, its residual variance in regression (Eq. 35.1) is the same as its
total variance. Thus, we can think of the optimal orthogonal portfolio as maximiz-
ing alpha given its residual variance among all orthogonal portfolios.
Given a benchmark portfolio with return Rp, the optimal orthogonal portfolio
with conditioning information is useful for active portfolio management. It might
seem natural for a manager with information Z to simply reinterpret the classical
analysis, where all the moments are the conditional moments given Z. This, in fact,
is the interpretation that much of the literature on active portfolio management has
used (e.g., Jorion 2003; Roll 1982). This approach produces conditionally mean
variance efficient portfolios given Z. However, as shown by Dybvig and Ross
(1985), a conditionally efficient portfolio is likely to be seen as inefficient from
the (unconditional) perspective of a client without access to the information Z.
The optimal orthogonal portfolio describes the active portfolio bets that deliver
optimal performance from the client’s perspective.
Let S2(R) be the maximum squared Sharpe ratio obtained by the efficient-with-
respect-to-Z portfolio and let Rc be the optimal orthogonal portfolio with respect to
Rp and information Z.
Proposition 4 Law of conservation of squared Sharpe ratios. For a given zero-beta
or risk-free rate, if S2s ¼ S2(R) is the maximum squared Sharpe ratio obtained by
35 Optimal Orthogonal Portfolios with Conditioning Information 987
portfolios x(Z) and Rc is the optimal orthogonal portfolio with respect to Rp and
information Z at that zero-beta rate, then S2s ¼ S2p + S2c , where S2i denotes the squared
Sharpe ratio of portfolio i.
Proof: See Appendix 1.
Proposition 4 shows that if we test the hypothesis that a portfolio is efficient with
respect to Z using versions of the test statistic in Eq. 35.3, as developed in Ferson
and Siegel (2009), then the role of the optimal orthogonal portfolio with informa-
tion Z is analogous to the role of the optimal orthogonal portfolio in the case of the
classical test statistics. The Sharpe ratio of the optimal orthogonal portfolio with
conditioning information indicates how far the tested benchmark portfolio is from
the efficient-with-respect-to-Z frontier. The squared Sharpe ratio of the optimal
orthogonal portfolio is the numerator of the test statistic. This numerator and thus
the test statistic is zero only if the tested portfolio is efficient in the sample, and it
grows larger as the tested portfolio is further from efficiency.
In the second case, no risk-free asset exists. Here, we again set g0 ¼ 3.8 % to pick
a point on the mean variance boundary, and we do not allow the portfolio to take
a position in a risk-free asset. We do allow the lagged Treasury return as condi-
tioning information in Z, which highlights the information in lagged Treasury
returns about the future risky asset returns. In the third case, there is
a conditionally risk-free return that is contained in Z. Here, we use the lagged
Treasury return in the conditioning information, and we allow the portfolio to trade
the subsequent Treasury return in addition to the other risky assets. The subsequent
Treasury return is not really known ex ante as the formula assumes, but its
correlation with the lagged return is 0.92 during our sample, and this allows us to
implement the time-varying risk-free rate example in a somewhat realistic way.
(In reality, there is no ex ante risk-free asset given the importance of inflation risk.)
In practical terms, this example highlights the effects of “market timing,” or
varying the amount of “cash” in the portfolio, in addition to varying the allocation
among the risky assets.
Table 35.1 summarizes the results. The rows show results for the benchmark
index (Market), three equity portfolios, and the government bond return. The
CAPM a refers to the intercept in the regression (1), of the portfolio returns in
excess of the Treasury bill returns, on the excess return of the market index. The
small stock portfolio has the largest alpha, at 3.95 % per year, while the growth
stock portfolio has a negative alpha. The symbol su refers to the standard deviation
of the regression residuals. The small stock portfolio has the largest su or
nonmarket risk, at more than 25 % per year.
The bottom four rows of Panel A summarize the optimal orthogonal portfo-
lios when the market index is the benchmark. The fixed-weight portfolio Rc uses
no conditioning information. Its alpha is larger than any of the separate assets, at
4.66 % per year, and its residual standard deviation is also relatively large, at
16.5 % per year. Since the portfolio is orthogonal to the market index, its
residual standard deviation is the same as its total standard deviation, or vola-
tility of return. The ratio of the alpha to the residual volatility is known as the
appraisal ratio (Treynor and Black 1973) or the information ratio (Grinold and
Kahn 1992). Optimal orthogonal portfolios try to maximize the square of this
ratio. The fixed-weight portfolio Rc delivers an information ratio of 0.282,
substantially larger than those of the small stock or value portfolios, at 0.155
and 0.177, respectively, and also larger than the bond portfolio, which has an
information ratio of 0.183 by virtue of its relatively small volatility.
Table 35.1 summarizes performance statistics for the optimal orthogonal port-
folios with conditioning information. There are three versions with (1) a fixed risk-
free rate, (2) no risk-free rate, and (3) a time-varying conditional risk-free rate. The
information ratios in all three cases are larger than those of the individual assets or
the fixed-weight active portfolio, which illustrates the potential value of using
conditioning information explicitly in a portfolio management context (see also,
Chiang 2009). The improvement over fixed weights is modest for the case with no
risk-free asset. However, the information ratio is about three times as large, at
0.955, in the example with a time-varying risk-free rate. This illustrates the
35
Xc(Z) portfolios:
Fixed Rf 0.189 0.115 0.304
No Rf 0.189 0.088 0.277
Varying Rf 0.182 0.909 1.191
Annual returns on portfolios of common stocks and long-term US government bonds cover the 1931–2007 period. The CAPM a refers to the intercept in the
regression Eq. 35.1, of the portfolio return in excess of a zero-beta parameter, on that of the broad value-weighted stock market index (market index). The
symbol su refers to the standard deviation of the regression residuals. The fixed portfolio, Rc, ignores the conditioning information. The alphas and residual
standard deviations are annual percentage units. The information ratio is the ratio a/ su
989
990 W.E. Ferson and A.F. Siegel
potential usefulness of interest rate information and the ability to hold “cash” as
a function of market conditions, in a portfolio management context.
The averages over time of the optimal orthogonal portfolios’ weights on the
risky assets are shown in the right-hand columns of Panel A. The weights are
normalized to sum to 1.0. The weights xc of the optimal orthogonal portfolios
combine with the benchmark (whose weights, xp, are 100 % in the market index)
to determine an efficient portfolio. The overall efficient portfolio weights xs
therefore vary over time and depend on how xc and xp are combined, which is
determined by the coefficients A and B in Eqs. 35.9 and 35.10. The estimated
weights are as follows. In the fixed risk-free rate case, xs ¼ 0.60 xc + 0.40 xp. In the
no risk-free rate case, xs ¼ 0.23 xc + 0.77 xp. In the time-varying risk-free rate
case, xs ¼ 0.44 xc + 0.56 xp. Thus, the efficient portfolio is formed as a convex
combination of the benchmark and the active portfolio, with reasonable weights
in each case.
The four right-hand columns of Panel A show that two of the active portfolios
take short positions in the market benchmark, indicating an optimal tilt away from
the market index. The fixed-weight portfolio Rc takes an extreme short
position of 126 %, while on average the portfolio using the conditioning
information but no risk-free asset takes a position of 21 % in the market index.
These short positions finance large long positions in the US government bond and
also long positions (in most cases) in small stocks, value stocks, and growth stocks.
It is interesting that all the portfolios tilt positively, although by small amounts, into
growth stocks even though growth stocks have negative CAPM alphas. This occurs
because of the correlations among the asset classes.
All the versions of the optimal orthogonal portfolio suggest strong tilts into
government bonds. The bond tilt is the most extreme for the fixed-weight solution,
at 141 %, and is relatively modest, at 43.2 %, for the portfolio assuming a time-
varying conditional risk-free rate. This makes sense, as that portfolio can hold
short-term Treasuries in addition to government bonds. The large weights in bonds
reflect various features of the data and the value of the zero-beta rate. With larger
values for the zero-beta rate, provided that the rate remains below the mean of the
global minimum-variance portfolio of risky assets, the optimal orthogonal portfo-
lios become more aggressive as the target expected return of the efficient-with-
respect-to-Z portfolio increases.
The squared Sharpe ratios in Panel B of Table 35.1 indicate how far the stock
market index is from efficiency. The squared Sharpe ratio for the market is 0.189,
measured relative to the fixed zero-beta rate of 3.8 %. The market portfolio’s
squared Sharpe ratio is slightly smaller, at 0.182, for returns measured in excess
of a time-varying risk-free rate. This is because the negative covariance between the
risk-free rate and stock returns increases the variance of the excess return. For the
fixed-weight orthogonal portfolio Rc, the squared Sharpe ratio is 0.079, and for
the portfolios using conditioning information, it varies between 0.088 and 0.909.
According to the law of conservation of squared Sharpe ratios in Proposition 4,
the sum of the index and optimal orthogonal portfolios’ squared Sharpe ratios is
the squared slope of the tangency from the given zero-beta rate to the relevant
35 Optimal Orthogonal Portfolios with Conditioning Information 991
Portfolio Weight
Growth Small
0.5
Value
0.0
−0.5
SP500
−1.0
1930 1950 1970 1990 2010
Year
1.0
Bonds
0.8
0.6
Portfolio Weight
0.4 Small
0.2
0.0
−0.2
Growth Value SP500
−0.4
−0.6
1930 1950 1970 1990 2010
Year
Fig. 35.2 Optimal orthogonal portfolio weights: the time-varying risk-free rate case
35.6 Conclusions
R ¼ mðZ Þ þ e:
The noise term e is assumed to have conditional mean zero given Z and
nonsingular conditional covariance matrix S = e ðZ Þ. The conditional expected return
vector is m(Z) ¼ E(R|Z). Let the 1 N row vector x0 (Z) ¼ (x1(Z), . . . , xN(Z)) denote
the portfolio share invested in each of the N risky assets, investing (or borrowing) at
0
the risk-free rate the amount 1 x ðZÞ1, where 1 (1 ,. . .,1)0 denotes the column
vector of ones. We allow for a conditional risk-free asset returning Rf ¼ Rf (Z).
994 W.E. Ferson and A.F. Siegel
0
The return on the portfolio is Rs ¼ Rf þ x ðZÞ R Rf 1 , with unconditional
expectation and variance as follows:
n 0 o
ms ¼ E Rf þ E x ðZ Þ mðZ Þ Rf 1 ,
s2s ¼ E R2s m2s ¼ E E R2s jZÞ m2s ,
h 0 i n 0 o
s2s ¼ E R2f þ E x ðZÞQ1 xðZÞ þ 2E Rf x ðZÞ mðZÞ Rf 1 m2s , (35.17)
and
n 0 o
c E R2f mðZÞ Rf 1 Q mðZÞ Rf 1 :
Theorem 1 Given a target unconditional expected return ms, N risky assets, instru-
ments Z, and a conditional risk-free asset with rate Rf ¼ Rf (Z) that may vary over
time, the unique portfolio having minimum unconditional variance is determined
by the weights
m s E Rf þ ’
xs ðZ Þ ¼ Rf Q mðZ Þ Rf 1
z (35.18)
¼ ðc þ 1Þms þ b Rf Q mðZ Þ Rf 1 ,
ms Rf
xs ðZ Þ ¼ Q mðZ Þ Rf 1
z
2
and with optimal portfolio variance s2s ¼ 1z
z ms Rf .
Proof Our objective is to minimize, over the choice of
xs(Z), the
portfolio variance
0
Var(Rs) subject to E(Rs) ¼ ms, where Rs ¼ Rf þ x ðZ Þ R Rf 1 and the variance is
given by Eq. 35.17. We form the Lagrangian:
0 0
L½xðZÞ ¼ E x ðZ ÞQ1 xðZ Þ þ 2E Rf x ðZÞ mðZÞ Rf 1
0
þ2lE ms Rf x ðZ Þ mðZ Þ Rf 1
and proceed using a perturbation argument. Let q(Z) ¼ x(Z) + dy(Z), where x(Z) is
the conjectured optimal solution, y(Z) is any regular function of Z, and d is a scalar.
Optimality of x(Z) follows when the partial derivative of L[q(Z)] with respect to d is
identically zero when evaluated at d ¼ 0. Thus,
0
0 ¼ E y ðZÞ Q1 xðZÞ þ Rf l mðZÞ Rf 1
for all functions y(Z), which implies that Q1 xðZ Þ þ Rf l mðZ Þ Rf 1 ¼ 0
almost surely in Z. Solve this expression for x(Z) to obtain Eq. 35.18, where the
Lagrange multiplier l is evaluated by solving for the target mean, ms. The expression
for the optimal portfolio variance follows by substituting the optimal weight function
into Eq. 35.17. Formulas for fixed Rf then follow directly. QED.
When the risk-free asset’s return is time varying and contained in the informa-
tion set Z at the beginning of the portfolio formation period, the conditional mean
variance efficient boundary varies over time with the value of Rf (Z) along with the
conditional asset means and covariances . In this case, a zero-beta parameter, g0,
may be chosen to fix a point on the unconditionally efficient-with-respect-to-
Z boundary. The choice of the zero-beta parameter corresponds to the choice of
a target unconditional expected return ms. For a given value of g0, the target mean
maximizes the squared Sharpe ratio (ms g0)2/s2s along the mean variance bound-
ary, which implies ms ¼ (a + bg0)/(b + cg0).
When there is a risk-free asset that is constant over time, the unconditionally
efficient-with-respect-to-Z boundary is linear (a degenerate hyperbola) and reaches
the risk-free asset at zero risk. In this case, we use g0 ¼ Rf and can obtain any ms
larger or smaller than Rf, levering the efficient portfolio up or down with positions
in the risk-free asset.
When there is no risk-free asset, we define portfolio s by letting x0 ¼ x0 (Z) ¼ [x1(Z),
. . ., xN(Z)] denote the shares invested in each of the N risky assets, with the
constraint that the weights sum to 1.0 almost surely in Z. The return on this
portfolio, Rs ¼ x0 (Z)R, has expectation and variance as follows:
h 0 i
ms ¼ E x ðZ ÞmðZ Þ ,
n 0 o
s2s ¼ E x ðZÞL1 xðZÞ m2s ,
996 W.E. Ferson and A.F. Siegel
n h 0 io1 h 1
L ¼ LðZÞ E RR Z
0
¼ mðZ Þm ðZÞ þ S
= e ðZÞ :
0
!
1 LmðZÞ
d2 ¼ E ,
10 L1
and
" 0
! #
0 L1 1 L
d3 ¼ E m ð Z Þ L 0 mðZÞ :
1 L1
Theorem 2 (Ferson and Siegel 2001, Theorem 3) Given N risky assets and no risk-
free asset, the unique portfolio having minimum unconditional variance and uncon-
ditional expected return ms is determined by the weights:
0 0
!
0 1L m d2 0 L1 1 L
xs ðZ Þ ¼ 0 þ s m ðZ Þ L 0
1 L1 d 3 1 L1
0 0
!
1L 0 L1 1 L
¼ 0 þ ½ðc þ 1Þms þ bm ðZ Þ L 0 , (35.19)
1 L1 1 L1
Proofs of Propositions
We begin with Proposition 3. To see that Eqs. 35.15 and 35.16 are equivalent, use
s
the fact that efficiency of Rs implies that mp ¼ g0 þ sps2 ðms g0 Þ and thus mssg
2
0
s s
m g
¼ psps 0 . Next, given any portfolio Rq 6¼ Rc, we will show that a2q/s2q < a2c /s2c .
Beginning with Eq. 35.16, we compute:
and
s2p s2p s2s s2ps
s2c ¼ 2 :
s2p sps
s
The efficiency of Rs implies that mc ¼ g0 þ sscs2 ðms g0 Þ, mp ¼ g0 þ sps2 ðms g0 Þ,
s s s
and mq ¼ g0 þ sq2s ðms g0 Þ. Substituting these expressions and using the fact that
s
scp ¼ 0, which follows from (35.16), we compute:
2 2
a2c a2q mc go þ mp go scp =sp 2 mq go þ mp go sqp =sp 2
¼
s2c s2q s2c s2q
0 2 1
s 2 sqs s2p sps sqp
B C
¼ ðms g0 Þ2 @ 4cs2 A
ss sc s4s s4p s2q
0 1
s2p s2s s2ps s2qs s4p þ s2ps s2qp 2sqs s2p sps sqp
¼ ðms g0 Þ 2 @ A
s4s s2p s4s s4p s2q
ðms g0 Þ2 4 2 2
¼ sp ss sq s2p s2q s2ps s2qs s4p s2ps s2qp þ 2s2p sqs sps sqp :
ss sp sq
4 4 2
We now use the fact that s2sp < s2s s2p (because, by assumption, Rp and Rs are not
perfectly correlated) to see that
998 W.E. Ferson and A.F. Siegel
ðm g Þ2
¼ s2 2 04 s2p s2q s2s s2q s2sp s2p s2qs s2qp s2s þ 2sqs sqp sps 0
sp sq ss
where the final inequality follows from recognizing that the variance-covariance
terms in parentheses are equal to the determinant of the (necessarily nonnegative
definite) covariance matrix of (Rp,Rq,Rs). This establishes the maximal property of Rc.
To show uniqueness, note further that the inequality will be strict (and we will have
a2c/s2c a2q/s2q > 0) unless we have both of the following conditions corresponding to
the two inequalities in the final calculation: (1) sqp ¼ 0 so that Rq and Rp are
orthogonal, and (2) the covariance matrix of (Rp,Rq,Rs) is singular so that Rq is
a linear combination of Rp and Rs. However, there is only one portfolio orthogonal
to Rp that can be formed as a linear combination lRp + (1 l)Rs, and this solution
is Rc. This establishes Proposition 3, which holds in the case of both Theorem 1 and
Theorem 2, that is, whether or not there is a conditionally risk-free asset.
The expressions in Proposition 2 for the optimal weights, xc(Z), follow from
substituting portfolio weights from Theorem 1 (if there exists a conditional risk-free
asset that may be time varying) or Theorem 2 (otherwise) into Eq. 35.15 and noting
that the constants A and B represent the combining portfolio weights implied by
(35.15) as Rc ¼ ARs + BRp. Substituting, we see that Eq. 35.15 implies that the
portfolio weight function Axs(Z) + Bxp(Z) generates returns Rc, completing the
proof of Proposition 2. QED.
Proof of Proposition 4
2 !
mp g0 ðmc g0 Þ2 ðms g0 Þ2 s2ps s2cs
S2p þ S2c ¼ þ ¼ þ 2 :
s2p s2c s4s s2p sc
Next, substituting for scs and s2c from the above expressions, we verify that this
expression reduces to S2s . QED.
Appendix 2: Methodology
are the optimal generalized method of moments (GMM, see Hansen 1982)
estimators. The covariance matrix of the residuals is used as the estimate of
S
= e ðZÞ , which is assumed to be constant. These are the maximum likelihood
estimates (MLE) under joint normality of (R,Z). In general, the conditional
covariance matrix of the returns given Z will be time varying as a function
of Z, as in conditional heteroskedasticity. Ferson and Siegel (2003) model
conditional heteroskedasticity in alternative ways and find using parametric
bootstrap simulations that this increases the tendency of the efficient-with-
respect-to-Z portfolio weights to behave conservatively in the face of extreme
realizations of Z.
The optimal orthogonal portfolio weights in Table 35.1 and Fig. 35.1 are
estimated from Eqs. 35.7 to 35.8 in the text where, in the time-varying risk-free
rate case, the Treasury bill return is assumed to be conditionally risk-free in
Eq. 35.8. The benchmark portfolio xp is a vector with a 1.0 in the place of the
market index and zeros elsewhere. The matrix Q is estimated by using the MLE
estimates of m(Z) and S= e ðZ Þ in the function given by Eq. 35.13. The parameter mp is
estimated as the sample excess return on the market index, and g0 is the sample
mean of the Treasury return, 3.8 %.
is obviously worse if the original sample has fewer observations. If the data are
contaminated with measurement errors, in contrast, the extent of the true variability
can be overstated. Even with large sample sizes, the bootstrap can be unreliable. For
example, if the true distribution has infinite variance, the bootstrap distribution for
the sample mean is inconsistent (Athreya 1987).
With a parametric bootstrap, we can sometimes do better than with
a nonparametric bootstrap, where “do better” means, for example, obtain more
accurate confidence intervals (e.g., Andrews et al. 2006). The idea of the parametric
bootstrap is to use some of the parametric structure of the data. This might be as
simple as assuming the form of the probability distribution. For example, assuming
that the data are independent and normally distributed, we can generate artificial
samples from a normal distribution using the sample mean and variance as the
parameters. This is not exactly the right thing to do, because we should be sampling
from a population with the true parameter values, not their estimated values. But, if
the estimates of the mean and variance are good enough, we should be able to
obtain reliable inferences.
To illustrate and further suggest the flexibility of the parametric bootstrap,
consider an example, similar to the setting in our paper, where we have
a regression of stock returns on a vector of lagged instruments, Z, which are highly
persistent over time. Obviously, sampling from the Z’s randomly with replacement
would destroy their strong time-series dependence. Time-series dependence can be
accommodated in a nonparametric way by using a block bootstrap. Here, we sample
randomly a block of consecutive observations, where the block length is set to
capture the extent of memory in the data.
In order to capture the time-series dependence of the lagged Z in a parametric
bootstrap, we can model the lagged instruments as vector AR(1), for example,
retaining the estimator of the AR(1) coefficient and the model residual, which we
call the shocks, Uz. Regressing the future stock returns on the lagged instruments,
we retain the regression coefficient and the residuals, which we call the shocks, Ur.
We generate a sample of artificial data, with the same length as the original sample,
as follows. We concatenate the shocks as v ¼ (Uz,Ur). Resampling rows from v,
randomly with replacement, retains the covariances between the stock return
shocks and the instrument Z shocks. This can be important for capturing features
like the lagged stochastic regressor bias described by Stambaugh (1999). Drawing
an initial row from v, we take the Uz shock and add it to the initial value of the
Z (perhaps, drawn from the unconditional sample distribution) to produce the first
lagged instrument vector, Zt–1. We draw another row from v and construct the first
observation of the remaining data as follows. The stocks’ returns are formed by
adding the Ur shock to bzt–1, where b is the “true” regression coefficient that defines
the conditional expected return. The contemporaneous values of the Zs are formed
by multiplying the VAR coefficient by Zt–1 and adding the shock, Uz. The next
observation is generated by taking the previous contemporaneous value of the Z as
Zt–1 and repeating the process. In this way, the Z values are built up recursively,
which captures their strong serial correlation.
35 Optimal Orthogonal Portfolios with Conditioning Information 1001
References
Abhyankar, A., Basu, D., & Streme, A. (2007). Portfolio efficiency and discount factor bounds
with conditioning information: An empirical study. Journal of Banking and Finance, 31,
419–437.
Andrews, D. W. K., Lieberman, O., & Marmer, V. (2006). Higher order improvements of the
parametric bootstrap for long-memory gaussian processes. Journal of Econometrics, 133,
673–702.
Athreya, K. B. (1987). Bootstrap of the mean in the infinite variance case. Annals of Statistics, 15,
724–731.
Bekaert, G., & Liu, J. (2004). Conditioning information and variance bounds on pricing kernels.
Review of Financial Studies, 17, 339–378.
Campbell, J., Lo, A., & MacKinlay, A. C. (1987). The econometrics of financial markets.
Princeton: Princeton University Press.
Chiang, E. (2012). Modern portfolio management with conditioning information. Working paper.
University of North Carolina, Charlotte.
Dybvig, P., & Ross, S. (1985). Differential information and performance measurement using
a security market line. Journal of Finance, 40, 383–399.
Efron, B. (1979). Bootstrap methods: Another looks at the jackknife. Annals of Statistics, 7, 1–26.
Efron, B., & Tibshirani, R. J. (1993). Introduction to the bootstrap. New York: Chapman and Hall.
Fama, E. F. (1973). A note on the market model and the two-parameter model. Journal of Finance,
28, 1181–1185.
Ferson, W. E., & Siegel, A. F. (2001). The efficient use of conditioning information in portfolios.
Journal of Finance, 56, 967–982.
Ferson, W. E., & Siegel, A. F. (2003). Stochastic discount factor bounds with conditioning
information. Review of Financial Studies, 16, 567–595.
Ferson, W. E., & Siegel, A. F. (2009). Testing portfolio efficiency with conditioning information.
Review of Financial Studies, 22, 2735–2758.
Ferson, W. E., Sarkissian, S., & Simin, T. (2003). Spurious regressions in financial economics?
Journal of Finance, 58, 1393–1414.
Ferson, W. E., Siegel, A. F., & Xu, P. (2006). Mimicking portfolios with conditioning information.
Journal of Financial and Quantitative Analysis, 41, 607–635.
Frost, P. A., & Savarino, J. E. (1988). For better performance: Constrain portfolio weights. Journal
of Portfolio Management, 15, 29–34.
Gibbons, M. R., Ross, S. A., & Shanken, J. (1989). A test of the efficiency of a given portfolio.
Econometrica, 57, 1121–1152.
Grinold, R. C., & Kahn, R. N. (1992). Active portfolio management (2nd ed.). New York: McGraw
Hill.
Hansen, Lars P. (1982). Large Sample Properties of Generalized Method of Moments Estimators.
Econometrica, 50, 1029–1054.
Hansen, L. P., & Richard, S. F. (1987). The role of conditioning information in deducing testable
restrictions implied by dynamic asset pricing models. Econometrica, 55(3), 587–613.
Jagannathan, R., & Ma, T. (2003). Risk reduction in large portfolios: Why imposing the wrong
constraints helps. Journal of Finance, 58, 1651–1684.
Jobson, J. D., & Korkie, B. (1982). Potential performance and tests of portfolio efficiency. Journal
of Financial Economics, 10, 433–466.
Jorion, P. (2003). Portfolio optimization with tracking error constraints. Financial Analysts
Journal, 59, 70–82.
Kandel, S. A. (1984). The likelihood ratio test statistic of mean variance efficiency without
a riskless asset. Journal of Financial Economics, 13, 575–592.
Kandel, S. A., & Stambaugh, R. F. (1987). On correlations and inferences about mean-variance
efficiency. Journal of Financial Economics, 18, 61–90.
1002 W.E. Ferson and A.F. Siegel
Kandel, S. A., & Stambaugh, R. F. (1989). A mean-variance framework for tests of asset pricing
models. Review of Financial Studies, 2, 125–156.
MacKinlay, A. C. (1995). Multifactor models do not explain deviations from the CAPM. Journal
of Financial Economics, 38, 3–28.
Markowitz, H. M. (1952). Portfolio selection. Journal of Finance, 7, 77–91.
Michaud, R. O. (1989). The Markowitz optimization enigma: Is ‘optimized’ optimal? Financial
Analysts Journal, 45, 31–42.
Roll, R. R. (1977). A critique of the asset pricing theory’s tests – part 1: On past and potential
testability of the theory. Journal of Financial Economics, 4, 129–176.
Roll, R. R. (1980). Orthogonal portfolios. Journal of Financial and Quantitative Analysis, 15,
1005–1023.
Roll, R. R. (1982). A mean variance analysis of tracking error. Journal of Portfolio Management,
18, 13–22.
Roll, R. R. (1985). A note on the geometry of Shanken’s CSR T2 test for mean/variance efficiency.
Journal of Financial Economics, 14, 349–357.
Shanken, J. (1987). Multivariate proxies and asset pricing relations: Living with the Roll critique.
Journal of Financial Economics, 18, 91–110.
Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk.
Journal of Finance, 19, 425–442.
Siegel, A. F., & Woodgate, A. (2007). Performance of portfolios optimized with estimation error.
Management Science, 53, 1005–1015.
Stambaugh, R. F. (1999). Predictive regressions. Journal of Financial Economics, 54, 375–421.
Treynor, J. L., & Black, F. (1973). How to use security analysis to improve portfolio selection.
Journal of Business, 46, 66–86.
Multifactor, Multi-indicator Approach
to Asset Pricing: Method and 36
Empirical Evidence
Contents
36.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1004
36.2 The MIMIC Model and the Tests of CAPM and APT . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1005
36.2.1 The MIMC Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1005
36.2.2 The Testing Model of CAPM by the MIMIC Approach . . . . . . . . . . . . . . . . . . . . 1007
36.2.3 The Testing Model of APT by the MIMIC Approach . . . . . . . . . . . . . . . . . . . . . . . 1008
36.3 Test of CAPM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1008
36.4 Tests of APT by MIMIC Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1012
36.4.1 Macroeconomic Variables as the Indicators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1012
36.4.2 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1013
36.5 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1020
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1022
Abstract
This paper uses a multifactor, multi-indicator approach to test the capital asset
pricing model (CAPM) and the arbitrage pricing theory (APT). This approach is
able to solve the measuring problem in the market portfolio in testing CAPM;
and it is also able to directly test APT by linking the common factors to the
macroeconomic indicators. Our results from testing CAPM support
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1003
DOI 10.1007/978-1-4614-7750-1_36,
# Springer Science+Business Media New York 2015
1004 C.-F. Lee et al.
Keywords
Capital asset pricing model, CAPM • Arbitrage pricing theory • Multifactor
multi-indicator approach • MIMIC • Measurement error • LISREL approach •
Ordinary least square, OLS • General least square, GLS • Maximum-likelihood
estimation, MLE
36.1 Introduction
Roll (1977) has shown that the capital asset pricing model (CAPM) can never be
tested unless the market portfolio is capable of being measured and identified.
However, the market portfolio is actually unobservable. Stated differently, since
the market portfolio is subject to measurement error, Sharpe (1964), Lintner
(1965), and Mossin’s (1966) type of CAPM can never be tested directly. In contrast,
the test of Ross’s (1976, 1977) arbitrage pricing theory (APT) does not rely upon the
identifications of the market portfolio or the true factors. Nevertheless, Shanken
(1982) argues that Ross’s contention that APT is inherently more easily tested is
questionable. If we can directly link these unobservable factors to some observable
indicators, the Shanken criticism of the test of APT can be avoided or reduced.
Fortunately, a multiple indicators and multiple causes (MIMIC) model, pro-
posed by Zellner (1970), Goldberger (1972a, b), Joreskog and Goldberger (1975),
and others, is an attractive methodology in dealing with the problem of
unobservable variables. Goldberger (1972b) conceptually described that structural
equation model is a combination of factor analysis and econometrics model.
Goldberger (1972b) and Joreskog and Goldberger (1975) develop a structural
equation model with multiple indicators and multiple causes of a single latent
variable, MIMIC model, and obtain maximum likelihood estimates of parameters.
The MIMIC model displays a mixture of econometric and factor analysis themes.
This concept has successfully been used to test economic models, such as the
structure of price expectation and inflationary expectation (see Turnovsky 1970;
Lahiri 1976). In addition, structural equation model has also been used in finance-
related studies. Titman and Wessels (1988), Chang et al. (2009), and Yang
et al. (2009) apply structural equation models (e.g., LISREL approach and
MIMIC model) in determining capital structure decision. Maddala and
36 Multifactor, Multi-indicator Approach to Asset Pricing 1005
Nimalendran (1996) use structural equation model to examine the effect of earnings
surprises on stock prices, trading volumes, and bid-ask spread. However, the
structural equation model, especially MIMIC model, has not been used in capital
asset pricing determination.
The purpose of this paper is twofold: (i) to use the MIMIC model to reexamine
CAPM and (ii) to use the MIMIC model to investigate the relationship between the
factors in APT and the macroeconomic indicators directly. APT is attractive to both
academicians and practitioners, because the model allows more than one factor.
However, to date the practical applications of APT are still limited since previous
studies in testing the model do not directly link the factors to the indicators. If the
linkage between the factors and the indicators can be derived, practical applications
will be much improved.
The outline of this paper is as follows. In Sect. 36.2, the MIMIC model is
reviewed and CAPM and APT in terms of the MIMIC model are demonstrated.
Section 36.3 shows how MIMIC can be used to test CAPM, and Sect. 36.4
investigates the MIMIC applied to test APT. Finally, a brief summary is contained
in Sect. 36.5.
36.2 The MIMIC Model and the Tests of CAPM and APT
Suppose that a system has k unobservable latent variables z ¼ (z1, . . .,zk)0 , p observable
exogenous indicators x ¼ (x1,. . ., xp)0 , and m observable endogenous variables
y ¼ (y1,. . ., ym)0 .1 The specification of this extended MIMIC model of Jöreskog and
Goldberger (1975) is as follows. The latent factors z are linearly determined, subject to
disturbances e ¼ (e1, . . .,ek)0 , by observable exogenous indicators x:2
z ¼ ax þ e, (36.1)
where
2 3
a11 , , a1p
6 7
6 7
a¼6
6 77 is a k p matrix:
4 5
ak1 , , akp
1
Fogler et al. (1981) and Chen et al. (1983) indirectly link the factors extracted from the APT to
economic indicators. Jöreskog and Goldberger (1975) have shown that this kind of indirect
estimation method is not as efficient as the direct estimation method to be explored in this section.
2
Here, we use different terminologies in defining the factors and indicators compared with those
used in traditional MIMIC model.
1006 C.-F. Lee et al.
y ¼ bax þ be þ u ¼ hx þ v, (36.3)
h ¼ ba: (36.4)
v ¼ be þ u, (36.5)
There are two types of restrictions on the reduced form: (i) the m p regression
coefficient matrix h has rank k, the m p components of h being expressed in terms
of kp + mk elements of a and b, and (ii) the m m residual covariance matrix Ω
satisfies a factor analysis model with k common factors, the m(m + 1)/2 distinct
elements of Ω being expressed in terms of the k + km + m elements of s2, b, and y2.
The first restriction, which is the same as the simultaneous equation model, is
familiar to econometricians. The second restriction, which is the same as the factor
analysis model, is familiar to psychometricians. In Eq. 36.5, e, b, and u are regarded
as the common factors, the factor loadings, and the unique disturbances in the factor
analysis model, respectively.
We observe that the reduced-form parameters remain unchanged, when any
column, say j, of b is multiplied by a scalar and the j th row of a and sj are both
divided by the same scalar. To remove this indeterminacy of the model, we can
normalize the model through (i) s2, or (ii) b, or (iii) a. After normalization, the
maximum likelihood estimation (MLE) procedure can be used to obtain consistent
estimators of the elements in parameters a, b, and y2 (see Attfield 1983; Chen 1981;
Joreskog and Goldberger 1975; and others). In the following, we demonstrate how
to apply the MIMIC model to test CAPM and APT.
er i ¼ bier m þ e
ui
i ¼ 1, . . . , N, (36.7)
er m ¼ er m þ e em,
where
er i ¼ the realized excess return (total return less risk-free rate) on security i in
a deviation form
er m ¼ the realized excess return of the NYSE Composite Index
er m ¼ the unobservable excess return on the market portfolio
In this special one-factor case, we remove the indeterminacy by setting
the coefficient on er m equal to one. Equation 36.7 is a simultaneous equation
model, in which there are N equations linking the individual security
(or portfolio) return to the unobservable true market return and one equation linking
the unobservable true market portfolio return to the realized return of the NYSE
composite index. After obtaining the estimated betas from simultaneous equation
system of (36.7), a cross-sectional regression of the security return against its risk
(b) will be used to test CAPM or to estimate the riskless rate and the market risk
premium as follows:
r it ¼ ^a 0t þ ^a 1t bi , (36.8)
1008 C.-F. Lee et al.
where
rit ¼ the excess return on security i at time t
^
a 0t ¼ the estimate of the intercept which is supposed to be zero
^
a 1t ¼ the estimate of the market risk premium
Four different estimation procedures will be used to estimate Eq. 36.8. They are:
(i) stationary OLS, (ii) nonstationary OLS, (iii) GLS, and (iv) MLE.3
The testing model of APT, in terms of the MIMIC model, can be rewritten as
follows:
where
ef j ¼ the jth unobservable factor
eI h ¼ the hth macroeconomic indicator, h ¼ 1, ,p
For convenience and easy explanation, each factor ef is assumed to have
different set of explained indicators eI ’s. Note that there are N return equations
plus k factor equations in the simultaneous equation system (36.9). The LISREL
computer program of Jöreskog and Sörbon (1981) is used to estimate the parame-
ters, a and b, in Eq. 36.9. A cross-sectional regression is also used to test APT and to
estimate the riskless rate and the factor risk premia by regressing the security return
against its risks, b’s. The a’s coefficients in Eq. 36.9 can be used to explain the
relationship between factors and indicators.
This section tests CAPM using the market model and the MIMIC model described in
Sect. 36.2. The objective is to investigate whether the MIMIC method yields
a different inference from the market model. Nineteen industry common stock
portfolios are formed with the same manner used by Schipper and Thompson
(1981), and Stambough (1982).4 The return on a portfolio is the arithmetic average
of returns for firms on the CRSP monthly tape with the appropriate two-digit SEC
code for the given month. The tests use returns from the period 1963–1982, and this
total period is divided into two equal subperiods: (i) subperiod 1, 1963–1972 and
3
The terminologies stationary OLS and nonstationary OLS have been used by Friend and West-
field (1980). The GLS and MLE methods have been discussed by Litzenberger and
Ramaswamy (1979).
4
See Chen (1981).
36 Multifactor, Multi-indicator Approach to Asset Pricing 1009
(ii) subperiod 2, 1973–1982. Portfolios are formed primarily because they provide
a convenient way to limit the computational dimensions of the MIMIC method. As
mentioned by Stambough (1982), industry portfolios also allow rejection of CAPM
due to the presence of additional industry-related variables in the risk-return relation.
Table 36.1 indicates the number of securities in each portfolio, the SEC codes,
and betas calculated from the market model and also from the MIMIC model. When
the MIMIC model was used to estimate betas by the portfolio excess returns in
subperiod 1, convergent problems were encountered during minimization so that
the result is inappropriate. Consequently, raw returns in deviation form on the
portfolios are used to estimate betas for subperiod 1. The betas estimated from
the MIMIC model are very close to those from the market model in both periods.
This evidence supports Stambough’s discovery that inferences about CAPM are
very insensitive to alternative market indexes.
Table 36.2 presents return-risk trade-off from the cross-sectional relationship in
which the average monthly excess portfolio returns (monthly portfolio returns less
monthly return on 3-month Treasury bills) is regressed on a beta estimated either
from the MIMIC model or the market model from two different 120-month periods.
Four different estimation procedures are used to estimate the intercepts and the
market risk premia. The OLS method presents two sets of t-statistics shown in
the parentheses under the same relevant regression coefficients. The first set
(denoted S) assumes that the regression coefficients are constant or stationary
over each 120-month period. The second set (denoted NS) of the OLS t-statistics
allows the nonstationarities of the regression coefficients by computing the
cross-sectional regression coefficients in each month and deriving the appropriate
standard errors from the time series 120 estimates of the OLS regression coeffi-
cients. The GLS and MLE methods also permit the nonstationary coefficients.
Thus, their t-statistics are derived from the same procedure in the OLS
(NS) method. Although the OLS and the GLS estimators are biased and inconsistent
due to measurement error in beta (see Litzenberger and Ramaswamy 1979), the
maximum likelihood estimators are consistent simply because MLE takes care of
measurement error in beta.
In Table 36.2, the coefficients in the NS regression of OLS, GLS, and MLE are
obviously characterized by much larger standard errors so that they lose any
significance shown in the stationary OLS regression.5
However, different estimated betas cause little changes in return-risk relation-
ships. From the results of the OLS stationary method, there exists a significant
return-risk relationship in subperiod 1, but not in subperiod 2 in both MIMIC and
market models. The poor return-risk relationship in subperiod 2 may be due to the
poor performance of CAPM in determining a pricing relation. In the next section,
APT will be used to examine an alternative pricing relationship. Even though the
null hypothesis of CAPM that a0 ¼ 0 cannot be rejected at the 5 % level in all four
cases, all coefficients are positive. In addition, the intercept in subperiod 2, about
5
The similar results were also found in the Friend and Westfield’s (1980) study of co-skewness.
1010 C.-F. Lee et al.
Table 36.1 Industry portfolio SEC codes, number of firms, and estimated betas
Portfolio description SEC code # of firmsEstimated betas
Period 1 Period 2
Market MIMC Market MIMC
1972 1982 model model model model
1. Mining 10–14 56 71 1.056 1.009 0.922 0.916
2. Food and beverages 20 75 51 0.894 0.841 0.803 0.803
3. Textile and apparel 22,23 58 45 1.264 1.220 1.081 1.082
4. Paper products 26 30 30 1.030 1.029 0.910 0.909
5. Chemical 28 87 83 0.949 0.947 0.847 0.846
6. Petroleum 29 28 22 0.706 0.792 0.745 0.740
7. Stone, clay, glass 32 43 31 1.050 1.106 1.045 1.045
8. Primary metals 33 56 49 1.136 1.193 0.932 0.930
9. Fabricated metals 34 45 46 1.145 1.155 1.102 1.102
10. Machinery 35 93 104 1.234 1.231 1.104 1.104
11. Appliance and 36 87 82 1.384 1.401 1.179 1.180
elec. equip.
12. Transport. equip. 37 64 50 1.209 1.275 1.150 1.151
13. Misc. manufactrng. 38,39 64 59 1.375 1.314 1.197 1.198
14. Railroads 40 18 11 1.294 1.229 0.899 0.895
15. Other transport. 41, 42, 44 34 35 1.335 1.447 1.203 1.203
45, 47
16. Utilities 49 138 152 0.443 0.467 0.564 0.562
17. Department stores 53 35 28 1.149 1.104 1.125 1.125
18. Other retail trades 50–52, 103 97 1.144 1.088 1.123 1.124
54–59
19. Banking, finance, 60–67 184 240 0.968 1.021 1.069 1.068
real estate
This table presents the number of firms and betas in each industry portfolio. The numbers of firms
are obtained in the end of 1972 and the end of 1982. The betas shown in the first column are
estimated from the market model, while those in the second column are estimated from the MIMIC
model. Period 1 represents the sample period from 1963 to 1972 and period 2 represents the sample
period from 1973 to 1982
4.3 % annual rate, is too high. This is consistent with prior tests of the traditional
version of CAPM.
All nonstationary estimates of a0 and a1 in OLS, GLS, and MLE in Table 36.2
are insignificant. Because of the low test power for all nonstationary procedures
(the standard errors are too high), in the following, only the magnitudes of the
estimated coefficients are discussed. The MLE and GLS estimates of a0 are much
lower and much closer to zero than the corresponding OLS estimates in all four cases.
In addition, the MLE estimates of a1 is greater than the corresponding GLS estimates
and is closer to the realized market risk premia in all four cases. The realized market
risk premia are 0.751 % and 0.636 % monthly for periods 1 and 2, respectively.
36 Multifactor, Multi-indicator Approach to Asset Pricing 1011
This evidence proves that the MLE estimator in the return-risk cross-sectional
regressions is more appropriate than OLS or GLS estimator in testing CAPM.
In sum, we have proposed an alternative estimator of betas by the MIMIC model
in which measurement error in a market portfolio is allowed. Nevertheless, this
reasonable alternative method does not gain much from the traditional OLS
estimator. However, some interesting results have surfaced. This evidence supports
Stambough’s conclusion that the tests of CAPM are insensitive to different market
indexes. In return-risk cross-sectional regressions, our evidence shows that the
MLE estimator is more appropriate than the OLS or GLS estimator due to
measurement error in beta. From these two interesting results, we conclude that
measurement error on beta is more serious than measurement error on the market
portfolio in testing CAPM.
1012 C.-F. Lee et al.
This section tests APT using the MIMIC model demonstrated in Sect. 36.2. The
objective is to investigate that (i) the proper number of factors is used to explain the
data and (ii) the relationships between factors and indicators which are measured by
macroeconomic variables. The same 19 industry portfolios described in previous
section are used here. The macroeconomic variables are selected from those most
likely related to common stock returns. In the following, the indicators selected in
this study will be discussed.
They are price level as measured by consumer price index (CPI) and the real gross
national product less corporate profit (GNP). They find that hedge portfolios offer
meaningful hedging potential in portfolio-formation period. In addition, CAPM or
the market model indicates that the return on a security or a portfolio most likely
co-move with the return on the market portfolio.
In summary, the variables most likely correlated with a stock return would be
classified as five categories: (1) money supply, (2) real production, (3) inflation,
(4) interest rate, and (5) market return. Further, Brigham (1982) decomposes a risk
premium into maturity risk premium and default risk premium. Thus, these two
indicators are also included in our study. According to the above discussion, the
following 11 variables are selected as the indicators.
1. Return on the market portfolio (RM): the return on NYSE common stock
composite index.
2. Transaction volume (VL): the change rate in the transaction volume (shares)
for all of the NYSE common stocks.
3. Real riskless rate (RF): the real interest rate on 3-month Treasury bills.
4. Maturity risk premium (MP): the difference between the real interest rates on
long-term Treasury bonds (10 or more years) and on 3-month Treasury bills.
5. Default risk premium (DP): the difference between the real interest rates on
new AA corporate bonds and on 3-month Treasury bills.
6. Consumer price index inflation rate (CPI): the change rate in urban consumer
price index for all items.
7. Money supply (M2): the real change rate in money stock as measured by M2
(M1 + time deposits).
8. Velocity of money supply (PI/M2): the ratio of personal income to money
supply M2. This is an alternative measure of money supply.
9. Real industrial production (IP): the change rate in real total industrial
productions.
10. Real auto production (IPA): the change rate in real automotive products.
11. Real home production (IPH): the change rate in real home goods.
Since the automobile and housing industries generally lead the rest of the
economy, the last two indicators, IPA and IPH, are used to catch up the first two
biggest industries. The reason to select industrial production instead of GNP in this
study is that all other indicators are published monthly while GNP is published
quarterly. Since the industrial production is a very good proxy for GNP, we sacrifice
GNP measure to gain the number of time periods.
After carefully selecting the indicator candidates, the time lag or leading problem
needs to be solved. The correlation coefficients between the market portfolio
return and the other indicators with lags and leadings of zero to 5 months for both
periods were examined. All indicators are contemporaneously correlated with the
1014 C.-F. Lee et al.
Table 36.3 Eigenvalue as a percentage of the first eigenvalue for 19 industry portfolios:
1963–1982
Factor PRC ALP SCF ULS
Panel A: Period: 1963–1972
1 100 % 100 % 100 % 100 %
2 3.8 4.4 3.9 2.9
3 3.1 1.5 2.6 1.7
4 2.1 1.0 2.3 0.8
5 1.9 0.7 2.0 0.7
6 1.6 0.6 1.9 0.6
Panel B: Period: 1973–1982
1 100 % 100 % 100 % 100 %
2 7.9 6.9 7.1 7.2
3 2.1 2.8 2.9 1.8
4 1.3 1.0 1.9 0.9
5 1.0 0.6 1.6 0.5
6 1.0 0.4 1.4 0.4
This table shows the eigenvalues as a percentage of the first eigenvalue. Four methods are used to
determine the number of factors in APT model. They are principal component analysis (PRC),
alpha factor analysis (ALP), simple common factor analysis (SCF), and unweighted least squares
method (ULS)
market portfolio except three real production indicators. The real production
indicators follow the market portfolio return with a lag of 2 or 3 months.
Therefore, in this study, all indicators are contemporaneous except three real
production indicators which is a 2-month lag.
Before APT is directly tested by the MIMIC model, factor analysis is prelimi-
narily used to determine the number of factors in both periods. Table 36.3 shows the
eigenvalues as a percentage of the first eigenvalue. Clearly, it is only one factor in
period 1, while it is perhaps two factors in period 2 by “scree” test described in
Chapter 4 of Wei (1984).6 Consequently, at most a two-factor model is enough to
explain the historical data. Three alternative MIMIC models are proposed to test
APT: (i) one-factor 11-indicator model, (ii) one-factor six-indicator model, and
(iii) two-factor six-indicator model. When the two one-factor models are used to
test APT in period 1, there is little difference between 11-indicator and six-indicator
model. Thus, only six indicators are used in the two-factor model to save the
computer time.7
The structural coefficients of APT in the MIMIC model for period 1 are reported
in Table 36.4a. In both one-factor models, only the stock market-related variables,
6
In his dissertation, Wei (1984) has shown that the “scree” test is a powerful test in identifying the
number of relevant factors in the APT. By using simulation study, Wei has shown that Roll and
Ross’s (1980) ML method in estimating factors are inferior to methods listed in Table 36.3.
7
It is very expensive to run LISREL program, especially for more than two factor models.
36 Multifactor, Multi-indicator Approach to Asset Pricing 1015
the market return (RM), and the transaction volume (VL) are significant at the 5 %
level. From this evidence, if the pricing relation in period 1 is a one-factor APT, this
common factor would be most likely related to only the stock market-related
indicators, namely, the market portfolio return and the transaction volume. Other
indicators may be correlated with this single common factor, but they are obviously
not as important as the stock market-related indicators. Now, let us closely examine
other indicators with an absolute t-value of greater than one for 11-indicator model.
The real riskless interest rate (RF), CPI inflation rate (CPI), and the real auto
production (IPA) are negatively correlated with this common factor, while the
velocity of money supply (PI/M2) is positively related to this common factor.
If we regard this common factor as a proxy of market portfolio because most of
the weight is on the market portfolio, then, except for real auto production, this
evidence supports previous studies done on the relationship between common stock
returns and other indicators (Keran 1971; Homa and Jaffee 1971; Hamburger and
Kochin 1972; Pesando 1974; Bodie 1976; Nelson 1976; Fama and Schwert 1977;
and others). However, there is no previous study which examines the relationship
between stock returns and the real auto production.
Some might argue that weak relationship between non-stock market indicators
and the common factor is due to multicollinearities among the indicators. There-
fore, six of the 11 indicators are selected to represent each category indicator. They
are the market portfolio return (RM), the transaction volume (VL), real riskless
interest rate (RF), CPI inflation rate (CPI), money supply (M2), and the real total
industrial production (IP). This is the one-factor six-indicator model. The result of
this model is shown in Table 36.4a column 2. The result of this one-factor
six-indicator model is very close to that of the one-factor 11-indicator model. As
mentioned before, only the stock market-related indicators are significantly corre-
lated with the common factor. Real riskless interest rate, inflation, money supply,
and real production are all negatively but insignificantly related to the common
factor. For factor equation, the 11-indicator model has only a little high R-square
than the six-indicator model. They are 0.5611 and 0.5412, respectively. Comparing
the betas of these two one-factor models in Table 36.4a, they are very highly
correlated with a correlation coefficient of about 1.000. In addition, the average
R-square of each return equation in both one-factor models is the same with a value
of 0.811. Up to this point, there is not much difference between the one-factor
11-indicator and the one-factor six-indicator models. Later on, the cross-sectional
regression will be used to double check this result.
Even though we have already discussed that the appropriate model is a one-factor
model for period 1 by the scree test of factor analysis, we want to use a two-factor
model to double check whether the second factor is significant or not.
A predetermined two-factor six-indicator model will be used to test APT. Factor
analysis is employed to classify these six indicators into two groups. The first group
includes the market portfolio and the transaction volume, while the second group
includes other four indicators: real riskless interest rate (RF), inflation rate (CPI),
money supply (M2), and real total industrial production (IP). The two-factor result
shown in Table 36.4a columns 3 and 4 displays that only the first factor is
1016 C.-F. Lee et al.
significantly related to RM and VL, whereas the second factor is very insignifi-
cantly correlated with the second group indicators. This is also evident by exam-
ining from the second betas in the table. All of the second betas are insignificant.
Furthermore, the first beta coefficient is very highly correlated with the betas in both
one-factor models with both correlation coefficients of 0.996. This is further
supported by the R-square criteria. The average R-square of each return equation in
the one-factor model is 0.811, while 0.844 in the two-factor model. If the adjusted
R-square is used as the criteria, the increase in R-square will be insignificant. Thus,
a one-factor APT is good enough to explain the first period data. We will reconfirm
this argument by the cross-sectional regression.
Table 36.5 Panel A reports the return-risk cross-sectional relationship in period 1
for APT in the MIMIC model. Both one-factor models have a very similar result.
Their adjusted R-squares are the same of 0.258. The intercepts are both insignifi-
cantly different from zero but positive (recall that the LHS variable is excess return).
The factor risk premia for both one-factor models are positive and significant. This is
exactly the result that should have been concluded from APT. On the other hand, the
intercept in the two-factor model is much higher than those in the one-factor model,
and the two factor risk premia are both insignificant.8 The adjusted R-square is a little
lower than those in the one-factor models. From the results of the MIMIC model and
the cross-sectional regressions, it is probable that one-factor APT with the market
portfolio and the transaction volume as the determinants of the single factor is the
appropriate pricing model for period 1. Comparing the MIMIC CAPM and CAPM in
Table 36.2 with this one-factor model, the MIMIC CAPM is closer to the one-factor
model. In addition, the factor risk premium is closer to the realized market risk
premium than that in the MIMIC CAPM model.
Now, let us examine APT in period 2. The structural coefficients of APT in the
MIMIC model for period 2 are represented in Table 36.4b. Because six indicators
do not make much difference from 11 indicators in the one-factor model in period 1,
only the 11-indicator model is used to test APT for the one-factor model in period 2
in order to save the computer time. From the one-factor model, the result is similar
to that in period 1. The stock market return and the transaction volume are the most
significant indicators. However, inflation and the real auto production are also
significant even just at the marginal level. This reinforces our suspicion that there
may be more than one factor in period 2.
Let us closely examine the indicators with an absolute t-value greater than one.
Real interest riskless rate, inflation, money supply, and real auto production are all
negatively correlated with the common factor, while the velocity of money supply is
positively correlated with this common factor. Overall, this result again supports
previous studies. Now, let us turn to the two-factor six-indicator model. Follow the
same procedures done for period 1. The result is displayed in Table 36.4b columns
2 and 3. The market portfolio (RM) and the transaction volume (LV) are significantly
related to the first factor. However, the real auto production is significantly correlated
with the second factor; this time even only at the marginal level. Comparing other
indicators with those in period 1, real riskless interest rate is again negatively related
to the second factor, but inflation rate and money supply are positively related to the
second factor in the period. From the result of the relationship between the factors
and the indicators, the second factor is still important even though less important than
8
The loss of the significance of the first factor risk premium is due to the multicollinearity problem.
36 Multifactor, Multi-indicator Approach to Asset Pricing 1019
Table 36.5 Return-risk cross-sectional relationships of APT in the MIMIC model: 1963–1982
2
^
a0 ^
a1 ^
a2 R
Model Panel A: Period 1, 1963–1972
One-factor 0.159 0.543* 0.258
11-indicator (0.710) (2.693)
One-factor 0.159 0.558* 0.258
6-indicator (0.710) (2.692)
Two-factor 0.507 0.206 0.071 0.257
6-indicator (1.219) (0.529) (1.221)
Panel B: Period 2, 1973–1982
One-factor 0.370 0.285 0.002
11-indicator (1.328) (0.967)
Two-factor 0.040 0.684* 0.362* 0.215
6-indicator (0.150) (2.377) (2.066)
This table presents the return-risk cross-sectional relationship for the APT in the MIMIC model.
The cross-sectional relationship between return and risk can be written as Ri ¼ ^ a0 þ ^a 1 b1 þ ^
a 2 b2
þ ee i, i¼1,. . .,19. Panel A reports the relationship in period 1 and Panel B reports the relationship in
period 2. The average monthly returns are multiplied by 100 before regressing. *represents
significant at the 5 % level
the first factor for period 2. This can also be checked by the significance of betas in
Table 36.4b. Ten out of the 19 s factor betas are significant although the relationship
is not as strong as those in the first factor betas. Because the first factor is correlated
with stock market-related indicator, this factor can be regarded as a proxy of the
market portfolio. The correlation coefficient between the beta in the first factor and
the beta in the MIMIC CAPM is very high with a coefficient of 0.923. Further, the
average R-square of each return equation in the one-factor model is 0.818, while
0.885 in the two-factor model. When the adjusted R-square is used as a criteria, the
increase in R-square should not be trivial and will be significant. In addition, the
R-square for the second factor equation is as high as 0.1836. All of these results
indicate that the second factor should be important. We further check whether the
second factor is important or not by the cross-sectional regression.
The risk-return cross-sectional relationship of APT in the MIMIC model for
period 2 is shown in Table 36.5 Panel B. For the one-factor model, both the
intercept and the factor risk premium are insignificant. The intercept is too high
with an annual rate of 4.3 %, while the factor risk premium is far below the realized
market risk premium with a monthly rate of 0.636 %. The adjusted R-square is
negative. On the other hand, the intercept for the two-factor model is insignificant
and very near to zero, while the two factor premia are significant with the first factor
premium very close to the realized market risk premium. The adjusted R-square is
as high as 0.215. The results of APT in the MIMIC model confirm the scree test of
the factor analysis and the poor performance of CAPM in previous section for
period 2. Furthermore, comparing the result in Table 36.5 Panel A with that in
Table 36.2, we can conclude that the two-factor APT outperforms the one-factor
APT, the MIMIC CAPM, and CAPM in period 2. But the one-factor APT does
1020 C.-F. Lee et al.
a better job than the two-factor APT in period 1, and there is not much difference
among the one-factor APT, the MIMIC CAPM, and CAPM. This evidence supports
Ross’s argument that APT is more general than CAPM because APT allows more
than one factor in the pricing relation.
It will be interesting to see what will happen if market variables (the market
portfolio and transaction volume) are excluded from the model. Table 36.6 shows
the structural coefficients of the one-factor APT without market variables.9 The
b coefficients (factor loadings) shown in Table 36.6 for both periods are very close
to those estimated from the one-factor APT with market variables shown in
Tables 36.4a and 36.4b, respectively.10 However, the results from the factor
equation have dramatically changed.
For period 1, the R-square is only 0.0611 which is much lower than 0.5611 from
the one-factor APT with market variables. This result indicates that the market
variables play the major role in the pricing behavior during period 1. When the
market variables excluded from the model, among nine indicators, only the two
money supply variables, M2 and PI/M2, are positively related to the unique
common factor at the marginal level. For period 2, the R-square of the factor
equation for the model without market variables is 0.3915 which is higher than
the one in period 1. This evidence denotes that, in addition to the market variables,
some other macroeconomic indicators play a relatively important role in the pricing
behavior during period 2. Among the nine nonmarket variables, real risk-free rate
(RF), default risk premium (DF), and inflation (CPI) are significantly negatively
related to the common factor, while the velocity of money supply (PI/M2) and the
real auto production (IPA) are significantly positively related to the common factor.
All of the relationships are as we expect. The results from the model without market
variables also confirm our previous evidence that the 1963–1972 data can be
described by a one-factor APT with the market variables as its indicators, while
the 1973–1982 data should be explained by more than one-factor APT.
This paper bases on parts of Wei’s (1984) dissertation using a MIMIC approach to test
CAPM and APT. The results support the conclusion that APT outperforms CAPM,
especially for the period from 1973 to 1982. The beta estimated from the MIMIC
model by allowing measurement error on the market portfolio does not significantly
improve the OLS beta. However, the MLE estimator does a better job than the OLS
and GLS estimators in the cross-sectional regressions because the MLE estimator
takes care of the measurement error in beta. Therefore, the measurement error problem
9
Only the one-factor APT is used to investigate the difference between the models shown in
Table 36.4 and in Table 36.6.
10
If we normalize the one-factor 11-indicator model for period 2 shown in Table 36.4b by setting
b1 ¼ 1.00, it is easily seen that the b coefficients of one-factor model shown in Tables 36.4b and
36.6 column 2 are very similar.
36 Multifactor, Multi-indicator Approach to Asset Pricing 1021
Table 36.6 Structural coefficients of the one-factor APT in the MIMIC model without market
variables: 1963–1982
a’s Coefficients
Indicator 1963–1972 1973–1982
RF 25.541(1.26) 18.666(3.39)*
MP 26.061(1.18) 1.067(0.17)
DP 4.350(0.23) 30.745(2.68)*
CPI 25.989(1.32) 23.l46(3.52)*
M2 3.652(1.85)* 1.630(0.98)
PI/M2 45.126(1.91)* 110.380(4.22)*
IP 0.932(0.97) 0.124(0.16)
I PA 0.046(0.57) 0.214(2.18)*
IPH 0.298(1.06) 0.120(0.33)
R-square 0.0611 0.3915
b’s Coefficients
Industry 1963–1972 1973–1982
1 1.000(–) 1.000(–)
2 0.833(16.45)* 0.930(10.10)*
3 1.209(16.20)* 1.270(10.03)*
4 1.019(14.70)* 1.045(9.99)*
5 0.939(17.43)* 0.981(10.11)*
6 0.785(10.53)* 0.800(9.64)*
7 1.097(15.33)* 1.214(10.41)*
8 1.183(15.79)* 1.063(9.64)*
9 1.146(18.11)* 1.283(10.63)*
10 1.221(18.25)* 1.282(10.53)*
11 1.389(16.58)* 1.386(10.60)*
12 1.265(17.74)* 1.349(10.50)*
13 1.302(17.16)* 1.407(10.63)*
14 1.218(13.19)* 1.006(8.96)*
15 1.436(14.05)* 1.404(9.92)*
16 0.463(7.66)* 0.632(8.23)*
17 1.094(14.13)* 1.329(9.84)*
18 1.078(16.46)* 1.318(10.47)*
19 1.012(14.70)* 1.233(9.95)*
This table shows estimated coefficients of the one-factor APT in the MIMIC model without market
variables. The structural model is shown as follows. *represents significant at the 5 % level
er i ¼ bi1ef 1 þ bi2ef 2 þ e
ui, i ¼ 1, . . . , 19
ef j ¼ a1 ðRFÞ þ a2 ðMPÞ þ a3 ðDPÞ
þ a4 ðCPI Þ þ a5 ðM2Þ þ a6 ðPI=M2Þ þ a7 ðIPÞ
þ a8 ðIPAÞ þ a9 ðIPH Þ þ ee j ,
on beta is more serious than that on the market portfolio. This evidence supports
Stambough’s (1982) argument that the inference about the tests of CAPM is insensi-
tive to alternative market indexes. When the one-factor APT with market variables is
compared with the model without market variables, we found that the market variables
1022 C.-F. Lee et al.
References
Attfield, C. (1983). Consistent estimation of certain parameters in the unobservable variable model
when there is a specification error. Review of Economics and Statistics, 65, 164–167.
Bodie, Z. (1976). Common stocks as a hedge against inflation. Journal of Finance, 31, 459–740.
Brigham, S. (1983). Fundamentals of financial management (3rd ed.). Chicago: Dryden Press.
Brigham, E. F., & Houston, J. F. (2012). Fundamentals of financial management (13th ed.).
Singapore: Cengage Learning.
Chang, C., Lee, A. C., & Lee, C. F. (2009). Determinants of capital structure choice: A structural
equation modeling approach. Quarterly Review of Economics and Finance, 49, 197–213.
Chen, C. (1981). The EM approach to the multiple indicators and multiple causes model via the
estimation of the latent variable. Journal of the American Statistical Association, 76, 702–708.
Chen, N. F., Roll, R., & Ross, S. A. (1983). Economic forces and the stock market: Testing the APT
and alternative asset pricing theories. Working paper, University of Chicago.
Cohn, R., & Lessard, D. (1981). Are markets efficient? Tests of alternative hypothesis. On the
effect of inflation on stock prices: International evidence. Journal of Finance, 36, 277–289.
Fama, E. (1981). Stock returns, real activity, inflation, and money. The American Economic
Review, 71, 545–565.
Fama, E., & Schwert, G. (1977). Asset returns and inflation. Journal of Financial Economics, 5,
115–146.
Fogler, H. R., John, K., & Tipton, J. (1981). Three factor interest rate differentials and stock
groups. Journal of Finance, 36, 323–336.
Friend, I., & Westerfield, R. (1980). Co-skewness and capital asset pricing. Journal of Finance, 35,
897–913.
Gargett, D. (1978). The link between stock prices and liquidity. Financial Analysis Journal, 34, 50–54.
Geske, R., & Roll, R. (1983). The fiscal and monetary linkage between stock returns and inflation.
Journal of Finance, 38, 1–33.
Goldberger, A. (1972a). Maximum-likelihood estimation of regressions containing unobservable
independent variables. International Economic Review, 13, 1–15.
Goldberger, A. (1972b). Structural equation methods in the social sciences. Econometrica, 40,
979–1001.
Granger, C., & Newbold, P. (1974). Spurious regression in econometrics. Journal of
Econometrics, 2, 111–120.
Gultekin, N. (1983). Stock market returns and inflation: Evidence from other countries. Journal of
Finance, 38, 49–65.
Hamburger, M., & Kochin, L. (1972). Money and stock prices: The channels of influence. Journal
of Finance, 27, 231–249.
Homa, K., & Jaffee, D. (1971). The supply of money and common stock prices. Journal of
Finance, 26, 1045–1066.
Joreskog, K., & Goldberger, A. (1975). Estimation of a model with multiple indicators and multiple
causes of a single latent variable. Journal of the American Statistical Association, 70, 631–639.
Joreskog, K., & Sorbom, D. (1981). LISREL V: Analysis of linear structural relationships by the
method of maximizing likelihood, user’s guide. Chicago: National Educational Resources.
Keran, M. (1971). Expectations, money, and stock market. Review, Federal Reserve Bank of
St. Louis, 53, 16-31.
Lahiri, K. (1976). Inflationary expectations: Their formation and interest rate effects.
The American Economic Review, 66, 124–131.
36 Multifactor, Multi-indicator Approach to Asset Pricing 1023
Lintner, J. (1965). The valuation of risk assets and the selection of risky investments in stock
portfolios and capital budgets. Review of Economics and Statistics, 47, 13–37.
Litzenberger, R., & Ramaswamy, K. (1979). The effects of personal taxes and dividends on capital
asset prices: Theory and empirical evidence. Journal of Financial Economics, 7, 163–195.
Maddala, G. S., & Nimalendran, M. (1996). Error-in-variables problems in financial models.
In G. S. Maddala & C. R. Rao (Eds.), Handbook of statistics. New York: Elsevier Science.
Mossin, J. (1966). Equilibrium in a capital asset market. Econometrica, 34, 768–783.
Nelson, C. (1976). Inflation and rate of returns on common stocks. Journal of Finance, 31, 471–483.
Nelson, C. (1979). Recursive structure in U.S. income, prices and output. Journal of Political
Economy, 87, 1307–1327.
Pesando, J. (1974). The supply of money and common stock prices: Further observations on
econometric evidence. Journal of Finance, 29, 909–921.
Roll, R. (1977). A critique of the asset pricing theory’s tests. Journal of Financial Economics, 4,
129–176.
Roll, R., & Ross, S. (1980). An empirical investigation of the arbitrage pricing theory. Journal of
Finance, 35, 1073–1103.
Ross, S. (1976). The arbitrage theory of capital asset pricing. Journal of Economic Theory, 13, 341–360.
Ross, S. (1977). Return, risk, and arbitrage. In I. Friend & J. L. Bicksler (Eds.), Risk and return in
finance (Vol. I). Ballinger Pub. Co.: Cambridge, MA.
Schipper, K., & Thompson, R. (1981). Common stocks as hedges against shifts in the consumption
on investment opportunity set. Journal of Business, 54, 305–328.
Shanken, J. (1982). The arbitrage pricing theory: Is it testable? Journal of Finance, 37, 1129–1140.
Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk.
Journal of Finance, 19, 425–442.
Solnik, B. (1983). The relation between stock prices and inflationary expectations: The interna-
tional evidence. Journal of Finance, 38, 35–48.
Stambough, R. (1982). On the exclusion of assets from tests of the two-parameter model:
A sensitivity analysis. Journal of Financial Economics, 10, 237–268.
Titman, S., & Wessels, R. (1988). The determinants of capital structure choice. Journal of
Finance, 43, 1–19.
Turnovsky, S. (1970). Empirical evidence on the formation of price expectations. Journal of the
American Statistical Association, 65, 1441–1454.
Wei, K. (1984). The arbitrage pricing theory versus the generalized intertemporal capital asset
pricing model: Theory and empirical evidence. Unpublished Ph.D. dissertation, University of
Illinois.
Yang, C. C., Lee, C. F., Gu, Y. X., & Lee, Y. W. (2009). Co-determinantion of capital structure
and stock returns – A LISREL approach: An empirical test of Taiwan stock markets. Quarterly
Review of Economics and Finance, 50, 222–233.
Zellner, A. (1970). Estimation of regression relationships counting un-observable variables.
International Economic Review, 11, 441–454.
Binomial OPM, Black–Scholes OPM,
and Their Relationship: Decision Tree 37
and Microsoft Excel Approach
John C. Lee
Contents
37.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1026
37.2 Call and Put Options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1027
37.3 One-Period Option Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1028
37.4 Two-Period Option Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1033
37.5 Using Microsoft Excel to Create the Binomial Option Trees . . . . . . . . . . . . . . . . . . . . . . . . . 1036
37.6 Black–Scholes Option Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1040
37.7 Relationship Between the Binomial OPM and the Black–Scholes OPM . . . . . . . . . . . . . 1044
37.8 Decision Tree Black–Scholes Calculation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1047
37.9 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1047
Appendix 1: Excel VBA Code: Binomial Option Pricing Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1047
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1059
Abstract
This chapter will first demonstrate how Microsoft Excel can be used to
create the decision trees for the binomial option pricing model. At the same
time, this chapter will discuss the binomial option pricing model in a less
mathematical fashion. All the mathematical calculations will be taken care by
the Microsoft Excel program that is presented in this chapter. Finally, this
chapter also uses the decision tree approach to demonstrate the relationship
between the binomial option pricing model and the Black–Scholes option
pricing model.
J.C. Lee
Center for PBBEF Research, North Brunswick, NJ, USA
e-mail: johnleejohnlee@yahoo.com; leeleeassociates@gmail.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1025
DOI 10.1007/978-1-4614-7750-1_37,
# Springer Science+Business Media New York 2015
1026 J.C. Lee
Keywords
Binomial option pricing model • Decision trees • Black–Scholes option pricing
model • Call option • Put option • Microsoft Excel • Visual Basic for Applica-
tions • VBA • Put-call parity • Sigma • Volatility • Recursive programming
37.1 Introduction
The binomial option pricing model derived by Rendleman and Barter (1979)
and Cox et al. (1979) is one the most famous models used to price options. Only
the Black–Scholes model (1973) is more famous. One problem with learning the
binomial option pricing model is that it is computationally intensive. This results in
a very complicated formula to price an option.
The complexity of the binomial option pricing model makes it a challenge to
learn the model. Most books teach the binomial option model by describing
the formula. This is not very effective because it usually requires the learner to
mentally keep track of many details, many times to the point of information
overload. There is a well-known principle in psychology that the average number
of things that a person can remember at one time is seven.
This chapter will first demonstrate the power of Microsoft Excel. It will do this
by demonstrating that it is possible to create large decision trees for the
binomial pricing model using Microsoft Excel. A ten-period decision tree would
require 2,047 call calculations and 2,047 put calculations. This chapter will also
show the decision tree for the price of a stock and the price of a bond, each requiring
2,047 calculations. Therefore, there would be 2,047 * 4 ¼ 8,188 calculations for
a complete set of ten-period decision trees.
Secondly, this chapter will present the binomial option model in a less mathe-
matical matter. It will try to make it so that the reader will not have to keep track of
many things at one time. It will do this by using decision trees to price call and put
options.
Finally, this chapter will show the relationship between the binomial option
pricing model and the Black–Scholes option pricing model.
This chapter uses a Microsoft Excel workbook called binomialBS_OPM.xls
that contains the VBA code to create the decision trees for the binomial option
pricing model. The VBA code is published in Appendix 1. The password for the
workbook is bigsky for those who want to study the VBA code. E-mail me at
JohnLeeExcelVBA@gmail.com and indicate that the password is “bigsky” to get
a copy of this Microsoft Excel workbook.
Section 37.2 discusses the basic concepts of call and put options. Section 37.3
demonstrates the one-period call and put option pricing models. Section 37.4
presents the two-period option pricing model. Section 37.5 demonstrates how to
use the Microsoft Excel workbook binomialBS_OPM.xls to create the decision trees
for an n-period binomial option pricing model. Section 37.6 demonstrates the use
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1027
Fig. 37.1 Value of AMZN Value of AMZN Call Option Strike Price = $200
call option 900
800
700
600
Value
500
400
300
200
100
0
0 200 300 400 500 600 700 800 900 1000
Strike Price
of the Black–Scholes model. Section 37.7 shows the relationship between the
binomial option pricing model and the Black–Scholes option pricing model.
Section 37.8 demonstrates how to use the Microsoft Excel workbook
binomialBS_OPM.xls to demonstrate the relationship between the binomial option
pricing model and the Black–Scholes option pricing model.
A call option gives the owner the right but not the obligation to buy the
underlying security at a specified price. The price in which the owner can buy the
underlying price is called the exercise price. A call option becomes valuable when
the exercise price is less than the current price of the underlying stock price.
For example, a call option on an AMZN stock with an exercise price of $200
when the stock price of an Amazon stock is $250 is worth $50. The reason
it is worth $50 is because a holder of the call option can buy the AMZN stock at
$200 and then sell the AMZN stock at the prevailing price of $250 for a profit of
$50. Also, a call option on an AMZN stock with an exercise price of $300 when the
stock price of an AMZN stock is $150 is worth $0.
A put option gives the owner the right but not the obligation to sell the
underlying security at a specified price. A put option becomes valuable when the
exercise price is more than the current price of the underlying stock price.
For example, a put option on an AMZN stock with an exercise price of $200
when the stock price of an AMZN stock is $150 is worth $50. The reason it is
worth $50 is because a holder of the put option can buy the AMZN stock at the
prevailing price of $150 and then sell the AMZN stock at the put price of $200
for a profit of $50. Also, a put option on an AMZN stock with an exercise price of
$200 when the stock price of the AMZN stock is $250 is worth $0.
Figures 37.1 and 37.2 are charts showing the value of call and put options of the
above GE stock at varying prices.
1028 J.C. Lee
Fig. 37.2 Value of AMZN Value of AMZN Put Option Strike Price = $200
put option 250
200
150
Value
100
50
0
0 50 100 150 200 250 300 350 400 450
Strike Price
What should be the value of these options? Let’s look at a case where we are only
concerned with the value of options for one period. In the next period, a stock price
can either go up or go down. Let’s look at a case where we know for certain that an
AMZN stock with a price of $200 will either go up 5 % or go down 5 % in the next
period and the exercise after one period is $200. Figures 37.3, 37.4, and 37.5 show
the decision tree for the AMZN stock price, the AMZN call option price, and the
AMZN put option price, respectively.
Let’s first consider the issue of pricing an AMZN call option. Using a one-period
decision tree, we can illustrate the price of an AMZN stock if it goes up 5 % and the
price of a stock AMZN if it goes down 5 %. Since we know the possible ending
values of the AMZN stock, we can derive the possible ending values of a call
option. If the stock price increases to $210, the price of the AMZN call option will
then be $10 ($210 $200). If the AMZ stock price decreases to $190, the value of
the call option will worth $0 because it would be below the exercise price of $200.
We have just discussed the possible ending value of an AMZN call option in
period 1. But what we are really interested in is what the value is now of the
AMZN call option knowing the two resulting values of the AMZN call option.
To help determine the value of a one-period AMZN call option, it’s useful to
know that it is possible to replicate the resulting two states of the value of the
AMZN call option by buying a combination of stocks and bonds. Below is the
formula to replicate the situation where the price increases to $210. We will assume
that the interest rate for the bond is 3 %:
210S þ 1:03B ¼ 10
190S þ 1:03B ¼ 0
We can use simple algebra to solve for both S and B. The first thing that we need
to do is to rearrange the second equation as follows:
1:03B ¼ 190S
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1029
Period 0 Period 1
10
??
Fig. 37.4 AMZN call option
price 0
Period 0 Period 1
0
??
Fig. 37.5 AMZN put option
price 10
210S þ ð190SÞ ¼ 10
20S ¼ 10
S ¼ 0:5
We can solve for B by substituting the value 0.05 for S in the first equation:
210ð0:5Þ þ 1:03B ¼ 10
105 þ 1:03B ¼ 10
1:03B ¼ 95
B ¼ 92:23
Therefore, from the above simple algebraic exercise, we should at period 0 buy
0.05 shares of AMZN stock and borrow 9.223 at 3 % to replicate the payoff
of the AMZN call option. This means the value of an AMZN call option should
be 0.5 * 200 92.23 ¼ 7.77.
If this were not the case, there would then be arbitrage profits. For example, if
the call option were sold for $30, there would be a profit of 22.23. This would result
in the increase in the selling of the AMZN call option. The increase in the supply of
AMZN call options would push the price down for the call options. If the call option
were sold for $5, there would be a saving of $2.77. This saving would result in the
increase demand for the AMZN call option. This increase demand would result in
the price of the call option to increase. The equilibrium point would be $7.77.
1030 J.C. Lee
Using the above mentioned concept and procedure, Benninga (2000) has derived
a one-period call option model as
where
id
qu ¼
ð1 þ i Þðu d Þ
ui
qd ¼
ð1 þ i Þðu d Þ
u ¼ increase factor
d ¼ down factor
i ¼ interest rate
If we let i ¼ r, p ¼ (r d)/(u d), 1 p ¼ (u r)/(u d), R ¼ 1/(1 + r),
Cu ¼ Max[S(1 + u) X, 0], and Cd ¼ Max[S(1 + d) X, 0], then we have
where
Cu ¼ call option price after increase
Cd ¼ call option price after decrease
Equation 37.2 is identical to Eq. 6B.6 in Lee et al. (2000, p. 234).1
Below calculates the value of the above one-period call option where the strike
price, X, is $200 and the risk-free interest rate is 3 %. We will assume that the price
of a stock for any given period will either increase or decrease by 5 %:
X ¼ $200
S ¼ $200
u ¼ 1:05
d ¼ 0:95
R ¼ 1 þ r ¼ 1 þ 0:03
p ¼ ð1:03 0:95Þ=ð1:05 0:95Þ
C ¼ ½0:8ð10Þ þ 0:2ð0Þ=1:03 ¼ $7:77
Therefore, from the above calculations, the value of the call option is $7.77.
Figure 37.6 shows the resulting decision tree for the above call option.
1
Please note that in Lee et al. (2000, p. 234) u ¼ 1 + percentage of price increase and
d ¼ 1 – percentage of price increase.
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1031
Like the call option, it is possible to replicate the resulting two states of the value
of the put option by buying a combination of stocks and bonds. Below is the
formula to replicate the situation where the price decreases to $190:
210S þ 1:03B ¼ 0
190S þ 1:03B ¼ 10
We will use simple algebra to solve for both S and B. The first thing we will do is
to rewrite the second equation as follows:
1:03B ¼ 10 190S
The next thing to do is to substitute the above equation to the first put option
equation. Doing this would result in the following:
210S þ 10 190S ¼ 0
Now let us solve for B by putting the value of S into the first equation. This is
shown below:
210ð0:5Þ þ 1:03B ¼ 0
1:03B ¼ 105
B ¼ 101:94
From the above simple algebra exercise, we have S ¼ 0.5 and B ¼ 101.94.
This tells us that we should in period 0 lend $101.94 at 3 % and sell 0.5 shares of
stock to replicate the put option payoff for period 1. And the value of the AMZN put
option should be 200(0.5) + 101.94 ¼ 1.94.
Using the same arbitrage argument that we used in the discussion of the call
option, 0.194 has to be the equilibrium price of the put option.
As with the call option, Benninga (2000) has derived a one-period put option
model as
where
id
qu ¼
ð1 þ i Þðu d Þ
ui
qd ¼
ð1 þ i Þðu d Þ
u ¼ increase factor
d ¼ down factor
i ¼ interest rate
where
Pu ¼ put option price after increase
Pd ¼ put option price after decrease
Below calculates the value of the above one-period put option where the strike
price, X, is $20 and the risk-free interest rate is 3 %:
From the above calculation, the put option pricing decision tree would look like
the following.
Figure 37.7 shows the resulting decision tree for the above put option.
There is a relationship between the price of a put option and the price of all call
option. This relationship is called the put-call parity. Equation 37.5 shows the
relationship between the price of a put option and the price of a call option:
P ¼ C þ X=R S (37.5)
where
C ¼ call price
X ¼ strike price
R ¼ 1 + interest rate
S ¼ stock price
Period 0 Period 1
0
1.940
Fig. 37.7 AMZN put option
10
price
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1033
The following uses the put-call parity to calculate the price of the AMZN put
option:
P ¼ $7:77 þ $200=ð1:03Þ $200
¼ 7:77 þ 194:17 200
¼ 1:94
We now will look at pricing options for two periods. Figure 37.8 shows the
stock price decision tree based on the parameters indicated in the last section.
This decision tree was created based on the assumption that a stock price will either
increase by 5 % or decrease by 5 %.
How do we price the value of a call and put option for two periods?
The highest possible value for our stock based on our assumption is $220.5.
We get this value first by multiplying the stock price at period 0 by 105 % to get the
resulting value of $210 of period 1. We then again multiply the stock price in
period 1 by 105 % to get the resulting value of $220.5. In period two, the value
of a call option when a stock price is $220.5 is the stock price minus the exercise
price, $220.5 $200, or $20.5. In period two, the value of a put option when a stock
price is $220.5 is the exercise price minus the stock price, $200 $220.5,
or $20.5. A negative value has no value to an investor so the value of the put option
would be $0.
The lowest possible value for our stock based on our assumptions is $180.5.
We get this value first by multiplying the stock price at period 0 by 95 %
(decreasing the value of the stock by 5 %) to get the resulting value of $190 of
period 1. We then again multiply the stock price in period 1 by 95 % to get the
resulting value of $180.5. In period two, the value of a call option when a stock
price is $180.5 is the stock price minus the exercise price, $180.5 $200, or
$19.5. A negative value has no value to an investor so the value of a call option
would be $0. In period two, the value of a put option when a stock price is $18.05 is
the exercise price minus the stock price, $200 $180.5, or $19.5. We can derive the
call and put option values for the other possible values of the stock in period 2 in
the same fashion.
Figures 37.9 and 37.10 show the possible call and put option values for period 2.
We cannot calculate the value of the call and put options in period 1 the
same way we did in period 2 because it’s not the ending value of the stock.
In period 1, there are two possible call values. One value is when the stock price
increased, and one value is when the stock price decreased. The call option decision
tree shown in Fig. 37.9 shows two possible values for a call option in period 1. If we
just focus on the value of a call option when the stock price increases from period
one, we will notice that it is like the decision tree for a call option for one period.
This is shown in Fig. 37.11.
1034 J.C. Lee
0.0
0.0
0.5
0.5
20.5000
0
Fig. 37.11 AMZN call
option
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1035
Using the same method for pricing a call option for one period, the price of a call
option when stock price increases from period 0 will be $15.922. The resulting
decision tree is shown in Fig. 37.12.
In the same fashion, we can price the value of a call option when a stock
price decreases. The price of a call option when a stock price decreases from
period 0 is $0. The resulting decision tree is shown in Fig. 37.13.
In the same fashion, we can price the value of a call option in period 0.
The resulting decision tree is shown in Fig. 37.14.
20.5000
15.9220
0
0
0
0
Fig. 37.13 AMZN call
option
20.5000
15.9220
0.0000
12.3670
0.0000
0.0000
0.0000
Fig. 37.14 AMZN call
option
1036 J.C. Lee
We can calculate the value of a put option in the same manner as we did
in calculating the value of a call option. The decision tree for a put option is
shown in Fig. 37.15.
In the previous section, we priced the value of a call and put option by pricing
backwards, from the last period to the first period. This method of pricing call and
put options will work for any n-period. To price the value of a call option for
two periods required seven sets of calculations. The number of calculations
increases dramatically as n increases. Table 37.1 lists the number of calculations
for specific number of periods.
After two periods, it becomes very cumbersome to calculate and create the
decision trees for a call and put option. In the previous section, we saw that
calculations were very repetitive and mechanical. To solve this problem,
this chapter will use Microsoft Excel to do the calculations and create the
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1037
decision trees for the call and put options. We will also use Microsoft Excel to
calculate and draw the related decision trees for the underlying stock and bond.
To solve this repetitive and mechanical calculation of the binomial option
pricing model, we will look at a Microsoft Excel file called binomialBS_OPM.
xls. We will use this Microsoft Excel workbook to produce four decision trees
for the GE stock that was discussed in the previous sections. The four decision
trees are:
1. Stock price
2. Call option price
3. Put option price
4. Bond price
This section will demonstrate how to use the binomialBS_OPM.xls Excel file to
create the four decision trees. Figure 37.16 shows the Excel file binomialBS_OPM.
xls after the file is opened. Pushing the button shown in Fig. 37.16 will get the
dialog box shown in Fig. 37.17.
The dialog box shown in Fig. 37.17 shows the parameters for the binomial
option pricing model. These parameters are changeable. The dialog box in
Fig. 37.17 shows the default values.
Pushing the calculate button shown in Fig. 37.17 will produce the four decision
trees shown in Figs. 37.18, 37.19, 37.20, and 37.21.
The table at the beginning of this section indicated 31 calculations were required
to create a decision tree that has four periods. This section showed four decision
trees. Therefore, the Excel file did 31 * 4 ¼ 121 calculations to create the four
decision trees.
Benninga (2000, p. 260) has defined the price of a call option in a binomial
option pricing model with n-periods as
Xn h i
n i ni
C¼ qu qd max Sð1 þ uÞi ð1 þ dÞni , 0 (37.6)
i¼0
i
and the price of a put option in a binomial option pricing model with n-periods as
X n h i
n i ni
P¼ qu qd max X Sð1 þ uÞi ð1 þ dÞni , 0 (37.7)
i¼0
i
Lee et al. (2000, p. 237) have defined the pricing of a call option in a binomial
option pricing model with n-period as
1 X n
n! h i
C¼ n pk ð1 pÞnk max 0, ð1 þ uÞk ð1 þ dÞnk , S X (37.8)
R k¼0 k!ðn k!Þ
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1039
The definition of the pricing of a put option in a binomial option pricing model
with n-period would then be defined as
1 X n
n! h i
P¼ n pk ð1 pÞnk max 0, X ð1 þ uÞk ð1 þ dÞnk , S (37.9)
R k¼0 k!ðn kÞ!
1040 J.C. Lee
The most famous option pricing model is the Black–Scholes option pricing model.
In this section, we will demonstrate the usage of the Black–Scholes option pricing
model. In latter sections, we will demonstrate the relationship between the binomial
option pricing model and the Black–Scholes pricing model. The Black–Scholes
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1041
model prices European call and put options. The Black–Scholes model for
a European call option is
where
C ¼ call price
S ¼ stock price
r ¼ risk-free interest rate
T ¼ time to maturity of option in years
N ¼ standard normal distribution
s ¼ stock volatility
s2
lnðS=XÞ þ r þ T
2
d1 ¼ pffiffiffi
s T
pffiffiffi
d2 ¼ d1 s T
Let’s manually calculate the price of a European call option in terms of Eq. 37.10
with the following parameter values, S ¼ 200, X ¼ 200, r ¼ 3 %, T ¼ 4, and
s ¼ 20 %.
Solution
s2 :22
lnðS=XÞ þ r þ T lnð200=200Þ þ :03 þ ð4Þ
2 2 ð:03 þ :02Þ 4 :2
d1 ¼ pffiffiffi ¼ pffiffiffi ¼ ¼ ¼ :5,
s T :2 4 :4 :4
pffiffiffi
d2 ¼ :5 :2 4 ¼ :1
P ¼ C S þ XerT
We can use either the binomial model or Black–Scholes to price an option. They
both should result in similar numbers. If we look at the parameters in both models,
we will notice that the binomial model has an Increase Factor (U), a Decrease
Factor (D), and n-period parameters that the Black–Scholes model does not have.
We also notice that the Black–Scholes model has the s and T parameters that the
binomial model does not have. Benninga (2008) suggests the following translation
between the binomial and Black–Scholes parameters:
pffiffiffiffi pffiffiffiffi
Dt ¼ T=n R ¼ erDt U ¼ es Dt D ¼ es Dt
RD UR
qU ¼ , qD ¼
RðU DÞ RðU DÞ
where
U ¼ 1 + percentage of price increase
D ¼ 1 percentage of price increase
R ¼ 1 + interest rate
We will now use the BinomialBS_OPM.xls Excel file to calculate the binomial and
Black–Scholes call and put values illustrated in Sect. 37.5. Notice that in Fig. 37.22
the Binomial Black–Scholes Approximation check box is checked. Checking this
box will cause T and Sigma parameters to appear and will adjust the Increase
Factor – u and Decrease Factor – d parameters. The adjustment was done as
indicated in Sect. 37.7.
Notice in Figs. 37.23 and 37.24 that the binomial option pricing model value
does not agree with the Black–Scholes option pricing model. The binomial OPM
value will get very close to the Black–Scholes OPM value once the binomial
parameter n gets very large. Benninga (2008) demonstrated that the
binomial value will be close to the Black–Scholes when the binomial n parameter
gets larger than 500.
37.9 Summary
This chapter demonstrated, with the aid of Microsoft Excel and decision trees, the
binomial option model in a less mathematical fashion. This chapter allowed the
reader to focus more on the concepts by studying the associated decision trees,
which were created by Microsoft Excel. This chapter also demonstrates that using
Microsoft Excel releases the reader from the computation burden of the binomial
option model.
This chapter also published the Microsoft Excel Visual Basic for Application
(VBA) code that created the binomial option decision trees. This allows for those
who are interested in studying the many advance Microsoft Excel VBA program-
ming concepts that were used to create the decision trees. One major computer
science programming concept used by the Excel VBA program in this chapter is
recursive programming. Recursive programming is the ideal of a procedure
calling itself many times. Inside the procedure, there are statements to decide
when not to call itself.
This chapter also used decision trees to demonstrate the relationship between the
binomial option pricing model and the Black–Scholes option pricing model.
It is important to note that the thing that makes Microsoft Excel powerful is
that it offers a powerful professional programming language called Visual
Basic for Applications (VBA). This section shows the VBA code that
1048 J.C. Lee
generated the decision trees for the binomial option pricing model. This code is
in the form frmBinomiaOption. The procedure cmdCalculate_Click is the first
procedure to run.
’/**************************************************
’/ Relationship Between the Binomial OPM
’/ and Black-Scholes OPM:
’/ Decision Tree and Microsoft Excel Approach
’/
’/ by John Lee
’/ JohnLeeExcelVBA@gmail.com
’/ All Rights Reserved
’/**************************************************
Option Explicit
Dim mwbTreeWorkbook As Workbook
Dim mwsTreeWorksheet As Worksheet
Dim mwsCallTree As Worksheet
Dim mwsPutTree As Worksheet
Dim mwsBondTree As Worksheet
Dim mdblPFactor As Double
Dim mBinomialCalc As Long
Dim mCallPrice As Double ’jcl 12/8/2008
Dim mPutPrice As Double ’jcl 12/8/2008
’/**************************************************
’/Purpose: Keep track the numbers of binomial calc
’/**************************************************
Property Let BinomialCalc(l As Long)
mBinomialCalc ¼ l
End Property
Property Get BinomialCalc() As Long
BinomialCalc ¼ mBinomialCalc
End Property
Property Set TreeWorkbook(wb As Workbook)
Set mwbTreeWorkbook ¼ wb
End Property
Property Get TreeWorkbook() As Workbook
Set TreeWorkbook ¼ mwbTreeWorkbook
End Property
Property Set TreeWorksheet(ws As Worksheet)
Set mwsTreeWorksheet ¼ ws
End Property
Property Get TreeWorksheet() As Worksheet
Set TreeWorksheet ¼ mwsTreeWorksheet
End Property
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1049
End If
End Sub
Private Sub txtTimeT_Change()
’jcl 12/8/2008
On Error Resume Next
If Me.chkBinomialBSApproximation Then
Me.txtBinomialU ¼ Exp(Me.txtSigma * Sqr(Me.txtTimeT /
Me.txtBinomialN))
Me.txtBinomialD ¼ Exp(-Me.txtSigma * Sqr(Me.txtTimeT /
Me.txtBinomialN))
End If
End Sub
Private Sub UserForm_Initialize()
With Me
.txtBinomialS ¼ 20
.txtBinomialX ¼ 20
.txtBinomialD ¼ 0.95
.txtBinomialU ¼ 1.05
.txtBinomialN ¼ 4
.txtBinomialr ¼ 0.03
.txtSigma ¼ 0.2
.txtTimeT ¼ 4
Me.chkBinomialBSApproximation ¼ False
End With
chkBinomialBSApproximation_Click
Me.Hide
End Sub
Sub BinomialOption()
Dim wbTree As Workbook
Dim wsTree As Worksheet
Dim rColumn As Range
Dim ws As Worksheet
Set Me.TreeWorkbook ¼ Workbooks.Add
Set Me.BondTree ¼ Me.TreeWorkbook.Worksheets.Add
Set Me.PutTree ¼ Me.TreeWorkbook.Worksheets.Add
Set Me.CallTree ¼ Me.TreeWorkbook.Worksheets.Add
Set Me.TreeWorksheet ¼ Me.TreeWorkbook.Worksheets.Add
Set rColumn ¼ Me.TreeWorksheet.Range("a1")
With Me
.BinomialCalc ¼ 0
.PFactor ¼ Me.txtBinomialr
.CallTree.Name ¼ "Call Option Price"
.PutTree.Name ¼ "Put Option Price"
.TreeWorksheet.Name ¼ "Stock Price"
.BondTree.Name ¼ "Bond"
1052 J.C. Lee
End With
DecisionTree rCell:¼rColumn, nPeriod:¼Me.txtBinomialN
+ 1, _
dblPrice:¼Me.txtBinomialS, sngU:¼Me.txtBinomialU, _
sngD:¼Me.txtBinomialD
DecitionTreeFormat
TreeTitle wsTree:¼Me.TreeWorksheet, sTitle:¼"Stock
Price "
TreeTitle wsTree:¼Me.CallTree, sTitle:¼"Call Option
Pricing"
TreeTitle wsTree:¼Me.PutTree, sTitle:¼"Put Option
Pricing"
TreeTitle wsTree:¼Me.BondTree, sTitle:¼"Bond Pricing"
Application.DisplayAlerts ¼ False
For Each ws In Me.TreeWorkbook.Worksheets
If Left(ws.Name, 5) ¼ "Sheet" Then
ws.Delete
Else
ws.Activate
ActiveWindow.DisplayGridlines ¼ False
ws.UsedRange.NumberFormat ¼ "#,##0.0000_);(#,##0.0000)"
End If
Next
Application.DisplayAlerts ¼ True
Me.TreeWorksheet.Activate
End Sub
Sub TreeTitle(wsTree As Worksheet, sTitle As String)
wsTree.Range("A1:A5").EntireRow.Insert (xlShiftDown)
With wsTree
With.Cells(1)
.Value ¼ sTitle
.Font.Size ¼ 20
.Font.Italic ¼ True
End With
With.Cells(2, 1)
.Value ¼ "Decision Tree"
.Font.Size ¼ 16
.Font.Italic ¼ True
End With
With.Cells(3, 1)
.Value ¼ "Price ¼ " & Me.txtBinomialS & _
",Exercise ¼ " & Me.txtBinomialX & _
",U ¼ " & Format(Me.txtBinomialU, "#,##0.0000") & _
",D ¼ " & Format(Me.txtBinomialD, "#,##0.0000") & _
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1053
End Sub
Sub BondDecisionTree(rPrice As Range, arCell As Variant,
iCount As Long)
Dim rBond As Range
Dim rPup As Range
Dim rPDown As Range
Set rBond ¼ Me.BondTree.Cells(rPrice.Row, rPrice.
Column)
Set rPup ¼ Me.BondTree.Cells(arCell(iCount - 1).Row,
arCell(iCount - 1).Column)
Set rPDown ¼ Me.BondTree.Cells(arCell(iCount).Row,
arCell(iCount).Column)
If rPup.Column ¼ Me.TreeWorksheet.UsedRange.Columns.
Count Then
rPup.Value ¼ (1 + Me.txtBinomialr) ^ (rPup.Column - 1)
rPDown.Value ¼ rPup.Value
End If
With rBond
.Value ¼ (1 + Me.txtBinomialr) ^ (rBond.Column - 1)
.Borders(xlBottom).LineStyle ¼ xlContinuous
End With
rPDown.Borders(xlBottom).LineStyle ¼ xlContinuous
With rPup
.Borders(xlBottom).LineStyle ¼ xlContinuous
.Offset(1, 0).Resize((rPDown.Row - rPup.Row), 1). _
Borders(xlEdgeLeft).LineStyle ¼ xlContinuous
End With
End Sub
Sub PutDecisionTree(rPrice As Range, arCell As Variant,
iCount As Long)
Dim rCall As Range
Dim rPup As Range
Dim rPDown As Range
Set rCall ¼ Me.PutTree.Cells(rPrice.Row, rPrice.Column)
Set rPup ¼ Me.PutTree.Cells(arCell(iCount - 1).Row,
arCell(iCount - 1).Column)
Set rPDown ¼ Me.PutTree.Cells(arCell(iCount).Row,
arCell(iCount).Column)
If rPup.Column ¼ Me.TreeWorksheet.UsedRange.Columns.
Count Then
rPup.Value ¼ WorksheetFunction.Max(Me.txtBinomialX -
arCell(iCount - 1), 0)
rPDown.Value ¼ WorksheetFunction.Max(Me.txtBinomialX -
arCell(iCount), 0)
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1055
End If
With rCall
’12/10/2008
If Not Me.chkBinomialBSApproximation Then
.Value ¼ (Me.PFactor * rPup + (1 - Me.PFactor) * rPDown) /
(1 + Me.txtBinomialr)
Else
.Value ¼ (Me.qU * rPup) + (Me.qD * rPDown)
End If
Me.PutPrice ¼.Value ’12/8/2008
.Borders(xlBottom).LineStyle ¼ xlContinuous
End With
rPDown.Borders(xlBottom).LineStyle ¼ xlContinuous
With rPup
.Borders(xlBottom).LineStyle ¼ xlContinuous
.Offset(1, 0).Resize((rPDown.Row - rPup.Row), 1). _
Borders(xlEdgeLeft).LineStyle ¼ xlContinuous
End With
End Sub
Sub CallDecisionTree(rPrice As Range, arCell As Variant,
iCount As Long)
Dim rCall As Range
Dim rCup As Range
Dim rCDown As Range
Set rCall ¼ Me.CallTree.Cells(rPrice.Row, rPrice.
Column)
Set rCup ¼ Me.CallTree.Cells(arCell(iCount - 1).Row,
arCell(iCount - 1).Column)
Set rCDown ¼ Me.CallTree.Cells(arCell(iCount).Row,
arCell(iCount).Column)
If rCup.Column ¼ Me.TreeWorksheet.UsedRange.Columns.
Count Then
With rCup
.Value ¼ WorksheetFunction.Max(arCell(iCount - 1) - Me.
txtBinomialX, 0)
.Borders(xlBottom).LineStyle ¼ xlContinuous
End With
With rCDown
.Value ¼ WorksheetFunction.Max(arCell(iCount) - Me.
txtBinomialX, 0)
.Borders(xlBottom).LineStyle ¼ xlContinuous
End With
End If
With rCall
1056 J.C. Lee
End Sub
’/**************************************************
’/Purpse: To calculate the price value of every state
of the binomial
’/ decision tree
’/**************************************************
Sub DecisionTree(rCell As Range, nPeriod As Integer, _
dblPrice As Double, sngU As Single, sngD As Single)
Dim lIteminColumn As Long
If Not nPeriod ¼ 1 Then
’Do Up
DecisionTree rCell:¼rCell.Offset(0, 1), nPeriod:¼
nPeriod - 1, _
dblPrice:¼dblPrice * sngU, sngU:¼sngU, _
sngD:¼sngD
’Do Down
DecisionTree rCell:¼rCell.Offset(0, 1), nPeriod:¼
nPeriod - 1, _
dblPrice:¼dblPrice * sngD, sngU:¼sngU, _
sngD:¼sngD
End If
lIteminColumn ¼ WorksheetFunction.CountA(rCell.
EntireColumn)
If lIteminColumn ¼ 0 Then
rCell ¼ dblPrice
Else
If nPeriod <> 1 Then
rCell.EntireColumn.Cells(lIteminColumn + 1) ¼ dblPrice
Else
rCell.EntireColumn.Cells(((lIteminColumn + 1) * 2) -
1) ¼ dblPrice
Application.StatusBar ¼ "The number of binomial calcs are
: " & Me.BinomialCalc _ & " at cell " & rCell.EntireColumn.
Cells(((lIteminColumn + 1) * 2) - 1).Address
End If
End If
Me.BinomialCalc ¼ Me.BinomialCalc + 1
End Sub
Function BS_D1() As Double
Dim dblNumerator As Double
Dim dblDenominator As Double
On Error Resume Next
dblNumerator ¼ VBA.Log(Me.txtBinomialS / Me.
txtBinomialX) + _
37 Binomial OPM, Black–Scholes OPM, and Their Relationship 1059
References
Benninga, S. (2000). Financial modeling. Cambridge: MIT Press.
Benninga, S. (2008). Financial modeling. Cambridge: MIT Press.
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of
Political Economy, 31, 637–659.
Cox, J., Ross, S. A., & Rubinstein, M. (1979). Option pricing: A simplified approach. Journal of
Financial Economics, 7, 229–263.
Daigler, R. T. (1994). Financial futures and options markets concepts and strategies. New York:
Harper Collins.
Jarrow, R., & TurnBull, S. (1996). Derivative securities. Cincinnati: South-Western College
Publishing.
Lee, J. C. (2001). Using Microsoft excel and decision trees to demonstrate the binomial option
pricing model. Advances in Investment Analysis and Portfolio Management, 8, 303–329.
Lee, C. F. (2009). Handbook of quantitative finance. New York: Springer.
Lee, C. F., & Lee, A. C. (2006). Encyclopedia of finance. New York: Springer.
Lee, C. F., Lee, J. C., & Lee, A. C. (2000). Statistics for business and financial economics. New
Jersey: World Scientific.
Lee, J. C., Lee, C. F., Wang, R. S., & Lin, T. I. (2004). On the limit properties of binomial
and multinomial option pricing models: Review and integration. In C. F. Lee (Ed.), Advances
in quantitative analysis of finance and accounting new series (Vol. 1). Singapore: World
Scientific.
Rendleman, R. J., Jr., & Barter, B. J. (1979). Two-state option pricing. Journal of Finance, 34(5),
1093–1110.
Walkenbach, J. (2010). Excel 2010 power programming with VBA. Indianapolis: Wiley.
Wells, E., & Harshbarger, S. (1997). Microsoft excel 97 developer’s handbook. Redmond:
Microsoft Press.
Dividend Payments and Share
Repurchases of US Firms: 38
An Econometric Approach
Alok Bhargava
Contents
38.1 Introduction and Background . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1063
38.2 Specific Hypotheses, Financial Databases, and the Formulation of
Econometric Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1065
38.3 Analytical Framework for Modeling the Dividends and
Share Repurchases Interrelationships . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1067
38.3.1 Background Issues . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1067
38.3.2 Some Conceptual Aspects of Dividends and Share Repurchase
Interrelationships . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1067
38.4 Processing the Compustat Database from the United States . . . . . . . . . . . . . . . . . . . . . . . . . 1068
38.5 Comprehensive Empirical Models for Dividends and Share Repurchase
Interrelationships . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1069
38.6 The Econometric Framework for Addressing Simultaneity and
Between-Firm Heterogeneity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1071
38.7 Descriptive Statistics from the Annual Compustat Database . . . . . . . . . . . . . . . . . . . . . . . . 1072
38.8 Results from Simple Dynamic Models for Dividends and
Share Repurchase Controlling for Between-Firm Heterogeneity . . . . . . . . . . . . . . . . . . . . 1075
38.9 Results from Comprehensive Dynamic Models for Dividends and
Share Repurchases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1078
38.9.1 Results from Dynamic and Static Models for Firms’
Dividends per Share . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1078
38.9.2 Results from Dynamic and Static Models for Firms’
Share Repurchases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1080
38.9.3 Results from Dynamic and Static Models for Dividends
per Share with Share Repurchases and Intangible Assets Included as
Explanatory Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1082
38.9.4 Results from Dynamic and Static Models for Share Repurchases with
Intangible Assets Included as Explanatory Variables . . . . . . . . . . . . . . . . . . . . . . 1085
A. Bhargava
School of Public Policy, University of Maryland, College Park, MD, USA
e-mail: bhargava@umd.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1061
DOI 10.1007/978-1-4614-7750-1_38,
# Springer Science+Business Media New York 2015
1062 A. Bhargava
Abstract
The analyses of dividends paid by firms and decisions to repurchase their own
shares require an econometric approach because of the complex dynamic
interrelationships. This chapter begins by, first, highlighting the importance
of developing comprehensive econometric models for these interrelation-
ships. It is common in finance research to spell out “specific hypotheses”
and conduct empirical research to investigate validity of the hypotheses.
However, such an approach can be misleading in situations where variables
are simultaneously determined as is often the case in financial applications.
Second, financial and accounting databases such as Compustat are complex
and contain useful longitudinal information on variables that display
considerable heterogeneity across the firms. Empirical analyses of financial
databases demand the use of econometric and computational methods in
order to draw robust inferences. For example, using longitudinal data on the
same US firms, it was found that dividends were neither “disappearing”
nor “reappearing” but were relatively stable in the period 1992–2007
Bhargava (Journal of the Royal Statistical Society A, 173, 631–656, 2010).
Third, the econometric methodology tackled the dynamics of relationships
and investigated endogeneity of certain explanatory variables. Identification
of the model parameters is achieved in such models by exploiting the
cross-equations restrictions on the coefficients in different time periods.
Moreover, the estimation entails using nonlinear optimization methods
to compute the maximum likelihood estimates of the dynamic random
effects models and for testing statistical hypotheses using likelihood ratio
tests. For example, share repurchases were treated as endogenous
explanatory variable in the models for dividend payments, and dividends
were treated as endogenous variables in the models for share repurchases.
The empirical results showed that dividends are decided quarterly at the
first stage, and higher dividends payments lowered share repurchases
by firms that are made at longer intervals. These findings cast
some doubt on evidence for the simple “substitution” hypothesis
between dividends and share repurchases. The appendix outlines some of
the econometric estimation techniques and tests that are useful for research
in finance.
Keywords
Compustat database • Corporate policies • Dividends • Dynamic random effects
models • Econometric methodology • Endogeneity • Maximum likelihood •
Intangible assets • Model formulation • Nonlinear optimization • Panel data •
Share repurchases
38 Dividend Payments and Share Repurchases of US Firms 1063
employed. This chapter discusses several difficulties in the previous literature and
presents findings on the interrelationships between dividends and share repurchases
by over 2,000 US industrial firms (see Bhargava 2010).
First, it is essential to model the time structure of the interrelationships between
dividends and share repurchases using longitudinal data on firms. While Lintner
(1956) argued that firms may be reluctant to alter dividend rates, firms often adjust
dividends per share to the market circumstances. Thus, classification of firms
between dividend “payers” and “non-payers” is not a concrete one since a majority
of firms may pay dividends during some periods. For example, using eight 2-yearly
averages of dividends paid by a panel of 3,154 US industrial firms in the period
1992–2007 (i.e., for 1992–1993, 1994–1995, 1996–1997, 1998–1999, 2000–2001,
2002–2003, 2004–2005, 2006–2007), Bhargava (2010) found that 42 % of the firms
never paid dividends, 33 % paid dividends in all 8 periods, and 25 % paid dividends
in 1–7 periods. Moreover, dividends per share ranged from $0.00 to $67.61 so that it
is appropriate to model the payments via continuous random variables (Fama and
Babiak 1968). Similarly, firms’ share repurchase are continuous variables and can be
defined using the variables “Treasury stock,” “purchase of common stock,” and “sale
of common stock.” Using longitudinal data on 2,907 firms in 8 periods, 20 % of firms
made no repurchases, 5 % repurchased shares in all 8 periods, and 75 % made 1–7
repurchases (Bhargava 2010). Thus, models with continuous dependent variables for
dividends and share repurchases are suitable for analyses of payout decisions.
Moreover, it is essential to use longitudinal data on the same firms for analyzing
the interrelationships between dividends and share repurchases.
Second, from a methodological standpoint, the estimation of dynamic models
from longitudinal data covering large number of firms observed for a few time
periods requires the treatment of lagged dependent variables as “endogenous”
variables that are correlated with firm-specific random effects (Anderson and
Hsiao 1981; Bhargava and Sargan 1983); this issue will be discussed below and
in the Appendix. Despite the popularity of Lintner (1956) type models in finance
literature, such issues have seldom been addressed by researchers and critically
affect consistency properties of the estimated model parameters. Moreover, it is
informative to model the actual magnitudes of dividends and share repurchases,
since these are likely to be related to firm characteristics such as size, earnings, and
debt. It would also be useful to consider their joint determination; models for
continuous dependent variables using longitudinal data can handle situations
where some explanatory variables are correlated with errors affecting the models
(Bhargava and Sargan 1983; Bhargava 1991).
Third, there has been a discussion in the finance literature regarding possible
“substitution” between dividends payments and share repurchases by firms. While
some researchers have argued that these payout methods may not be substitutes
(e.g., John and Williams 1985), substitution of dividends by share repurchases
(or vice versa) are consistent with certain economic formulations (Miller and
Modigliani (1961). For example, Grullon and Michaely (2002) interpreted negative
correlations between forecasting errors in changes in dividends and share
repurchases as evidence of substitution. However, these issues demand an
38 Dividend Payments and Share Repurchases of US Firms 1065
using existing databases such as Compustat for the United States. However, the
hypotheses driving the economic and finance investigation are often based on
untested assumptions and/or “stylized facts” (Granger 1992). By contrast, assump-
tions invoked in biomedical sciences reflect the knowledge accumulated from
previous experimental studies, and hypotheses are tested using the new data.
Further, in social sciences research, most phenomena being explained result
from the interactions between economic and social factors. Because different
dimensions of the relationships merit different emphasis depending on the context,
social science researchers are forced to address several hypotheses simultaneously.
Typically, this is done via formulation of comprehensive models where, for exam-
ple, the systematic part of the relationships incorporates the relevant variables. The
stochastic properties of the dependent variable and error terms are also tackled in
the formulation of comprehensive econometric models.
The analyses of databases such as Compustat entail the modeling of accounting,
finance, and economic variables so that financial research often demands econo-
metric modeling of variables that are jointly determined. Moreover, economic
theory is helpful in identifying variables and often implies restrictions on certain
model parameters. Thus, investigation of hypotheses in finance applications entails
the development of comprehensive econometric models within which several
hypotheses can be embedded and tested using the estimated model parameters.
This is in contrast with the approach in biomedical sciences where trials are
designed to investigate specific hypotheses, and researchers are not concerned
with interdependence between the variables. Thus, the currently popular approach
in finance of spelling out a few specific hypotheses needs to be augmented by
development of comprehensive econometric models reflecting the hypotheses. It is
only after the estimation of comprehensive econometric models from databases
such as Compustat that investigators can satisfactorily test the validity of specific
hypotheses using statistical procedures.
As an illustration of the importance of methodological aspects for drawing infer-
ences, it would be helpful to reappraise the work by Grullon and Michaely (2002)
claiming empirical support for the “substitution” hypothesis between
dividend payments and share repurchases. First, as noted in the Introduction, the
correlations between dividend payments and share repurchases estimated from
Compustat data were positive so the interrelationships are likely to be complex.
Second, if dividends and share repurchases are substitutes, then they are likely to be
simultaneously determined; modeling the interrelationships would require the use of
econometric methods for handling endogenous explanatory variables in longitudinal
analyses. Also, longitudinal data need to cover the same firms over time in order to
shed light on the interrelationships. While such issues were not considered by Grullon
and Michaely (2002), the authors explained forecasting errors in changes in dividends
based on share repurchases and interpreted negative correlations as evidence for
substitution hypothesis. However, dividends are announced quarterly, whereas share
repurchases are made at 2–3-yearly intervals. Thus, the stochastic properties of vari-
ables cast doubt on adequacy of the models that explain a higher frequency variable
such as dividends by the slowly changing share repurchases. In fact, it is likely that
38 Dividend Payments and Share Repurchases of US Firms 1067
dividend payments that are made on a quarterly basis affect share repurchases but not
vice versa. As shown below, firms’ share repurchases are constrained by dividend
payments and repurchases are relatively flexible (Bhargava 2010). Overall, from
a methodological standpoint, it is important in financial analyses to develop compre-
hensive models prior to the testing of specific hypotheses.
Di t ¼ ai t þ b Pi t þ d Di t1 þ ui t ði ¼ 1, 2, . . . , H; t ¼ 2, 3, . . . , TÞ (38.1)
ui t ¼ di þ vi t (38.2)
where, di are firm-specific random effects that follow some distribution (e.g., normal)
with zero mean and constant variance, and vi t are distributed with zero mean and
constant variance (Anderson and Hsiao 1981; Bhargava and Sargan 1983). More-
over, one can invoke the general assumption that ui t’s are drawings from
a multivariate distribution with a symmetric and positive definite dispersion matrix.
Third, even when the same firms are retained in the sample, there is large intra-
and interfirm variation in financial variables. In the context of dividends, some
problems can be tackled by expressing variables in per share terms (Fama and
Babiak 1968). However, as recognized in the early statistical literature by Pearson
(1897) and Neyman (1952), transforming variables into ratios can induce spurious
correlations. Such practices were criticized by Kronmal (1993), Bhargava (1994)
developed a likelihood ratio statistic for testing the restrictions on coefficients that
enable expressing variables as ratios in empirical models. While research in finance
often expresses variables in ratio forms, such transformation can affect the magni-
tudes and signs of estimated regression coefficients. It is important to conduct
robustness check to ensure that transformation of variables do alter the conclusions.
Finally, there is often high inter- and intra-firm variation in variables such as
share repurchases, assets, and debt that are measured in millions of dollars in the
Compustat database. Transformations to natural logarithms can reduce internal varia-
tion in the data. For example, share repurchase can be expressed in dollars and then
transformed into natural logarithms with zero dollar values set equal to one. This
procedure facilitates the estimation of model parameters using numerical optimization
techniques and obviates the need for arbitrarily truncating large values of variables
(Bhargava 2010).
A dynamic random effects model for dividends per share adjusted for “stock splits”
using cumulative adjustment factor (#27), with Compustat item numbers next to the
variables, can be written as (Bhargava 2010):
ðDividends per share; #26, #27Þi t ¼ a0 þ a1 ðEarnings; #18, #17Þi t þ a2 ln ðTotal assets; #6Þi t
þ a3 ðMarket-to-book value; #199, #25, #60Þi t þ a4 ln ðLong term debt; #9Þi t
þ a5 ln ðShort-term investments; #193Þi t þ a6 ðTime dummy period 3Þi t
þ a7 ðTime dummy period 4Þi t þ a8 ðTime dummy period 5Þi t
þ a9 ðTime dummy period 6Þi t þ a10 ðTime dummy period 7Þi t
þ a11 ðTime dummy period 8Þi t þ a12 ðDividends per share; #26, #27Þi t1
þ ui t ði ¼ 1, 2, . . . , H; t ¼ 2, 3, . . . , 8Þ
(38.3)
In the empirical model in Eq. 38.3, it is recognized that firms sometimes split
their shares especially if share prices are high. Earnings are defined as “income
before extraordinary items” (#18) minus 0.6 times “special items” (#17), measured
in million of dollars (Skinner 2008). Income before extraordinary items is the firms’
incomes taking into account all expenses, while special items reflect adjustments
1070 A. Bhargava
for debts and losses. “Total assets” (#6), “long-term debt” (#9), and “short-term
investments” (#193) were expressed in dollars and then converted into natural
logarithms. The “market-to-book value” variable was constructed by multiplying
“share price” (#199) by number of “common shares outstanding” (#25) and
dividing this product by “common equity total” (#60). “Dividends per share” in
period (t1) will be referred to as “lagged dependent variable.” Models of the type in
Eq. 38.3 were also estimated for dividends expressed as ratios to firms’ book values,
while dropping the explanatory variable market-to-book value from Eq. 38.3.
It is assumed in Eq. 38.3 that 8-time observations at 2-yearly intervals
{1992–1993, 1994–1995, 1996–1997, 1998–1999, 2000–2001, 2002–2003,
2004–2005, 2006–2007} were available on H firms. Moreover, as explained in
Sect. 38.6 and the Appendix, initial values of the dependent variable in time period 1
were modeled using a “reduced form” equation that includes a separate coefficient,
c0, for the constant term. Thus, the model in Eq. 38.3 includes separate coefficients
of the constant term for each of the 8-time periods. This formulation allows the
variables in the model to have different means in the 8-time periods and is useful if
there are trends in the dependent variables in the observation period.
Further, ui t’s are error terms that can be decomposed in the simple random effect
fashion as in Eq. 38.2. The variance of vi t is the “within” (or intra-) firm variance, and
[var (di)/var (vit)] is the “between-to-within” variance ratio that can be estimated when
errors are decomposed as in Eq. 38.2. In another version of the model in Eq. 38.3,
firms’ intangible assets (Compustat item #33) were included in the model, though
greater numbers of observations were missing for this variable. Also, firms’ share
repurchases were introduced as potentially endogenous explanatory variables to test if
they influenced dividend payments. Last, static versions of the model in Eq. 38.3 that
excluded the lagged dividends were estimated to assess robustness of the results.
The model for share repurchases was similar to that in Eq. 38.3 except that
dividends per share were included as a potentially endogenous explanatory variable:
Note that share repurchases were defined as the change in “Treasury stock”
(#226), and if this was less than or equal to zero for the firm, the nonnegative
difference between “purchase of common stock” (#115) and “sale of common
stock” (#108) was used (Fama and French 2001). This definition covered the
38 Dividend Payments and Share Repurchases of US Firms 1071
situation where firms “retire” shares from Treasury stock so that when the change in
Treasury stock was less than or equal to zero, nonnegative difference between
purchase of common stock and sale of common stock was used as the measure of
repurchases. Banyi et al. (2008) have advocated the use of purchase of common stock
as a measure of repurchases, and this variable was also modeled to assess robustness
of the results. Finally, alternative models for share repurchases were estimated with
intangible assets included as an explanatory variable, and also where dividends per
share was replaced by dividends expressed as ratios to firms’ book values.
The methodology used for estimation of dynamic and static random effect models,
where some explanatory variables are endogenous, was developed in Bhargava and
Sargan (1983) and Bhargava (1991). Let the dynamic model be given by:
X
m X
n1
yi t ¼ z i j gj þ x1 i j t bj
j¼1 j¼1
X
n
þ x2 i j t bj þ a yi t1 þ ui t ði ¼ 1, 2, . . . , H; t ¼ 2, 3, . . . , TÞ
j¼n1 þ1
(38.5)
where z’s are time-invariant variables, x1 and x2 are, respectively, n1 exogenous and
n2 endogenous time-varying variables (n1 + n2 ¼ n). In the model for share
repurchases, for example, dividends per share is potentially an endogenous time-
varying variable; unobserved factors affecting share repurchases, reflected in firm-
specific random effects (di) in Eq. 38.2, can influence dividend payments. For
exposition purposes, first assuming that all time-varying variables are exogenous
so that n2 ¼ 0 and subscripts on the x variables can be dropped, the dynamic model
can be written in a simultaneous equations framework by defining a “reduced form”
equation for initial observations that do not include any endogenous explanatory
variables, and a “triangular” system of (T-1) “structural” equations for the
remaining periods (Bhargava and Sargan 1983):
X
m n X
X T
yi 1 ¼ z i j zj þ uj k xi j k þ ui 1 ði ¼ 1, . . . , HÞ (38.6)
j¼1 j¼1 k¼1
And
0 0 0
B Y þ Cz 0 þ Cx X ¼ U
Z
ðT 1ÞxT TxH ðT 1Þx m m x H ðT 1Þx nT nTxH ðT 1ÞxH (38.7)
1072 A. Bhargava
The Compustat annual data on firms for the period 1992–2007 were processed for
the analyses (Bhargava 2010). As is customary in the dividends literature, financial
firms with Standard Industry Classification (SIC) codes between 6,000 and 6,999
and utility firms with SIC codes between 4,900 and 4,949 were dropped since there
are regulations on dividend payments by such firms (Fama and French 2001).
Moreover, observations were retained on firms that were in the Compustat database
for at least 14 years, and 2-yearly averages were created at 8-time points, i.e., for
{1992–1993, 1994–1995, 1996–1997, 1998–1999, 2000–2001, 2002–2003,
2004–2005, 2006–2007}. An alternative data set created 3-yearly averages at
5-time points, i.e.,{1992–1994, 1995–1997, 1998–2000, 2001–2003, 2004–2006},
though the analyses of the 2-yearly averages led to more robust parameter estima-
tion. Averaging over 2 years led to a sample of 3,290 industrial firms observed in
8-time periods. Also, estimation of dynamic models requires “balanced” panels,
i.e., where firms are included at all eight time points. While it is possible to analyze
“unbalanced” panels using static models in software packages such as Stata (2008),
it is difficult to address endogeneity issues in the analyses.
The sample means and standard deviations of 2-year averages of variables are
reported in Table 38.1 for some of the years with Compustat annual item numbers
noted next to the variables. The mean dividends per share were approximately
$0.32 for the period 1992–1999, and declined slightly to $0.27 for 2000–2003, and
increased to $0.39 in 2006–2007. While a simple paired t-test on the set of firms’
observations showed significant (P < 0.05) increase in dividends per share between
38
Table 38.1 Sample means and standard deviations of 2-yearly averages of selected variables from the Compustat database for up to 3,290 US industrial firms
in 1992–2007
1992–1993 1998–1999 2002–2003 2006–2007
Variable Mean SD Mean SD Mean SD Mean SD
Income before extraordinary items (#18), $ million 53.99 317.55 138.05 702.98 124.09 1,015.57 309.43 1,818.94
Special items (#17), $ million 15.99 215.30 1.56 372.44 55.10 552.09 28.20 731.89
Treasury common stock (#226), $ million 37.08 398.01 108.79 852.50 194.16 1,461.74 406.76 2,994.53
Purchase of common stock (#115), $ million 10.39 66.72 54.54 282.34 51.85 349.32 221.96 1,538.69
Sale of common stock (#108), $ million 21.66 112.12 31.07 160.90 25.69 104.93 72.96 1,022.77
Share repurchases (#226, #115, #108), $ million 9.95 111.13 44.44 316.76 49.49 467.39 174.77 1,564.85
Dividends (#21), $ million 32.40 166.22 63.22 291.33 63.09 376.21 113.18 629.87
Dividends per share (#26), $ 0.304 0.902 0.298 0.882 0.281 1.272 0.390 1.095
Total assets (#6), $ million 2,292.84 13,195.74 3,845.47 23,638.40 5,491.57 33,186.31 8,083.80 61,202.42
Cumulative adjustment factor (#27) 2.48 4.48 1.41 1.24 1.18 0.72 1.03 0.89
Share price (#199), $ 17.55 19.61 19.40 27.03 17.21 25.29 23.55 42.62
Long-term debt (#9), $ million 429.26 2,417.02 695.61 3,664.86 1,070.41 6,085.70 1,519.05 11,825.75
Dividend Payments and Share Repurchases of US Firms
Short-term investments (#193), $ million 88.31 1,181.56 173.91 2,767.14 238.97 3,735.60 368.76 8,385.86
Intangible assets (#33), $ million 99.91 746.35 279.68 1,279.77 651.12 4,629.00 980.21 5,507.11
Common shares outstanding (#25), million 39.53 133.66 84.22 279.68 159.55 1,682.32 177.41 1,684.17
Common equity total (#60), $ million 541.70 2,214.41 958.11 3,917.14 1,351.86 6,265.71 1,985.57 8,140.76
Percentages of firms paying dividendsa 36.50 – 40.79 – 39.89 – 44.00 –
Percentages of firms making repurchasesa 26.29 – 49.82 – 37.93 – 43.46 –
Means and standard deviations of variables using 2-yearly averages at four time points for up to 3,290 US industrial firms observed for at least 14 years in
1992–2007 (Source: Bhargava 2010); specific number of firms included depends on the extent of unavailable information for each variable in each period;
Compustat items and units of measurements are listed next to variables
a
Percentages are based on firms’ with nonzero payments
1073
1074 A. Bhargava
450
400
350
300
250
$million
200
150
100
50
0
1992–93 1994–95 1996–97 1998–99 2000–01 2002–03 2004–05 2006–07
Year
Treasury common stock, (#226) Purchase of common stock, (#115)
Sale of common stock, (#108) Share repurchases, (#226, #115, #108)
Dividends, (#21)
Fig. 38.1 Sample means of dividends and components of share repurchases by US industrial
firms for 1992–2007. Share repurchases were calculated as the change in Treasury stock (#226),
and if this was less than or equal to zero, the nonzero difference between purchase of common
stock (#115) and sale of common stock (#108) was used (Source: Bhargava 2010)
1992–1993 and 2006–2007, the pattern over time was more complex partly because
of tax changes in 2003. Mean share repurchases increased from $10 million in
1992–1993 to $175 million in 2006–2007.
Figure 38.1 plots the sample means (in $ millions) of Treasury stock, purchase
of common stock, sale of common stock, dividends, and share repurchases.
There were upward trends in Treasury stock, share repurchases, and purchases
of common stock. There was an increase in dividends in the period from 2002–2003
to 2006–2007, where dividends increased from $63 million to $113 million.
This could be due to reductions in 2003 in tax rates on dividends (Julio and
Ikenberry 2004).
Further, focusing on 2,880 firms with non-missing observations on dividends
and share repurchases, percentages of firms paying dividends and making share
repurchases in the 8-time periods, were {36.5, 39.9, 40.6, 40.8, 39.6, 39.9, 43.7,
44.0} and {26.3, 30.7, 38.2, 49.8, 48.1, 37.9, 38.1, 43.5}, respectively. While
percentages of firms making share repurchases increased from 26.3 in 1992–1993
to 43.5 in 2006–2007, the peak of 49.8 % was reached in the 1998–1999 period.
The mean share repurchases in 1998–1999 were 44.44 million, while they
38 Dividend Payments and Share Repurchases of US Firms 1075
The results from estimating simple dynamic random effects models for dividends
per share are in Table 38.2 for the sample of 3,113 industrial firms (Bhargava 2010).
The results are presented for the cases where firms paid nonzero dividends at least
in one of the eight time periods and where firms paid dividends in all 8-time periods.
The results for ratios of dividends to book value of firms are presented in the last
column of Table 38.2. A set of six dummy variables for time periods 3–8 was
included to account for differences in dividend payments in the 8-time periods. The
models were estimated by maximum likelihood and provide consistent and efficient
estimates of the parameters.
For the pooled sample, coefficients of the dummy variables in the model for
dividends per share were positive and significant (P < 0.05) in time periods 7 and
8 corresponding to 2004–2005 and 2006–2007, respectively. The positive coeffi-
cients were consistent with reported increases in dividends following reduction in
tax rates in 2003. Coefficients of the dummy variables were qualitatively similar for
the reduced sample of 1,827 firms that paid dividends at least in one of the eight
periods and in the case where 1,035 firms paid dividends in all eight periods.
Moreover, coefficients of the dummy variables for time periods 7 and
8 corresponding to 2004–2005 and 2006–2007, respectively, were approximately
twice as large for firms paying dividends than for firms that never paid dividends.
Thus, reduction in tax rates on dividends in 2003 appeared to have increased
dividends per share. These results also show the robustness of dynamic models in
situations where firms decide against paying dividends in some time periods, i.e., to
zero values of the dependent variables. This is not surprising since one is modeling
the deviations of dividends per share from an overall mean.
The coefficients of the lagged dependent variables for the first three cases
presented in Table 38.2 were close to 0.28. Moreover, the ratios of between-to-
within variances for the three cases, i.e., pooled sample, firms paying nonzero
dividends at least once, and firms paying dividends in all 8 periods were, 0.38,
0.33, and 0.45, respectively. These estimates were close despite reduction in sample
size from 3,113 firms in the pooled sample to 1,035 in the case of firms paying
nonzero dividends in all 8 periods. The within firm variances for the three cases
were 0.52, 0.88, and 1.07, respectively. Finally, in the last column, the results for
the ratio of dividends to firms’ book values were similar to those for the pooled
sample for dividends per share. While the estimated coefficient of lagged dependent
1076
Table 38.2 Maximum likelihood estimates of simple dynamic models for 2-yearly averages of dividend paid by US industrial firms in 1992–2007
Dependent variable:dividends/
Dependent variable: dividends per share (Compustat items #26/#27) book value (#21,#60)
Firms paying nonzero Firms paying nonzero
All firms dividends at least once dividends in all 8 periods All firms
Explanatory variables Coefficient SE Coefficient SE Coefficient SE Coefficient SE
Constant 0.172 0.015 0.294 0.018 0.421 0.031 0.018 0.004
Time period 3 dummy variable 0.017 0.020 0.029 0.026 0.029 0.046 0.001 0.006
Time period 4 dummy variable 0.033 0.020 0.058* 0.026 0.081 0.046 0.003 0.006
Time period 5 dummy variable 0.002 0.019 0.003 0.026 0.010 0.046 0.0002 0.006
Time period 6 dummy variable 0.017 0.019 0.030 0.026 0.082 0.046 0.00002 0.006
Time period 7 dummy variable 0.087* 0.020 0.148* 0.027 0.130* 0.046 0.005 0.006
* *
Time period 8 dummy variable 0.133 0.019 0.228 0.027 0.235* 0.046 0.025* 0.006
Lagged dependent variable 0.285* 0.009 0.283* 0.010 0.258* 0.013 0.290* 0.018
(between-to-within) variance ratio 0.375* 0.017 0.328* 0.019 0.451* 0.034 0.016* 0.007
Within variance 0.518 – 0.879 – 1.073 – 0.040 –
2 x (maximized log-likelihood function) 11,552.55 – 741.58 – 2,301.34 58,929.64
Number of firms 3,113 1,827 1,035 2,315
Dependent variable is dividends per share (#26) divided by cumulative adjustment factor (#27) (Source: Bhargava 2010); 8-time observations at 2-yearly
intervals on firms were used; slope coefficients and standard errors are reported
*
P < 0.05
A. Bhargava
38 Dividend Payments and Share Repurchases of US Firms 1077
Table 38.3 Maximum likelihood estimates of simple dynamic models for 2-yearly averages of
share repurchases of US industrial firms in 1992–2007
Dependent variable: ln (share repurchases $) (using Compustat
items #226, #115, #108)
Firms making
Firms making nonzero
nonzero repurchases repurchases in all
All firms at least once 8 periods
Explanatory variables Coefficient SE Coefficient SE Coefficient SE
Constant 3.216 0.061 4.057 0.153 4.376 0.587
Time period 3 dummy variable 1.008* 0.122 1.262* 0.205 0.054 0.192
Time period 4 dummy variable 2.428* 0.117 3.048* 0.200 0.067 0.193
Time period 5 dummy variable *
1.426 0.126 1.818* 0.208 0.128 0.197
Time period 6 dummy variable 0.095 0.117 0.083 0.203 0.540* 0.183
Time period 7 dummy variable *
0.738 0.122 *
0.938 0.208 0.017 0.192
Time period 8 dummy variable 1.730* 0.123 2.180* 0.205 0.171 0.194
Lagged dependent variable 0.362* 0.008 0.354* 0.009 0.754* 0.038
(between-to-within) variance 0.153* 0.010 0.066* 0.008 0.184* 0.092
ratio
Within variance 39.176 – 48.341 – 1.841 –
2 x (maximized log-likelihood 88,260.15 – 73,671.73 – 886.97
function)
Number of firms 2,907 2,329 127
Dependent variable share repurchases were calculated from Compustat items #226, #115, and
#108 and expressed in dollars and transformed into natural logarithms with zero value assigned to
zero purchases (source: Bhargava 2010); 8-time observations at 2-yearly intervals on the firms in
1992–2007 were used; slope coefficients and standard errors are reported
*
P < 0.05
variable was very close (0.29), the estimated between-to-within variance ratio was
0.016 that was lower than the corresponding estimate (0.38) in the model for
dividends per share.
The results from simple dynamic models for logarithms of US firms’ share
repurchases are in Table 38.3 for the three cases, i.e., pooled sample, firms making
at least one repurchase, and firms making repurchase in all 8-time periods. For the
pooled sample of 2,907 firms, coefficients of dummy variables for five of the time
periods were positive and statistically significant; coefficient of the dummy variable
for time period six corresponding to 2002–2003 was not significant. The results
were very similar for the case where 2,329 firms made repurchases at least once.
The dummy variables were generally insignificant in the third case for firms making
repurchases in all 8 periods; there were 127 firms in this group which was a small
sample size for estimating dynamic models.
The coefficients of lagged dependent variables in the three cases were 0.36, 0.35,
and 0.75, respectively; these estimates were significantly less than unity thereby
indicating that share repurchase series were stationary. The between-to-within
variance ratios for the three cases were 0.15, 0.07, and 0.18, respectively; within
1078 A. Bhargava
firm variances were quite large in these models. The higher coefficient of lagged
dependent variable in the model where 127 firms made repurchases in all 8 periods
may be due to the small sample size. Moreover, such firms may have been less
heterogeneous in some respects and had a smooth pattern of share repurchases.
Such issues can be systematically investigated by controlling for firm characteris-
tics and are addressed below.
The results from empirical models for dividends per share and share repurchases,
outlined in Eqs. 38.3 and 38.4, are presented in this section. Dynamic and static
random effects models were estimated and the results are initially presented for the
pooled samples and for the case where firms paid dividends or made repurchases at
least once during 8-time periods.
38.9.1 Results from Dynamic and Static Models for Firms’ Dividends
per Share
Table 38.4 presents the results from dynamic and static models for dividends per
share paid by US industrial firms; a set of six dummy variables for time periods was
included in the models though the coefficients are not reported (Bhargava 2010).
The firms’ total assets, long-term debt, and short-term investments were expressed
in dollars and converted to natural logarithms. Because firms’ earnings were
defined as the difference between income before extraordinary items and special
items, this variable sometimes assumed negative values and was not transformed
into logarithms.
For the dynamic model estimated using the pooled sample, coefficient of
earnings was estimated with a positive coefficient that was statistically significant.
The coefficients of earnings and total assets were also significant in the static model
for the pooled sample and in dynamic and static models for the subsample where
firms paid dividends at least once during the 8-time periods. These results provide
evidence that earnings and total assets of US firms were positively and significantly
associated with dividends per share. By contrast, Skinner (2008, Table 8) found
these variables to be generally insignificant predictors of the ratio of dividends to
total payout (dividends plus repurchases) using cross-sectional regressions and data
on 345 US firms. When some of the coefficients were statistically significant, they
were positive for some years and negative for others. The contradictory findings
were likely to be due to combining dividends and share repurchases into a single
variable and using estimation methods that are appropriate for cross-sectional
analyses. The results in Table 38.4, however, were consistent with evidence from
the European Union (von Eije and Megginson 2008), where static random effects
models were estimated using unbalanced panel data on approximately 3,000 firms.
38
Table 38.4 Maximum likelihood estimates from comprehensive dynamic and static random effects models for 2-yearly averages of dividends per share paid
by US industrial firms in 1992–2007
Dependent variable: dividends per share (adjusted for stock splits; Compustat items # 26/#27)
Dynamic model: all Dynamic model: firms paying Static model: firms paying at
firms Static model: all firms at least once least once
Explanatory variables: Coefficient SE Coefficient SE Coefficient SE Coefficient SE
Constant 9.383 2.664 9.169 0.464 9.357 1.306 11.091 0.817
Earnings (#18, #17), $million 0.0001* 0.00001 0.0001* 0.00001 0.0002* 0.00001 0.0001* 0.00001
* * *
ln (total assets) (#6), $ 0.058 0.019 0.060 0.003 0.054 0.010 0.072* 0.006
Market-to-book value (#199, #25, #60) 0.0004 0.004 0.0004 0.0004 0.0005 0.0008 0.001 0.001
ln (long-term debt) (#9), $ 0.0007 0.009 0.0008 0.0006 0.001 0.003 0.002 0.001
ln (short-term investments) (#193), $ 0.002 0.007 0.0006 0.0005 0.005* 0.002 0.002* 0.001
* *
Lagged dependent variable 0.201 0.021 – 0.193 0.014 –
(between-to-within) variance ratio 0.126* 0.014 – 0.114* 0.014 –
Dividend Payments and Share Repurchases of US Firms
38.9.2 Results from Dynamic and Static Models for Firms’ Share
Repurchases
Table 38.5 Maximum likelihood estimates from comprehensive dynamic and static random effects models for 2-yearly averages of share repurchases by US
industrial firms in 1992–2007
Dependent variable: ln (share repurchases $) (using Compustat items #226, #115, #108)
Dynamic model: all Static model: all Dynamic model: firms Static model: firms
firms firms repurchasing least once repurchasing at least once
Explanatory variables Coefficient SE Coefficient SE Coefficient SE Coefficient SE
Constant 99.008 6.428 114.142 7.960 116.688 7.620 147.576 9.851
Earnings (#18, #17), $million 0.0004* 0.0001 0.0008* 0.0001 0.0004* 0.0001 0.0007* 0.0001
* * *
ln (total assets) (#6), $ 0.613 0.041 0.703 0.051 0.776 0.052 0.961* 0.066
Market-to-book value (#199, #25, #60) 0.010* 0.004 0.010* 0.003 0.011* 0.005 0.012* 0.005
ln (long-term debt) (#9), $ 0.021* 0.008 0.020* 0.010 0.054* 0.012 0.053* 0.017
ln (short-term investments) (#193), $ 0.001 0.006 0.003 0.008 0.020 0.008 0.019 0.012
Dividends per share (#26/#27), $ 0.201* 0.057 0.139* 0.059 0.179* 0.078 0.153* 0.077
* *
Lagged dependent variable 0.404 0.011 – 0.393 0.011 –
(between-to-within) variance ratio 0.168* 0.012 – 0.014* 0.005 –
Dividend Payments and Share Repurchases of US Firms
In all the models in Table 38.5, firms’ market-to-book value ratio was estimated
with negative and significant coefficients in the dynamic and static models using the
pooled and disaggregated samples. Thus, firms with higher market-to-book value
ratio made lower share repurchases. Also, firms’ long-term debt was estimated with
negative and significant coefficients in dynamic models from the pooled sample and
from the subsample of firms making repurchases at least once. The short-run
elasticity from the pooled sample was 0.02, while the long-run elasticity was
0.03. Doubling of firms’ long-term debt predicted a 3 % decline in long-run share
repurchases. The coefficients of short-term investments were not significant in any
of the models in Table 38.5.
An important aspect of the results in Table 38.5 was that the estimated coeffi-
cients of dividends per share were negative and statistically significant in the
dynamic and static models for the pooled sample and for the subsample of firms
that made at least one repurchase. Thus, controlling for factors such as earnings and
size of the firm, firms paying higher dividends per share made significantly lower
share repurchases. The short-run effects of dividends per share on share repurchases
were 0.20 and 0.18, respectively, for the pooled sample and subsample of firms
making repurchases; the respective long-run effects were 0.33 and 0.30. The
relatively large magnitudes of these effects indicated that firms with higher divi-
dend payments were likely to make smaller share repurchases.
Last, exogeneity hypotheses for dividends per share were accepted at the 5 %
level using test statistics that were distributed as Chi-square variables with 8 of
freedom (5 % critical limit of Chi-square (8) ¼ 15.5). The values of Chi-square
statistics around 12 in Table 38.5 indicated that there might be some dependence in
the unobserved components of firms’ decisions to pay dividends and make share
repurchases though it was not significant at the 5 % level. Such hypotheses will be
tested below for the effects of share repurchases on firms’ dividend payments.
The results from models for dividends per share and ratios of dividends to firms’
book value, with share repurchases included as a potentially endogenous explana-
tory variable, are in Table 38.6; firms’ intangible assets were included in these
models. Because short-term investments were not significant predictors of share
repurchases in Table 38.5, this variable was dropped from the models for dividends
and share repurchases in Tables 38.6 and 38.7, respectively. There were greater
number of missing observations on intangible assets and sample sizes in Table 38.6
were lower than sample sizes in Table 38.4.
The noteworthy feature of the results for dividends per share in Table 38.6 was
that coefficients of share repurchases was estimated with a small negative coeffi-
cient (0.004) in the dynamic model that reached statistical significance at the 5 %
level. However, this coefficient from the static model was 0.0002 and was not
38
Table 38.6 Maximum likelihood estimates from comprehensive dynamic and static random effects models for 2-yearly averages of dividends per share paid
by US industrial firms in 1992–2007 with intangible assets and share repurchases included as explanatory variables
Dependent variable: dividends per share (#26/#27) Dependent variable: (dividends/book value) (#21/#60)
Dynamic model Static model Dynamic model Static model
Explanatory variables Coefficient SE Coefficient SE Coefficient SE Coefficient SE
Constant 11.348 0.844 10.855 0.612 0.077 0.027 0.068 0.010
Earnings (#18, #17), $million 0.0001* 0.00001 0.0001* 0.00001 0.00001* 0.000003 0.00001* 0.000001
* *
ln (total assets) (#6), $ 0.075 0.006 0.067* 0.004 0.005 0.002 0.005* 0.001
Market-to-book value (#199, #25, #60) 0.001 0.001 0.0004 0.0004 – –
ln (long-term debt) (#9), $ 0.0004 0.002 0.0003 0.001 0.00003 0.0005 0.0004* 0.0001
ln (intangible assets) (#33), $ 0.004* 0.001 0.003* 0.001 0.0002 0.0004 0.0001 0.0001
ln (share repurchases) (#226, #115, #108),$ 0.004* 0.002 0.0002 0.0005 0.0003 0.0004 0.0002 0.0001
Lagged dependent variable 0.191* 0.012 – 0.255* 0.023 –
(between-to-within) variance ratio 0.122* 0.013 – 0.012* 0.006 –
Dividend Payments and Share Repurchases of US Firms
Table 38.7 Maximum likelihood estimates from dynamic and static random effects models for 2-yearly averages of share repurchases by US industrial firms
in 1992–2007 with intangible assets included as an explanatory variable
Dependent variable: ln (Share repurchases $) (Using Compustat items #226, #115, #108)
Specification 1: Specification 2:
Dynamic model Static model Dynamic model Static model
Explanatory variables: Coefficient SE Coefficient SE Coefficient SE Coefficient SE
Constant 89.993 5.547 103.897 8.953 81.508 4.659 94.684 8.243
Earnings (#18, #17), $million 0.0004* 0.0001 0.001* 0.0001 0.0004* 0.0001 0.001* 0.0001
* * *
ln (total assets) (#6), $ 0.524 0.039 0.602 0.058 0.517 0.036 0.599* 0.058
* *
Market-to-book value (#199, #25, #60) 0.012 0.0004 0.011 0.003 – –
ln (long-term debt) (#9), $ 0.027* 0.009 0.019 0.013 0.028* 0.009 0.021* 0.013
* * *
ln (intangible assets) (#33), $ 0.044 0.003 0.044 0.011 0.046 0.004 0.046* 0.011
* *
Dividend per share (#26/#27), $ 0.217 0.054 0.183 0.058 – –
Dividends/Book value (#21/#60) – – 0.583* 0.248 0.537* 0.207
Lagged dependent variable 0.412* 0.012 – 0.412* 0.011 –
*
(between-to-within) variance ratio 0.171 0.014 – 0.166* 0.014 –
Within variance 28.123 – – 28.310 – –
2x (maximized log-likelihood function) 39,778.71 – 39,523.88 –
Chi-square test exogeneity of dividends (8df) 9.62 11.30 10.34 15.40*
Number of firms 1,429 1,429 1,418 1,418
Dependent variable share repurchases were based on items #226, #115, and #108 (Source: Bhargava 2010); Specification 1 included dividends per share as
explanatory variable, while Specification 2 included dividends expressed in terms of book values; 8-time observations at 2-yearly intervals on firms in
1992–2007 were used; slope coefficients and standard errors are reported; time dummies for 6 periods were included though their coefficients are not reported
*
P < 0.05
A. Bhargava
38 Dividend Payments and Share Repurchases of US Firms 1085
significant. Also, coefficients of share repurchases in the model for the ratio of
dividends to book value were statistically not different from zero. Exogeneity
hypotheses for share repurchases could not be rejected using the Chi-square statis-
tics in dynamic and static models for dividends per share and for the ratio of
dividends to firms’ book value. Because share repurchases in the comprehensive
models generally did not affect dividend payments, it seems likely that firms
decided their dividends payments at an earlier stage than share repurchases. By
contrast, dividends per share were significant predictors of share repurchases in
Table 38.5.
Another important finding in Table 38.6 was that intangible assets were esti-
mated with negative and significant coefficients in dynamic and static models for
dividends per share. The results for other explanatory variables were similar to
those presented in Table 38.4, thereby showing robustness of the estimates to
changes in model specification and to reductions in sample sizes due to greater
number of missing observations on intangible assets.
Table 38.7 presents the results for logarithms of share repurchases with intangible
assets included as explanatory variables in the models; in Specification 2, dividends
per share were replaced by the ratio of dividends to firms’ book values. The results
in Table 38.7 showed that intangible assets were positively and significantly
associated with share repurchases. This was in contrast with the results in Table 38.6
where intangible assets were negatively associated with dividends per share. Thus,
firms possessing higher intangible assets paid lower dividends per share and made
greater share repurchases; these results indicate the inappropriateness of modeling
the ratio of dividend payments to total payouts (Skinner 2008). Furthermore, while
intangible assets have received attention in the literature (Lev 2001), their effects on
dividends and share repurchases have not been rigorously investigated. For example,
it has been suggested that information technology firms may pay low (or zero)
dividends and make frequent repurchases in part because they have higher invest-
ments in research and development. A dummy variable was created using SIC
codes for information technology firms in manufacturing, transportation, and com-
munications sectors. However, the estimated coefficient was not significantly
different from zero. Thus, the reported coefficients in Table 38.7 appear not to
suffer from biases due to omission of such variables.
Further, dividends per share and the ratio of dividends to book value were
negatively and significantly associated with share repurchases, with large magni-
tudes of the estimated coefficients. Thus, the models with intangible assets as an
explanatory variable again showed that firms paying higher dividends made smaller
share repurchases. Also, exogeneity hypothesis for dividends-to-book value ratio
was close to rejection using the Chi-square statistic at 5 % level in the static model.
1086 A. Bhargava
The rejection of the null indicated that there was some dependence in the
unobserved factors affecting the firms’ decisions to pay dividends and make share
repurchases. For example, if firms made “unexpectedly” large share repurchases,
then such decisions in turn can affect dividend payments.
Finally, in view of the literature on choice of measures for share repurchases
(Banyi et al. 2008), the variable “purchase of common stock” was used as a proxy
for share repurchases, and dynamic and static models were estimated. The results
from modeling purchase of common stock were similar to those reported in
Tables 38.5 and 38.7 using the more complex definition of share repurchases.
However, greater numbers of observations were missing on purchase of common
stock variable and these were reduced in the more complex definition because
changes in Treasury stock were often positive so that the purchase of common stock
variable was not utilized. Overall, the results indicated that these two measures for
share repurchases yield similar results, perhaps because the models captured many
salient features of the interrelationships.
38.10 Conclusion
Appendix
The concentrated or “profile” log-likelihood functions of the model in Eq. 38.3 was
computed by a FORTRAN program and was optimized using the numerical scheme
1088 A. Bhargava
E04 JBF from Numerical Algorithm Group (1991). Note that the data were read
firm-by-firm and several second moment matrices were created. Thus, there was no
limit on the number of firms in the sample. The profile likelihood function, which
depends only on the structural form parameters of interest, was computed using the
second moment matrices. FORTRAN programs for the computations of the likeli-
hood functions are available on request from the author.
The likelihood functions were separately computed for the case where the
variance-covariance matrix of the errors was unrestricted and where the errors
were decomposed in a random effect fashion, as in Eq. 38.2. Likelihood ratio
tests were applied to discriminate between these two alternative formulations.
Assuming that the number of firms (H) is large but the number of time observations
is fixed, asymptotic standard errors of the parameters were obtained by approxi-
mating second derivatives of the functions at the maximum.
Further, assuming that n1 and n2 time-varying variables are exogenous and
endogenous, respectively, as in Eq. 38.5, one can tackle the correlation between
the errors u’s and x2’s. It is reasonable to assume in short panels that a variable such
as dividends per share in Eq. 38.5 may be correlated only with firm-specific random
effects di. Thus, the correlation pattern can be decomposed as:
x2 i j t ¼ lj di þ x 2i j t (38.9)
where x*2 i j t are uncorrelated with di, and di are randomly distributed variables
with zero mean and finite variance as in Eq. 38.2. This correlation pattern was
invoked by Bhargava and Sargan (1983) and has the advantage that deviations of
x2 i j t’s from their time means:
xþ 2 i j t ¼ x2 i j t x
2ij ðt ¼ 2, . . . , T; j ¼ n1 þ 1, . . . , n; i ¼ 1, . . . , HÞ
(38.10)
Where
X
T
x
2ij ¼ x2 i j t =T ðj ¼ n1 þ 1, . . . n; i ¼ 1, . . . , HÞ (38.11)
t¼1
X
m Xn1 X
T n2 X
X T
yi 1 ¼ z i j zj þ uj k x1i j k þ mj k xþ 2i j k þ ui 1 ði ¼ 1, . . . , HÞ
j¼1 j¼1 k¼1 j¼1 k¼2
(38.12)
38 Dividend Payments and Share Repurchases of US Firms 1089
Further, likelihood ratio tests can be applied to test for exogeneity of time means
of the n2 endogenous variables (x2) in Eq. 38.11 by testing if correlations between
the T errors (ui t) affecting the dependent variables and n2 time means in Eq. 38.11
are zero. For example, given 8-time observations (T ¼ 8), likelihood ratio statistic
for testing zero correlation between errors affecting share repurchases in Eq. 38.4
(n2 ¼ 1) and time means of dividends per share is distributed for large H as
a Chi-square variable with 8 degrees of freedom.
The identification of parameters is achieved via the n1 time-varying exogenous
variables in the model (e.g., Sargan 1958). For example, in the system representing
(T1) time periods in Eq. 38.7, only the exogenous explanatory variables in time
period t explain the dependent variable in period t. The remaining (T1) n2 vari-
ables are excluded from the t th equation and are used in the set of instrumental
variables for identifying the coefficients of endogenous variables. Sufficient con-
ditions for identification, exploiting the time structure of longitudinal models, were
developed in Bhargava and Sargan (1983). For example, each exogenous time-
varying variable in the model for share repurchases in Eq. 38.4 effectively provides
eight exogenous variables of which seven are excluded from the equations. The null
hypothesis that dividends per share do not affect share repurchases can be tested
using the estimated coefficient b12 of dividends per share in Eq. 38.4. Also, one can
test if share repurchases affect dividends per share by including repurchases as
a potentially endogenous variable in Eq. 38.3.
For static version of the models not containing lagged dependent variables, one
can use stepwise estimation procedures with equality restrictions on coefficients
across time periods (Bhargava 1991). Efficient instrumental variables estimators
were used to estimate model parameters of Eq. 38.4, assuming the correlation
patterns for x2 i j t as in Eq. 38.9 and without restricting variance-covariance
matrices of the errors, i.e., the T x T dispersion matrix was assumed to be symmetric
and positive definite but not of the simple random effects form as in Eq. 38.2. This
formulation has the advantage that the errors vi t in Eq. 38.2 may be serially
correlated which was likely to be the case in static models since the lagged
dependent variables were omitted. Exogeneity hypotheses can be tested in the
stepwise estimation procedures via Chi-square tests that are asymptotically equiv-
alent to likelihood ratio tests.
Finally, while one can use “fixed” effects estimators (with dummy variables for
each firm) to circumvent certain endogeneity problems, increase in the number of
parameters with sample size leads to the problem of “incidental parameters”
(Neyman and Scott 1948). For example, coefficient of the lagged dependent variable
cannot be consistently estimated due to the incidental parameters in fixed effects
models unless the number of time observation is large. Moreover, from a modeling
standpoint, the use of random effects models obviates the need for estimating the
coefficients of large numbers of dummy variables for firms thereby enhancing the
efficiency of estimates. Exogeneity hypotheses for variables such as dividends per
share can be tested in the model for share repurchases. If the null were rejected, then
the models were estimated under the appropriate assumption that dividends per share
were correlated with errors affecting the models for share repurchases.
1090 A. Bhargava
References
Anderson, T. W., & Hsiao, C. (1981). Estimation of dynamic models with error components.
Journal of the American Statistical Association, 76, 598–606.
Banyi, M., Dyl, E., & Kahle, K. (2008). Errors in estimating share repurchases. Journal of
Corporate Finance, 14, 460–474.
Bhargava, A. (1991). Identification and panel data models with endogenous regressors. Reviews of
Economic Studies, 58, 129–140.
Bhargava, A. (1994). Modeling the health of Filipino children. Journal of the Royal Statistical
Society A, 157, 417–432.
Bhargava, A. (2008). Randomized controlled experiments in health and social sciences. Econom-
ics and Human Biology, 6, 293–298.
Bhargava, A. (2010). An econometric analysis of the dividends and share repurchases by
U.S. firms. Journal of the Royal Statistical Society A, 173, 631–656.
Bhargava, A. (2012, in press). Executive compensation, share repurchases and investment expen-
ditures: Econometric evidence from U.S. firms. Review of Quantitative Finance and
Accounting.
Bhargava, A., & Sargan, J. D. (1983). Estimating dynamic random effects models from panel data
covering short time periods. Econometrica, 51, 1635–1660.
Brav, A., Graham, J. R., Harvey, C. R., & Michaely, R. (2005). Payout policy in the 21st century.
Journal of Financial Economics, 77, 483–527.
Chen, S. -S., & Wang, Y. (2012, in press). Financial constraints and share repurchases. Journal of
Financial Economics.
Compustat. (2008). Standard & Poor’s Compustat. http://www.compustat.com/
Cox, D. (1958). The planning of experiments. New York: Wiley.
De Angelo, H., De Angelo, L., & Stultz, R. (2006). Dividend policy and the earned/contributed
capital mix: A test of the life cycle theory. Journal of Financial Economics, 81, 227–254.
Duesenberry, J. (1949). Income, saving, and the theory of consumer behavior. Cambridge:
Harvard University Press.
ExecuComp. (2008). Standard & Poor’s ExecuComp. http://www.compustat.com/products.aspx?
id¼2147492873
Fama, E., & Babiak, H. (1968). Dividend policy: An empirical analysis. Journal of the American
Statistical Association, 63, 1132–1161.
Fama, E., & French, K. (2001). Disappearing dividends: Changing firm characteristics or lower
propensity to pay? Journal of Financial Economics, 60, 3–43.
Fenn, G. W., & Liang, N. (2001). Corporate payout policy and managerial stock incentives.
Journal of Financial Economics, 60, 45–72.
Granger, C. (1992). Evaluating economic theory. Journal of Econometrics, 51, 3–5.
Grullon, G., & Michaely, R. (2002). Dividends, share repurchases and the substitution hypothesis.
Journal of Finance, 57, 1649–1684.
Gu, F., & Wang, W. (2005). Intangible assets, information complexity, and analysts’ earnings
forecasts. Journal of Business Finance and Accounting, 32, 1673–1702.
Guay, W., & Harford, J. (2000). The cash-flow permanence and information content of dividend
increases versus repurchases. Journal of Financial Economics, 57, 385–415.
Jagannathan, M., Stephens, C., & Weisbach, M. (2000). Financial flexibility and the choice
between dividends and repurchases. Journal of Financial Economics, 57, 355–384.
John, K., & Williams, J. (1985). Dividends, dilution, and taxes: A signaling equilibrium. Journal
of Finance, 40, 1053–1070.
Jones, R., & Wu, Y. (2010). Executive compensation, earnings management and share holder
litigation. Review of Quantitative Finance and Accounting, 35, 1–20.
Julio, B., & Ikenberry, D. (2004). Reappearing dividends. Journal of Applied Corporate Finance,
16, 89–100.
38 Dividend Payments and Share Repurchases of US Firms 1091
Kahle, K. (2002). When a buyback isn’t a buyback: Open market repurchases and employee
options. Journal of Financial Economics, 63, 235–261.
Koopmans, T. C. (Ed.). (1950). Statistical inference in dynamic economic models. New York:
Wiley.
Kronmal, R. (1993). Spurious correlation and the fallacy of the ratio standard revisited. Journal of
the Royal Statistical Society A, 156, 379–392.
Lee, C., Wu, C., & Djarraya, M. (1987). A further empirical investigation of the dividend
adjustment process. Journal of Econometrics, 35, 267–285.
Lev, B. (2001). Intangibles. Washington, DC: Brookings Institution Press.
Lintner, J. (1956). Distributions of incomes of corporations among dividends, retained earnings,
and taxes. American Economic Review, 46, 97–113.
Miller, M., & Modigliani, F. (1961). Dividend policy, growth, and the valuation of shares. Journal
of Business, 34, 411–433.
Neyman, J. (1952). Lectures and conferences on mathematical statistics and probability (2nd ed.,
pp. 143–154). Washington, DC: U.S. Department of Agriculture.
Neyman, J., & Scott, E. (1948). Consistent estimates based on partially consistent observations.
Econometrica, 16, 1–32.
Numerical Algorithm Group. (1990). Numerical Algorithm Group library. Mark 13. Oxford:
Oxford University.
Pearson, K. (1897). Mathematical contributions to the theory of evolution- on a form of spurious
correlation which may arise when indices are used in the measurement of organs. Proceedings
of the Royal Society of London, 60, 489–497.
Sargan, J. D. (1958). The estimation of economic relationships using instrumental variables.
Econometrica, 26, 393–415.
Skinner, D. J. (2008). The evolving relation between earnings, dividends, and stock repurchases.
Journal of Financial Economics, 87, 582–609.
Stata. (2008). Stata version 10. College Station: StataCorp.
Stephens, C., & Weisbach, M. (1998). Actual share reacquisition in open-market repurchase
programs. Journal of Finance, 53, 313–333.
Tinbergen, J. (1939). The dynamics of share-price formation. Review of Economics and Statistics,
21, 153–160.
Von Eije, H., & Megginson, W. (2008). Dividends and share repurchases in the European Union.
Journal of Financial Economics, 89, 347–374.
Term Structure Modeling and Forecasting
Using the Nelson-Siegel Model 39
Jian Hua
Contents
39.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1094
39.2 Modeling the Term Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1094
39.3 Fitting the Yield Curve . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1095
39.3.1 Data Construction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1095
39.3.2 Estimation Procedure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1096
39.4 Forecasting the Yield Curve . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1096
39.5 Other Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1099
39.6 Generalized Duration Measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1100
39.7 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1101
Appendix 1: One-Step Estimation Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1101
Appendix 2: Two-Step Estimation Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1103
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1103
Abstract
In this chapter, we illustrate some recent developments in the yield curve
modeling by introducing a latent factor model called the dynamic Nelson-Siegel
model. This model not only provides good in-sample fit, but also produces
superior out-of-sample performance. Beyond Treasury yield curve, the model
can also be useful for other assets such as corporate bond and volatility.
Moreover, the model also suggests generalized duration components
corresponding to the level, slope, and curvature risk factors.
The dynamic Nelson-Siegel model can be estimated via a one-step procedure,
like the Kalman filter, which can also easily accommodate other variables of
interests. Alternatively, we could estimate the model through a two-step process
by fixing one parameter and estimating with ordinary least squares. The model is
flexible and capable of replicating a variety of yield curve shapes: upward
J. Hua
Baruch College (CUNY), New York, NY, USA
e-mail: jian.hua@baruch.cuny.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1093
DOI 10.1007/978-1-4614-7750-1_39,
# Springer Science+Business Media New York 2015
1094 J. Hua
sloping, downward sloping, humped, and inverted humped. Forecasting the yield
curve is achieved through forecasting the factors and we can impose either
a univariate autoregressive structure or a vector autoregressive structure on the
factors.
Keywords
Term structure • Yield curve • Factor model • Nelson-Siegel curve • State-space
model
39.1 Introduction
There have been major advances in theoretical term structure models, as well as
their econometric estimation. Two popular approaches to term structure modeling
are no-arbitrage models and equilibrium models. The no-arbitrage tradition focuses
on eliminating arbitrage opportunities by perfectly fitting the term structure at
a point in time, which is important for pricing derivatives; see Hull and White
(1990) and Health et al. (1992), among others. The equilibrium tradition models the
dynamics of the instantaneous rate typically through affine models so that yields at
other maturities can be derived under various assumptions about the risk premium
(e.g., Vasicek 1977; Cox et al. 1985; Duffie and Kan 1996).
For many finance questions, such as bond portfolio management, derivatives pricing,
and risk management, it is crucial to both produce accurate estimates of the current term
structure as well as forecast future interest rate dynamics. One class of models that has
one potential satisfactory answer to these questions is that of the Nelson-Siegel class of
models (see, Nelson and Siegel 1987). Here, we survey some recent developments of
this model. This model not only provides good in-sample fit, but also produces superior
out-of-sample performance. Moreover, the model also suggests generalized duration
components corresponding to the level, slope, and curvature risk factors.
Let Pt(t) be the date t price of a zero-coupon riskless bond that pays $1 in t periods.
Then,
In practice, yield curves, discount curves, and forward curves are not observed.
Instead, they must be estimated from observed bond prices. A popular approach is
to estimate forward rates at the observed maturities, and then, construct
unsmoothed yields by averaging appropriate estimated forward rates. These yields
exactly price the included bonds (see Fama and Bliss 1987).
The original Nelson and Siegel (1987) framework is a convenient and parsimo-
nious three-component exponential approximation. They work with the forward
39 Term Structure Modeling and Forecasting Using the Nelson-Siegel Model 1095
Loadings
0.6 β2 Loadings
0.4
0.2 β3 Loadings
0
0 20 40 60 80 100 120
τ (Maturity, in Months)
rate, which can be viewed as a constant plus a Laguerre function. Diebold and
Li (2006) made it dynamic, and thus, the corresponding yield curve is
1 elt 1 elt lt
yt ðtÞ ¼ b1t þ b2t þ b3t e þ et , (39.2)
lt lt
and call it DNS hereafter. b1 changes all yields uniformly, and it can be called the
level factor (Lt). b2 loads the short rate more heavily, its loading decays to zero as
maturity lengthens, and it can be called the slope factor (St). b3 loads the medium
term more heavily, its loading starts at zero and decays back to zero as maturity
increases, and it can be called the curvature factor (Ct). In Fig. 39.1, we plot the
factor loadings for a specific value of l. l determines the maturity at which the
medium-term (or the curvature factor) loading achieves its maximum, which effec-
tively controls the location of the hump of the curve. The DNS model is different
from the factor analysis, in which one estimates both the unobserved factors and
factor loadings. Here, the framework imposes structure on the factor loadings.
The standard way of measuring the term structure is by means of the spot rate curve
(or equivalently the yield-to-maturity) on zero-coupon bonds.1 The problem with
1
Zero coupon bonds are chosen to limit the “coupon effect,” which implies that two bonds that are
identical in every respect except for bearing different coupon rates can have a different yield-to-
maturity.
1096 J. Hua
zero-coupon yields is that these are not usually directly observed. For example, the
US Treasury Bills do not bear coupon, but they are only available for maturities of
1-year or less, and the longer maturity zero-coupon yields need to derived from
coupon bearing Treasury Notes and Bonds. A popular approach to resolve such an
issue is the bootstrapping procedure by Fama and Bliss (1987), which sequentially
extracts forward rates from bond prices with successively longer maturities and
then takes advantages of the interchangeable nature of the spot rate curve, discount
curve, and forward rate curve.
Because of the structure the model imposes, estimation can be achieved with high
precision. There are two popular approaches: the one-step and two-step procedures.
The one-step procedure can be achieved in two ways. One method is simply to
estimate the model by nonlinear least squares for each month t. The other way is to
transform the system into a state-space representation and estimate l and the factors
via a Kalman filter. Alternatively, for a two-step procedure, Diebold and Li (2006)
advocate prefixing the l, and estimating the factors via ordinary least squares (OLS).
The appendix provides more discussion of these two approaches.
As we can see in Fig. 39.2, the model is flexible and is capable of replicating
a variety of yield curve shapes: upward sloping, downward sloping, humped, and
inverted humped. Figure 39.3 displays the autocorrelations of the estimated factors
and residuals. The level factor displays high persistence and is, of course, positive.
In contrast, the slope and curvature factors are less persistent and assume both
positive and negative values.
Forecasting the yield curve can be achieved through forecasting the factors. We can
impose either a univariate autoregressive structure or a vector autoregressive
structure on the factors.
The yield forecasts based on underlying univariate AR(1) factor specifications
are
1 elt t 1 elt t
^y tþh=t ðtÞ ¼ b1, tþh=t þ b2, tþh=t þ b3, tþh=t elt t , (39.3)
lt t lt t
^ ^
c 1 þ ^g 1 b
b 1, tþh=t ¼ ^ 1t
^
b ¼ ^
c þ ^g ^
b (39.4)
2, tþh=t 2 2 2t
^
b ¼ ^ þ ^g ^
3, tþh=t c 3 3 b 3t
39 Term Structure Modeling and Forecasting Using the Nelson-Siegel Model 1097
Yield (Percent)
Yield (Percent)
7.8
9.5
9.4 7.7
9.3 7.6
9.2
7.5
9.1
9 7.4
8.9 7.3
0 20 40 60 80 100 120 0 20 40 60 80 100 120
Maturity (Months) Maturity (Months)
Yield Curve on 5/30/1997 Yield Curve on 8/31/1998
6.8 5.05
6.6
6.4 5
Yield (Percent)
Yield (Percent)
6.2
4.95
6
5.8
4.9
5.6
5.4 4.85
5.2
5 4.8
0 20 40 60 80 100 120 0 20 40 60 80 100 120
Maturity (Months) Maturity (Months)
Fig. 39.2 Selected fitted yield curves. This figure presents fitted yield curve for selected dates,
together with actual yields (Source: Diebold and Li 2006)
where
^ ^,
c þ ^g b
b tþh=t ¼ ^ t (39.6)
and b ¼ {b1,b2,b3,}0 .
The random walk model,
2
Diebold and Li (2006) conduct a more extensive comparisons with competing models.
1098 J. Hua
Autocorrelation of βˆ 1t Autocorrelation of εˆ 1t
1.0 0.2
0.8
0.1
Autocorrelation
Autocorrelation
0.6
0.4
0.0
0.2
0.0 −0.1
−0.2
−0.4 −0.2
5 10 15 20 25 30 35 40 45 50 55 60 5 10 15 20 25 30 35 40 45 50 55 60
Displacement Displacement
Autocorrelation of βˆ 2t Autocorrelation of εˆ 2t
1.0 0.2
0.8
0.1
Autocorrelation
Autocorrelation
0.6
0.4
0.0
0.2
0.0 −0.1
−0.2
−0.4 −0.2
5 10 15 20 25 30 35 40 45 50 55 60 5 10 15 20 25 30 35 40 45 50 55 60
Displacement Displacement
Autocorrelation of βˆ 3t Autocorrelation of εˆ 3t
1.0 0.2
0.8
0.1
Autocorrelation
Autocorrelation
0.6
0.4
0.0
0.2
0.0 −0.1
−0.2
−0.4 −0.2
5 10 15 20 25 30 35 40 45 50 55 60 5 10 15 20 25 30 35 40 45 50 55 60
Displacement Displacement
Fig. 39.3 Autocorrelation and residual autocorrelation. This figure presents sample autocorrela-
tions of the level, slope, and curvature factors, as well as the sample autocorrelations of AR
(1) models t to the three estimated factors, along with Barletts approximate 95 condence bands
(Source: Diebold and Li 2006)
We define forecast errors at t + h as ytþh ðtÞ ^y tþhjt ðtÞ. Table 39.1 reports the
root mean-squared errors of the out-of-sample performance of the DNS model
versus the random walk model. An error of 0.235 indicates a mean-squared error
of 23.5 basis points in the yield prediction. For 1 month ahead, the random walk is
39 Term Structure Modeling and Forecasting Using the Nelson-Siegel Model 1099
hard to beat, but matters improve dramatically. The DNS model with an AR
(1) specification produces smaller errors, so it clearly has advantages (see Diebold
and Li 2006 for more details). Remember here, we only illustrate the point forecast,
and it can be extended easily to interval forecasting since they are useful for risk
management.
Beyond the US Treasury yield curve, the Nelson-Siegel model has also shown
success in fitting and forecasting other assets and global yield curves (see Krishnan
et al. 2010; Diebold et al. 2008; Hua 2010). Since the model produces three factors
that capture the information in the entire term structure, we can analyze the
dynamic relationship among them. For example, Hua (2010) analyzes the dynamic
interaction between the credit spread term structure and equity option implied
volatility term structure.
Moreover, the DNS framework can be easily augmented with other variables
of interest. We could also analyze how macroeconomic variables are linked
with the interest rate term structure; see Auroba et al. (2006). This can be done
easily in a state-space framework. Frequently, yield variation reacts to
macrovariables, and macrovariables can also be impacted by the yields. Poten-
tially, we can improve our forecasts of one by incorporating information from
the other.
1100 J. Hua
Traditional interest rate risk management focuses on duration and duration man-
agement, which only considers parallel shifts of the yield curve. However, in
practice, nonparallel shifts do exist and these are a significant source of risk. The
DNS model presents a generalized duration component that corresponds to the
level, slope, and curvature risk factors.
We can define a bond duration measure as follows. Let the cash flows from bond
be C1,C2, . . .,CI, and define the associated maturities to be t1,t2,. . .,tI. We assume
that the yield curve is linear in some arbitrary factors f1, f2, and f3,
This is consistent with the DNS setup. Since the price of the bond can be
expressed as,
X
I
P¼ Ci eti yt ðti Þ , (39.9)
i¼1
X XI h i
I
∂P
dP ¼ dyi ðti Þ ¼ Ci eti yt ðti Þ ðti Þ dyt ðti Þ, (39.10)
i¼1
∂yt ðti Þ i¼1
( ) ( )
I
dP X 3 X 1 ti yt ðti Þ
X3 XI
¼ Ci e ti Bj ðti Þ df it ¼ wi ti Bj ðti Þ df it ,
P j¼1 i¼1
P j¼1 i¼1
(39.11)
X
I
Dj ¼ wi ti Bj ðti Þ: (39.12)
i¼1
39 Term Structure Modeling and Forecasting Using the Nelson-Siegel Model 1101
X
I
D1 ¼ wi ti
i¼1
XI
1 elti
D2 ¼ wi (39.13)
i¼1
l
X
I
1 elti
D3 ¼ wi ti elti :
i¼1
l
39.7 Summary
We have presented the DNS model as a capable and flexible model that can fit the
term structure of interest rates in-sample and predict them out-of-sample. It can
easily be extended to a different asset class or accommodate other variables into the
specifications. Given the advantages of the dynamic Nelson-Siegel model in many
dimensions, it is gaining in popularity. Recently, the Board of Governors of the
Federal Reserve System has started publishing the daily estimated factors of the
Treasury yield curve on their website.3
One potential drawback of the DNS model is that it does not rule out arbitrage
opportunities. Recently, an arbitrage free version of the model has been developed;
see Christensen et al. (2009, 2011).
We illustrate the one-step method to estimate the DNS model. If the dynamics of
betas (factors) follow a vector autoregressive process of first order, the model
3
The data can be found at http://www.federalreserve.gov/econresdata/researchdata.htm
1102 J. Hua
immediately forms a state-space system since ARMA state vector dynamics of any
order may be transformed into a state-space form. Thus, the transition equation is
0 1 0 10 1 0 1
b1, t mb1 a11 a12 a13 b1, t1 mb1 1, t
@ b2, t mb2 A ¼ @ a21 a22 a23 A@ b2, t1 mb2 A þ @ 2, t A, (39.14)
b3, t mb3 a31 a32 a33 b3, t1 mb3 3, t
yt ¼ Gft þ et ; (39.16)
state equation,
For optimality of the Kalman filter, we assume that the white noise transaction
and measurement errors be orthogonal to each other.
t 0 Q 0
WN ; , (39.18)
et 0 0 H
0
E f0 t ¼ 0, (39.19)
0
E f0 et ¼ 0: (39.20)
The state-space set up with the application of the Kalman filter delivers maximum-
likelihood estimates and smoothed underlying factors, where all parameters are
estimated simultaneously. Such representation also allows for heteroskedasticity,
missing data, or heavy-tailed measurement errors. Moreover, other useful variables,
such as macroeconomic variables, can be augmented into the state equation to
understand the dynamic interactions between the yield curve and the macroeconomy.
39 Term Structure Modeling and Forecasting Using the Nelson-Siegel Model 1103
The estimation for the two-step procedure requires us to choose a l value first. Once
the l is fixed, the values of the two regressors (factor loadings) can be computed, so
ordinary least squares can be applied to estimate the betas (factors) at each period t.
Doing so is not only simple and convenient, but also eliminates the potential for
numerical optimization challenges. The question is: What is the appropriate value
of l. Recall that l determines the maturity at which the loading achieves its
maximum. For example, for Treasury data, 2 or 3 years are commonly considered
medium term, so a simple average of the two is 30-month. The l value that
maximizes the loadings on the medium-term factor at exactly 30-month is
0.0609. For a different market, the medium maturity could be different, so the
corresponding l value could be different as well.
References
Aruoba, B. S., Diebold, F. X., & Rudebusch, G. D. (2006). The macroeconomy and yield curve:
A dynamic latent factor approach. Journal of Econometrics, 127, 309–338.
Christensen, J. H. E., Diebold, F. X., & Rudebusch, G. D. (2009). An arbitrage-free generalized
Nelson-Siegel term structure model. Econometrics Journal, 12, 33–64.
Christensen, J. H. E., Diebold, F. X., & Rudebusch, G. D. (2011). The affine arbitrage-free class of
Nelson-Siegel term structure models. Journal of Econometrics, 164, 4–20.
Cox, J. C., Ingersoll, J. E., & Ross, S. A. (1985). A theory of the term structure of interest rates.
Econometrica, 53, 385–407.
Diebold, F. X., Ji, L., & Li, C. (2006). A three-factor yield curve model: Non-affine structure,
systematic risk sources, and generalized duration. In L. R. Klein (Ed.), Macroeconomics,
finance and econometrics: Essays in memory of Albert Ando (pp. 240–274). Cheltenham:
Edward Elgar.
Diebold, F. X., & Li, C. (2006). Forecasting the term structure of government bond yields. Journal
of Econometrics, 127, 337–364.
Diebold, F. X., Li, C., & Yue, V. (2008). Global yield curve dynamics and interactions:
A generalized Nelson-Siegel approach. Journal of Econometrics, 146, 351–363.
Duffie, D., & Kan, R. (1996). A yield-factor model of interest rates. Mathematical Finance,
6, 379–406.
Fama, E., & Bliss, R. (1987). The information in long-maturity forward rates. American Economic
Review, 77, 680–692.
Heath, D., Jarrow, R., & Morton, A. (1992). Bond pricing and the term structure of interest rates:
A new methodology for contingent claims valuation. Econometrica, 60, 77–105.
Hua, J. (2010). Option implied volatilities and corporate bond yields: A dynamic factor approach
(Working Paper). Baruch College, City University of New York.
Hull, J., & White, A. (1990). Pricing interest-rate-derivative securities. Review of Financial
Studies, 3, 573–592.
Krishnan, C. N. V., Ritchken, P., & Thomson, J. (2010). Predicting credit spreads. Journal of
Financial Intermediation, 19(4), 529–563.
Nelson, C. R., & Siegel, A. F. (1987). Parsimonious modeling of yield curve. Journal of Business,
60, 473–489.
Vasicek, O. (1977). An equilibrium characterization of the term structure. Journal of Financial
Economics, 5, 177–188.
The Intertemporal Relation Between
Expected Return and Risk on Currency 40
Turan G. Bali and Kamil Yilmaz
Contents
40.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1106
40.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1109
40.3 Estimation Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1114
40.4 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1117
40.5 Time-Varying Risk Aversion in the Foreign Exchange Market . . . . . . . . . . . . . . . . . . . . . . 1125
40.6 Testing Merton’s (1973) ICAPM in Currency Market . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1130
40.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1133
Appendix 1: Maximum Likelihood Estimation of GARCH-in-Mean Models . . . . . . . . . . . . . . . 1135
Appendix 2: Estimation of a System of Regression Equations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1137
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1139
Abstract
The literature has so far focused on the risk-return trade-off in equity markets
and ignored alternative risky assets. This paper examines the presence and
significance of an intertemporal relation between expected return and risk in
the foreign exchange market. The paper provides new evidence on the
intertemporal capital asset pricing model by using high-frequency intraday
data on currency and by presenting significant time variation in the risk aversion
parameter. Five-minute returns on the spot exchange rates of the US dollar
vis-à-vis six major currencies (the euro, Japanese yen, British pound sterling,
Swiss franc, Australian dollar, and Canadian dollar) are used to test the existence
and significance of a daily risk-return trade-off in the FX market based on the
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1105
DOI 10.1007/978-1-4614-7750-1_40,
# Springer Science+Business Media New York 2015
1106 T.G. Bali and K. Yilmaz
GARCH, realized, and range volatility estimators. The results indicate a positive
but statistically weak relation between risk and return on currency.
Our empirical analysis relies on the maximum likelihood estimation of the
GARCH-in-mean models as described in Appendix 1. We also use the seem-
ingly unrelated (SUR) regressions and panel data estimation to investigate the
significance of a time-series relation between expected return and risk on
currency as described in Appendix 2.
Keywords
GARCH • GARCH-in-mean • Seemingly unrelated regressions (SUR) • Panel
data estimation • Foreign exchange market • ICAPM • High-frequency data •
Time-varying risk aversion • High-frequency data • Daily realized volatility
40.1 Introduction
Merton’s (1973) intertemporal capital asset pricing model (ICAPM) indicates that
the conditional expected excess return on a risky market portfolio is a linear
function of its conditional variance plus a hedging component that captures the
investor’s motive to hedge for future investment opportunities. Merton (1980)
shows that the hedging demand component becomes negligible under certain
conditions and the equilibrium relation between risk and return is defined as
Et ðRtþ1 Þ ¼ b Et s2tþ1 (40.1)
where Et(Rt+1) and Et(s2t+1) are, respectively, the conditional mean and variance of
excess returns on a risky market portfolio and b > 0 is the risk aversion parameter
of market investors. Equation 40.1 establishes the dynamic relation that investors
require a larger risk premium at times when the market is riskier.
Many studies investigate the significance of an intertemporal relation between
expected return and risk in the aggregate stock market. However, the existing
literature has not yet reached an agreement on the existence of a positive risk-
return trade-off for stock market indices.1 Due to the fact that the conditional mean
and volatility of the market portfolio are not observable, different approaches,
different data sets, and different sample periods used by previous studies in
estimating the conditional mean and variance are largely responsible for the
contradictory empirical evidence.
The prediction of Merton (1973, 1980) that expected returns should be related to
conditional risk applies not only to the stock market portfolio but also to any risky
portfolio. However, earlier studies have so far focused on the risk-return trade-off in
1
See French et al. (1987), Campbell (1987), Nelson (1991), Campbell and Hentschel (1992), Chan
et al. (1992), Glosten et al. (1993), Scruggs (1998), Harvey (2001), Goyal and Santa-Clara (2003),
Brandt and Kang (2004), Ghysels et al. (2005), Bali and Peng (2006), Christoffersen and Diebold
(2006), Guo and Whitelaw (2006), Lundblad (2007), and Bali (2008).
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1107
equity markets and ignored other risky financial assets. Although there are a few
studies testing the significance of a time-series relation between risk and return
in international equity markets, the focus is generally on the US stock market. It is
also important to note that earlier studies assume a constant risk-return trade-off and
ignore time variation in the risk aversion parameter b.2 This paper examines the
intertemporal relation between expected return and risk in currency markets. The
paper not only investigates ICAPM in the foreign exchange market but examines
the significance of time-varying risk aversion as well.
The foreign exchange market includes the trading of one currency against
another between large banks, central banks, currency speculators, multinational
corporations, governments, and other financial markets and institutions. The
FX market is an interbank or inter-dealer network first established in 1971 when
many of the world’s major currencies moved towards floating exchange rates. It is
considered an over-the-counter (OTC) market, meaning that transactions are
conducted between two counterparties that agree to trade via telephone or elec-
tronic network. Because foreign exchange is an OTC market where brokers/dealers
negotiate directly with one another, there is no central exchange or clearing house.3
The FX market has grown rapidly since the early 1990s. According to the
triennial central bank surveys conducted by the Bank for International Settlements
(BIS), the April 2007 data show an unprecedented rise in activity in traditional
foreign exchange markets compared to 2004. As shown in Table 40.1, average daily
turnover rose to US $3.1 trillion in April 2007, an increase of 69 % (compared to
April 2004) at current exchange rates and 63 % at constant exchange rates.4 Since
April 2001, average daily turnover in foreign exchange markets worldwide
(adjusted for cross-border and local double-counting and evaluated at April 2007
exchange rates) increased by 58 % and 69 % between two consecutive triennial
surveys. Comparing the average daily turnovers of US $500 billion in 1988 and US
$3.1 trillion in 2007 indicates that trading volume in FX markets increased by more
than five times over the past two decades.
The FX market has become the world’s largest financial market, and it is not
uncommon to see over US $3 trillion traded each day. By contrast, the New York
Stock Exchange (NYSE) – the world’s largest equity market with daily trading
volumes in the US $60–80 billion dollar range – is positively dwarfed when
compared to the FX market. Daily turnover in FX markets is now more than ten
times the size of the combined daily turnover on all the world’s equity markets.
2
A few exceptions are Chou et al. (1992), Harvey (2001), and Lettau and Ludvigson (2010).
3
As FX trading has evolved, several locations have emerged as market leaders. Currently, London
contributes the greatest share of transactions with over 32 % of the total trades. Other trading
centers – listed in order of volume – are New York, Tokyo, Zurich, Frankfurt, Hong Kong, Paris,
and Sydney. Because these trading centers cover most of the major time zones, FX trading is a true
24-h market that operates 5 days a week.
4
In addition to “traditional” turnover of US $3.1 trillion in global foreign exchange market, US
$2.1 trillion was traded in currency derivatives.
1108 T.G. Bali and K. Yilmaz
Table 40.1 Reported foreign exchange market turnover by currency paira (Daily averages in
April, in billions of US dollars and percent)
2001 2004 2007
Amount % share Amount % share Amount % share
US dollar/euro 354 30 503 28 840 27
US dollar/yen 231 20 298 17 397 13
US dollar/British pound 125 11 248 14 361 12
US dollar/Australian dollar 47 4 98 5 175 6
US dollar/Swiss franc 57 5 78 4 143 5
US dollar/Canadian dollar 50 4 71 4 115 4
US dollar/other 195 17 295 16 628 21
Euro/yen 30 3 51 3 70 2
Euro/sterling 24 2 43 2 64 2
Euro/Swiss franc 12 1 26 1 54 2
Euro/other 21 2 39 2 112 4
Other currency pairs 26 2 42 2 122 4
All currency pairs 1,173 100 1,794 100 3,081 100
a
Adjusted for local and cross-border double-counting
Even when combining the US bond and equity markets, total daily volumes still do
not come close to the values traded on the currency market.
The FX market is unique because of its trading volumes; the extreme liquidity
of the market; the large number of, and variety of, traders in the market; its
geographical dispersion; its long trading hours (24 h a day except on weekends);
the variety of factors that affect exchange rates; the low margins of profit compared
with other markets of fixed income (but profits can be high due to very large trading
volumes); and the use of leverage.
Earlier studies have so far focused on the US stock market when investigating
the ICAPM. However, with an average daily trading volume of US $3 trillion per
day, Forex is far and away the most enormous financial market in the world,
dwarfing the trading volumes of other markets. We contribute to the existing
literature by examining for the first time the significance of an intertemporal
relation between expected return and risk on currency. We also test whether
aggregate risk aversion in the FX market changes through time.
We utilize 5-min returns on the spot exchange rates of the US dollar vis-à-vis six
major currencies (the euro, Japanese yen, British pound sterling, Swiss franc,
Australian dollar, and Canadian dollar) to construct the daily returns, realized
volatility, and range volatility estimators. Then, using the intraday data-based
daily returns as well as the GARCH, realized, and range-based volatility measures,
we test for the presence and significance of a risk-return trade-off in the FX market.
By sampling the return process more frequently, we improve the accuracy of
the conditional volatility estimate and measure the risk-return relationship at the
daily level. When we assume a constant risk-return trade-off in currency
markets, we find a positive but statistically weak relation between expected return
and risk on currency.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1109
40.2 Data
5
Note that volume percentages should add up to 200 %; 100 % for all the sellers and 100 % for all
the buyers. As shown in Table 40.2, the market shares of seven major currencies add up to 180 %.
The remaining 20 % of the total (200 %) market turnover has been accounted by other currencies
from Europe and from other parts of the world.
1110 T.G. Bali and K. Yilmaz
because the trading activity in FX markets slows down substantially during the
weekends and the major US official holidays. Following Andersen et al. (2001b),
along with the weekends, we removed the following holidays from our sample:
Christmas (December 24–26), New Year’s (December 31–January 2), July 4th,
Good Friday, Easter Monday, Memorial Day, Labor Day, Thanksgiving Day, and
the day after. In addition to official holidays and weekends, we removed 3 days
(March 4, 2002, April 14, 2003, and January 30, 2004) from our sample as these
days contained the longest zero or constant 5-min return sequences that might
contaminate the daily return and variance estimates. As a result, we end up with
a total of 1,556 daily observations.
Panel A of Table 40.3 presents the mean, median, maximum, minimum,
standard deviation, skewness, kurtosis, and autoregressive of order one, AR(1),
statistics for daily returns on the six exchange rates. The standard errors of the
skewness and kurtosis estimates provide evidence that the empirical distributions of
returns on exchange rates are generally symmetric and fat tailed. More specifically,
the skewness measures are statistically insignificant for all currencies, except for
the Japanese yen. The kurtosis measures are statistically significant without any
exception. The Jarque-Bera, JB ¼ n[(S2/6) + (K–3)2/24], is a formal statistic
with the Chi-square distribution for testing whether the returns are normally
distributed, where n denotes the number of observations, S is skewness, and K is
kurtosis. The JB statistics indicate significant departures from normality for
the empirical return distributions of six exchange rates. As expected, daily
returns on exchange rates are not highly persistent, as shown by the negative AR
(1) coefficients which are less than 0.10 in absolute value. Although the
economic significance of the AR(1) coefficients is low, they are statistically
significant at the 5 % or 1 % level for all currencies, except for the British pound
and Australian dollar.
The daily intertemporal relation between expected return and risk on
currency is tested using the daily realized variance of returns on exchange rates.
In very early work, the daily realized variance of asset returns is measured
40
(continued)
1112
by the squared daily returns, where the asset return is defined as the natural
logarithm of the ratio of consecutive daily closing prices. A series of papers by
Andersen et al. (2001a, b, 2003, 2004) indicate that these traditional measures are
poor estimators of day-by-day movements in volatility, as the idiosyncratic
component of daily returns is large. They demonstrate that the realized volatility
measures based on intraday data provide a dramatic reduction in noise and a radical
improvement in temporal stability relative to realized volatility measures based on
daily returns. Andersen et al. (2003) show formally that the concept of realized
variance is, according to the theory of quadratic variation and under suitable
conditions, an asymptotically unbiased estimator of the integrated variance, and
thus it is a canonical and natural measure of daily return volatility.
Following the recent literature on integrated volatility, we use the high-
frequency intraday data to construct the daily realized variance of exchange
rates. To set forth notation, let Pt denote the time t (t 0) exchange rate with the
unit interval t corresponding to 1 day. The discretely observed time-series process
of logarithmic exchange rate returns with q observations per day, or a return horizon
of 1/q, is then defined by
where t ¼ 1/q, 2/q, . . . . We calculate the daily realized variance of exchange rates
using the intraday high-frequency (5-min) return data as
qX
i 1
VARrealized
t ¼ R2ðqÞ, ti=q (40.3)
i¼0
where qi,t is the number of 5-min intervals on day t and Ri,t is the logarithmic
exchange rate return in 5-min interval i on date t.
On a regular trading day, there are 288 5-min intervals. The exchange rate of the
most recent record in a given 5-min interval is taken to be the exchange rate of that
interval. A 5-min return is then constructed using the logarithmic exchange rate
difference for a 5-min interval. With 1,556 days in our full sample, we end up with
using a total of 1,556 288 ¼ 448,128 5-min return observations to calculate daily
return and variance estimates.
Panel B of Table 40.3 presents the summary statistics of the daily realized-
variances of exchange rate returns. The average daily realized variance is 6 105
for AUD/USD, 4.31 105 for CHF/USD, 4.07 105 for JPY/USD, 3.60 105
for CAD/USD, 3.47 105 for EUR/USD, and 2.77 105 for GBP/USD. These
measures correspond to an annualized volatility of 12.30 % for AUD/USD, 10.42 %
for CHF/USD, 10.13 % for JPY/USD, 9.52 % for CAD/USD, 9.35 % for
EUR/USD, and 8.35 % for GBP/USD. A notable point in Panel B is that the daily
realized variances are highly persistent, as shown by the AR(1) coefficients which
are in the range of 0.49–0.64. Consistent with Andersen et al. (2001a, b), the
distributions of realized variances are skewed to the right and have much thicker
tails than the corresponding normal distribution.
1114 T.G. Bali and K. Yilmaz
Market microstructure noises in transaction data such as the bid-ask bounce may
influence our risk measures based on the realized volatility and GARCH volatility
forecasts, even though the data we use contain very liquid financial time series
and thus are least subject to biases created by market microstructure effects.
An alternative volatility measure that utilizes information contained in the high-
frequency intraday data is Parkinson’s (1980) range-based estimator of the daily
integrated variance:
2
VARrange
t ¼ 0:361 ln Pmax
t ln Pmin
t (40.4)
where Pmax
t and Pmin
t are the maximum and minimum values of the exchange rate on
day t. Alizadeh et al. (2002) and Brandt and Diebold (2006) show that the
range-based volatility estimator is highly efficient, approximately Gaussian, and
robust to certain types of microstructure noise such as bid-ask bounce. In addition,
range data are available for many assets over a long sample period.
Panel C of Table 40.3 presents the summary statistics of the daily range variances
of exchange rate returns. The average daily range variance is 4.20 105 for
AUD/USD, 3.56 105 for CHF/USD, 3.15 105 for JPY/USD, 2.63 105 for
CAD/USD, 2.76 105 for EUR/USD, and 2.0 105 for GBP/USD. These
measures correspond to an annualized volatility of 10.29 % for AUD/USD, 9.47 %
for CHF/USD, 8.91 % for JPY/USD, 8.14 % for CAD/USD, 8.34 % for EUR/USD,
and 7.61 % for GBP/USD. These results indicate that the daily realized volatility
estimates are somewhat higher than the daily range volatilities. Another notable
point in Panel C is that the daily range variances are less persistent than the daily
realized variances. Specifically, the AR(1) coefficients are in the range of 0.09–0.34
for the daily range variances. Similar to our findings for the daily realized variances,
the distributions of range variances are skewed to the right and have much thicker
tails than the corresponding normal distribution.
The following GARCH-in-mean process is used with the conditional normal density
to model the intertemporal relation between expected return and risk on currency:
where Rt+1 is the daily return on exchange rates for period t+1, mt+1|t a + b st+1|t
2
is
the conditional mean for period t + 1 based on the information set up to time t,
et+1 ¼ zt+1 st+1|t is the error term with E(et+1) ¼ 0, st+1|t, is the conditional standard
deviation of daily returns on currency, and zt+1 N(0,1) is a random variable drawn
from the standard normal density and can be viewed as information shocks in the
FX market. st+1|t 2
is the conditional variance of daily returns based on the
information set up to time t denoted by Ot. The conditional variance, st+1|t 2
, follows
a GARCH(1,1) process as defined by Bollerslev (1986) to be a function of the last
period’s unexpected news (or information shocks), zt, and the last period’s variance,
st2. f(Rt+1;mt+1|t,st+1|t) is the conditional normal density function of Rt+1 with the
conditional mean of mt+1|t and conditional variance of st+1|t 2
. Our focus is
to examine the magnitude and statistical significance of the risk aversion parameter
b in Eq. 40.5.
Campbell (1987) and Scruggs (1998) point out that the approximate relationship
in Eq. 40.1 may be misspecified if the hedging term in ICAPM is important. To
make sure that our results from estimating Eq. 40.5 are not due to model misspeci-
fication, we added to the specifications a set of control variables that have been used
in the literature to capture the state variables that determine changes in the
investment opportunity set. Several studies find that macroeconomic variables
associated with business cycle fluctuations can predict the stock market.6 The
commonly chosen variables include Treasury bill rates, federal funds rate, default
spread, term spread, and dividend-price ratios. We study how variations in the fed
funds rate, default spread, and term spread affect the intertemporal risk-return
relation.7 Earlier studies also control for the lagged return in the conditional mean
specification.
We obtain daily data on the federal funds rate, 3-month Treasury bill, 10-year
Treasury bond yields, and BAA-rated and AAA-rated corporate bond yields from
the H.15 database of the Federal Reserve Board. The federal funds (FED) rate is the
interest rate at which a depository institution lends immediately available funds
(balances at the Federal Reserve) to another depository institution overnight. It is
a closely watched barometer of the tightness of credit market conditions in the
banking system and the stance of monetary policy. In addition to the fed funds rate,
we use the term and default spreads as control variables. The term spread (TERM) is
calculated as the difference between the yields on the 10-year Treasury bond and
the 3-month Treasury bill. The default spread (DEF) is computed as the difference
between the yields on the BAA-rated and AAA-rated corporate bonds. We test the
significance of the risk aversion parameter, b, after controlling for macroeconomic
variables and lagged return:
6
See Keim and Stambaugh (1986), Chen et al. (1986), Campbell and Shiller (1988), Fama and
French (1988, 1989), Campbell (1987, 1991), Ghysels et al (2005), and Guo and Whitelaw (2006).
7
We could not include the aggregate dividend yield (or the dividend-price ratio) because the data
on dividends are available only at the monthly frequency while our empirical analyses are based on
the daily data.
1116 T.G. Bali and K. Yilmaz
Earlier studies that investigate the daily risk-return trade-off generally rely on
the GARCH-in-mean methodology. In risk-return regressions, it is not common to
use the realized variance measures obtained from the intraday data. In this paper,
we first generate the daily realized variance based on the 5-min returns on exchange
rates and then estimate the following risk-return regression:
Rtþ1 ¼ a þ b Et VARrealized
tþ1 þ etþ1 (40.12)
where Rt+1 is the 1-day ahead return on exchange rate and Et[VARrealized
t+1 ] is proxied
by the lagged realized variance measure, i.e., Et[VARrealized
t+1 ]¼VARrealized
t defined in
Eq. 40.3. As reported in Panel B of Table 40.3, VARrealized t has significant
persistence measured by the first-order serial correlation that makes VARrealizedt
a reasonable proxy for the 1-day ahead expected realized variance. The slope
coefficient b in Eq. 40.12, according to Merton’s (1973) ICAPM, is the relative
risk aversion coefficient which is expected to be positive and statistically
significant.
To control for macroeconomic variables and lagged returns that may
potentially affect the fluctuations in the FX market, we estimate the risk aversion
coefficient, b, after controlling for the federal funds rate, term spread, default
spread, and lagged return:
Rtþ1 a þ b VARrealized
t þ l1 FEDt þ l2 DEFt þ l3 TERMt þ l4 Rt þ etþ1
(40.13)
Rtþ1 ¼ a þ b VARrange
t þ etþ1 (40.14)
Rtþ1 a þ b VARrange
t þ l1 FEDt þ l2 DEFt þ l3 TERMt þ l4 Rt þ etþ1
(40.15)
where VARrange
t is the Parkinson’s (1980) range-based estimator of the daily inte-
grated variance defined in Eq. 40.4.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1117
foreign
Rtþ1 a þ b s2tþ1jt þ l1 LIBORUS
t þ l2 LIBORt þ l3 Rt þ etþ1 (40.16)
foreign
Rtþ1 a þ b VARrealized
t þ l1 LIBORUS
t þ l2 LIBORt þ l3 Rt þ etþ1 (40.17)
Rtþ1 a þ b VARrange
t þ l1 LIBORUS foreign
t þ l2 LIBORt þ l3 Rt þ etþ1 (40.18)
Table 40.4 presents the maximum likelihood parameter estimates and the t-statistics
in parentheses for the GARCH-in-mean model. The risk aversion parameter (b) is
estimated to be positive for all currencies considered in the paper, but the parameter
estimates are not statistically significant, except for the British pound and
the Canadian dollar. Specifically, b is estimated to be 5.18 for the euro, 4.42 for
the Japanese yen, 29.07 for the British pound, 0.87 for the Swiss franc, 11.04 for
the Australian dollar, and 22.40 for the Canadian dollar. Based on the Bollerslev-
Wooldridge (1992) heteroscedasticity consistent covariance t-statistics reported in
Table 40.4, the risk aversion coefficient has a t-statistic of 1.83 for the Canadian
dollar and t-statistic of 1.77 for the British pound. Although we do not have
a strong statistical significance, we can interpret this finding as a positive risk-
return trade-off in the US/Canadian dollar and US dollar/lb exchange rate markets.
Overall, these results indicate a positive but statistically weak relation between
expected return and risk on currency.
8
Assuming that the interest rate is 5 % per annum in the US and 2 % per annum in Japan, the
uncovered interest rate parity predicts that the US dollar would depreciate against the Japanese yen
by 3 %.
1118 T.G. Bali and K. Yilmaz
Table 40.4 Daily risk-return trade-off in foreign exchange markets based on the GARCH-in-
mean model
Parameters EUR JPY GBP CHF AUD CAD
a 0.0005 0.0002 0.0009 0.0003 0.0008 0.0008
(1.46) (0.30) (2.25) (0.72) (2.26) (2.70)
b 5.1772 4.4206 29.065 0.8693 11.042 22.399
(0.47) (0.28) (1.77) (0.08) (1.28) (1.83)
g0 1.09 107 1.69 106 3.37 107 2.60 107 6.97 107 2.45 107
(0.65) (1.76) (2.32) (1.00) (0.92) (2.21)
g1 0.0301 0.0587 0.0430 0.0331 0.0541 0.0420
(4.25) (4.57) (4.51) (4.24) (3.81) (4.82)
g2 0.9672 0.8922 0.9443 0.9617 0.9318 0.9502
(116.10) (35.85) (73.13) (107.35) (50.04) (93.08)
The following GARCH-in-mean process is used with conditional normal density to model the
intertemporal relation between expected return and risk on currency
where Rt+1 is the daily return on exchange rates for period t+1, mt+1|t a+b st+1|t
2
is the conditional
mean for period t+1 based on the information set up to time t denoted by Ot, et+1 ¼ zt+1 st+1|t is the
error term with E(et+1) ¼ 0, st+1|t is the conditional standard deviation of daily returns on currency,
and zt+1 N(0,1) is a random variable drawn from the standard normal density and can be viewed
as information shocks in FX markets. s2t+1|t is the conditional variance of daily returns based on the
information set up to time t denoted by Ot. The conditional variance, st+1|t
2
, follows a GARCH(1,1)
process defined as a function of the last period’s unexpected news (or information shocks), zt, and
the last period’s variance, st2. The table presents the maximum likelihood parameter estimates and
the t-statistics in parentheses
Another notable point in Table 40.4 is the significance of volatility clustering. For
all currencies, the conditional volatility parameters (g1, g2) are positive, between zero
and one, and highly significant. The results indicate the presence of rather extreme
conditionally heteroskedastic volatility effects in the exchange rate process because
the GARCH parameters, g1 and g2, are found to be not only highly significant, but
also the sum (g1 + g2) is close to one for all exchange rates considered in the paper.
This implies the existence of substantial volatility persistence in the FX market.
Table 40.5 reports the daily risk aversion parameter estimates and their statistical
significance for each currency after controlling for macroeconomic variables and
lagged return. The risk-return coefficient estimates are similar to our earlier findings
in Table 40.4. The relationship between expected return and conditional risk is
positive but statistically weak for all exchange rates, except for the British pound
and the Canadian dollar where we have a risk aversion parameter of 36.51 with
t-stat. ¼1.89 for the British pound and 27.86 with t-stat. ¼ 2.15 for the Canadian
dollar. These results indicate that controlling for the hedging demand component of
the ICAPM does not alter our findings.
40
Table 40.5 Daily risk-return trade-off in foreign exchange markets based on the GARCH-in-mean model with control variables
Parameters EUR JPY GBP CHF AUD CAD
a 0.0003 0.0024 0.0012 0.0006 0.0007 0.0013
(0.13) (1.54) (0.73) (0.29) (0.39) (0.95)
b 6.9731 15.986 36.514 17.913 16.478 29.329
(0.68) (0.96) (1.89) (1.17) (1.56) (2.15)
l1 2.19 104 3.76 104 1.00 104 2.43 104 3.11 104 4.64 105
(0.66) (1.13) (0.35) (0.69) (0.90) (0.18)
l2 5.58 105 7.24 104 8.06 104 1.00 104 8.66 105 7.33 104
(0.07) (0.96) (1.13) (0.09) (0.10) (1.03)
l3 5.33 104 6.79 104 2.71 104 5.82 104 5.64 104 1.62 104
(1.45) (1.65) (0.82) (1.43) (1.27) (0.50)
l4 0.055 0.042 0.008 0.062 0.010 0.038
(2.04) (1.48) (0.30) (2.48) (0.35) (1.45)
g0 1.14 107 1.64 106 3.13 107 2.37 107 7.32 107 2.42 107
(0.99) (2.94) (2.30) (1.83) (2.63) (2.15)
g1 0.0311 0.0581 0.0408 0.0331 0.0566 0.0417
(4.56) (4.41) (4.52) (3.81) (4.03) (4.80)
g2 0.9660 0.8938 0.9475 0.9622 0.9285 0.9506
(97.28) (33.74) (77.76) (88.65) (53.11) (92.60)
The following GARCH-in-mean process is used with control variables to estimate the intertemporal relation between expected return and risk on currency
where FEDt, DEFt, and TERMt are macroeconomic variables that proxy for the hedging demand component of ICAPM, and Rt is the lagged daily return. The
federal funds rate (FED) is the interest rate at which a depository institution lends immediately available funds (balances at the Federal Reserve) to another
depository institution overnight. The term spread (TERM) is calculated as the difference between the yields on the 10-year Treasury bond and the 3-month
Treasury bill. The default spread (DEF) is computed as the difference between the yields on the BAA-rated and AAA-rated corporate bonds. The table
1119
presents the maximum likelihood parameter estimates and the t-statistics in parentheses
1120 T.G. Bali and K. Yilmaz
Table 40.5 shows that the slope coefficient (l4) on the lagged return is negative
for all currencies, but it is statistically significant only for the euro (with
t-stat. ¼ –2.04) and the Swiss franc (with t-stat. ¼ –2.48).9 We find a negative
but insignificant first-order serial correlation for the Japanese yen, British pound,
Australian dollar, and Canadian dollar.
Table 40.6 presents the parameter estimates and their Newey and West (1987)
adjusted t-statistics from the risk-return regressions with realized daily variance.
Panel A reports results without the control variables and tests whether the realized
variance obtained from the sum of squared 5-min returns can predict 1-day ahead
returns on exchange rates. The risk aversion parameter (b) is estimated to be
positive for five out of six currencies considered in the paper, but only two of
these parameter estimates are statistically significant at the 10 % level. Specifically,
b is estimated to be 11.39 for the euro, 7.91 for the Japanese yen, 7.91 for the British
pound, 13.78 for the Swiss franc, –1.09 for the Australian dollar, and 11.89 for the
Canadian dollar. Based on the Newey-West (1987) t-statistics reported in
Table 40.6, the Swiss franc has a risk aversion parameter of 13.78 (t-stat. ¼ 2.21)
and the euro has a risk aversion coefficient of 11.39 (t-stat. ¼ 1.77). These results
indicate that the daily realized variance measures obtained from intraday data
positively predict future returns on exchange rates, but the link between risk and
return is generally statistically insignificant.
Panel B of Table 40.6 presents the risk aversion coefficient estimates after
controlling for the federal funds rate, term spread, default spread, and lagged return.
Similar to our findings in Panel A, the risk aversion parameter is estimated to be
18.76 with t-stat. ¼ 2.42 for the euro, 8.77 with t-stat. ¼ 1.80 for the Japanese yen,
and 18.89 with t-stat. ¼ 2.70 for the Swiss franc, indicating a positive and
significant link between the realized variance and the 1-day ahead returns on the
US dollar/euro, US dollar/yen, and US dollar/Swiss franc exchange rates. There
is also a positive but statistically weak relation for the British pound and the
Canadian dollar.
Table 40.7 reports the parameter estimates and their Newey-West t-statistics
from the risk-return regressions with the daily range variance of Parkinson (1980).
As shown in both panels, with and without control variables, the risk aversion
parameter (b) is estimated to be positive but statistically insignificant, except for
the marginal significance of b for the Canadian dollar in Panel B. These results
provide evidence that the daily range volatility obtained from the intraday data
positively predict future returns on exchange rates, but there is no significant
relation between risk and return on currency.
The estimates in Tables 40.6 and 40.7 present a negative and significant
autocorrelation for the euro, Japanese yen, Swiss franc, and Canadian dollar.
The first-order autocorrelation coefficient is negative but statistically insignificant
for the British pound and the Australian dollar.
9
Jegadeesh (1990), Lehmann (1990), and Lo and MacKinlay (1990) provide evidence for the
significance of short-term reversal (or negative autocorrelation) in short-term stock returns.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1121
Table 40.6 Daily risk-return trade-off in foreign exchange markets based on the realized
variance
Panel A. Daily risk-return trade-off without control variables
Parameters EUR JPY GBP CHF AUD CAD
a 0.00073 0.00040 0.00042 0.00088 0.00027 0.00068
(2.84) (1.59) (1.77) (2.90) (0.77) (2.25)
b 11.393 7.9064 7.9145 13.777 1.0895 11.885
(1.77) (1.61) (1.07) (2.21) (0.21) (1.38)
Panel B. Daily risk-return trade-off with control variables
Parameters EUR JPY GBP CHF AUD CAD
a 0.00124 0.00257 0.00080 0.00169 0.00030 0.00074
(0.82) (1.58) (0.57) (0.96) (0.17) (0.49)
b 18.759 8.7656 10.429 18.886 0.0858 13.608
(2.42) (1.80) (1.30) (2.70) (0.01) (1.55)
l1 0.00028 0.00035 0.00025 0.00033 4.4 106 0.00007
(0.99) (1.06) (0.90) (1.03) (0.01) (0.25)
l2 0.00051 0.00112 0.00013 0.00076 0.00059 0.00041
(0.72) (1.33) (0.18) (0.91) (0.63) (0.60)
l3 0.00054 0.00054 0.00041 0.00065 0.00004 0.00012
(1.54) (1.36) (1.19) (1.67) (0.07) (0.37)
l4 0.068 0.048 0.013 0.077 0.017 0.071
(2.81) (1.81) (0.49) (3.00) (0.56) (2.88)
The following regression is estimated with and without control variables to test the significance of
the intertemporal relation between expected return and risk on currency
where VARtrealized is the daily realized variance computed as the sum of squared 5-min returns on
exchange rates. The table presents the parameter estimates and their Newey and West (1987)
t-statistics in parentheses
An interesting observation in Tables 40.5, 40.6, and 40.7 is that the slope
coefficients (l1, l2, l3) on the lagged macroeconomic variables are found to be
statistically insignificant, except for some marginal significance for the term spread
in the regressions with the Swiss franc. Although one would think that unexpected
news in macroeconomic variables could be viewed as risks that would be rewarded in
the FX market, we find that the changes in federal funds rate and term and default
spreads do not affect time-series variation in daily exchange rate returns. Our inter-
pretation is that it would be very difficult for macroeconomic variables to explain daily
variations in exchange rates. If we examined the risk-return trade-off at lower fre-
quency (such as monthly or quarterly frequency), we might observe significant impact
of macroeconomics variables on monthly or quarterly variations in exchange rates.
Panel A of Table 40.8 presents the maximum likelihood parameter estimates and
the t-statistics in parentheses for the GARCH-in-mean model with LIBOR rates for
the US and the corresponding foreign country. The risk aversion parameter (b) is
estimated to be positive for all currencies, but the parameter estimates are
1122 T.G. Bali and K. Yilmaz
Table 40.7 Daily risk-return trade-off in foreign exchange markets based on the range volatility
Panel A. Daily risk-return trade-off without control variables
Parameters EUR JPY GBP CHF AUD CAD
a 0.00051 0.00022 0.00038 0.00053 0.00053 0.00048
(2.54) (1.11) (2.15) (2.28) (2.15) (2.67)
b 6.6224 4.6907 7.8004 6.8730 4.4630 8.4727
(1.32) (1.15) (1.31) (1.56) (1.05) (1.61)
Panel B. Daily risk-return trade-off with control variables
Parameters EUR JPY GBP CHF AUD CAD
a 0.00179 0.00250 0.00088 0.00238 0.00042 0.00053
(1.22) (1.57) (0.65) (1.40) (0.23) (0.37)
b 8.6053 5.0792 8.3969 7.3411 5.3794 9.2256
(1.58) (1.24) (1.36) (1.52) (1.14) (1.69)
l1 0.00036 0.00030 0.00026 0.00045 0.00010 0.00005
(1.26) (0.95) (0.93) (1.41) (0.25) (0.16)
l2 0.00041 0.00106 0.00015 0.00049 0.00038 0.00033
(0.57) (1.29) (0.22) (0.59) (0.42) (0.48)
l3 0.00056 0.00049 0.00040 0.00070 0.00019 0.00009
(1.59) (1.25) (1.18) (1.81) (0.40) (0.28)
l4 0.067 0.051 0.012 0.078 0.026 0.067
(2.77) (1.89) (0.45) (3.13) (0.87) (2.71)
The following regression is estimated with and without control variables to test the significance of
the intertemporal relation between expected return and risk on currency
Rt+1 a + b VARrange
t + l1 FEDt + l2 DEFt + l3 TERMt + l4 Rt + et+1
statistically significant only for the British pound, Australian dollar, and
Canadian dollar. Specifically, b is estimated to be 30.87 for the British pound,
17.75 for the Australian dollar, and 32.90 for the Canadian dollar. Based on
the Bollerslev-Wooldridge heteroscedasticity consistent covariance t-statistics
reported in Table 40.8, the risk aversion coefficient has a t-statistic of 1.82 for the
British pound, 1.84 for the Australian dollar, and 2.36 for the Canadian dollar.
Although we do not have a strong statistical significance, we can interpret this finding
as a positive risk-return trade-off in the US dollar/British pound, US/Australian dollar,
and US/Canadian dollar markets. Overall, the results indicate a positive but statisti-
cally weak relation between expected return and risk on currency.
Another point worth mentioning in Panel A is that the slope coefficients on
the US LIBOR rate are estimated to be positive and statistically significant at the
5 % level for the euro, Japanese yen, and Swiss franc and significant at the 10 %
level for the Canadian dollar. As expected, the slope coefficients on the
LIBOR rates of the corresponding foreign country turn out to be negative but
statistically insignificant.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1123
Table 40.8 Daily risk-return trade-off in foreign exchange markets with LIBOR interest rates
Panel A. GARCH-in-mean
Parameters EUR JPY GBP CHF AUD CAD
a 0.0004 0.0011 0.0019 0.0004 0.0004 0.0006
(0.62) (1.82) (1.97) (0.31) (0.27) (0.58)
b 6.7962 13.422 30.871 18.036 17.753 32.898
(0.66) (0.84) (1.82) (1.33) (1.84) (2.36)
l1 2.16 104 2.86 104 1.41 3.30 104 2.57 104 1.95 104
105
(1.97) (2.53) (0.12) (2.19) (1.47) (1.66)
l2 2.61 104 8.94 1.83 3.44 3.35 4.81
104 104 104 104 104
(1.10) (1.32) (0.78) (1.13) (0.78) (1.73)
l3 0.055 0.042 0.008 0.062 0.010 0.039
(2.22) (1.62) (0.31) (2.46) (0.36) (1.48)
g0 1.15 107 1.64 106 3.11 2.64 107 7.35 107 2.44 107
107
(1.22) (1.30) (2.25) (0.73) (0.99) (2.13)
g1 0.0300 0.0575 0.0408 0.0338 0.0563 0.0424
(4.44) (4.49) (4.51) (3.72) (3.93) (4.87)
g2 0.9672 0.8946 0.9476 0.9608 0.9287 0.9498
(96.57) (32.83) (77.50) (79.92) (50.57) (92.31)
Panel B. Realized variance
Parameters EUR JPY GBP CHF AUD CAD
a 0.00074 0.00108 0.00150 0.00171 0.00118 0.00024
(1.08) (2.52) (1.61) (2.86) (0.77) (0.37)
b 17.558 10.272 10.499 18.730 0.708 15.199
(2.24) (1.95) (1.31) (2.63) (0.11) (1.69)
l1 2.5 104 3.1 104 3.14 4.5 104 4.6 106 1.4 104
106
(2.39) (2.47) (0.03) (2.96) (0.026) (1.00)
l2 3.3 1.6 2.2 104 6.5 1.6 104 3.0
104 103 104 104
(1.81) (2.01) (0.90) (2.19) (0.42) (1.03)
l3 0.069 0.048 0.013 0.078 0.016 0.073
(2.85) (1.79) (0.51) (3.10) (0.53) (2.91)
Panel C. Range variance
Parameters EUR JPY GBP CHF AUD CAD
a 0.00015 0.00079 0.00136 0.00090 0.00079 0.00018
(0.26) (2.13) (1.49) (1.86) (0.52) (0.28)
b 8.181 5.558 8.416 7.400 5.249 9.761
(1.49) (1.29) (1.37) (1.51) (1.03) (1.77)
l1 1.9 104 2.8 104 1.1 3.5 104 7.26 105 9.5 105
105
(1.92) (2.32) (0.09) (2.51) (0.46) (0.68)
(continued)
1124 T.G. Bali and K. Yilmaz
foreign
Rtþ1 a þ b s2tþ1jt þ l1 LIBORUS
t þ l2 LIBORt þ l3 Rt þ etþ1 ,
foreign
Rtþ1 a þ b VARrealized
t þ l1 LIBORUS
t þ l2 LIBORt þ l3 Rt þ etþ1 ,
Rtþ1 a þ b VARrange
t þ l1 LIBORUS foreign
t þ l2 LIBORt þ l3 Rt þ etþ1 ,
where LIBORUS
t and LIBORforeign
t are the LIBOR rates for the US and the corresponding foreign
country
Panel B of Table 40.8 reports the parameter estimates and their Newey-West
adjusted t-statistics from the risk-return regressions with daily realized variance
after controlling for the LIBOR rates and the lagged return. The results indicate
a positive and significant link between the realized variance and the 1-day ahead
returns on the euro, Japanese yen, Swiss franc, and Canadian dollar. There is also
a positive but statistically weak relation for the British pound.
Panel C of Table 40.8 shows the parameter estimates and their Newey-West
t-statistics from the risk-return regressions with the daily range variance of
Parkinson (1980). With LIBOR rates and the lagged return, the risk aversion
parameter (b) is estimated to be positive for all currencies but statistically signif-
icant only for the Canadian dollar. Overall, the results provide evidence that after
controlling for the interest rate differential of two countries, there is a positive but
statistically weak relation between risk and return on currency.
Similar to our earlier findings from the GARCH-in-mean model, Panels B and
C of Table 40.8 show that the slope coefficients on the US LIBOR rate are generally
positive, whereas the slopes on the corresponding foreign LIBOR rates are negative
with a few exceptions.
Many studies fail to identify a statistically significant intertemporal relation
between risk and return of the stock market portfolios. French et al. (1987) find
that the coefficient estimate is not significantly different from zero when they use
past daily returns to estimate the monthly conditional variance.10 Chan et al. (1992)
employ a bivariate GARCH-in-mean model to estimate the conditional variance,
10
When testing monthly risk-return trade-off, French et al. (1987) use the monthly realized
variance obtained from the sum of squared daily returns within a month.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1125
and they also fail to obtain a significant coefficient estimate for the United States.
Campbell and Hentchel (1992) use the quadratic GARCH (QGARCH) model of
Sentana (1995) to determine the existence of a risk-return trade-off within an
asymmetric GARCH-in-mean framework. Their estimate is positive for one
sample period and negative for another sample period, but neither is statistically
significant. Glosten et al. (1993) use monthly data and find a negative but statisti-
cally insignificant relation from two asymmetric GARCH-in-mean models. Based
on semi-nonparametric density estimation and Monte Carlo integration, Harrison
and Zhang (1999) find a significantly positive risk and return relation at a 1-year
horizon, but they do not find a significant relation at shorter holding periods such as
1 month. Using a sample of monthly returns and implied and realized volatilities for
the S and P 500 index, Bollerslev and Zhou (2006) find an insignificant
intertemporal relation between expected return and realized volatility, whereas
the relation between return and implied volatility turns out to be significantly
positive.
Several studies find that the intertemporal relation between risk and return is
negative (e.g., Campbell 1987; Breen et al. 1989; Turner et al. 1989; Nelson 1991;
Glosten et al. 1993; Harvey 2001; Brandt and Kang 2004). Some studies do provide
evidence supporting a positive and significant relation between expected return and
risk on stock market portfolios (e.g., Bollerslev et al. 1988; Scruggs 1998;
Ghysels et al. 2005; Bali and Peng 2006; Guo and Whitelaw 2006; Lundblad
2007; Bali 2008).
Merton’s (1973) ICAPM provides a theoretical model that gives a positive
equilibrium relation between the conditional first and second moments of excess
returns on the aggregate market portfolio. However, Abel (1988), Backus and
Gregory (1993), and Gennotte and Marsh (1993) develop models in which
a negative relation between expected return and volatility is consistent with equi-
librium. As summarized above, there has been a lively debate on the existence and
direction of a risk-return trade-off, and empirical studies are still not in agreement
for the stock market portfolios. The empirical results presented in Tables 40.4, 40.5,
40.6, 40.7, and 40.8 indicate that the intertemporal relation between expected return
and risk on currency is positive but in most cases statistically insignificant. Hence,
our findings from the FX market are in line with some of the earlier studies that
investigated the significance of a risk-return trade-off for the stock market.
Chou et al. (1992), Harvey (2001), and Lettau and Ludvigson (2010) suggest that
the risk-return relation for the stock market may be time varying. In the existing
literature, there is no study investigating the presence and significance of time-
varying risk aversion in the FX market. We have so far assumed a constant
risk-return trade-off in currency markets and found a positive but statistically
insignificant relation between expected return and risk on exchange rates.
1126 T.G. Bali and K. Yilmaz
11
Since the time-varying risk aversion coefficients from estimating Eqs. 40.12 and 40.13 with and
without control variables turn out to be very similar, we only report results from the full
specification of Eq. 40.13. Time-varying risk aversion estimates obtained from the parsimonious
specification of Eq. 40.12 are available from the authors upon request.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1127
Third panel in Fig. 40.1 shows that in the US dollar/lb sterling market, the risk
aversion is generally positive, but there is a long period of time in which we
observe a negative relation between expected return and risk in the US dollar/lb
market. Specifically, 872 out of 1,306 daily risk aversion estimates are positive, but
only 5 out of 872 are marginally significant. Similarly, only 46 out of 434 negative
risk aversion estimates turn out to be statistically significant at the 10 % level.
These results provide evidence that although there is a significant time variation in
the aggregate risk aversion, it is not clear whether the currency trade
generates a larger or smaller risk premium at times when the US dollar/lb FX
market is riskier.
The fourth panel in Fig. 40.1 indicates that in the Swiss franc market, the risk
aversion is estimated to be positive throughout the sample period (2002–2008),
except for a few months in 2006. Only 71 out of 1,306 daily risk aversion
estimates are negative, but none of them is statistically significant. 353 (467) out
of 1,235 positive risk aversion estimates turn out to be statistically significant at
least at the 5 % level (10 % level). These results indicate a positive and relatively
strong time-varying risk aversion, implying that the currency trade generates
a larger risk premium at times when the US dollar/Swiss franc trade becomes
riskier.
The fifth panel in Fig. 40.1 indicates that in the Australian dollar market
736 out of 1,306 daily risk aversion estimates are positive, but none of them is
statistically significant. Only 65 out of 570 negative risk aversion estimates
turn out to be marginally significant at the 10 % level. The figure indicates
a strong time-varying risk aversion, but there is no significantly positive or
negative relation between risk and return in the US/Australian dollar exchange
rate market.
The last panel in Fig. 40.1 demonstrates that in the US/Canadian dollar market,
for slightly more than half of the sample, the risk aversion is estimated to be
positive and slightly less than half of the sample it turns out to be negative.
However, based on the t-statistics of these risk aversion estimates, there is no
evidence for a significantly positive or negative link between expected return and
risk on currency. Only 35 out of 757 positive risk aversion coefficients and
only 46 out of 549 negative risk aversion parameters are found to be significant
at the 10 % level. Although there is a significant time-series variation in the
aggregate risk aversion, trading in the US/Canadian dollar FX market does not
provide clear evidence for a larger or smaller risk premium at times when the
market is riskier.
Figure 40.2 plots the estimated relative risk aversion parameters (b) and their
statistical significance over time from the rolling regressions with a fixed starting
date. Specifically, the first 250 daily return observations of exchange rates and
their realized variances (from 1/3/2002 to 1/7/2003) are used for estimation of the
relative risk aversion parameter for 1/8/2003. The sample is then extended by
adding one observation to the end (from 1/3/2002 to 1/8/2003), and the 1-day
ahead risk-return relation is measured for 1/9/2003. This recursive estimation
procedure is repeated until March 31, 2008.
1128 T.G. Bali and K. Yilmaz
t-statistic
0
2
2 0
−20
1 0
−40 −40
0
−2
−1
−2 −4
2003 2004 2005 2006 2007 2003 2004 2005 2006 2007
U.S.Dollar – British Pound U.S.Dollar – Swiss Franc
80 80
risk aversion risk aversion
t-statistic time -varying risk aversion t-statistic 60
0 20
t-statistic
1
t-statistic
3
0
0 −40 2
−20
−1 1
−80 −40
−2 0
−3 −1
2003 2004 2005 2006 2007 2003 2004 2005 2006 2007
20 20
3
2 0 0
t-statistic
t-statistic
2
1
−20 −20
0 1
−40
−40
−1 0 −60
−2
−1
−3
−4 −2
2003 2004 2005 2006 2007 2003 2004 2005 2006 2007
Fig. 40.1 Rolling regression estimates from the fixed length window of 250 days
Similar to our findings from the fixed rolling window regressions, Fig. 40.2
provides evidence for a significant time variation in the risk aversion estimates for
all currencies considered in the paper. The first panel in Fig. 40.2 shows that in the
US dollar/euro market, the aggregate risk aversion is positive with a few exceptions
in January 2003. Only 16 out of 1,306 risk aversion estimates are negative, but none
of these estimates is statistically significant based on the Newey-West t-statistics.
795 (870) out of 1,290 positive risk aversion estimates turn out to be statistically
significant at least at the 5 % level (10 % level). These results indicate a positive and
strong time-varying risk aversion in the US dollar/euro market.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1129
2.0 10 5
2.0
t-statistic
t-statistic
1.5 1.6 0
0
1.0 1.2 −5
0.8
0.5 −10
0.4
0.0
0.0
−0.5 −0.4
2003 2004 2005 2006 2007 2003 2004 2005 2006 2007
t-statistic
1.0 5 2
0
0
0.5 1
−5 −10
0.0 0
−10
−0.5 −1
2003 2004 2005 2006 2007 2003 2004 2005 2006 2007
15
3 10
2.0 10
t-statistic
t-statistic
5
2 1.5 5
0 1.0 0
1
−5 −5
0.5
0 −10
0.0
-1 −0.5
2003 2004 2005 2006 2007 2003 2004 2005 2006 2007
Fig. 40.2 Rolling regression estimates from the windows with fixed starting point
The second panel in Fig. 40.2 shows that in the US dollar/yen FX market, the
aggregate risk aversion is positive with a few exceptions from March to June 2004.
Only 68 out of 1,306 risk aversion estimates are negative and all of them are
statistically insignificant. 129 out of 1,238 positive risk aversion estimates turn
out to be marginally significant at the 10 % level. These results imply a positive but
statistically weak time-varying risk aversion in the US dollar/yen market.
The third panel in Fig. 40.2 depicts that in the pound sterling market, the risk
aversion is positive throughout the sample, except for a short period of time in 2003.
Only 90 out of 1,306 risk aversion estimates are negative, but they are not
statistically significant. Although there is a significant time variation in the risk
1130 T.G. Bali and K. Yilmaz
aversion and most of the risk-return coefficients are positive, only 41 out of 1,216
positive risk aversion estimates turn out to be significant at the 10 % level.
Therefore, it is not clear whether the currency trade generates a larger or smaller
risk premium at times when the US dollar/lb market is riskier.
The fourth panel in Fig. 40.2 provides evidence that in the Swiss franc market,
the risk aversion is estimated to be positive throughout the sample period
(2002–2008), except for a few days in January 2003. Only 20 out of 1,306 risk
aversion estimates are negative but statistically insignificant. 816 (934) out of 1,286
positive risk aversion estimates turn out to be statistically significant at least at the
5 % level (10 % level). These results suggest a positive and strong time-varying risk
aversion, implying that the currency trade generates a larger risk premium at times
when the US dollar/Swiss franc exchange rate market is riskier.
The fifth panel in Fig. 40.2 shows that in the Australian dollar market only
138 out of 1,306 risk aversion estimates are negative with no statistical significance
even at the 10 % level. Only 106 out of 1,168 positive risk aversion coefficients are
found to be marginally significant at the 10 % level. Although there is significant
time variation in the aggregate risk aversion, the results do not suggest a strong
positive or negative link between expected return and risk in the US/Australian
dollar market.
The last panel in Fig. 40.2 demonstrates that in the US/Canadian dollar market,
the risk aversion is estimated to be positive, except for a few days in May, October,
and November 2003. Similar to our earlier findings, only 41 out of 1,308 risk
aversion estimates are negative with very low t-statistics. However, based on the
statistical significance of positive risk aversion estimates, there is no evidence for
a strong positive link between expected return and risk on currency either.
Only 276 out of 1,265 positive risk aversion coefficients are found to be significant
at the 10 % level. Although there is a significant time-series variation in the
aggregate risk aversion, trading in the US/Canadian dollar FX market does not
provide clear evidence for a larger or smaller risk premium at times when the
market is riskier.
Merton’s (1973) ICAPM implies the following equilibrium relation between risk
and return for any risky asset i:
where r is the risk-free interest rate, mi r is the expected excess return on the risky
asset i, sim denotes the covariance between the returns on the risky asset i and the
market portfolio m, and six denotes a (1 k) row of covariances between the
returns on risky asset i and the k state variables x. A denotes the average relative risk
aversion of market investors, and B measures the market’s aggregate reaction to
shifts in a k-dimensional state vector that governs the stochastic investment
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1131
where the expected conditional covariance of individual exchange rates with the
currency market, Et(sim,t+1), is represented by the 1-day lagged realized covariance,
i.e., Et(sim,t+1) ¼ sim,t. Similarly, the expected conditional variance of the
currency market, Et(sm,t+12), is represented by the 1-day lagged realized variance,
i.e., Et(sm,t+12) ¼ sm,t2.12
The currency market portfolio is measured by the “value-weighted” average
returns on EUR, JPY, GBP, CHF, AUD, and CAD. The weights are obtained from
the “US Dollar Index.” Just as the Dow Jones Industrial Average reflects the general
state of the US stock market, the US Dollar Index (USDX) reflects the general
assessment of the US dollar. USDX does it through exchange rates averaging of US
dollar and six most tradable global currencies. The weights are 57.6 % for EUR,
13.6 % for JPY, 11.9 % for GBP, 9.1 % CAD, 4.2 % for AUD, and 3.6 % for CHF. In
our empirical analysis, daily returns on the currency market, Rm,t+1, are calculated by
multiplying daily returns on the six exchange rates by the aforementioned weights.
We estimate the system of Eq. 40.20 using an ordinary least square (OLS) as
well as a weighted least square method that allows us to place constraints on
coefficients across equations. We constrain the slope coefficient (A) on the lagged
realized variance-covariance matrix (sim,t, sm,t2) to the same value across all the
currencies for cross-sectional consistency. We allow the intercepts (Ci, Cm) to differ
across all the currencies. Under the null hypothesis of the ICAPM, the intercepts
should be jointly zero, and the common slope coefficient (A) should be positive and
statistically significant. We use insignificant estimates of A and the deviations from
zero of the intercept estimates as a test against the validity and sufficiency of
12
Daily realized covariances between the exchange rates and the currency market and daily
realized variance of the currency market are computed using 5-min returns in a day.
1132 T.G. Bali and K. Yilmaz
the ICAPM. In addition to the OLS panel estimates, we compute the t-statistics of
the parameter estimates accounting for heteroskedasticity and autocorrelation as
well as contemporaneous cross-correlations in the error terms. This estimation
methodology for the system of Eq. 40.20 can be regarded as an extension of the
seemingly unrelated regression (SUR) method.
Table 40.9 presents the OLS and SUR panel regression estimates of the
currency-specific intercepts, common slope coefficients on the lagged realized
variance-covariance matrix, and their t-statistics. The parameters and their
t-statistics are estimated using the daily returns on the currency market and the
six exchange rates. The last row reports the Wald statistics with p-values from
testing the joint hypothesis that all intercepts equal zero: H0:
C1 ¼ C2 ¼ . . . ¼ C6 ¼ Cm ¼ 0. A notable point in Table 40.9 is that the common
slope coefficient (A) is positive and statistically significant. Specifically, the risk
aversion coefficient on the realized variance-covariance matrix is estimated to be
23.33 with a t-statistic of 5.40. After correcting for heteroscedasticity, autocorrela-
tion, and contemporaneous cross-correlations, the SUR estimate of the risk aversion
coefficient turns out to be 15.80 with t-stat. ¼ 2.57. These results indicate
a positive and significant relation between risk and return on the currency market.
Another notable point in Table 40.9 is that for both the OLS and SUR estimates,
the Wald statistics reject the hypothesis that all intercepts equal zero. This implies
that the market risk alone cannot explain the entire time-series variation in
exchange rates.
According to the original ICAPM of Merton (1973), the relative risk aversion
coefficient (A) is restricted to the same value across all risky assets, and it is
positive and statistically significant. The common slope estimates in Table 40.9
provide empirical support for the positive risk-return trade-off.
We now test whether the slopes on (sim, sm2) are different across currencies.
We examine the sign and statistical significance of different slope coefficients
(Ai, Am) on (sim, sm2) in the following system of equations:
Ri, tþ1 ¼ Ci þ Ai sim, t þ ei, tþ1 ,
(40.21)
Rm, tþ1 ¼ Cm þ Am s2m, t þ em, tþ1
To determine whether there is a common slope coefficient (A) that corresponds
to the average relative risk aversion, we first estimate the currency-specific
slope coefficients (Ai, Am) and then test the joint hypothesis that H0:
A1 ¼ A 2 ¼ . . . ¼ A 6 ¼ Am .
Table 40.10 presents the OLS and SUR parameter estimates using daily returns
on the six exchange rates and the value-weighted currency market index. As
compared to Eq. 40.20, we have additional six-slope coefficients to estimate in
Eq. 40.21. As shown in Table 40.10, all of the slope coefficients (Ai, Am) are
estimated to be positive and highly significant without any exception. These
results indicate a positive and significant intertemporal relation between risk and
return on the currency market. We examine the cross-sectional consistency of the
intertemporal relation by testing the equality of slope coefficients based on the
Wald statistics. As reported in Table 40.10, the Wald statistics, from testing
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1133
Table 40.9 Testing Merton’s (1973) ICAPM with a common slope coefficient
OLS panel regression SUR panel regression
Intercept t-stat. Intercept t-stat.
AUD 0.00076 4.69 0.00063 3.15
EUR 0.00097 5.24 0.00077 3.53
GBP 0.00062 3.82 0.00049 2.95
CAD 0.00049 3.24 0.00041 2.77
CHF 0.00085 4.72 0.00066 2.98
JPY 0.00045 2.79 0.00032 1.79
Market 0.00078 4.50 0.00061 3.35
Risk aversion Slope t-stat. Slope t-stat.
23.33 5.40 15.80 2.57
H0: Intercepts ¼ 0 Wald p-value Wald p-value
52.20 0.00 17.98 0.0121
Entries report the OLS and SUR panel regression estimates based on the following system of
equations
Ri, tþ1 ¼ Ci þ A sim, t þ ei, tþ1 ,
Rm, tþ1 ¼ Cm þ A s2m, t þ em, tþ1 ,
where sim,t is the 1-day lagged realized covariance between the exchange rate and the currency
market. sm,t2 is the 1-day lagged realized variance of the currency market. A is a common slope
coefficient on the lagged realized variance-covariance matrix. (Ci, Cm) denotes currency-specific
intercepts for AUD, EUR, GBP, CAD, CHF, JPF, and the currency market. The last row reports the
Wald statistics with p-values from testing the joint hypothesis that all intercepts equal zero: H0:
C1 ¼ C2 ¼ . . . ¼ C6 ¼ Cm ¼ 0
the joint hypothesis that H0: A1 ¼ A2 ¼. . . ¼ A6 ¼ Am, is 1.41 for OLS and 5.24
for SUR, which fail to reject the null hypothesis. These results indicate the
equality of positive slope coefficients across all currencies, which empirically
validates the ICAPM.
40.7 Conclusion
There is an ongoing debate in the literature about the intertemporal relation between
stock market risk and return and the extent to which expected stock returns are
related to expected market volatility. Recently, some studies have provided evi-
dence for a significantly positive link between risk and return in the aggregate stock
market, but the risk-return trade-off is generally found to be insignificant and
sometimes even negative. This paper is the first to investigate the presence and
significance of an intertemporal relation between expected return and risk in the
foreign exchange market. The paper provides new evidence on the ICAPM by
using high-frequency intraday data on currency and by presenting significant time
variation in the risk aversion parameter. We utilize daily and 5-min returns on the
spot exchange rates of the US dollar vis-à-vis six major currencies (the euro, Japanese
yen, British pound sterling, Swiss franc, Australian dollar, and Canadian dollar)
1134
Table 40.10 Testing Merton’s (1973) ICAPM with different slope coefficients
OLS panel regression SUR panel regression
Intercept t-stat. Slope t-stat. Intercept t-stat. Slope t-stat.
AUD 0.00084 2.95 28.51 2.08 0.00072 2.85 21.01 1.90
EUR 0.00080 2.80 17.12 1.83 0.00071 3.13 13.62 2.06
GBP 0.00060 2.35 22.19 1.73 0.00030 1.38 4.60 0.46
CAD 0.00063 3.04 35.92 2.47 0.00041 2.23 15.67 1.36
CHF 0.00082 2.66 21.92 2.10 0.00070 2.80 17.32 2.26
JPY 0.00048 1.89 25.25 2.06 0.00050 2.30 26.29 2.77
Market 0.00075 3.08 21.92 2.31 0.00059 3.14 14.53 2.29
H0: Intercepts ¼ 0 Wald p-value Wald p-value
51.40 0.00 15.15 0.0341
H0: Equal slopes Wald p-value Wald p-value
1.41 0.9655 5.24 0.5131
Entries report the OLS and SUR panel regression estimates based on the following system of equations
where sim,t is the 1-day lagged realized covariance between the exchange rate and the currency market. s2m,t is the 1-day lagged realized variance of the
currency market. (Ai, Am) denotes currency-specific slope coefficients on the lagged realized variance-covariance matrix. (Ci, Cm) denotes currency-specific
intercepts for AUD, EUR, GBP, CAD, CHF, JPF, and the currency market. The Wald statistics with p-values are reported from testing the joint hypothesis that
all intercepts equal zero: H0:C1 ¼ C2 ¼ . . . ¼ C6 ¼ Cm ¼ 0. The last row presents the Wald statistics from testing the equality of currency-specific slope
coefficients H0: A1 ¼ A2 ¼ . . . ¼ A6 ¼ Am
T.G. Bali and K. Yilmaz
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1135
and test the existence and significance of a risk-return trade-off in the FX market
using the GARCH, realized, and range-based volatility measures. The maximum
likelihood parameter estimates of the GARCH-in-mean model and the risk-return
regressions with daily realized and range volatility estimators indicate that the
intertemporal relation between risk and return is generally positive but statistically
weak in the FX market.
We provide strong evidence on the time variation of risk aversion parameters for
all currencies considered in the paper. However, the direction of a relationship
between expected return and risk is not clear. The results indicate a positive but
not strong time-varying risk aversion in the US dollar/euro exchange rate market.
The risk-return regressions with realized variance provide evidence for a positive
but statistically weak risk aversion estimates in the US dollar/yen market.
Although there is a significant time variation in risk aversion estimates for the
British pound, it is not clear whether the currency trade generates a larger or smaller
risk premium at times when the US dollar/lb market is riskier. The risk aversion
parameter is estimated to be positive but marginally significant throughout the
sample period for the Swiss franc, implying that the currency trade generally yields
a larger risk premium at times when the US dollar/Swiss franc market is riskier. For
most of the sample, the risk-return coefficients are estimated to be positive but
statistically insignificant for the Canadian dollar, suggesting that the intertemporal
relation between risk and return is flat for the US/Canadian dollar market.
Modeling and estimating the volatility of financial time series has been high on
the agenda of financial economists since the early 1980s. Engle (1982) put forward
the Autoregressive Conditional Heteroskedastic (ARCH) class of models for
conditional variances which proved to be extremely useful for analyzing financial
return series. Since then an extensive literature has been developed for modeling the
conditional distribution of stock prices, interest rates, exchange rates, and futures
prices. Following the introduction of ARCH models by Engle (1982) and their
generalization by Bollerslev (1986), there have been numerous refinements of this
approach to estimating conditional volatility. Most of the refinements have been
driven by empirical regularities in financial data.
Engle (1982) introduces ARCH(p) model:
"
#
1 1 Rtþ1 m 2
f Rtþ1 ; m; stþ1jt ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi exp (40.25)
2ps2tþ1jt 2 stþ1jt
where Rt+1 is the daily return for period t+1, m ¼ a is the constant conditional mean,
et+1 ¼ zt+1 st+1|t is the error term with E(et +1) ¼ 0, st+1|t is the conditional standard
deviation of daily returns, and zt+1 N(0,1) is a random variable drawn from the
standard normal density and can be viewed as information shocks or unexpected
news in the market. st+1|t2 is the conditional variance of daily returns based on the
information set up to time t denoted by Ot. The conditional variance, st+1|t2, follows
an ARCH(p) process which is a function of the last period’s unexpected news
(or information shocks). f(Rt+1;m,st+1|t) is the conditional normal density function of
Rt+1 with the conditional mean of m and conditional variance of st+1|t2.
Given the initial values of et and, the parameters in Eqs 40.22 and 40.24 can be
estimated by maximizing the log-likelihood function over the sample period.
The conditional normal density in Eq. 40.25 yields the following log-likelihood
function:
n
n n 1X Rtþ1 m 2
LogLARCH ¼ lnð2pÞ lnstþ1jt (40.26)
2 2 2 t¼1 stþ1jt
Bollerslev (1986) extends the original work of Engle (1982) and defines the
current conditional variance as a function of the last period’s unexpected news as
well as the last period’s conditional volatility:
Rtþ1 a þ etþ1 (40.27)
etþ1 ¼ ztþ1 stþ1jt , ztþ1 N ð0; 1ÞEðetþ1 Þ ¼ 0 (40.28)
E e2tþ1 jOt ¼ s2tþ1jt ¼ g0 þ g1 e2t þ g2 s2t (40.29)
where the conditional variance, st+1|t2, in Eq. 40.29 follows a GARCH(1,1) process
as defined by Bollerslev (1986) to be a function of the last period’s unexpected
news (or information shocks), zt, and the last period’s variance, s2t . The parameters
in Eqs. 40.27, 40.28, and 40.29 are estimated by maximizing the conditional
log-likelihood function in Eq. 40.26.
Engle et al. (1987) introduce the ARCH-in-mean model in which the conditional
mean of financial time series is defined as a function of the conditional variance. In
our empirical investigation of the ICAPM for exchange rates, we use the following
GARCH-in-mean process to model the intertemporal relation between expected
return and risk on currency
Rtþ1 a þ b s2tþ1jt þ etþ1 (40.30)
etþ1 ¼ ztþ1 stþ1jt ; ztþ1 N ð0; 1Þ;
(40.31)
Eðetþ1 Þ ¼ 0
E e2tþ1 jOt ¼ s2tþ1jt ¼ g0 þ g1 e2t þ g2 s2t (40.32)
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1137
"
#
1 1 Rtþ1 mtþ1jt 2
f Rtþ1 ; mtþ1jt ; stþ1jt ¼ qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi exp (40.33)
2ps2tþ1jt 2 stþ1jt
where Rt+1 is the daily return on exchange rates for period t+1, mt+1|t a + b st+1|t
2
is
the conditional mean for period t+1 based on the information set up to time t,
et+1 ¼ zt+1 st+1|t is the error term with E(et+1) ¼ 0, st+1|t is the conditional standard
deviation of daily returns on currency, and zt+1 N(0,1) is a random variable drawn
from the standard normal density and can be viewed as information shocks in the FX
market. st+1|t
2
is the conditional variance of daily returns based on the information set
up to time t denoted by Ot. The conditional variance, st+1|t 2
, follows a GARCH(1,1)
process as defined by Bollerslev (1986) to be a function of the last period’s
unexpected news (or information shocks), zt, and the last period’s variance,
st2. f(Rt+1;mt+1|t,st+1|t) is the conditional normal density function of Rt+1 with the
conditional mean of mt+1|t and conditional variance of st+1|t
2
.
Given the initial values of et and, the parameters in Eqs. 40.30 and 40.32 can be
estimated by maximizing the log-likelihood function over the sample period. The
conditional normal density in Eq. 40.33 yields the following log-likelihood function
n n
LogLARCH ¼ lnð2pÞ lnstþ1jt
2 2
2 (40.34)
1 X Rtþ1 mtþ1jt
n
2 t¼1 stþ1jt
where the conditional mean mt+1|t a + b s2t+1|t has two parameters and the
conditional variance st+1|t2 ¼ g0 + g1et2 + g2s2t has three parameters. Maximizing
the log-likelihood in Eq. 40.34 yields the parameter estimates (a, b, g0, g1, g2).
The interested reader may wish to consult Enders (2009), Chap. 3, and Tsay
(2010), Chap. 3, for comprehensive analysis of ARCH/GARCH models and their
maximum likelihood estimation. Chapter 3 in Enders (2009) provides a detailed
coverage of the basic ARCH and GARCH models, as well as the GARCH-in-mean
processes and multivariate GARCH in some detail. Chapter 3 in Tsay (2010) pro-
vides a detailed coverage of the ARCH, GARCH, GARCH-M, the exponential
GARCH, and Threshold GARCH models.
y i ¼ X i b i þ ui (40.35)
yi ¼ Xi bi þ Pi ei , (40.38)
with
2 1=2 3
1 r2i 0 0 ... 0
6 7
6 r 1 r2 1=2
... 07
6 i i
1 0 7
6 2 7
6 ri 1 r2i 1=2 ri 0 ... 07
6
Pi ¼ 6 7 (40.39)
: 7
6 7
6 : 7
6 7
4 5
: 2 1=2
rN1
i 1 ri rN2
i rN3
i ... 1
Under this setup, Parks presents a consistent and asymptotically efficient three-
step estimation technique for the regression coefficients. The first step uses single
equation regressions to estimate the parameters of autoregressive model. The
second step uses single equation regressions on transformed equations to estimate
the contemporaneous covariances. Finally, the Aitken estimator is formed using the
estimated covariance,
^ ¼ XT O1 X 1 XT O1 y
b (40.40)
References
Abel, A. B. (1988). Stock prices under time-varying dividend risk: An exact solution in an infinite-
horizon general equilibrium model. Journal of Monetary Economics, 22, 375–393.
Alizadeh, S., Brandt, M. W., & Diebold, F. X. (2002). Range-based estimation of stochastic
volatility models. Journal of Finance, 57, 1047–1092.
Andersen, T. G., Bollerslev, T., & Diebold, F. X. (2004). Parametric and nonparametric volatility
measurement. In Y. Aı̈t-Sahalia & L. P. Hansen (Eds.), Handbook of financial econometrics.
Amsterdam: North Holland.
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Ebens, H. (2001a). The distribution of realized
stock return volatility. Journal of Financial Economics, 61, 43–76.
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Labys, P. (2001b). The distribution of realized
exchange rate volatility. Journal of the American Statistical Association, 96, 42–55.
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Labys, P. (2003). Modeling and forecasting
realized volatility. Econometrica, 71, 579–626.
Backus, D. K., & Gregory, A. W. (1993). Theoretical relations between risk premiums and
conditional variance. Journal of Business and Economic Statistics, 11, 177–185.
Bali, T. G. (2008). The intertemporal relation between expected returns and risk. Journal of
Financial Economics, 87, 101–131.
Bali, T. G., & Engle, R. F. (2010). The intertemporal capital asset pricing model with dynamic
conditional correlations. Journal of Monetary Economics, 57, 377–390.
Bali, T. G., & Peng, L. (2006). Is there a risk-return tradeoff? Evidence from high frequency data.
Journal of Applied Econometrics, 21, 1169–1198.
Bollerslev, T. (1986). Generalized autoregressive conditional heteroscedasticity. Journal of
Econometrics, 31, 307–327.
Bollerslev, T., & Domowitz, I. (1993). Trading patterns and prices in the interbank foreign
exchange market. Journal of Finance, 48, 1421–1443.
Bollerslev, T., Engle, R. F., & Wooldridge, J. M. (1988). A capital asset pricing model with time-
varying covariances. Journal of Political Economy, 96, 116–131.
Bollerslev, T., & Wooldridge, J. M. (1992). Quasi-maximum likelihood estimation and inference
in dynamic models with time varying covariances. Econometric Reviews, 11, 143–172.
Bollerslev, T., & Zhou, H. (2006). Volatility puzzles: A simple framework for gauging return-
volatility regressions. Journal of Econometrics, 131, 123–150.
Brandt, M. W., & Diebold, F. X. (2006). A no-arbitrage approach to range-based estimation of
return covariances and correlations. Journal of Business, 79, 61–73.
Brandt, M. W., & Kang, Q. (2004). On the relationship between the conditional mean and
volatility of stock returns: A latent VAR approach. Journal of Financial Economics, 72,
217–257.
Breen, W., Glosten, L. R., & Jagannathan, R. (1989). Economic significance of predictable
variations in stock index returns. Journal of Finance, 44, 1177–1189.
Campbell, J. Y. (1987). Stock returns and the term structure. Journal of Financial Economics, 18,
373–399.
Campbell, J. Y. (1991). A variance decomposition for stock returns. Economic Journal, 101,
157–179.
1140 T.G. Bali and K. Yilmaz
Campbell, J. Y., & Hentchel, L. (1992). No news is good news: An asymmetric model of changing
volatility in stock returns. Journal of Financial Economics, 31, 281–318.
Campbell, J. Y., & Shiller, R. (1988). The dividend-price ratio and expectations of future
dividends and discount factors. Review of Financial Studies, 1, 195–228.
Chan, K. C., Karolyi, G. A., & Stulz, R. M. (1992). Global financial markets and the risk premium
on U.S. equity. Journal of Financial Economics, 32, 137–167.
Chen, N.-F., Roll, R., & Ross, S. A. (1986). Economic forces and the stock market. Journal of
Business, 59, 383–403.
Chou, R., Engle, R. F., & Kane, A. (1992). Measuring risk aversion from excess returns on a stock
index. Journal of Econometrics, 52, 201–224.
Christoffersen, P. F., & Diebold, F. X. (2006). Financial asset returns, direction-of-change,
forecasting, and volatility dynamics. Management Science, 52, 1273–1287.
Enders, W. (2009). Applied econometric time series (3rd ed.). New York: Wiley.
Engle, R. F. (1982). Autoregressive conditional heteroscedasticity with estimates of the variance
of United Kingdom inflation. Econometrica, 50, 987–1007.
Engle, R. F., Lilien, D. M., & Robins, R. P. (1987). Estimating time varying risk premia in the term
structure: The ARCH M model. Econometrica, 55, 391–407.
Fama, E. F., & French, K. R. (1988). Dividend yields and expected stock returns. Journal of
Financial Economics, 22, 3–25.
Fama, E. F., & French, K. R. (1989). Business conditions and expected returns on stocks and
bonds. Journal of Financial Economics, 25, 23–49.
French, K. R., Schwert, G. W., & Stambaugh, R. F. (1987). Expected stock returns and volatility.
Journal of Financial Economics, 19, 3–29.
Gennotte, G., & Marsh, T. A. (1993). Variations in economic uncertainty and risk premiums on
capital assets. European Economic Review, 37, 1021–1041.
Ghysels, E., Santa-Clara, P., & Valkanov, R. (2005). There is a risk-return tradeoff after all.
Journal of Financial Economics, 76, 509–548.
Glosten, L. R., Jagannathan, R., & Runkle, D. E. (1993). On the relation between the expected
value and the volatility of the nominal excess returns on stocks. Journal of Finance, 48,
1779–1801.
Goyal, A., & Santa-Clara, P. (2003). Idiosyncratic risk matters! Journal of Finance, 58, 975–1008.
Guo, H., & Whitelaw, R. F. (2006). Uncovering the risk-return relation in the stock market.
Journal of Finance, 61, 1433–1463.
Harrison, P., & Zhang, H. (1999). An investigation of the risk and return relation at long horizons.
Review of Economics and Statistics, 81, 399–408.
Harvey, C. R. (2001). The specification of conditional expectations. Journal of Empirical Finance,
8, 573–638.
Jegadeesh, N. (1990). Evidence of predictable behavior of security returns. Journal of Finance, 45,
881–898.
Keim, D. B., & Stambaugh, R. F. (1986). Predicting returns in the stock and bond markets. Journal
of Financial Economics, 17, 357–390.
Lehmann, B. (1990). Fads, martingales, and market efficiency. Quarterly Journal of Economics,
105, 1–28.
Lettau, M., & Ludvigson, S. C. (2010). Measuring and modeling variation in the risk-return
tradeoff. In Handbook of Financial Econometrics by Y. Ait-Sahalia & L. P. Hansen (Eds.),
Vol 1, pp. 617–690. Amsterdam: North Holland.
Lo, A. W., & MacKinlay, A. C. (1990). When are contrarian profits due to stock market
overreaction? Review of Financial Studies, 3, 175–205.
Lundblad, C. (2007). The risk return tradeoff in the long run: 1836–2003. Journal of Financial
Economics, 85, 123–150.
Merton, R. C. (1973). An intertemporal capital asset pricing model. Econometrica, 41, 867–887.
Merton, R. C. (1980). On estimating the expected return on the market: An exploratory investi-
gation. Journal of Financial Economics, 8, 323–361.
40 The Intertemporal Relation Between Expected Return and Risk on Currency 1141
Contents
41.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1144
41.2 Traditional Estimation Methods of VaR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1146
41.3 Quantile Regression . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1147
41.4 Autoregressive Quantile Regression Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1149
41.4.1 The QAR Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1149
41.4.2 Nonlinear QAR and Copula-Based Quantile Models . . . . . . . . . . . . . . . . . . . . . . . . 1151
41.4.3 The CaViaR Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1153
41.5 Quantile Regression of Conditional Heteroskedastic Models . . . . . . . . . . . . . . . . . . . . . . . . . 1153
41.5.1 ARCH Quantile Regression Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1153
41.5.2 GARCH Quantile Regression Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1155
41.6 An Empirical Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1159
41.6.1 Data and the Empirical Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1159
41.6.2 Model Specification Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1160
41.6.3 Estimated VaRs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1163
41.6.4 Performance of the ARCH Quantile Regression Model . . . . . . . . . . . . . . . . . . . . . 1165
41.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1166
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1167
Abstract
This paper studies quantile regression (QR) estimation of Value at Risk (VaR).
VaRs estimated by the QR method display some nice properties. In this paper,
different QR models in estimating VaRs are introduced. In particular, VaR
We thank Prof. C.-F. Lee for helpful comments. This project is partially supported by Boston
College Research fund.
Z. Xiao (*)
Department of Economics, Boston College, Chestnut Hill, MA, USA
e-mail: xiaoz@bc.edu
H. Guo • M.S. Lam
Bertolon School of Business, Salem State University, Salem, MA, USA
e-mail: hguo@salemstate.edu; miranda.lam@salemstate.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1143
DOI 10.1007/978-1-4614-7750-1_41,
# Springer Science+Business Media New York 2015
1144 Z. Xiao et al.
Keywords
ARCH • Copula • GARCH • Non-normality • QAR • Quantile regression • Risk
management • Robust estimation • Time series • Value at risk
41.1 Introduction
The Value at Risk (VaR) is the loss in market value over a given time horizon that is
exceeded with probability t, where t is often set at 0.01 or 0.05. In recent years,
VaR has become a popular tool in the measurement and management of financial
risk. This popularity is spurred both by the need of various institutions for managing
risk and by government regulations (see Blankley et al., 2000; Dowd 1998, 2000;
Saunders 1999). VaR is an easily interpretable measure of risk that summarizes
information regarding the distribution of potential losses. In requiring publicly
traded firms to report risk exposure, the Securities and Exchange Commission
(SEC) lists VaR as a disclosure method “expressing the potential loss in future
earnings, fair values, or cash flows from market movements over a selected period
of time and with a selected likelihood of occurrence.”
Estimation of VaR has attracted much attention from researchers (Duffie and
Pan (1997); Wu and Xiao (2002); Guo et al. (2007)). Many existing methods of VaR
estimation in economics and finance are based on the assumption that financial
returns have normal (or conditional normal) distributions (usually with ARCH or
GARCH effects). Under the assumption of a conditionally normal return distribu-
tion, the estimation of conditional quantiles is equivalent to estimating conditional
volatility of returns. The massive literature on volatility modeling offers a rich
source of parametric methods of this type. However, there is accumulating evidence
that financial time series and return distributions are not well approximated by
Gaussian models. In particular, it is frequently found that market returns display
negative skewness and excess kurtosis. Extreme realizations of returns can
adversely affect the performance of estimation and inference designed for Gaussian
conditions; this is particularly true of ARCH and GARCH models whose
41 Quantile Regression and Value at Risk 1145
higher volatility, only large negative return shocks indicate higher Value at Risk.
GARCH models treat both large positive and negative return shocks as indications of
higher volatility. VaRs estimated by the ARCH/GARCH quantile regression model,
while predicting higher volatility in the ARCH/GARCH component, assign a much
bigger weight to the large negative return shock than the large positive return shock.
The resulting estimated VaRs seem to be closer to reality.
In this chapter, we study quantile regression estimation of VaR. Different quantile
models in estimating VaR are introduced in this paper. In particular, Value at Risk
analysis based on quantile regression of the QAR models, copula models, ARCH
models, GARCH models, and the CaViaR models is systematically introduced. To
highlight the importance of quantile regression method in estimating VaR, we apply
the quantile regression techniques to estimate VaR in International Equity Markets.
Numerical evidence indicates that quantile regression is a robust estimation method
for VaR.
This chapter is organized as follows. We introduce the traditional VaR estima-
tion methods and quantile regression in Sect. 41.2. The quantile autoregression
(QAR) models are given in Sect. 41.3; nonlinear QAR models based on copula and
the CaViaR models are also introduced. Section 41.4 introduces quantile regression
estimation on ARCH and GARCH models. Section 41.5 contains an empirical
application of quantile regression estimation of VaRs. Section 41.6 concludes.
For a time series of returns on an asset, {rt}nt¼1, the t (or 100t%) VaR at time t,
denoted by VaRt, is defined by
where F t1 denotes the information set at time t1, including past values of returns
and possibly the value of some covariates Xt.
If we assume that the time series of returns are modeled by
r t ¼ mt þ st et ,
where rt is the return of an asset at time t and mt, st 2 F t1. The random variables et are
martingale difference sequences. The Conditional Value at Risk of rt given F t1 is
VaRt ðtÞ ¼ mt þ st Qe ðtÞ,
where Qe(t) denotes the Unconditional Value at Risk of the error term et. Assuming
conditional normality, the 5 % VaR at time t can be computed as
where mt and st are the conditional mean and conditional volatility for rt.
41 Quantile Regression and Value at Risk 1147
where the parameter l is called the decay factor (1 > l > 0). Conceptually l should
be estimated using a maximum likelihood approach. RiskMetrics simply set it
optimally at 0.94 for daily data and 0.97 for monthly data. Our analysis is on
weekly data and we set l at 0.95.
There are extensive empirical evidences supporting the use of ARCH and
GARCH models in conditional volatility estimation. Bollerslev et al. (1992) pro-
vide a nice overview of the issue. Sarma et al. (2000) showed that at the 5 % level,
an AR(1)-GARCH(1,1) model is a preferred model under the conditional normality
assumption. The AR(1)-GARCH(1,1) model is specified:
r tþ1 ¼ a0 þ a1 r t þ 2tþ1 2tþ1 I t N 0; s2t
Quantile regression was introduced by Koenker and Bassett (1978) and has
received a lot of attention in econometrics and statistics research in the past two
decades. The quantile function of a scalar random variable Y is the inverse of its
distribution function. Similarly, the conditional quantile function of Y given X is the
inverse of the corresponding conditional distribution function, i.e.,
QY tjXÞ ¼ F1
Y ðtjXÞ ¼ inf fy : FY ðyjXÞ tg,
where FY(yjX) ¼ P(Y yjX). By definition, the t VaR at time t is the t-th
conditional quantile of rt giving information at time t1.
Consider a random variable Y characterized by its distribution function F(y), the
t-th quantile of Y is defined by
If we have a random sample {y1, . . ., yn} from the distribution F, the t-th sample
quantile can be defined as
^ Y ðtÞ ¼ inf yF
Q ^ ðyÞ t ,
^ is the empirical distribution function of the random sample. Note that the
where F
above sample quantile may be found by solving the following minimization
problem:
2 3
X X
min 4 t j y t bj þ ð1 tÞjyt bj5: (41.4)
b2<
t2ft:yt bg t2ft:yt <bg
Koenker and Bassett (1978) studied the analogue of the empirical quantile
function for the linear models and generalized the concept of quantiles to the
regression context.
If we consider the linear regression model
0
Y t ¼ b X t þ ut , (41.5)
where ut are iid mean zero random variables with quantile function Qu(t) and Xt are
k-by-1 vector of regressors including an intercept term and lagged residuals, then,
conditional on the regressor Xt, the t-th quantile of Y is a linear function of Xt:
where b(t)0 ¼ (b1 + Qu(t), b2, , bk). Koenker and Bassett (1978) show that the
t-th conditional quantile of Y can be estimated by an analogue of Eq. 41.4:
^
^ Y ðtjXt Þ ¼ X0 b
Q t t ðtÞ
where
2 3
X 0
X 0
^ ðtÞ ¼ arg min 4
b t y t x t b þ ð1 tÞyt xt b5 (41.6)
b2<k t2ft:yt xt bg t2ft:yt <xt bg
is called as the regression quantiles. Let rt(u) ¼ u(tI(u < 0)), then
X
^ ðtÞ ¼ arg min
b
0
rt yt xt b :
b2<k t
behavior of the conditional distribution of the response near the specified quantile.
As a result, the estimated coefficient vector ^y ðtÞ is not sensitive to outlier obser-
vations. Such a property is especially attractive in financial applications since many
financial data like, say, portfolio returns (or log returns) are usually heavy tailed and
thus not normally distributed.
The quantile regression model has a mathematical programming representation
which facilitates the estimation. Notice that the optimization problem (Eq. 41.6)
may be reformulated as a linear program by introducing “slack” variables to
represent the positive and negative parts of the vector of residuals (see Koenker
and Bassett (1978) for a more detailed discussion). Computation of the regression
quantiles by standard linear programming techniques is very efficient. It is also
straightforward to impose the nonnegativity constraints on all elements of g.
Barrodale and Roberts (1974) proposed the first efficient algorithm for L1-
estimation problems based on modified simplex method. Koenker and d’Orey
(1987) modified this algorithm to solve quantile regression problems. For very
large quantile regression problems, there are some important new ideas which
speed up the performance of computation relative to the simplex approach under-
lying the original code. Portnoy and Koenker (1997) describe an approach that
combines some statistical preprocessing with interior point methods and achieves
faster speed over the simplex method for very large problems.
this implies, of course, that the right-hand side of Eq. 41.7 is monotonically
increasing in t. In the above QAR model, the autoregressive coefficients may be
t-dependent and thus can vary over different quantiles of the conditional
1150 Z. Xiao et al.
distribution. Consequently, the conditioning variables not only shift the location of
the distribution of Yt but also may alter the scale and shape of the conditional
distribution. The QAR models play a useful role in expanding the modeling
territory of the classical autoregressive time-series models, and the classical
AR(p) model can be viewed as a special case of QAR by setting yj(t) (j ¼ 1, . . ., p)
to constants.
Koenker and Xiao (2006) studied the QAR model. The QAR model can be
estimated by
X
^y ðtÞ ¼ min rt Y t y> Xt , (41.8)
y
t
where Xt ¼ (1, Yt1, . . ., Ytp)> and y(t) ¼ (y0(t), y1(t), . . ., yp(t))>; they show that
under regularity assumptions, the limiting distribution of the QAR estimator is
given by
pffiffiffi^
n y ðtÞ yðtÞ ) N 0, tð1 tÞO1 1
1 O0 O1 ,
V t ¼ y0 ðU t Þ Ey0 ðU t Þ þ ½y1 ðU t Þ Ey1 ðU t ÞY t1 þ þ yp ðU t Þ Eyp ðU t Þ Y tp ;
More complicated functional forms with nonlinearity can be considered for the
conditional quantile function if we are interested in the global behavior of the time
series. If the t-th conditional quantile function of Yt is given by
where Xt is the vector containing lagged Ys, we may estimate the vector of
parameters y(t) (and thus the conditional quantile of Yt) by the following nonlinear
quantile regression:
X
min rt ðY t H ðXt , yÞÞ: (41.10)
y
t
1X
f t QY t ðtjXt Þ H_ y ðXt , yðtÞÞH_ y ðXt , yðtÞÞ> ! V ðtÞ,
P
V n ð tÞ ¼
n t
1X _
Hy ðXt , yðtÞÞH_ y ðXt , yðtÞÞ> ! OðtÞ,
P
On ðtÞ ¼
n t
and
1 X _
pffiffiffi Hy ðxt , yðtÞÞY t ðett Þ ) N ð0, tð1 tÞOðtÞÞ,
n t
where V(t) and O(t) are non-singular; then under appropriate assumptions, the
nonlinear QAR estimator ^y ðtÞ defined as solution of Eq. 41.10 is root-n consistent
and
pffiffiffi^
n y ðtÞ yðtÞ ) N 0, tð1 tÞV ðtÞ1 OðtÞV ðtÞ1 : (41.11)
For any t 2 (0, 1), solving t ¼ Pr[Yt < y|Yt1 ¼ x] C1(F(x), F(y)) for
y (in terms of t), we obtain the t-th conditional quantile function of Yt given
Yt1 ¼ x:
QY t tjxÞ ¼ F1 C1
1 ðt; F ðxÞÞ ,
QY t tjxÞ ¼ F1 ðhðx; a; bÞ, bÞ H ðx; yÞ: (41.12)
one can easily verify that the t-th conditional quantile function of Ut given ut1 is
h i1=a
QUt tjut1 Þ ¼ ta=ð1þaÞ 1 ua
t1 þ 1 :
See Bouyé and Salmon (2008) for additional examples of copula-based condi-
tional quantile functions.
Although the quantile function specification in the above representation assumes
the parameters to be identical across quantiles, we may permit the estimated
parameters to vary with t, thus extending the original copula-based QAR models
to capture a wide range of systematic influences of conditioning variables on the
conditional distribution of the response. By varying the choice of the copula
specification, we can induce a wide variety of nonlinear QAR(1) dependence, and
41 Quantile Regression and Value at Risk 1153
the choice of the marginal enables us to consider a wide range of possible tail
behavior as well. In many financial time-series applications, the nature of the
temporal dependence varies over the quantiles of the conditional distribution.
Chen et al. (2009) studied asymptotic properties of the copula-based nonlinear
quantile autoregression.
X
p X
q
Qut ðtjF t1 Þ ¼ b0 þ bi Quti ðtjF ti1 Þ þ aj ‘ Xtj , (41.13)
i¼1 j¼1
where Xtj 2 F tj, F tj is the information set at time tj. A natural choice of Xtj is
the lagged u. When we choose Xtj ¼ jutjj, we obtain GARCH-type CaViaR
models:
X
p X
q
Qut ðtjF t1 Þ ¼ b0 þ bi Quti ðtjF ti1 Þ þ aj utj :
i¼1 j¼1
X
p
Qut ðtjF t1 Þ ¼ b0 þ bi Quti ðtjF ti1 Þ:
i¼1
where 0 < g0 < 1, g1, . . ., gp 0 and et are iid(0,1) random variables with pdf f (·)
and CDF F (·). Let Zt ¼ (1,jut1j, . . .,jutqj)> and g(t) ¼ (g0F1(t), g1F1(t), . . .,
gqF1(t))>; the conditional quantiles of ut is given by
Qut ðtjF t1 Þ ¼ g0 ðtÞ þ g1 ðtÞjut1 j þ þ gp ðtÞutp ¼ gðtÞ> Z t
where g ¼ (g0, g1, , gq)>. The asymptotic behavior of the above quantile
regression estimator is given by Koenker and Zhao (1996). In particular,
suppose that ut is given by model (Eq. 41.14), f is bounded and continuous, and
f(F1(t)) > 0 for any 0 < t < 1. In addition, if Ejutj2+d < 1, then the regression
quantile ^g ðtÞ of Eq. 41.15 has the following Bahadur representation:
pffiffiffi S1 1 X n
nð^g ðtÞ gðtÞÞ ¼ 11 pffiffiffi Z> y ðett Þ þ op ð1Þ
f F ð tÞ n t¼1 t t
0
where S1 ¼ EZtZt/st and ett ¼ etF1(t). Consequently,
!
pffiffiffi tð1 tÞ 1 1 0
nð^g ðtÞ gðtÞÞ ¼ N 0, 1 2 S1 S0 S1 , with S0 ¼ EZ t Z t :
f F ð tÞ
where Xt ¼ (1, Yt1, . . .,Ytp)>, a ¼ (a0, a1, . . ., ap)>, and ut is a linear ARCH(p)
process given by model (Eq. 41.14). The conditional quantiles of Yt is then given by
ARCH models are easier to estimate, but cannot parsimoniously capture the
persistent influence of long past shocks comparing to the GARCH models. How-
ever, quantile regression GARCH models are highly nonlinear and thus compli-
cated to estimate. In particular, the quantile estimation problem in GARCH models
corresponds to a restricted nonlinear quantile regression, and conventional
nonlinear quantile regression techniques are not directly applicable.
Xiao and Koenker (2009) studied quantile regression estimation of the following
linear GARCH(p, q) model:
ut ¼ st et , (41.18)
st ¼ b0 þ b1 st1 þ þ bp stp þ g1 jut1 j þ þ gq utq : (41.19)
Let F t1 represents information up to time t1; the t-th conditional quantile of
ut is given by
where Zt ¼ (1, st1, . . .,stp,jut1j, . . .,jutqj)> and y(t)> ¼ (b0, b1, . . ., bp,
g1, . . ., gq)F1(t).
Since Zt contains stk(k ¼ 1, , q) which in turn depends on unknown
parameters y ¼ (b0, b1, . . ., bp, g1, . . ., gq), we may write Zt as Zt(y) to emphasize
the nonlinearity and its dependence on y. If we use the following nonlinear quantile
regression
X
min rt ut y> Z t ðyÞ , (41.21)
y
t
focus on the specified quantile to find the best local estimate for the conditional
quantile. Let
X
1
st ¼ a0 þ aj utj : (41.22)
j¼1
and
X
1
Qut ðtjF t1 Þ ¼ a0 ðtÞ þ aj ðtÞutj ,
j¼1
∨
Qut (t | Ft−1 ) = Qut (t | Ft−1 ) + Op (m / n ).
However, Monte Carlo evidence indicates that the simple sieve approximation
does not directly provide a good estimator for the GARCH model, but it serves as an
adequate preliminary estimator. Since the first step estimation focuses on the global
model, it is desirable to use information over multiple quantiles in estimation.
41 Quantile Regression and Value at Risk 1157
h i>
a ¼ ½a1 ; . . . ; am ; q1 ; . . . ; qK > , a ðt1 Þ> , . . . , e
p¼ e a ðtK Þ> ,
where g ¼ [q1, . . ., qK]> and a ¼ [1, a1, a2, . . ., am]>; we consider the following
estimator for the vector a that combines information over the K quantile estimates
based on the restrictions aj ðtÞ ¼ aj Qet ðtÞ:
X
m
st ¼ e
e a0 þ a j utj :
e
j¼1
the two-step estimator of y(t)> ¼ (b0(t), b1(t), . . ., bp(t), g1(t), . . ., gq(t)) is then
_
given by the solution of Eq. 41.26, y ðtÞ, and the t-th conditional quantile of ut can
be estimated by
Let e
a ðtÞ be the solution of Eq. 41.24; then under appropriate assumptions, we have
a ðtÞ aðtÞ
2 ¼ Op ðm=nÞ,
ke (41.27)
" #
1 X n
xt x> 1 X n
Dn ¼ t
, S n ð tÞ ¼ xt xT Y 2 ðutt Þ,
n t¼mþ1 st n t¼mþ1 t t
where g0 and a0 are the true values of vectors g ¼ [q1, . . ., qK]> and a ¼ [1, a1, a2,
. . ., am]>, and
2 3
0 0
6 1 0 7
Jm ¼ 6
4⋮
7
⋱ ⋮5
0 1
where
2 !3
y ðutt Þ
6 D1
n xt t1 1 7
6 f e F1e ðt1 Þ 7
pffiffiffi 1 X n 6
6
7
7
nðp pÞ ¼ pffiffiffi ! 7 þ op ð1Þ,
n t¼mþ1 6
6 7
4 ytm ðuttm Þ 5
D1 x t 1
n
f e F e ð tm Þ
41 Quantile Regression and Value at Risk 1159
and the two-step estimator ^y ðtÞ based on Eq. 41.26 has asymptotic representation:
( )
pffiffiffi^ 1 1 1 X pffiffiffi
n y ðtÞ yðtÞ ¼ 1 O pffiffiffi Z t yt ðutt Þ þ O1 G nðe a aÞ þ op ð1Þ,
f e Fe ðtÞ n t
X
p
Zt
G¼ yk Ck , Ck ¼ E ðjutk1 j; . . . ; jutkm jÞ :
k¼1
st
In this section, we apply the quantile regression method to five major world equity
market indexes. The data used in our application are the weekly return series, from
September 1976 to June 2008, of five major world equity market indexes: the US
S&P 500 Composite Index, the Japanese Nikkei 225 Index, the UK FTSE
100 Index, the Hong Kong Hang Seng Index, and the Singapore Strait Times
Index. The FTSE 100 Index data are from January 1984 to June 2008. Table 41.1
reports some summary statistics of the data.
The mean weekly returns of the five indexes are all over 0.1 % per week, with the
Hang Seng Index producing an average return of 0.23 % per week, an astonishing
increase in the index level over the sample period. In comparison, the average
return of Nikkei 225 index is only 0.1 %. The Hang Seng’s phenomenal rise does
not come without risk. The weekly sample standard deviation of the index is
3.76 %, the highest of the five indexes. In addition, over the sample period the
Hang Seng suffered four larger than 15 % drop in weekly index level, with
maximum loss reaching 35 %, and there were 23 weekly returns below 10 %!
As has been documented extensively in the literature, all five indexes display
negative skewness and excess kurtosis. The excess kurtosis of Singapore Strait
Times Index reached 19.31, to a large extent driven by the huge 1 week loss of
47.47 % during the 1987 market crash. The autocorrelation coefficients for all five
indexes are fairly small. The Hang Seng Index seems to display the strongest
autocorrelation with the AR(1) coefficient equal to 0.0891.
We consider an AR-linear ARCH model in the empirical analysis. Thus, the
return process is modeled as
where
ut ¼ st et , st ¼ g0 þ g1 jut1 j þ þ gq utq ,
0
Qut ðtjF t1 Þ ¼ gðtÞ Zt
0
gðtÞ ¼ g0 ðtÞ, g1 ðtÞ, . . . , gq ðtÞ , and Zt ¼ 1; jut1 j; . . . ; utq :
0
For each time series, we first conduct model specification analysis and choose the
appropriate lags for the mean equation and the quantile ARCH component. Based on
the selected model, we use Eq. 41.28 to obtain a time series of residuals. The residuals
are then used in the ARCH VaR estimation using a quantile regression.
We conduct sequential tests for the significance of the coefficients on lags. The
inference procedures we use here are asymptotic inferences. For estimation of the
covariance matrix, we use the robust HAC (Heteroskedastic and Autocorrelation
Consistent) covariance matrix estimator of Andrews (1991) with the data-
dependent automatic bandwidth parameter estimator recommended in that paper.
First of all, we choose the lag length in the autoregression,
r t ¼ a0 þ a1 r t1 þ þ as r ts þ ut ,
41 Quantile Regression and Value at Risk 1161
using a sequential test of significance on lag coefficients. The maximum lag length
that we start with is s ¼ 9, and the procedure is repeated until a rejection occurs.
Table 41.2 reports the sequential testing results for the S&P 500 index. The
t-statistics of all the coefficients are listed for nine rounds of the test. We see that
the t-statistic of the coefficient with the maximum number of lags does not become
significant until s ¼ 1, the ninth round. The preferred model is an AR(1) model. The
selected mean equations for all five indexes are reported in Table 41.4.
Our next task is to select the lag length in the ARCH effect
ut ¼ g0 þ g1 jut1 j þ þ gq utq et :
dF1 ðtÞ 1
¼ 1 ; (41.29)
dt f F ðtÞ
h 00
i1=3
hH S ¼ z2=3
a 1:5sðtÞ=s ðtÞ n1=3 ,
where za satisfies F(za) ¼ 1 a/2 for the construction of 1a confidence intervals.
In the absence of additional information, s(t) is just the normal density. Starting
with qmax ¼ 10, a sequential test was conducted and results for the 5 % VaR model
of the S&P 500 Index are reported in Table 41.3. We see that in the fourth round,
the t-statistic on lag 7 becomes significant. The sequential test stops here, and it
suggests that ARCH(7) is appropriate.
Based on the model selection tests, we decide to use the AR(1)-ARCH(7)
regression quantile model to estimate 5 % VaR for the S&P 500 index. We also
conduct similar tests on the 5 % VaR models for other four indexes. To conserve
space we do not report the entire testing process in the paper. Table 41.4 provides
a summary of the selected models based on the tests. The mean equations
1162
Table 41.2 VaR model mean specification test for the S&P 500 Index
Round 1st 2nd 3rd 4th 5th 6th 7th 8th 9th
a0 3.3460 3.3003 3.2846 3.3248 3.2219 3.7304 3.1723 3.0650 3.8125
a1 1.6941 1.7693 1.8249 1.9987 1.9996 2.0868 2.1536 2.097 2.2094
a2 1.2950 1.3464 1.1555 1.0776 0.0872 1.3106 1.2089 1.0016
a3 0.9235 0.9565 0.9774 1.5521 0.8123 0.8162 0.9553
a4 1.0414 1.0080 0.9947 1.0102 0.9899 0.1612
a5 0.7776 0.7642 0.7865 0.8288 0.7662
a6 0.2094 0.5362 0.7166 0.8931
a7 1.5594 1.5426 1.5233
a8 0.8926 0.8664
a9 0.3816
This table reports the test results for the VaR model mean equation specification for the S&P 500 Index. The number of lags in the AR component of the ARCH
model is selected according to the sequential test. The table reports the t-statistic for the coefficient with the maximum lag in the mean equation
Z. Xiao et al.
41 Quantile Regression and Value at Risk 1163
Table 41.3 5 % VaR model ARCH specification test for the S&P 500 Index
Round 1st 2nd 3rd 4th
g0 16.856 15.263 17.118 15.362
g1 2.9163 3.1891 3.2011 3.1106
g2 1.9601 2.658 2.533 2.321
g3 1.0982 1.0002 0.9951 1.0089
g4 0.6807 0.8954 1.1124 1.5811
g5 0.7456 0.8913 0.9016 0.9156
g6 0.3362 0.3456 0.4520 0.3795
g7 1.9868 2.0197 1.8145 2.1105
g8 0.4866 0.4688 1.5631
g9 1.2045 1.0108
g10 1.1326
This table reports the test results for the 5 % VaR model specification for the S&P 500 Index. The
number of lags in the volatility component of the ARCH model is selected according to the test.
The table reports the t-statistic for the coefficient with the maximum lag in the ARCH equation
generally have one or two lags, except the Hang Seng Index, which has a lag of
3 and displays more persistent autoregressive effect.
For the ARCH equations, at least six lags are needed for the indexes.
The estimated parameters for the mean equations for all five indexes are reported in
Table 41.5. The constant term for the five indexes is between 0.11 % for the Nikkei
and 0.24 % for the Hang Seng. As suggested by Table 41.1, the Hang Seng seems to
display the strongest autocorrelation, and this is reflected in the four lags chosen by
the sequential test. Table 41.6 reports the estimated quantile regression ARCH
parameters for the 5 % VaR model:
USA – S&P 500 Index. The estimated 5 % VaRs generally range between 2.5 %
and 5 %, but during very volatile periods they could jump over 10 %, as what
happened in October 1987. During high-volatility periods, there is high variation
in estimated VaRs.
1164 Z. Xiao et al.
Table 41.6 Estimated ARCH equation parameters for the 5 % VaR model
Parameter S&P 500 Nikkei 225 FTSE 100 Hang Seng Singapore ST
g0 0.0351 0.0421 0.0346 0.0646 0.0428
(0.0016) (0.0023) (0.0013) (0.0031) (0.0027)
g1 0.2096 0.0651 0.0518 0.1712 0.1119
(0.0711) (0.0416) (0.0645) (0.0803) (0.0502)
g2 0.1007 0.1896 0.0588 0.0922 0.1389
(0.0531) (0.0415) (0.0665) (0.0314) (0.0593)
g3 0.0101 0.1109 0.0311 0.2054 0.0218
(0.0142) (0.0651) (0.0242) (0.0409) (0.0379)
g4 0.1466 0.0528 0.0589 0.0671 0.1102
(0.0908) (0.0375) (0.0776) (0.0321) (0.0903)
g5 0.0105 0.0987 0.0119 0.0229 0.1519
(0.0136) (0.0448) (0.0123) (0.0338) (0.0511)
g6 0.0318 0.0155 0.0876 0.0359 0.0311
(0.0117) (0.0297) (0.0412) (0.0136) (0.0215)
g7 0.2323 0.1123
(0.0451) (0.0517)
This table reports the estimated parameters of the ARCH equation for the 5 % VaR model for the
five global indexes. The standard errors are in parentheses under the estimated parameters
Japan – Nikkei 225 Index. The estimated VaR series is quite stable and remains at
the 4 % and the 7 % level from 1976 till 1982. Then the Nikkei 225 Index took
off and appreciated about 450 % over the next 8 years, reaching its highest level
at the end of 1989. This quick rise in stock value is accompanied by high risk,
manifested here by the more volatile VaR series. In particular, the VaRs fluctu-
ated dramatically, ranging from a low of 3 % to a high of 15 %. This volatility in
VaR may reflect both optimistic market outlook at times and worry about high
valuation and the possibility of a market crash. That crash did come in 1990, and
41 Quantile Regression and Value at Risk 1165
10 years later, the Nikkei 225 Index still hovers around at a level which is about
half off the record high in 1989. The 1990s is far from a rewarding decade for
investors in the Japanese equity market. Average weekly 5 % VaR is about 5 %,
and the variation is also very high.
UK – FTSE 100 Index. The 5 % VaR is very stable and averages about 3 %. They
stay very much within the 2–4 % band, except on a few occasions, such as the
1987 global market crash.
Hong Kong – Hang Seng Index. The Hang Seng Index produces an average return
of 0.23 % per week. The Hang Seng’s phenomenal rise does not come without
risk. We mentioned above that the weekly sample standard deviation of the
index is 3.76 %, the highest of the five indexes. In addition, the Hong Kong stock
market has had more than its fair share of the market crashes.
Singapore – Strait Times Index. Interestingly, the estimated VaRs display a pattern
very similar to that of the UK FTSE 100 Index, although the former is generally
larger than the latter. The higher risk in the Singapore market did not result in
higher return over the sample period. Among the five indexes, the Singapore
market suffered the largest loss during the 1987 crash, a 47.5 % drop in a week.
The market has since recovered much of the loss. Among the five indexes, the
Singapore market only outperformed the Nikkei 225 Index over this period.
percentage (probability) levels. The GARCH model with the normality assumption,
being a good volatility model, is not able to produce good VaR estimates. The
quantile regression model does not assume normality and is well suited to hand
negative skewness and heavy tails.
41.7 Conclusion
References
Andrews, D. W. K. (1991). Heteroskedasticity and autocorrelation consistent covariance matrix
estimation. Econometrica, 59, 817–858.
Barrodale, I., & Roberts, F. D. K. (1974). Solution of an overdetermined system of equations in the
l1 norm. Communications of the ACM, 17, 319–320.
Blanley, A., Lamb, R., & Schroeder, R. (2000). Compliance with SEC disclosure requirements
about market risk. Journal of Derivatives, 7, 39–50.
Bofinger, E. (1975). Estimation of a density function using order statistics. Australian Journal of
Statistics, 17, 1–7.
Bollerslev, T., Chou, R. Y., & Kroner, K. F. (1992). ARCH modeling in finance. Journal of
Econometrics, 52, 5–59.
Bouyé, E., & Salmon, M. (2008). Dynamic copula quantile regressions and tail area dynamic
dependence in forex markets [Manuscript]. Coventry: Financial Econometrics Research Cen-
tre, Warwick Business School.
Chen, X., Koenker, R., & Xiao, Z. (2009). Copula-based nonlinear quantile autoregression. The
Econometrics Journal, 12, 50–67.
Dowd, K. (1998). Beyond value at risk: The new science of risk management. England: Wiley.
Dowd, K. (2000). Assessing VaR accuracy. Derivatives Quarterly, 6, 61–63.
Duffie, D., & Pan, J. (1997). An overview of value at risk. Journal of Derivatives, 4, 7–49.
Engle, R. F., & Manganelli, S. (1999). CAViaR: Conditional autoregressive value at risk by
regression quantiles (Working paper). San Diego: University of California.
Engle, R. F., & Manganelli, S. (2004). CAViaR: Conditional autoregressive value at risk by
regression quantiles (Working paper). San Diego: University of California.
Guo, H., Wu, G., & Xiao, Z. (2007). Estimating value at risk for defaultable bond portfolios by
regression quantile. Journal of Risk Finance, 8(2), 166–185.
Hall, P., & Sheather, S. J. (1988). On the distribution of a studentized quantile. Journal of Royal
Statistical Society B, 50, 381–391.
Koenker, R., & Bassett, G. (1978). Regression quantiles. Econometrica, 84, 33–50.
Koenker, R., & d’Orey, V. (1987). Computing regression quantiles. Applied Statistics, 36,
383–393.
Koenker, R., & Xiao, Z. (2006). Quantile autoregression. Journal of the American Statistical
Association, 101(475), 980–1006.
Koenker, R., & Zhao, Q. (1996). Conditional quantile estimation and inference for ARCH models.
Econometric Theory, 12, 793–813.
Portnoy, S. L., & Koenker, R. (1997). The gaussian hare and the Laplacian tortoise: Computability
of squared-error vs. absolute-error estimator. Statistical Science, 12, 279–300.
Sarma, M., Thomas, S., & Shah, A. (2000). Performance evaluation of alternative VaR models
(Working paper). Indira Gandhi Institute of Development Research.
Saunders, A. (1999). Financial institutions management: A modern perspective, Irwin series in
finance. McGraw-Hill, New York.
Sheather, S. J., & Maritz, J. S. (1983). An estimate of the asymptotic standard error of the sample
median. Australian Journal of Statistics, 25, 109–122.
Siddiqui, M. (1960). Distribution of quantiles in samples from a bivariate population. Journa of
Research Natural Bureau Standards, 64B, 145–150.
Welsh, A. (1987). Asymptotically efficient estimation of the sparsity function at a point. Statistics
and Probability Letters, 6, 427–432.
Wu, G., & Zhijie Xiao (2002). An analysis of risk measures. Journal of Risk, 4(4), 53–75.
Xiao, Z., & Koenker, R. (2009). Conditional quantile estimation and inference for GARCH
models. Journal of the American Statistical Association, 104(488), 1696–1712.
Earnings Quality and Board Structure:
Evidence from South East Asia 42
Kin-Wai Lee
Contents
42.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1170
42.2 Prior Research and Hypotheses Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1174
42.2.1 Equity Ownership of Outside Directors . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1174
42.2.2 Board Independence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1175
42.2.3 Equity-Based Compensation of Outside Directors and
Control Divergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1176
42.2.4 Board Independence and Control Divergence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1176
42.3 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1177
42.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1177
42.4.1 Descriptive Statistics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1177
42.4.2 Discretionary Accruals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1178
42.4.3 Accrual Quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1181
42.4.4 Earnings Informativeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1182
42.4.5 Robustness Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1183
42.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1184
Appendix 1: Discretionary Accruals . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1184
Appendix 2: Accruals Quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1186
Appendix 3: Earnings Informativeness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1188
Appendix 4: Adjusting for Standard Errors in Panel Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1189
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1192
Abstract
Using a sample of listed firms in Southeast Asian countries, this paper examines
the association among board structure and corporate ownership structure in
affecting earnings quality. I find that the negative association between separation
K.-W. Lee
Division of Accounting, Nanyang Business School, Nanyang Technological University,
Singapore, Singapore
e-mail: akwlee@ntu.edu.sg
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1169
DOI 10.1007/978-1-4614-7750-1_42,
# Springer Science+Business Media New York 2015
1170 K.-W. Lee
of control rights from cash flow rights and earnings quality varies systematically
with board structure. I find that the negative association between separation of
control rights from cash flow rights and earnings quality is less pronounced in
firms with high equity ownership by outside directors. I also document that in
firms with high separation of control rights from cash flow rights, those firms
with higher proportion of outside directors on the board have higher earnings
quality. Overall, my results suggest that outside directors’ equity ownership and
board independence are associated with better financial reporting outcome,
especially in firms with high expected agency costs arising from misalignment
of control rights and cash flow rights.
The econometric method employed is regressions of panel data. In a panel
data setting, I address both cross-sectional and time-series dependence. Gow
et al. (2010, The Accounting Review 85(2), 483–512) find that in the presence of
both cross-sectional and time-series dependence, the two-way clustering method
which allows for both cross-sectional and time-series dependence produces
well-specified test statistics. Following Gow et al. (2010, The Accounting
Review 85(2), 483–512), I employ the two-way clustering method where the
standard errors are clustered by both firm and year in my regressions of panel
data. Johnston and DiNardo (1997, Econometrics method. New York: Mc-Graw
Hill) and Greene (2000, Econometrics analysis. Upper Saddle River: Prentice-
Hall) are two econometric textbooks that contain a detailed discussion of the
econometrics issues relating to panel data.
Keywords
Earnings quality • Board structure • Corporate ownership structure • Panel data
regressions • Cross-sectional and time-series dependence • Two-way clustering
method of standard errors
42.1 Introduction
In Asia, corporate ownership concentration is high, and many listed firms are
mainly controlled by a single large shareholder (La Porta et al. 1999; Claessens
et al. 2000). Asian firms also show a high divergence between control rights and
cash flow rights, which allows the largest shareholder to control a firm’s opera-
tions with a relatively small direct stake in its cash flow rights. Control is often
increased beyond ownership stakes through pyramid structures, cross-holdings
among firms, and dual class shares (Claessens et al. 2000). It is argued that
concentrated ownership facilitates transactions in weak property rights environ-
ment by providing the controlling shareholders the power and incentive to nego-
tiate and enforce contracts with various stakeholders (Shleifer and Vishny 1997).
As a result of concentrated ownership, the main agency problem in listed firms
in Asia is the conflict of interest between the controlling shareholder and
minority shareholder. Specifically, controlling shareholder has incentives to
42 Earnings Quality and Board Structure: Evidence from South East Asia 1171
First, prior research generally finds that in developed economies such as the United
States and the United Kingdom, there is a positive association between board
independence and earnings quality (Dechow et al. 1996; Klein 2002; Peasnell
et al. 2005). However, there is limited evidence on the effect of board independence
on the financial accounting process in Asia. My paper attempts to fill this gap.
Second, recent research on the monitoring incentives of the board suggests that
equity ownership of outside directors plays an important role in mitigating mana-
gerial entrenchment (Perry 2000; Ryan and Wiggins 2004). An implication of this
stream of research is that even in firms with high board independence, entrenched
managers can weaken the monitoring incentives of outside directors by reducing
their equity-based compensation. In other words, board independence that is not
properly augmented with incentive compensation may hamper the monitoring
effectiveness of independent directors over management.
My sample consists of 2,875 firm-year observations during the period
2004–2008 in five Asian countries comprising Indonesia, Malaysia, the Philippines,
Singapore, and Thailand. These countries provide a good setting to test the
governance potential of the board of directors because shareholders in these
countries typically suffer from misaligned managerial incentives, ineffective legal
protection, and underdeveloped markets for corporate control (La Porta et al. 1999;
Claessens et al. 2000; Fan and Wong 2002).
I measure earnings quality with three financial reporting metrics:
(i) discretionary accruals, (ii) mapping of accruals to cash flow, and (iii) informa-
tiveness of reported earnings. My results are robust across alternative earnings
quality metrics. I find that earnings quality is higher when outside directors have
higher equity ownership. This result suggests that internal monitoring of the quality
and credibility of accounting information is improved through aligning share-
holders’ and directors’ incentives. Consistent with the monitoring role of outside
directors (Fama and Jensen 1983), I also find that earnings quality is positively
associated with the proportion of outside directors on the board. This result supports
the notion that outside directors have incentives to be effective monitors in order to
maintain the value of their reputational capital. Consistent with prior studies
(Fan and Wong 2002; Haw et al. 2004), I also document that earnings quality is
negatively associated with the separation of control rights from cash flow rights of
the largest ultimate shareholder.
More importantly, I document that the negative association between separation
of control rights from cash flow rights and earnings quality is less pronounced in
firms with high equity ownership by outside directors. This result suggests equity
ownership improves the incentives and monitoring intensity of outside directors in
firms with high expected agency costs arising from the divergence of control rights
from cash flow rights. Furthermore, my result indicates the negative association
between separation of control rights from cash flow rights and earnings quality is
mitigated by the higher proportion of outside directors on the board. This result
provides evidence supporting the corporate governance role of outside directors in
42 Earnings Quality and Board Structure: Evidence from South East Asia 1173
Recent research examines the compensation structure of outside directors, who play
an important monitoring role over management’s actions. The central theme in this
body of research is that incentive compensation leading to share ownership
improves the outside directors’ incentives to monitor. Mehran (1995) finds firm
performance is positively associated with the proportion of directors’ equity-based
compensation. Perry (2000) finds that the likelihood of CEO turnover following
poor performance increases when directors receive higher equity-based compensa-
tion. Shivdasani (1993) finds that probability of a hostile takeover is negatively
associated with the percentage of shares owned by outside directors in target firms.
He interprets this finding as suggesting that board monitoring may substitute for
monitoring from the market of corporate control. Hermalin and Weisbach (1988)
and Gillette et al. (2003) develop models where incentive compensation for
directors increases their monitoring efforts and effectiveness. Ryan and Wiggins
(2004) find that directors in firms with entrenched CEOs receive a significantly
smaller proportion of compensation in the form of equity-based awards. Their
result suggests that entrenched CEOs use their position to influence directors’
42 Earnings Quality and Board Structure: Evidence from South East Asia 1175
Although compliance with the code of corporate governance is not legally mandatory,
listed companies are required to explain deviations from the recommendations of the
code of corporate governance.1
I posit that in the waves of corporate governance reform in emerging markets in
the early 2000s and the guidelines of country-specific code of corporate governance
in emphasizing the importance of board independence, there is heightened aware-
ness among outside directors on their increased monitoring responsibilities. To the
extent that outside directors in listed firms in Asia perform a corporate governance
role, I predict that:
H2 Earnings quality is positively associated with the proportion of outside direc-
tors on the board.
1
To illustrate, the Singapore Exchange Listing Rules require “listed companies to describe in the
annual reports their corporate governance practices with specific reference to the principles of the
Code, as well as disclose and explain any deviation from any guideline of the Code. Companies are
also encouraged to make a positive confirmation at the start of the corporate governance section of
the annual report that they have adhered to the principles and guidelines of the Code, or specify
each area of non-compliance. Many of these guidelines are recommendations for companies to
disclose their corporate governance arrangements.”
42 Earnings Quality and Board Structure: Evidence from South East Asia 1177
board are likely to mitigate the negative effects of the separation of control rights
from cash flow rights on earnings quality. Thus, I predict that:
H4 The negative association between separation of control rights from cash flow
rights and earnings quality is mitigated by the proportion of outside directors.
42.3 Data
I begin with the Worldscope database to identify listed firms in five Asian
countries comprising Indonesia, Malaysia, the Philippines, Singapore, and Thai-
land during the period 2004–2008. I exclude financial institutions because of their
unique financial structure and regulatory requirements. I eliminate observations
with extreme values of control variables such as return-on-assets and leverage.
I obtain stock price data from the Datastream database. I obtain annual reports for
the period 2004–2008 from the Global Report database and company websites.
The sample consists of 617 firms for 2,875 firm-year observations during the
period 2004–2008 in five Asian countries comprising Indonesia, Malaysia, the
Philippines, Singapore, and Thailand.
I collect data on the board characteristics such as board size, the number of
independent directors, and equity ownership of directors from the annual report.
I also examine the annual report to trace the ultimate owners of the firms.
The procedure of identifying ultimate owners is similar to the one used in La
Porta et al. (1999).2 In this study, I measure earnings quality with three financial
reporting metrics: (i) discretionary accruals, (ii) mapping of accruals to cash flow,
and (iii) informativeness of reported earnings.
Appendix 1 contains detailed description on the econometric method.
42.4 Results
Table 42.1 presents the descriptive statistics. Mean absolute discretionary accrual as
a proportion of lagged assets is 0.062. Mean equity ownership of outside directors
(computed as common stock and stock options held by outside directors divided by
2
In summary, an ultimate owner is defined as the shareholder who has the determining voting
rights of the company and who is not controlled by anyone else. If a company does not have an
ultimate owner, it is classified as widely held. To economize on the data collection task, the
ultimate owner’s voting right level is set at 50 % and not traced any further once that level exceeds
50 %. Although a company can have more than one ultimate owner, we focus on the largest
ultimate owner. We also identify the cash flow rights of the ultimate owners. To facilitate the
measurement of the separation of cash flow and voting rights, the maximum cash flow rights level
associated with any ultimate owner is also set at 50 %. However, there is no minimum cutoff level
for cash flow rights.
1178 K.-W. Lee
Table 42.1 Descriptive statistics. The sample consists of 617 firms for 2,875 firm-year
observations during the period 2004–2008 in five Asian countries comprising Indonesia, Malaysia,
the Philippines, Singapore, and Thailand
Mean 25th percentile Median 75th percentile Standard deviation
DISCAC 0.062 0.009 0.038 0.085 0.053
AQ 0.068 0.029 0.035 0.063 0.037
EBC (%) 2.041 0.837 1.752 2.663 1.035
OUTDIR 0.489 0.206 0.385 0.520 0.217
VOTE 1.198 1.000 1.175 1.326 0.638
BOARDSIZE 7 5 8 10 3
CEODUAL 0.405 0 0 1 –
LNASSET 11.722 9.867 11.993 13.078 2.115
MB 1.851 0.582 1.272 2.195 0.833
LEV 0.261 0.093 0.211 0.335 0.106
ROA 0.086 0.027 0.0705 0.109 0.071
DISCAC ¼ absolute value of discretionary accruals estimated based on the modified Jones model
AQ ¼ accrual quality measured by Dechow and Dichev’s (2002) measure of mapping of accruals
to past, present, and future cash from operations
DIROWN ¼ common stock and stock options held by outside directors divided by number of
ordinary shares outstanding in the firm
OUTDIR ¼ proportion of outside directors on the board
VOTECASH ¼ voting rights divided by cash flow rights of the largest controlling shareholder
CEODUAL ¼ a dummy variable that equals 1 if the CEO is chairman of board and 0 otherwise
BOARDSIZE ¼ number of directors on the board
LNASSET ¼ natural logarithm of total assets
MB ¼ market value of equity divided by book value of equity
LEV ¼ long-term debt divided by total assets
ROA ¼ net profit after tax divided by total assets
number of ordinary shares outstanding in the firm) is 2.04 %. The mean board size
and proportion of outside directors on the board are 7 and 0.489, respectively. CEO
chairs the board in 40 % of the firms. Consistent with Fan and Wong’s (2002) study
of East Asian economies, the firms in my sample also have high divergence of
control rights from cash flow rights (mean VOTE ¼ 1.198).
Table 42.2 Regressions of unsigned discretionary accruals. The sample consists of 617 firms
for 2,875 firm-year observations during the period 2004–2008 in five Asian countries comprising
Indonesia, Malaysia, the Philippines, Singapore, and Thailand. The dependent variable is absolute
discretionary accruals computed based on the modified Jones model. All variables are defined in
Table 42.1. The t-statistics (in parentheses) are adjusted based on standard errors clustered by firm
and year (Petersen 2009). The symbols *, **, and *** denote statistical significance at the 10 %,
5 %, and 1 % levels (two-tailed), respectively
Predicted sign 1 2
EBC 0.2513 (3.07)*** 0.2142 (2.86)***
OUTDIR 0.1862 (2.41)** 0.1053 (2.19)**
VOTE + 0.8173 (2.85)*** 0.9254 (2.94)***
VOTE *EBC 0.4160 (2.73)***
VOTE * OUTDIR 0.1732 (2.29)**
BOARDSIZE +/ 0.0359 (1.57) 0.0817 (1.42)
CEODUAL + 0.4192 (1.61) 0.2069 (1.45)
LNASSET 0.5311 (8.93)*** 0.5028 (8.01)***
MB + 0.1052 (4.25)*** 0.2103 (3.02)***
LEV + 1.2186 (5.38)*** 1.1185 (5.19)***
ROA +/ 3.877 (4.83)*** 4.108 (4.72)***
Adjusted R2 12.5 % 14.1 %
Table 42.3 Regressions of signed discretionary accruals. The sample consists of 617 firms for
2,875 firm-year observations during the period 2004–2008 in five Asian countries comprising
Indonesia, Malaysia, the Philippines, Singapore, and Thailand. In column (1), the sample consists
of firms with income-increasing discretionary accruals, and the dependent variable is positive
discretionary accruals. In column (2), the sample consists of firms with income-decreasing
discretionary accruals, and the dependent variable is negative discretionary accruals. All variables
are defined in Table 42.1. All regressions contain dummy control variables for country, year, and
industry. The t-statistics (in parentheses) are adjusted based on standard errors clustered by firm
and year (Petersen 2009). The symbols *, **, and *** denote statistical significance at the 10 %,
5 %, and 1 % levels (two-tailed), respectively
(1) (2)
Positive DISCAC Negative DISCAC
EBC 0.1865 (2.21)** 0.2017 (2.09)**
OUTDIR 0.1172 (2.08)** 0.1302 (2.11)**
VOTE 0.8103 (3.25)*** 0.6735 (2.23)**
VOTE *EBC 0.3952 (2.49)*** 0.3064 (2.05)**
VOTE * OUTDIR 0.1732 (2.13)** 0.1105 (2.01)**
BOARDSIZE 0.0533 (1.27) 0.0681 (1.09)
CEODUAL 0.1860 (1.32) 0.1562 (1.26)
LNASSET 0.7590 (5.22)*** 0.4463 (6.12)***
MB 0.2019 (2.08)** 0.1085 (1.93)**
LEV 0.9781 (3.20)*** 0.7701 (2.10)**
ROA 3.087 (4.13)*** 4.253 (3.62)***
Adjusted R2 13.8 % 12.4 %
I partition the sample into two groups based on the sign of the firms’ discretion-
ary accruals. Table 42.3 column (1) presents the results using the subsample of
firms with income-increasing discretionary accruals. Results indicate firms with
higher equity ownership by outside directors, higher board independence, and
lower divergence of control rights from cash flow rights, have lower income-
increasing discretionary accruals. More importantly, I find that outside directors’
equity ownership and proportion of outside directors mitigate the propensity of
firms with high separation of control rights from cash flow rights to make higher
income-increasing discretionary accruals. Table 42.3 column (2) presents the
results using the subsample of firms with income-decreasing discretionary accruals.
I find firms with higher equity ownership, higher board independence, and lower
divergence of control rights from cash flow rights, have lower income-decreasing
discretionary accruals. Furthermore, I find that equity ownership by outside direc-
tors and proportion of outside directors mitigate the propensity of firms with high
separation of control rights from cash flow rights to make higher income-decreasing
discretionary accruals.
In summary, when outside directors have equity ownership and when board
independence is high, firms have both lower income-increasing and income-
decreasing discretionary accruals, apparently mitigating earnings management
42 Earnings Quality and Board Structure: Evidence from South East Asia 1181
Table 42.4 Regressions of accrual quality. The sample consists of 617 firms for 2,875 firm-
year observations during the period 2004–2008 in five Asian countries comprising Indonesia,
Malaysia, the Philippines, Singapore, and Thailand. The dependent variable is AQ measured by
Dechow and Dichev’s (2002) measure of mapping of accruals to past, present, and future cash
from operations with higher values of AQ denoting better accrual quality. All variables are defined
in Table 42.1. All regressions contain dummy control variables for country, year, and industry. The
t-statistics (in parentheses) are adjusted based on standard errors clustered by firm and year
(Petersen 2009). The symbols *, **, and *** denote statistical significance at the 10 %, 5 %,
and 1 % levels (two-tailed), respectively
Predicted sign 1 2
EBC + 0.3725 (3.11)*** 0.2133 (3.26)***
OUTDIR + 0.2049 (2.15)** 0.1557 (2.86)***
VOTE 0.5108 (3.74)*** 0.4352 (3.05)***
VOTE *EBC + 0.1751 (2.80)***
VOTE * OUTDIR + 0.1163 (2.09)**
BOARDSIZE +/ 0.1003 (1.29) 0.0642 (1.17)
CEODUAL +/ 0.1890 (0.83) 0.2173 (1.56)
LNASSET + 3.2513 (5.11)*** 2.8764 (4.82)***
OPERCYCLE 3.2941 (4.75)*** 3.0185 (5.01)***
NETPPE + 1.1802 (3.35)*** 0.9926 (3.72)***
STDCFO 1.2981 (2.83)*** 1.8344 (3.32)***
STDSALE 1.0203 (2.77)*** 0.7845 (2.09)**
NEGEARN 0.8306 (2.02)** 1.0345 (1.77)*
Adjusted R2 9.2 % 10.5 %
both on the upside and downside. For firms with greater separation of control
rights from cash flow rights of controlling shareholders, those with high
equity ownership by outside directors and those with high proportion of outside
directors have lower income-increasing and lower income-decreasing discretionary
accruals.
Table 42.5 Regressions of returns on earnings. The sample consists of 617 firms for 2,875
firm-year observations during the period 2004–2008 in five Asian countries comprising Indonesia,
Malaysia, the Philippines, Singapore, and Thailand. The dependent variable (RET) is 12-month
cumulative raw return ending 3 months after the fiscal year-end. All regressions contain dummy
control variables for country, year, and industry. The t-statistics (in parentheses) are adjusted based
on standard errors clustered by firm and year (Petersen 2009). The symbols *, **, and *** denote
statistical significance at the 10 %, 5 %, and 1 % levels (two-tailed), respectively
Predicted sign 1 2
EARN + 1.1735 (3.85)*** 1.2811 (3.62)***
EARN *EBC + 0.3122 (2.87)*** 0.2983 (2.80)***
EARN * OUTDIR + 0.1094 (2.13)** 0.1105 (2.08)**
EARN * VOTE 0.6817 (3.72)*** 0.7019 (3.50)***
EARN * VOTE *EBC + 0.2602 (2.83)***
EARN * VOTE * OUTDIR + 0.1925 (2.15)**
EARN * BOARDSIZE +/ 0.0836 (1.50) 0.0801 (1.22)
EARN * CEODUAL +/ 0.0405 (1.42) 0.0215 (1.30)
EARN * LNASSET + 0.2011 (2.89)*** 0.3122 (3.07)***
EARN * MB + 0.1573 (1.80)* 0.1806 (1.81)*
EARN * LEV 0.7814 (2.03)** 0.6175 (2.84)***
N 3,172 3,172
Adjusted R2 11.8 % 13.3 %
Table 42.5 presents the regression results on earnings informativeness. The coeffi-
cient EARN* DIROWN is positive and significant, indicating the greater the equity
ownership by outside directors, the higher informativeness of reported earnings.
The coefficient EARN*OUTDIR is positive and significant, implying that firms
42 Earnings Quality and Board Structure: Evidence from South East Asia 1183
As a sensitivity analysis, I repeat all my tests at the economy level. The economy-
by-economy results indicate that earnings quality is positively associated with
equity ownership by outside directors and board independence and negatively
associated with the separation of cash flow rights from control rights. More
importantly, the mitigating effects of equity ownership and board independence
on the association between separation of cash flow rights from control rights and
earnings quality are not concentrated in any given economy. Year-by-year regres-
sions yield qualitatively similar results, suggesting my inferences are not time-
period specific.
As a robustness test, I follow Haw et al. (2004) to include legal institutions
that protect minority shareholder rights (proxied by legal tradition, minority
shareholder rights, efficiency of judicial system, or disclosure system) and extrale-
gal institutions (proxied by the effectiveness of competition law, diffusion of the
press, and tax compliance) in my tests. I continue to document firm-specific internal
governance mechanisms, namely, outside directors’ equity ownership and board
independence, still matter in constraining management opportunism over the finan-
cial reporting process, especially in firms with high expected agency costs arising
from the divergence between control rights and cash flow rights. Thus, my results
suggest that there is an incremental role for firm-specific internal governance
mechanisms, beyond country-level institutions, in improving the quality of finan-
cial information by mitigating insiders’ entrenchment.
1184 K.-W. Lee
42.5 Conclusion
discretionary accruals. Specifically, I estimate the following model for each country
in each year at the one-digit SIC industry:
where:
ACC ¼ total accruals, which are calculated as net income minus operating cash
flows scaled by beginning-of-year total assets.
LAG1ASSET ¼ total assets at beginning of the fiscal year.
CHGSALE ¼ sales change, which is net sales in year t less net sales in year
t 1, scaled by beginning-of-year-t total assets.
CHGREC ¼ change in accounts receivables scaled by beginning-of-year-t total
assets.
PPE ¼ gross property, plant, and equipment in year t scaled by beginning-of-year-t
total assets.
I use the residuals from the annual cross-sectional country-industry regression
model in (A1) as the modified Jones model discretionary accruals.
I use the following regression model to test the association between discretion-
ary accruals and board structure:
where:
DISCAC ¼ absolute value of discretionary accruals estimated based on the
modified Jones model (see Eq. 42.1).
DIROWN ¼ common stock and stock options held by outside directors divided by
number of ordinary shares outstanding in the firm.
OUTDIR ¼ proportion of outside directors on the board.
VOTE ¼ control rights divided by cash flow rights of the largest controlling
shareholder.
BOARDSIZE ¼ number of directors on the board.
CEODUAL ¼ a dummy variable that equals 1 if the CEO is chairman of board and
0 otherwise.
LNASSET ¼ natural logarithm of total assets.
MB ¼ market value of equity divided by book value of equity.
LEV ¼ long-term debt divided by total assets.
ROA ¼ net profit after tax divided by total assets.
Country Controls ¼ a set of country dummy variables.
Year Controls ¼ a set of year dummy variables.
1186 K.-W. Lee
If high equity ownership for outside directors improves the board monitoring of
managerial discretion over the financial accounting process, I predict the coefficient
b1 to be negative. Similarly, a negative coefficient for b2 suggests that board
independence curtails managerial opportunism on financial reporting. A positive
coefficient b3 indicates greater separation of control rights from cash flow rights of
the largest controlling shareholder induces greater earnings management. I predict
the positive association between absolute discretionary accruals and the separation
of control rights from cash flow rights to be less pronounced in firms with high
equity ownership by outside directors. Thus, I expect coefficient b4 to be negative.
Furthermore, I predict the positive association between absolute discretionary
accruals and the separation of control rights from cash flow rights to be less
pronounced in firms high proportion of outside directors on the board. Thus,
I expect coefficient b5 to be negative.
Other board characteristics include the total number of directors (BOARDSIZE)
and CEO-chairman duality (CEODUAL). The evidence is mixed on whether board
size and CEO duality impairs board effectiveness. Thus, ex ante, there is no
prediction on the sign on both variables. The model controls for the effects of
firm size, growth opportunities, and leverage on discretionary accruals. Large firms
have greater external monitoring, have more stable operations and stronger control
structures, and hence report smaller abnormal accruals (Dechow and Dichev 2002).
Firm size (LNASSET) is measured based on book value of total assets. Because
discretionary accruals are higher for firms with higher growth opportunities,
I employ the market-to-book equity (MB) ratio to control for the effect of growth
opportunities on discretionary accruals (Kothari et al. 2005). I also include financial
leverage (LEV), defined as long-term debt divided by total assets, to control for the
managerial discretion over the financial accounting process to mitigate constraints
of accounting-based debt covenants (Smith and Watts 1992). To control for the
effect of firm performance on discretionary accruals, I include firm profitability
(ROA), defined as net income divided by total assets. Finally, I include
country dummy variables to capture country-specific factors that may affect the
development of capital markets and financial accounting quality. I include dummy
variables for years and industries to control for time effect and industry effects,
respectively.
My second proxy for earnings quality is accruals quality. Dechow and Dichev
(2002) propose a measure of earnings quality that captures the mapping of current
accruals into last-period, current-period, and next-period cash flows. Francis
et al. (2005) find that this measure (which they term accrual quality) is associated
with measures of cost of equity capital. My measure of accrual quality is based on
Dechow and Dichev’s (2002) model relating current accruals to last-period,
current-period, and next-period cash flows:
42 Earnings Quality and Board Structure: Evidence from South East Asia 1187
where:
TCAj,t ¼ firm j’s total current accruals in year t ¼ DCAj,t – DCLj,t – DCASHj,t
+ DSTDj,t
Assets ¼ firm j’s average total assets in year t 1 and year t
CFOj,t ¼ cash flow from operations in year t is calculated as net income less total
accruals (TA) where:
TAj,t ¼ DCAj,t – DCLj,t – DCASHj,t + DSTDj,t – DEPNj,t where
DCAj,t ¼ firm j’s change in current assets between year t 1 and year t
DCLj,t ¼ firm j’s change in current liabilities between year t 1 and year t
DCASHj,t ¼ firm j’s change in cash between year t 1 and year t
DSTDj,t ¼ firm j’s change in debt in current liabilities between year t 1 and
year t
DEPNj,t ¼ firm j’s change in depreciation and amortization expense in year t
I estimate Eq. 42.3 for each one-digit SIC industry for each country-year
combination. These estimations yield firm- and year-specific residuals, ejt, which
form the basis for the accrual quality metric. AQ is the standard deviation of firm j’s
estimated residuals multiplied by 1. Hence, large values of AQ correspond to high
accrual quality.
I employ the following model to test the association between accrual quality and
board characteristics:
where:
AQ ¼ the standard deviation of firm j’s residuals from a regression of
current accruals on lagged, current, and future cash flows from operations.
I multiply the variable by 1 so that higher AQ measure denotes higher accrual
quality.
OPERCYCLE ¼ log of the sum of the firm’s days accounts receivable and days
inventory.
NETPPE ¼ ratio of the net book value of PP&E to total assets.
STDCFO ¼ standard deviation of the firm’s rolling 5-year cash flows from
operations.
STDSALE ¼ standard deviation of the firm’s rolling 5-year sales revenue.
NEGEARN ¼ the firm’s proportion of losses over the prior 5 years.
All other variables are previously defined.
1188 K.-W. Lee
If high equity ownership for outside directors improves the board monitoring of
managerial discretion over the financial accounting process, I predict coefficient b1
to be positive. Similarly, a positive coefficient for b2 suggests that higher board
independence is associated with higher accrual quality. If greater agency costs arise
from the higher separation of control rights from cash flow rights of the largest
controlling shareholder, coefficient b3 should be negative. I predict that the nega-
tive association between accrual quality and the separation of control rights from
cash flow rights is mitigated in firms with high equity ownership by outside
directors. Thus, I expect coefficient b4 to be positive. Furthermore, the negative
effect of the separation of control rights from cash flow on rights accrual quality
should be attenuated in firms with high proportion of outside directors on the board.
Thus, I expect coefficient b5 to be positive.
In Eq. 42.4, the control variables include innate determinants of accrual quality.
Briefly, Dechow and Dichev (2002) find that accrual quality is positively associated
with firm size and negatively associated with cash flow variability, sales variability,
operating cycle, and incidence of losses. Firm size is measured by the natural
logarithm of total assets (LNASSET). Operating cycle (OPERCYCLE) is the log
of the sum of the firm’s days accounts receivable and days inventory. Capital
intensity, NETPPE, is proxied by the ratio of the net book value of PP&E to total
assets. Cash flow variability (STDCFO) is the standard deviation of the firm’s
rolling 5-year cash flows from operations. Sales variability (STDSALE) is the
standard deviation of the firm’s rolling 5-year sales revenue. Incidence of negative
earnings realizations, NEGEARN, is measured as the firm’s proportion of losses
over the prior 5 years.
where:
RET ¼ 12-month cumulative raw return ending 3 months after the fiscal year-end.
EARN ¼ net income for year t, scaled by the market value of equity at the end of t 1.
42 Earnings Quality and Board Structure: Evidence from South East Asia 1189
and se(b) is the standard error of the coefficients based on their empirical
distribution. When there is no cross-regression (time-series) dependence, this
approach yields consistent estimates of the standard error of the coefficients as
T goes to infinity.
Two common variants of the Fama-MacBeth approach appear in the
accounting literature. The first variant, FM-i, involves estimating firm- or
portfolio-specific time-series regressions with inferences based on the
cross-sectional distribution of coefficients. This modification of the Fama-
MacBeth approach is appropriate if there is time-series dependence but not
cross-sectional dependence. However, FM-i is frequently used when cross-
sectional dependence is likely, such as when returns are the dependent
variable.
The second common variant of the FM-t approach, FM-NW, is intended to
correct for serial correlation in addition to cross-sectional correlation. FM-NW
modifies FM-t by applying a Newey-West adjustment in an attempt to correct
for serial correlation.
Gow et al. (2010) suggest two reasons to believe that FM-NW may not
correct for serial correlation. First, FM-NW involves applying Newey-West to
a limited number of observations, a setting in which Newey-West is known to
perform poorly. Second, FM-NW applies Newey-West to a time-series of
coefficients, whereas the dependence is in the underlying data.
(iv) Z2 Statistic
The Z2-t (Z2-i) statistic is calculated using t-statistics from separate cross-
sectional (time-series) regressions for each time period (cross-sectional unit)
and is given by the expression:
t 1X T
Z2 ¼ , where t ¼ ^t t , (42.7)
seðtÞ T t¼1
se(t) is the standard error of the t-statistics based on their empirical distribu-
tion, and T is the number of time periods (cross-sectional units) in the sample.
Gow et al. (2010) suggest Z2 may suffer from cross-regression dependence
in the same way as the Fama-MacBeth approach does.
(v) One-Way Cluster-Robust Standard Errors
A number of studies in our survey use cluster-robust standard errors, with
clustering either along a cross-sectional dimension (e.g., analyst, firm, indus-
try, or country) or along a time-series dimension (e.g., year); we refer to
the former as CL-i and the latter as CL-t. Cluster-robust standard errors
42 Earnings Quality and Board Structure: Evidence from South East Asia 1191
References
Ashbaugh, H., LaFond, R., & Mayhew, B. W. (2003). Do nonaudit services compromise auditor
independence? Further evidence. The Accounting Review, 78(3), 611–639.
Beasley, M. S. (1996). An empirical analysis of the relation between the board of director
composition and financial statement fraud. The Accounting Review, 71, 443–465.
Claessens, S., Djankov, S., & Lang, L. H. P. (2000). The separation of ownership and control in
East Asian corporations. Journal of Financial Economics, 58, 81–112.
Claessens, S., Djankov, S., Fan, J. P. H., & Lang, L. H. P. (2002). Disentangling the incentive and
entrenchment effects of large shareholdings. Journal of Finance, 57(2), 2741–2771.
Dechow, P. M., & Dichev, I. (2002). The quality of accruals and earnings: The role of accrual
estimation errors. The Accounting Review, 77, 35–59.
Dechow, P. M., Sloan, R. G., & Sweeney, A. P. (1996). Causes and consequences of earnings
manipulation: An analysis of firms subject to enforcement actions by the SEC. Contemporary
Accounting Research, 13, 1–36.
Fama, E., & Jensen, M. (1983). Agency problems and residual claims. Journal of Law and
Economics, 26, 327–349.
Fama, E., & MacBeth, J. (1973). Risk, return, and equilibrium: Empirical tests. The Journal of
Political Economy, 81(3), 607–636.
Fan, J., & Wong, T. J. (2002). Corporate ownership structure and the informativeness of account-
ing earnings in East Asia. Journal of Accounting & Economics, 33, 401–425.
Fan, J., & Wong, T. J. (2005). Do external auditors perform a corporate governance role in
emerging markets? Journal of Accounting Research, 43(1), 35–72.
Francis, J., Lafond, R., Olsson, P., & Schipper, K. (2005). The market pricing of accrual quality.
Journal of Accounting & Economics, 39, 295–327.
Gillette, A., Noe, T., & Rebello, M. (2003). Corporate board composition, protocols and voting
behavior: Experimental evidence. Journal of Finance, 58, 1997–2032.
Gow, I. D., Ormazabal, G., & Taylor, D. J. (2010). Correcting for cross-sectional and time series
dependence in accounting research. The Accounting Review, 85(2), 483–512.
Greene, W. (2000). Econometrics analysis. Upper Saddle River: Prentice-Hall.
Haw, I., Hu, B., Hwang, L., & Wu, W. (2004). Ultimate ownership, income management, and legal
and extra-legal institutions. Journal of Accounting Research, 42, 423–462.
Hermalin, B., & Weisbach, M. (1988). The determinants of board composition. Rand Journal of
Economics, 19, 589–606.
Johnson, S., Boone, P., Breach, A., & Friedman, E. (2000). Corporate governance in the Asian
financial crisis. Journal of Financial Economics, 58, 141–186.
Johnston, J., & DiNardo, J. (1997). Econometrics method. New York: Mc-Graw Hill.
Klein, A. (2002). Audit committee, board of director characteristics, and earnings management.
Journal of Accounting and Economics, 33(3), 375–400.
Kothari, S. P., Leone, A. J., & Wasley, C. E. (2005). Performance matched discretionary accruals.
Journal of Accounting and Economics, 39, 163–197.
La Porta, R., Lopez-de-Silanes, F., Shleifer, A., & Vishny, R. (1998). Law and finance. Journal of
Political Economy, 106, 1113–1155.
La Porta, R., Lopez-De-Silanes, F., & Shleifer, A. (1999). Corporate ownership around the world.
Journal of Finance, 54, 471–518.
Lang, M., Lins, K., & Miller, D. (2004). Concentrated control, analyst following, and valuation:
Do analysts matter most when investors are protected least? Journal of Accounting Research,
42(3), 589–623.
42 Earnings Quality and Board Structure: Evidence from South East Asia 1193
Lee, K. W. (2007). Corporate voluntary disclosure and the separation of cash flow rights from
control rights. Review of Quantitative Finance and Accounting, 28, 393–416.
Lee, K. W., Lev, B., & Yeo, H. H. (2007). Organizational structure and earnings management.
Journal of Accounting, Auditing and Finance, 22(2), 293–391.
Lee, K. W., Lev, B., & Yeo, H. H. (2008). Executive pay dispersion, corporate governance and
firm performance. Review of Quantitative Finance and Accounting, 30, 315–338.
Lee, C. F., Lee, K. W., & Yeo, H. H. (2009). Investor protection and convertible debt design.
Journal of Banking and Finance, 33(6), 985–995.
Leuz, C., Nanda, D., & Wysocki, P. (2003). Earnings management and investor protection: An
international comparison. Journal of Financial Economics, 69, 505–527.
Mehran, H. (1995). Executive compensation structure, ownership, and firm performance. Journal
of Financial Economics, 38(2), 163–184.
Morck, R., Shleifer, A., & Vishny, R. W. (1988). Management ownership and market valuation:
An empirical analysis. Journal of Financial Economics, 20, 293–315.
Newey, W., & West, D. (1987). A simple, positive semi-definite, heteroskedasticity and autocor-
relation consistent covariance matrix. Econometrica, 55(3), 703–708.
Peasnell, K., Pope, P., & Young, S. (2005). Board monitoring and earnings management:
Do outside directors influence abnormal accruals? Journal of Business Finance & Accounting,
32(7/8), 1311–1345.
Perry, T. (2000). Incentive compensation for outside directors and CEO turnover. Working paper,
Arizona State University.
Petersen, M. (2009). Estimating standard errors in finance panel data sets: Comparing approaches.
Review of Financial Studies, 22(1), 435–480.
Rosenstein, S., & Wyatt, J. (1990). Outside directors, board independence, and shareholder wealth.
Journal of Financial Economics, 26, 175–192.
Ryan, H., & Wiggins, R. (2004). Who is whose pocket? Director compensation, board indepen-
dence, barriers to effective monitoring. Journal of Financial Economics, 73, 497–524.
Shivdasani, A. (1993). Board composition, ownership structure, and hostile takeovers. Journal of
Accounting & Economics, 16, 167–188.
Shleifer, A., & Vishny, R. (1997). A survey of corporate governance. Journal of Finance,
52, 737–783.
Smith, C., & Watts, R. (1992). The investment opportunity set and corporate financing, dividend
and compensation policies. Journal of Financial Economics, 32, 263–292.
Warfield, T., Wild, J. J., & Wild, K. (1995). Managerial ownership, accounting choices, and
informativeness of earnings. Journal of Accounting and Economics, 20, 61–91.
Weisbach, M. (1988). Outside directors and CEO turnover. Journal of Financial Economics,
20, 431–460.
White, H. (1980). A heteroskedasticity-consistent covariance matrix estimator and a direct test for
heteroskedasticity. Econometrica, 48(4), 817–838.
Rationality and Heterogeneity of Survey
Forecasts of the Yen-Dollar Exchange 43
Rate: A Reexamination
Contents
43.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1196
43.2 Background: Testing Rationality in the Foreign Exchange Market . . . . . . . . . . . . . . . . . . . 1199
43.2.1 Why Test Rational Expectations with Disaggregated Survey
Forecast Data? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1200
43.2.2 Rational Reasons for the Failure of the Rational Expectations Hypothesis
Using Disaggregated Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1201
43.3 Description of Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1202
43.4 Empirical Tests of Rationality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1203
43.4.1 Joint Tests of Unbiasedness and Weak Efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . 1204
43.4.2 Pretests for Rationality: The Stationarity of the Forecast Error . . . . . . . . . . . . . 1212
43.4.3 Univariate Tests for Unbiasedness . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1218
43.4.4 Unbiasedness Tests Using Error Correction Models . . . . . . . . . . . . . . . . . . . . . . . . . 1222
43.4.5 Explicit Tests of Weak Efficiency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1228
43.5 Micro-homogeneity Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1230
43.5.1 Ito’s Heterogeneity Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1239
43.6 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1241
Appendix 1: Testing Micro-homogeneity with Survey Forecasts . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1244
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1245
R. Cohen
University of Hawaii Economic Research Organization and Economics, University of Hawaii at
Manoa, Honolulu, HI, USA
e-mail: afrc2@cbpp.uaa.alaska.edu
C.S. Bonham (*)
College of Business and Public Policy, University of Alaska Anchorage, Anchorage, AK, USA
e-mail: bonham@hawaii.edu
S. Abe
Faculty of Policy Studies, Doshisha University, Kyoto, Japan
e-mail: sabe@mail.doshisha.ac.jp
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1195
DOI 10.1007/978-1-4614-7750-1_43,
# Springer Science+Business Media New York 2015
1196 R. Cohen et al.
Abstract
This chapter examines the rationality and diversity of industry-level forecasts of
the yen-dollar exchange rate collected by the Japan Center for International
Finance. In several ways we update and extend the seminal work by Ito (1990,
American Economic Review 80, 434–449). We compare three specifications for
testing rationality: the “conventional” bivariate regression, the univariate regres-
sion of a forecast error on a constant and other information set variables, and an
error correction model (ECM). We find that the bivariate specification, while
producing consistent estimates, suffers from two defects: first, the conventional
restrictions are sufficient but not necessary for unbiasedness; second, the test has
low power. However, before we can apply the univariate specification, we must
conduct pretests for the stationarity of the forecast error. We find a unit root in
the 6-month horizon forecast error for all groups, thereby rejecting unbiasedness
and weak efficiency at the pretest stage. For the other two horizons, we find
much evidence in favor of unbiasedness but not weak efficiency. Our ECM
rejects unbiasedness for all forecasters at all horizons. We conjecture that these
results, too, occur because the restrictions test sufficiency, not necessity.
We extend the analysis of industry-level forecasts to a SUR-type structure
using an innovative GMM technique (Bonham and Cohen 2001, Journal of
Business & Economic Statistics, 19, 278–291) that allows for forecaster cross-
correlation due to the existence of common shocks and/or herd effects. Our
GMM tests of micro-homogeneity uniformly reject the hypothesis that fore-
casters exhibit similar rationality characteristics.
Keywords
Rational expectations • Unbiasedness • Weak efficiency • Micro-homogeneity •
Heterogeneity • Exchange rate • Survey forecasts • Aggregation bias • GMM •
SUR
43.1 Introduction
identical information sets, we must use individual survey forecasts to avoid the
aggregation bias inherent in the use of mean or median forecasts.
We use data from the same source as Ito (1990), the seminal study recognizing
the importance of using individual data to test rationality hypotheses about the
exchange rate. To achieve stationarity of the realizations and forecasts (which each
have a unit root), Ito (1990) followed the conventional specification at the time of
subtracting the current realization from each. These variables are then referred to as
being in “change” form. To test unbiasedness he regressed the future rate of
depreciation on the forecasted return and tested the joint restrictions that the
intercept equalled zero and the slope coefficient equalled one. At the industry
level he found approximately twice as many rejections (at the 1 % level) at the
longest horizon (6 months) than at the two shorter horizons (1 and 3 months).
We extend Ito’s analysis in two principal respects: the specification of unbiased-
ness tests and inference in tests for micro-homogeneity of forecasters. One problem
with the change specification of unbiasedness tests is that, since there is much more
variation in the change in the realization than in the forecast, there is a tendency to
under-reject the part of the joint hypothesis that the coefficient on the forecast
equals one. This is precisely what we would expect in tests of variables which are
near random walks.
Second, and more fundamentally, Ito’s (1990) bivariate (joint) regression test of
unbiasedness is actually a test of sufficiency, not necessity as well as sufficiency.
Following Holden and Peel (1990), the necessary and sufficient condition for
unbiasedness is a mean zero forecast error. This is tested in a univariate regression
by imposing a coefficient of unity on the forecast and testing the restriction that the
intercept equals zero. This critique applies whether or not the forecast and realiza-
tion are integrated in levels. However, when the realization and forecast are
integrated in levels, we must conduct a pretest to determine whether the forecast
error is stationary. If the forecast and realization are both integrated and
cointegrated, then a necessary and sufficient condition for unbiasedness is that
intercept and slope in the cointegrating regression (using levels of the realization
and forecast) are zero and one, respectively. We test this hypothesis using Liu and
Maddala’s (1992) method of imposing the (0, 1) vector, then testing the “restricted”
cointegrating residual for stationarity.1,2
Third, we use the result from Engle and Granger (1987) that cointegrated vari-
ables have an error correction representation. First, we employ the specification and
unbiasedness restrictions originally proposed by Hakkio and Rush (1989). How-
ever, the unbiasedness tests using the ECM specification produce more rejections
over industry groups and horizons than the univariate or bivariate specifications.
1
If in addition the residuals from the cointegrating regression are white noise, this supports a type
of weak efficiency.
2
Pretesting the forecast error for stationarity is a common practice in testing the RNMEH, but the
only study we know of that applies this practice to survey forecasts of exchange rates is Osterberg
(2000), and he does not test for a zero intercept in the cointegrating regression.
1198 R. Cohen et al.
We conjecture that one possible explanation for this apparent anomaly is that,
similar to the joint restrictions in the bivariate test, the ECM restrictions test
sufficient conditions for unbiasedness, while the univariate restriction only tests
a necessary and sufficient condition. Thus, the ECM has a tendency to over-reject.
We then respecify the ECM, so that only the necessary and sufficient conditions are
tested. We compare our results to those obtained using the sufficient conditions
represented by the joint restrictions as well as the necessary and sufficient condition
represented by the univariate restriction.
The second direction in which we extend Ito’s (1990) analysis has to do with
testing for differences among forecasters’ ability to produce rational predictions.3
We recognize, as does Ito, that differences among forecasters over time indicate
that at least some individuals form biased forecasts. (The converse does not
necessarily hold, since a failure to reject micro-homogeneity could conceivably
be due to the same degree of irrationality of each individual in the panel.) Ito’s
heterogeneity test is a single-equation test of deviations of individual forecasts from
the mean forecast, where the latter may or may not be unbiased. In contrast, we test
for differences in individual forecast performance using a micro-homogeneity test,
i.e., testing for equal coefficients across the system of individual univariate
rationality equations.
In our tests for micro-homogeneity, we expect cross-forecaster error correlation
due to the possibility of common macro shocks and/or herd effects in expectations.
To this end, we incorporate two innovations not previously used by investigators
studying survey data on exchange rate expectations. First, in our micro-
homogeneity tests, we use a GMM system with a variance-covariance matrix that
allows for cross-sectional as well as moving average and heteroscedastic errors.
Here we follow the widely used practice of modeling the individual regression
residuals as an MA process of order h-1, where h is the number of periods in the
forecast horizon. However, no other researchers have actually tested whether an
MA process of this length is required to model the cross-sectional behavior of
rational forecast errors. Thus, second, to investigate the nature of the actual MA
processes, we use Pesaran’s (2004) CD test to examine the statistical significance of
the cross-sectional dependence of forecast errors, both contemporaneous and
lagged.
The organization of the rest of the chapter is as follows: in Sect. 43.2 we review
some fundamental issues in testing rationality in the foreign exchange market. In
Sects. 43.3 and 43.4, we conduct various rationality tests on the JCIF data.
Section 43.5 contains our micro-homogeneity tests. Section 43.6 summarizes and
discusses areas for future research.
3
Market microstructure theories assume that there is a minimum amount of forecaster (as well as
cross-sectional forecast) diversity. Also, theories of exchange rate determination that depend upon
the interaction between chartists (or noise traders) and fundamentalists by definition require
a certain structure of forecaster heterogeneity.
43 Rationality and Heterogeneity of Survey Forecasts 1199
The Rational Expectations Hypothesis (REH) assumes that economic agents know
the true data-generating process (DGP) for the forecast variable. This implies that
the market’s subjective probability distribution of the variable is identical to the
objective probability distribution, conditional on a given information set, Ft.
Equating first moments of the market, Em ðstþh jFt Þ, and objective, Eðstþh jFt Þ,
distributions,
4
It is important to note that the result from one type of rationality test does not have implications
for the results from any other types of rationality tests. In this chapter we test for unbiasedness and
weak efficiency, leaving the more stringent tests of efficiency with respect to publicly available
information for future analysis.
1200 R. Cohen et al.
effect on exchange rate movements. (See Cavaglia et al. 1994.) More generally,
monetary models of the exchange rate (in which the UIP condition is embedded),
which assume model-consistent (i.e., rational) expectations with risk neutrality,
generally have not performed well empirically, especially in out-of-sample fore-
casting. (See, e.g., Bryant 1995.) One would like to be able to attribute the model
failure to some combination of a failure of the structural assumptions (including
risk neutrality) and a failure of the expectational assumption.
Beginning with Frankel and Froot (1987) and Froot and Frankel (1989), much of
the literature examining exchange rate rationality in general, and the decomposition
of deviations from the RNEMH in particular, has employed the representative agent
assumption to justify using the mean or median survey forecast as a proxy for the
market’s expectation. In both studies, Frankel and Froot found significant evidence
of irrationality. Subsequent research has found mixed results. Liu and Maddala
(1992, p. 366) articulate the mainstream justification for using aggregated forecasts
in tests of the REH. “Although . . .data on individuals are important to throw light
on how expectations are formed at the individual level, to analyze issues relating to
market efficiency, one has to resort to aggregates.” In fact, Muth’s (1961, p. 316)
original definition of rational expectations seemed to allow for the possibility
that rationality could be applied to an aggregate (e.g., mean or median) forecast.
‘. . . [E]xpectations of firms (or, more generally, the subjective probability distri-
bution of outcomes) tend to be distributed, for the same information set, about the
predictions of the theory (or the “objective” probability distribution of outcomes).’
(Emphasis added.)
However, if individual forecasters have different information sets, Muth’s def-
inition does not apply. To take the simplest example, the (current) mean forecast is
not in any forecaster’s information set, since all individuals’ forecasts must be made
before a mean can be calculated. Thus, current mean forecasts contain private
information (see MacDonald 1992) and therefore cannot be tested for rationality.5
Using the mean forecast may also result in inconsistent parameter estimates.
Figlewski and Wachtel (1983) were the first to show that, in the traditional bivariate
5
A large theoretical literature relaxes Muth’s assumption that all information relevant for forming
a rational forecast is publicly available. Instead, this literature examines how heterogeneous
individual expectations are mapped into an aggregate market expectation, and whether the latter
leads to market efficiency. (See, e.g., Figlewski 1978, 1982, 1984; Kirman 1992; Haltiwanger and
Waldman 1989.) Our paper focuses on individual rationality but allows for the possibility of
synergism by incorporating not only heteroscedasticity and autocorrelation consistent standard
errors in individual rationality tests but also cross-forecaster correlation in tests of micro-
homogeneity. The extreme informational requirement of the REH led Pesaran and Weale (2006)
to propose a weaker form of the REH that is based on the (weighted) average expectation using
only publicly available (i.e., common) information.
43 Rationality and Heterogeneity of Survey Forecasts 1201
Other than a failure to process available information efficiently, there are numerous
explanations for a rejection of the REH. One set of reasons relates to measurement
error in the individual forecast. Researchers have long recognized that forecasts of
economic variables collected from public opinion surveys should be less informed
than those sampled from industry participants. However, industry participants,
while relatively knowledgeable, may not be properly motivated to devote the
time and resources necessary to elicit their best responses. The opposite is also
possible.7 Having devoted substantial resources to produce a forecast of the price of
a widely traded asset, such as foreign exchange, forecasters may be reluctant to
reveal their true forecast before they have had a chance to trade for their own
account.8
Second, some forecasters may not have the symmetric quadratic loss function
embodied in typical measures of forecast accuracy, e.g., minimum mean
squared error. (See Zellner 1986; Stockman 1987; Batchelor and Peel 1998.)
In this case, the optimal forecast may not be the MSE. In one scenario, related to
6
The extent to which private information influences forecasts is more controversial in the foreign
exchange market than in the equity or bond markets. While Chionis and MacDonald (1997)
maintain that there is little or no private information in the foreign exchange market, Lyons
(2002) argues that order flow explains much of the variation in prices. To the extent that one agrees
with the market microstructure emphasis on the importance of the private information embodied in
dealer order flow, the Figlewski-Wachtel critique remains valid in the returns regression.
7
Elliott and Ito (1999) show that, although a random walk forecast frequently outperforms the JCIF
survey forecasts using an MSE criterion, survey forecasts generally outperform the random walk,
based on an excess profits criterion. This supports the contention that JCIF forecasters are properly
motivated to produce their best forecasts.
8
To mitigate the confidentiality problem in this case, the survey typically withholds individual
forecasts until the realization is known or (as with the JCIF) masks the individual forecast by only
reporting some aggregate forecast (at the industry and total level) to the public.
1202 R. Cohen et al.
the incentive aspect of the measurement error problem, forecasters may have
strategic incentives involving product differentiation.9
In addition to strategic behavior, another scenario in which forecasters may
deviate from the symmetric quadratic loss function is simply to maximize trading
profits. This requires predicting the direction of change, regardless of MSE.10
Third, despite their best efforts, forecasters may find it difficult to distinguish
between a temporary and permanent shift in the DGP. This difficulty underlies at
least three theories of rational forecast errors: the peso problem, learning about past
regime changes, and bubbles.
Below we conduct tests for structural change in estimated unbiasedness coeffi-
cients. When unbiasedness cannot be rejected, the structural change test may show
certain subperiods in which unbiasedness did not hold. In the obverse case, when
unbiasedness can be rejected, the structural change test may show certain sub-
periods in which unbiasedness cannot be rejected. Either situation would lend some
support to the theories attributing bias to the difficulty of distinguishing temporary
from permanent shifts.
Every 2 weeks, the JCIF in Tokyo conducts telephone surveys of yen/dollar exchange rate
expectations from 44 firms. The forecasts are for the future spot rate at horizons of
1 month, 3 months, and 6 months. Our data cover the period May 1985 to March 1996.
This data set has very few missing observations, making it close to a true panel. For
reporting purposes, the JCIF currently groups individual firms into four industry catego-
ries: (1) banks and brokers, (2) insurance and trading companies, (3) exporters, and (4) life
insurance companies and importers. On the day after the survey, the JCIF announces
overall and industry average forecasts. (For further details concerning the JCIF database,
see the descriptions in Ito (1990, 1994), Bryant (1995), and Elliott and Ito (1999).)
Figure 43.1 shows that, over the sample period (one of flexible exchange rates and
no capital controls), the yen appreciated dramatically relative to the dollar, from
a spot rate of approximately 270 yen/dollar in May 1985 to approximately
90 yen/dollar in March 1996. The path of appreciation was not steady, however.
In the first 2 years of the survey alone, the yen appreciated to about 140 per dollar.
9
Laster et al. (1999) called this practice “rational bias.” Prominent references in this growing
literature include Lamont (2002), Ehrbeck and Waldmann (1996), and Batchelor and Dua
(1990a, b, 1992). Because we have access only to forecasts at the industry average level, we
cannot test the strategic incentive hypotheses.
10
See Elliott and Ito (1999), Boothe and Glassman (1987), LeBaron (2000), Leitch and Tanner
(1991), Lai (1990), Goldberg and Frydman (1996), and Pilbeam (1995). This type of loss function
may appear to be relevant only for relatively liquid assets such as foreign exchange, but not for
macroeconomic flows. However, the directional goal is also used in models to predict business
cycle turning points. Also, trends in financial engineering may lead to the creation of derivative
contracts in macroeconomic variables, e.g., CPI futures.
43 Rationality and Heterogeneity of Survey Forecasts 1203
220
200
180
160
140
120
100
st
80
85/05/29 88/03/30 91/02/26 94/01/11 96/11/12
The initial rapid appreciation of the yen is generally attributed to the Plaza meeting in
September 1985, in which the Group of Five countries decided to let the dollar
depreciate, relative to the other currencies. At the Louvre meeting in February 1987,
the Group of Seven agreed to stabilize exchange rates by establishing soft target
zones. These meetings may well be interpreted as unanticipated regime changes,
since, as we will see below, forecasters generally underestimated the rapid appreci-
ation following the Plaza meeting, then overestimated the value of the yen following
the Louvre meeting. Thus, forecasts during these periods may have been subject to
peso and learning problems. The period of stabilization lasted until about 1990, when
yen appreciation resumed and continued through the end of the sample period.
Early studies of the unbiasedness aspect of rationality regressed the level of the
realization on the level of the forecast, testing the joint hypothesis that the intercept
equalled zero and the slope equalled one.11 However, since many macroeconomic
variables have unit roots, and realizations and forecasts typically share a common
stochastic trend, a rational forecast will be integrated and cointegrated with the target
series. (See Granger 1991, pp. 69–70.) According to the modern theory of regressions
11
The efficiency aspect of rationality is sometimes tested by including additional variables in the
forecaster’s information set, with corresponding hypotheses of zero coefficients on these variables.
See, e.g., Keane and Runkle (1990) for a more recent study using the level specification and
Bonham and Cohen (1995) for a critique of Keane and Runkle’s integration accounting.
1204 R. Cohen et al.
with integrated processes (see, inter alia Banerjee et al. 1993), conventional OLS
estimation and inference produce a slope coefficient that is biased toward one and,
therefore, a test statistic that is biased toward accepting the null of unbiasedness. The
second generation studies of unbiasedness addressed this inference problem by
subtracting the current realization from the forecast as well as the future realization,
transforming the levels regression into a “changes” regression. In this specification of
stationary variables, unbiasedness was still tested using the same (0, 1) joint hypothesis
as in the levels regression. (Ito (1990) is an example of this methodology.) However, an
implication of Engle and Granger (1987) is that the levels regression is now interpreted
as a cointegrating regression, with conventional t-statistics following nonstandard
distributions which depend on nuisance parameters. After establishing that the reali-
zation and forecast are integrated and cointegrated, we perform two types of rationality
tests. The first is a “restricted cointegration” test due to Liu and Maddala (1992). This is
a cointegration test imposing the (0, 1) restriction on the levels regression.
It is significant that, if realization and forecast are cointegrated, Liu and
Maddala’s (1992) technique is equivalent to regressing a stationary forecast error
on a constant and then testing whether the coefficient equals zero (to test unbiased-
ness) and/or whether the residuals are white noise (to test a type of weak efficiency).
Pretests for unit roots in the realization, forecast, and forecast error are required for
at least three reasons. First, univariate tests of unbiasedness are invalid if the
forecast error is not stationary. Second, following Holden and Peel (1990), we
show below (in Sect. 43.4.1.1) that nonrejection of the joint test in the bivariate
regression is sufficient but not necessary for unbiasedness, since the joint test is also
an implicit test of weak efficiency with respect to the lagged forecast error. A zero
intercept in the (correctly specified) univariate test is a necessary as well as
sufficient condition for unbiasedness. Third, the Engle and Granger (1987) repre-
sentation theorem proves that a cointegrating regression such as the levels joint
regression (Eq. 43.2 below) has an error correction form that includes both
differenced variables and an error correction term in levels. Under the joint null,
the error correction term is the forecast error. While the change form of the bivariate
regression, is not, strictly speaking, misspecified (since the regressor subtracts st,
e
not st1 , from ste), the ECM specification may produce a better fit to the data and,
therefore, a more powerful test of the unbiasedness restrictions. We conduct such
tests using a form of the ECM due to Hakkio and Rush (1989).
E(st+h − st)
Holden and Peel (1990) show that this result is a sufficient, though not a necessary,
condition for unbiasedness. The intuition for the lack of necessity comes
from interpreting the right-hand side of the bivariate unbiasedness regression
as a linear combination of two potentially unbiased forecasts: a constant
equal to the unconditional mean forecast plus a variable forecast, i.e.,
st + h st ¼ (1 bi,h) E(sei;t;h st) + bi,h(sei;t;h st) + ei,t,h. Then the intercept is
ai,h ¼ (1 bi,h) E(sei;t;h st). The necessary and sufficient condition for
unbiasedness is that the unconditional mean of the subjective expectation
E[sei;t;h st] equal the unconditional mean for the objective expectation
E[st + h st]. However, this equality can be satisfied without ai,h being equal to
zero, i.e., bi,h ¼ 1.
Figure 43.2 shows that an infinite number of ai,h, bi,h estimates are
consistent with unbiasedness. The only constraint is that the regression line inter-
sects the 45 ray from the origin where the sample mean of the forecast and target
are equal. Note that, in the case of differenced variables, this can occur at the origin,
so that ai,h ¼ 0, but bi,h is unrestricted (see Fig. 43.3). It is easy to see why
unbiasedness holds: in Figs. 43.2 and 43.3 the sum of all horizontal deviations
from the 45 line to the regression line, i.e., forecast errors, equal zero. However,
when ai,h 6¼ 0, and ai,h 6¼ (1 bi,h) E(sei;t;h st), there is bias regardless of the
value of bi,h. See Fig. 43.4, where the bias, E(st + h sei;t;h ), implies systematic
underforecasts.
1206 R. Cohen et al.
E(st+h − st)
0
E(s ei,t,h−st)
45⬚
0 s ei,t,h−st
st+h − st
E(st+h − st)
bias
To investigate the rationality implications of different values for ai,h and bi,h, we
follow Clements and Hendry (1998) and rewrite the forecast error in the bivariate
regression framework of Eq. 43.2 as
i, t, h ¼ stþh sei, t, h ¼ ai, h þ bi, h 1 sei, t, h st þ ei, t, h (43.3)
A special case of weak efficiency occurs when the forecast and forecast error are
uncorrelated, i.e.,
h i 2
E i, t, h sei, t, h st ¼ 0 ¼ ai, h E sei, t, h st þ bi, h 1 E sei, t, h st
h i
þ E ei, t, h sei, t, h st
(43.4)
Thus, satisfaction of the joint hypothesis (ai,h, bi,h) ¼ (0,1) is also sufficient for weak
efficiency with respect to the current forecast. However, it should be noted that Eq. 43.4
may still hold even if the joint hypothesis is rejected. Thus, satisfaction of the joint
hypothesis represents sufficient conditions for both unbiasedness and this type of weak
efficiency, but necessary conditions for neither.
If bi,h ¼ 1, then, whether or not ai,h ¼ 0, the variance of the forecast error equals
the variance of the bivariate regression residual, since then var(i,t,h) ¼
(bi, h 1)2var(sei;t;h st) + var(ei,t,h) + 2(bi,h 1)cov[(sei;t;h st),ei,t,h] ¼ var(ei,t,h).
Figure 43.4 illustrates this point. Mincer and Zarnowitz (1969) required only that
bi, h ¼ 1 in their definition of forecast efficiency. If in addition to bi,h ¼ 1, ai,h ¼ 0, then
the mean square forecast error also equals the variance of the forecast. Mincer and
Zarnowitz emphasized that as long as the loss function is symmetric, as is the case
with a minimum mean square error criterion, satisfaction of the joint hypothesis
implies optimality of forecasts.
with horizon. There are only two rejections at the 5 % level for each of the two
shorter horizons. However, the ai,h’s are all rejected at the 6.7 % significance level
for the 6-month horizon. The test results for the bi,h’s follow the opposite pattern
with respect to horizon. The null that bi,h ¼ 1 is rejected for all groups at the
1-month horizon, but only for the exporters at the 3-month horizon. There are no
rejections at the 6-month horizon. Thus, it appears that the pattern of rejection of the
joint hypothesis is predominantly influenced by tests of whether the slope coeffi-
cient equals one. In particular, tests of the joint hypothesis at the 1-month horizon
are rejected due to failure of this type of weak efficiency, not simple unbiasedness.
For this reason, Mincer and Zarnowitz (1969) and Holden and Peel (1990)
suggest that, if one begins by testing the joint hypothesis, rejections in this first
stage should be followed by tests of the simple unbiasedness hypothesis in a second
stage. Only if unbiasedness is rejected in this second stage should one conclude that
forecasts are biased. For reasons described below (in Sect. 43.4.2), our treatment
eliminates the first stage, so that unbiasedness and weak efficiency are separately
assessed using the forecast error as the dependent variable.
Finding greater efficiency at the longer horizon is unusual, because forecasting
difficulty is usually thought to increase with horizon. However, the longer horizon
result may not be as conclusive as the bi,h statistics suggest. For all tests at all
horizons, in only one case can the null hypothesis that bi,h equals zero not be rejected.
Thus, for the longer two horizons (with just the one exception for exporters at the
43 Rationality and Heterogeneity of Survey Forecasts 1209
3-month horizon), hypothesis testing cannot distinguish between the null hypotheses
that bi,h equals one or zero. Therefore, we cannot conclude that weak efficiency with
respect to the current forecast holds while unbiasedness may not. The failure to
precisely estimate the slope coefficient also produces R2s that are below 0.05 in all
regressions.12 The conclusion is that testing only the joint hypothesis has the potential
to obscure the difference in performance between the unbiasedness and weak effi-
ciency tests. This conclusion is reinforced by an examination of Figs. 43.5, 43.6, and
43.7, the scatterplots, and regression lines for the bivariate regressions.13 All three
scatterplots have a strong vertical orientation. With this type of data, it is easy to find
the vertical midpoint and test whether it is different from zero. Thus, (one-parameter)
tests of simple unbiasedness are feasible. However, it is difficult to fit a precisely
12
As we report in Sect. 43.5, this lack of power is at least consistent with the failure to reject micro-
homogeneity at all three horizons.
13
Note that, for illustrative purposes only, we compute the expectational variable as the four-group
average percentage change in the forecast. However, recall that, despite the failure to reject micro-
homogeneity at any horizon, the Figlewski-Wachtel critique implies that these parameter esti-
mates are inconsistent in the presence of private information. (See the last paragraph in this
subsection.)
1210 R. Cohen et al.
estimated regression line to this scatter, because the small variation in the forecast
variable inflates the standard error of the slope coefficient. This explains why the
bi,h’s are so imprecisely estimated that the null hypotheses that bi,h ¼ 1 and 0 are
simultaneously not rejected. This also explains why the R2s are so low. Thus,
examination of the scatterplots also reveal why bivariate regressions are potentially
misleading about weak efficiency as well as simple unbiasedness. Therefore, in
contrast to both Mincer and Zarnowitz (1969) and Holden and Peel (1990), we prefer
to separate tests for unbiasedness from tests for (all types of) weak efficiency at the
initial stage. This obviates the need for a joint test. In the next section, we conduct
such tests, making use of cointegration between forecast and realization where
it exists.14
More fundamentally, the relatively vertical scatter of the regression observations
around the origin is consistent with an approximately unbiased forecast of a random
14
However, in the general case of biased and/or inefficient forecasts, Mincer and Zarnowitz (1969,
p. 11) also viewed the bivariate regression ‘as a method of correcting the forecasts . . . to improve
[their] accuracy . . . Theil (1966, p. 33) called it the “optimal linear correction.”’ That is, the
correction would involve (1) subtracting ai,h and then (2) multiplying by 1/bi,h. Graphically, this is
a translation of the regression line followed by a rotation, until the regression line coincides with
the 45 line.
43 Rationality and Heterogeneity of Survey Forecasts 1211
0.05
−0.05
−0.1
−0.15
−0.15 −0.1 −0.05 0 0.05 0.1
s ei,t,h−st
st+h − st
0.2
0.15
0.1
0.05
−0.05
−0.1
−0.15
−0.2
Fig. 43.6 Actual versus −0.2 −0.15 −0.1 −0.05 0 0.05 0.1 0.15 0.2
expected depreciation, s ei,t,h−st
3-month-ahead forecast
1212 R. Cohen et al.
0.2
0.1
−0.1
−0.2
−0.3
−0.3 −0.2 −0.1 0 0.1 0.2 0.3
s ei,t,h−st
walk-in exchange rate levels.15 In Figs. 43.8, 43.9, and 43.10, we observe
a corresponding time series pattern of variation between the forecasts and realiza-
tions in return form. As Bryant lamented in reporting corresponding regressions
using a shorter sample from the JCIF, “the regression. . .is. . .not one to send home
proudly to grandmother” (Bryant 1995, p. 51). He drew the conclusion that “ana-
lysts should have little confidence in a model specification [e.g., uncovered interest
parity] setting [the average forecast] exactly equal to the next-period value of the
model. . .[M]odel-consistent expectations. . .presume a type of forward-looking
behavior [e.g., weak efficiency] that is not consistent with survey data on
expectations” (Bryant 1995, p. 40).
To test the null hypothesis of a unit root, we estimate the augmented Dickey-Fuller
(1979) (ADF) regression
X p
Dytþ1 ¼ a þ byt þ gt þ yk Dytþ1k þ etþ1 (43.5)
k¼1
15
Other researchers (e.g., Bryant 1995) have found similar vertical scatters for regressions where
the independent variable, e.g., the forward premium/discount ft,h st, the “exchange risk pre-
mium” ft,h st+h, or the difference between domestic and foreign interest rates (i i*), exhibits
little variation.
43 Rationality and Heterogeneity of Survey Forecasts 1213
0.06
0.04
0.02
−0.02
−0.04
−0.06
−0.08 sei,t,h − st
st+h − st
−0.1
86/05/14 91/01/16 96/10/15
0.2
0.15
0.1
0.05
−0.05
−0.1
−0.15
sei,t,h − st
Fig. 43.9 Actual and st+h − st
expected depreciation, −0.2
3-month-ahead forecast 86/05/14 91/01/16 96/07/30
where y is the level and first difference of the spot exchange rate, the level and first
difference of each group forecast, the residual from the (unrestricted) cointegrating
regression, and the forecast error (i.e., the residual from the “restricted”
cointegrating equation). The number of lagged differences to include in Eq. 43.5
1214 R. Cohen et al.
0.15
0.1
0.05
−0.05
−0.1
−0.15
−0.2 sei,t,h − st
st+h − st
−0.25
86/05/14 90/11/27 96/05/15
is chosen by adding lags until a Lagrange multiplier test fails to reject the null
hypothesis of no serial correlation (up to lag 12). We test the null hypothesis of
a unit root (i.e., b ¼ 0) with the ADF t and z tests. We also test the joint null
hypothesis of a unit root and no linear trend (i.e., b ¼ 0 and g ¼ 0).
As can be seen in Tables 43.4, 43.5, and 43.6, we fail to reject the null of
a unit root in the log of the spot rate in two of the three unit root tests (the
exception being the joint null), but we reject the unit root in the h th difference for
all three horizons. We conclude that the log of the spot rate is integrated of order
one. Similarly, we conclude that the log of the forecast of each spot rate is
integrated of order one. Thus, we can conduct cointegration tests on the
spot rate and each corresponding forecast. The null of a unit root in the
(unrestricted) residual in the “cointegrating regression” is rejected at the 10 %
level or less for all groups and horizons except group three (exporters) at the
6-month horizon. Thus, we can immediately reject unbiasedness for the latter
group and horizon. Next, since a stationary forecast error is a necessary condition
for unbiasedness, we test for unbiasedness (as well as) and weak efficiency in
levels using Liu and Maddala’s (1992) method of “restricted cointegration.” This
specification imposes the joint restriction ai,h ¼ 0, bi,h ¼ 1 on the bivariate
regression
and tests whether the residual (the forecast error) is nonstationary. In a bivariate
regression, any cointegrating vector is unique. Therefore, if we find that the forecast
43 Rationality and Heterogeneity of Survey Forecasts 1215
errors are stationary, then the joint restriction is not rejected, and (0, 1) must be the
unique cointegrating vector.16 The advantage of the one-step restricted
cointegration is that if the joint hypothesis is true, then tests which impose this
cointegrating vector have greater power than those which estimate a cointegrating
vector. See, e.g., Maynard and Phillips (2001).
Note that the Holden and Peel (1990) critique does not apply in the I(1) case,
because the intercept cannot be an unbiased forecast of a nonstationary variable.
Thus, the cointegrating regression line of the level realization on the level forecast
16
It is also possible to estimate the cointegrating parameters and jointly test whether they are zero
and one. A variety of methods, such as those due to Saikkonen (1991) or Phillips and Hansen
(1990), exist that allow for inference in cointegrated bivariate regressions.
1216 R. Cohen et al.
must have both a ¼ 0 and b ¼ 1 for unbiasedness to hold. This differs from
Fig. 43.3, the scatterplot in differences, where ai,h ¼ 0 but bi,h 6¼ 1. Intuitively,
the reason for the difference in results is that the scatterplot in levels must lie in the
first quadrant, i.e., no negative values of the forecast or realization.
At the 1-month horizon, the null of a unit root in the residual of the restricted
cointegrating regression (i.e., the forecast error) is rejected at the 1 % level for all
groups. We find nearly identical results at the 3-month horizon; the null of a unit
root in the forecast error is rejected at the 5 % level for all groups. Thus, for these
regressions we can conduct rationality tests by regressing the forecast error on
a constant (hypothesized equal to zero for unbiasedness) and other information set
variables (whose coefficients are hypothesized equal to zero for efficiency). (Recall
just above that we failed to reject the null of a unit root in the unrestricted residual
43 Rationality and Heterogeneity of Survey Forecasts 1217
for the 6-month forecasts of exporters.) Now, in the case of the restricted residual,
the other three groups failed to reject a unit root at the 10 % level in two out of three
of the unit root tests.17 (See Figs. 43.11, 43.12, and 43.13.) Thus, in contrast to the
results for the two shorter horizons, at the 6-month horizon, the evidence is clearly
in favor of a unit root in the forecast error for all four groups. Therefore, we reject
the null of simple unbiasedness because a forecast error with a unit root cannot be
mean zero. In fact, given our finding of a unit root in the forecast errors, rationality
tests regressing the forecast error on a constant and/or other information set
variables would be invalid.
17
As expected, exporters failed to reject at the 10 % level in all three tests.
1218 R. Cohen et al.
220
200
180
160
140
120
100
80
80 100 120 140 160 180 200 220 240 260
sei,t,h
where i,t,h is the forecast error of individual i, for an h-period-ahead forecast made
at time t. The results are reported in Tables 43.7 and 43.8. For the 1-month horizon,
unbiasedness cannot be rejected at conventional significance levels for any group.
For the 3-month horizon, unbiasedness is rejected only for exporters (at a p-value of
0.03). As we saw in the previous subsection, rationality is rejected for all groups at
the 6-month horizon, due to nonstationary forecast errors.18
In these unbiasedness tests, as well as all others, it is possible that coefficient
estimates for the entire sample are not stable over subsamples. The lower panels of
Tables 43.7 and 43.8 contain results of the test for equality of intercepts in four
equal subperiods, each consisting of approximately 75 biweekly forecasts:
18
The direction of the bias for exporters is negative; that is, they systematically underestimate the
value of the yen, relative to the dollar. Ito (1990) found the same tendency using only the first two
years of survey data (1985–1987). He characterized this depreciation bias as a type of “wishful
thinking” on the part of exporters.
43 Rationality and Heterogeneity of Survey Forecasts 1219
220
200
180
160
140
120
100
80
80 100 120 140 160 180 200 220 240 260
sei,t,h
st+h
260
240
220
200
180
160
140
120
100
For both 1- and 3-month horizons, all four forecaster groups undervalued the yen
in the first and third subperiods. This is understandable, as both these subperiods
were characterized by overall yen appreciation. (See Fig. 43.1.) Evidently, fore-
casters underestimated the degree of appreciation. Exporters were the only group to
undervalue the yen in the last subperiod as well, although that was not one of overall
yen appreciation. This is another perspective on the “wishful thinking” of
exporters.19
The main difference between the two horizons is in the significance of the test
for structural breaks. For the 1-month horizon, the estimates of the individual break
dummies generally do not reach statistical significance, and the test for their
19
Ito (1994) conducted a similar analysis for the aggregate of all forecasters, but without an
explicit test for structural breaks.
43 Rationality and Heterogeneity of Survey Forecasts 1221
equality rejects only for the exporters. Thus, the exporters’ bias was not constant
throughout the sample. In contrast, for the 3-month horizon, the test for no struc-
tural breaks is rejected at the 5 % level for all groups, even though unbiasedness
itself is rejected for the full sample only for exporters. Even setting aside the bias
and variability of exporters’ forecasts, our structural break tests allow us to con-
clude that there is considerably more variation around roughly zero mean forecast
errors at the longer horizon. This probably reflects the additional uncertainty
inherent in longer-term forecasts.20
20
This is consistent with the finding of nonstationary forecast errors for all groups at the 6-month
horizon.
1222 R. Cohen et al.
According to this specification of the ECM, the change in the spot rate is
a function of the change in the forecast, interpreted as a short-run effect, and the
current forecast error, interpreted as a long-run adjustment to past disequilibria. ai,h,
the coefficient of the error correction term, represents the fraction of the forecast
error observed at t-h that is corrected by time t. A negative coefficient indicates
a stabilizing adjustment of expectations. This formulation of the ECM has the
advantage that the misspecification (due to omitted variable bias) of the regression
of the differenced future spot rate on the differenced current forecast can be gauged
by the statistical significance of the error correction term.21
The regressors include the smallest number of lagged dependent variables
required such that we do not reject the hypothesis that the residuals are white
noise. We impose gi,h ¼ 1 when “restricted” cointegration of st + h and sei;t;h is not
rejected. Recall that 1- and 3-month forecast errors were found to be stationary, so it
was for these two horizons that estimation of the simple unbiasedness equation was
possible. Although it would be valid to estimate the ECM at the 6-month horizon
using the (unrestricted) stationary cointegrating residual (i.e., for all groups but
exporters), we elect not to, because the nonstationarity of the forecast error itself
implies a failure of the unbiasedness restrictions.22
Then, as first asserted by Hakkio and Rush (1989), the unbiasedness restriction is
represented by the joint hypothesis that ai,h ¼ bi,h ¼ 1 and all d and coefficients
equal zero.23 (The hypothesized coefficient on the error correction term of 1
21
Zacharatos and Sutcliffe (2002) note that the inclusion of the contemporaneous spot forecast
(in their paper, the forward rate) as a regressor assumes that the latter is weakly exogenous; that is,
deviations from unbiasedness are corrected only by movements in the realized spot rate. These
authors prefer a bivariate ECM specification, in which the change in the future spot rate and the
change in the contemporaneous forecast are functions of an error correction term and lags of the
dependent variables. However, Zivot (2000) points out that if the spot rate and forecast are
contemporaneously correlated, then our single-equation specification does not make any assump-
tions about the weak exogeneity of the forecast.
22
Our empirical specification of the ECM also includes an intercept. This will help us to determine
whether there are structural breaks in the ECM.
23
Since we include an intercept, we also test the restriction that the intercept equals zero – both
individually and as part of the joint unbiasedness hypothesis.
43 Rationality and Heterogeneity of Survey Forecasts 1223
reflects the unbiasedness requirement that the entire forecast error is corrected
within the forecast horizon h.) We also test unbiasedness without including lagged
dependent variables but incorporating robust standard errors which allow for
generalized serial correlation and heteroscedasticity. This allows comparison with
the univariate and bivariate unbiasedness equations.
First, we compare the ECM results to the joint unbiasedness restrictions in the
change regressions, using robust standard errors in both cases. Although the esti-
mated coefficient of the error correction term is generally negative, indicating a stable
error correction mechanism,24 the coefficient does not reach a 5 % significance level
in any of the regressions. Thus, there is little evidence that the error correction term
plays a significant role in the long-run dynamics of exchange rate changes. The ECM
test results are nearly identical to the joint unbiasedness test results in Table 43.9. In
both specifications, unbiasedness is rejected for three of four groups at the 1-month
horizon and not rejected for three of four groups at the 3-month horizon. However,
even though the EC term is not (individually) significant in the ECMs, it does provide
explanatory power, relative to the joint unbiasedness specification. The R2s in the
ECM, while never more than 0.044, still are greater than in the joint unbiasedness
specification, typically by factors of three to five (Tables 43.10, 43.11, and 43.12).
Second, we compare the ECM results to the univariate simple unbiasedness
regressions, again using robust standard errors in both cases. The ECM unbiased-
ness restrictions are rejected at a 5 % level more often than in the simple unbiased-
ness tests. Whereas the only rejection of simple unbiasedness at the shorter two
horizons is for exporters at the 3-month horizon, the ECM restrictions are rejected
for three out of four groups at the 1-month horizon as well as for exporters at the
3-month horizon.
While it is uncontroversial that, for testing unbiasedness, the ECM is preferred to
the conventional bivariate specification in returns, it is not at all clear that the ECM
is preferred to the simple univariate test of unbiasedness. Can the more decisive
rejections of unbiasedness using the ECM versus the simple univariate specification
be reconciled?25
One way to proceed is to determine whether the unbiasedness restrictions
imposed on the ECM are necessary as well as sufficient, as is the case for the
simple unbiasedness test, or just sufficient, as is the case for the bivariate unbiased-
ness test. Thus, it is possible that the stronger rejections of unbiasedness in the ECM
specification are due to the implicit test of weak efficiency with respect to the
current forecast. That is, the Holden and Peel (1990) critique applies to the Hakkio
and Rush (1989) test in Eq. 43.9, as well as the joint unbiasedness test in the returns
regression. Setting bi,h, the coefficient of the contemporaneous differenced forecast,
equal to one produces an ECM in which the dependent variable is the forecast error:
24
The only exception is for exporters at the 1-month horizon.
25
The standard errors in the univariate regression are about the same as those for the ECM.
(By definition, of course, the R2s for the univariate regression equal zero.)
1224 R. Cohen et al.
st sei, t, h ¼ 1 þ ai, h st sei, th, h (43.10)
Thus, in the ECM the necessary and sufficient condition for unbiasedness
is that ai,h equals 1.26 Table 43.9 contains tests of this conjecture. Here the
joint hypothesis that the intercept equals zero and ai,h equals minus one produces
exactly the same results as in the simple unbiasedness tests.27 It is interesting that,
even when we can decouple the test for weak efficiency with respect to the current
forecast from the unbiasedness test, the test of unbiasedness using this
ECM specification still requires weak efficiency with respect to the current
forecast error.28
26
Since we estimate the restricted ECM with an intercept, unbiasedness also requires the intercept
to be equal to zero.
27
Since the intercept in Eq. 43.10 is not significant in any regression, the simple hypothesis that ai,h
equals one also fares the same as the simple unbiasedness tests.
28
For purposes of comparison with both the bivariate joint and simple unbiasedness restrictions,
we have used the ECM results using the robust standard errors. In all cases testing the ECM
restrictions using F-statistics based on whitened residuals produces rejections of all restrictions,
simple and joint, except a zero intercept. Hakkio and Rush (1989) found similarly strong rejections
of Eq. 43.9, where the forecast was the forward rate.
29
Notice that the first two sets of weak efficiency variables include the mean forecast, rather than
the individual group forecast. Our intention is to allow a given group to incorporate information
from other groups’ forecasts via the prior mean forecast. This requires an extra lag in the
information set variables, relative to a contemporaneously available variable such as the realized
exchange rate depreciation.
43 Rationality and Heterogeneity of Survey Forecasts 1229
1. Single and cumulative lags of the mean forecast error (lagged one period):
X
hþ7
stþh sei, t, h ¼ ai, h þ bi, tþhk stþhk sem, tþhk, h þ ei, t, h (43.11)
k¼hþ1
2. Single and cumulative lags of mean expected depreciation (lagged one period):
X
hþ7
stþh sei, t, h ¼ ai, h þ bi, tþhk sem, tþhk, h stk þ ei, t, h (43.12)
k¼hþ1
X
hþ6
stþh sei, t, h ¼ ai, h þ bi, tþhk ðstþhk stk Þ þ ei, t, h (43.13)
k¼h
For each group and forecast horizon, we regress the forecast error on the
most recent seven lags of the information set variable, both singly and
cumulatively. We use a Wald test of the null hypothesis ai,h ¼ bi,t + h k ¼ 0 and
report chi-square test statistics, with degrees of freedom equal to the number of
regressors excluding the intercept. If we were to perform only simple regressions
(i.e., on each lag individually), estimates of coefficients and tests of significance
could be biased toward rejection due to the omission of relevant variables. If we
were to perform only multivariate regressions, tests for joint significance could be
biased toward nonrejection due to the inclusion of irrelevant variables. It is also
possible that joint tests are significant but individual tests are not. This will be the
case when the linear combination of (relatively uncorrelated) regressors spans the
space of the dependent variable, but individual regressors do not.
In the only reported efficiency tests on JCIF data, Ito (1990) separately regressed
the forecast error (average, group, and individual firm) on a single lagged forecast
error, lagged forward premium, and lagged actual change. He found that, for the
51 biweekly forecasts between May 1985 and June 1987, rejections increased from
a relative few at the 1- or 3-month horizons to virtual unanimity at the 6-month
horizon. When he added a second lagged term for actual depreciation, rejections
increased “dramatically” for all horizons.
The second type of weak efficiency tests uses the Breusch (1978)-Godfrey
(1978) LM test for the null of no serial correlation of order k ¼ h or greater, up
to order k ¼ h + 6, in the residuals of the forecast error regression, Eq. 43.11.
Specifically, we estimate
X
h1 Xhþ6
^e i, t, h ¼ ai, h þ bi, k stþhk sei, tk, h þ fi, l^e i, tl, h þ i, t, h (43.14)
k¼1 l¼h
1230 R. Cohen et al.
and test the null hypothesis H0 : bi,h ¼ . . . ¼ bi,h + 6 ¼ 0 for h ¼ 2,6.30,31 Results for
all efficiency tests for the 1- and 3-month horizons are presented in Tables 43.13,
43.14, 43.15, 43.16, 43.17, 43.18, 43.19, and 43.20. (Recall that the nonstationarity
of the forecast errors at the 6-month horizon is an implicit rejection of weak
efficiency.) For each group, horizon, and variable, there are seven individual
tests, i.e., on a single lag, and six joint tests, i.e., on multiple lags. These 13 tests
are multiplied by four groups times two horizons times three weak efficiency
variables for a total of 312 efficiency tests.
Using approximately nine more years of data than Ito (1990), we find many
rejections. In some cases, nearly all single lag tests are rejected, yet few, if any, joint
tests are rejected. (See, e.g., expected depreciation at the 3-month horizon.) In other
cases, nearly all joint tests are rejected, but few individual tests. (See, e.g., actual
depreciation at the 3-month horizon.) Remarkably, all but one LM test for serial
correlation at a specified lag produces a rejection at less than a 10 % level, with
most at less than a 5 % level. Thus, it appears that the generality of the alternative
hypothesis in the LM test permits it to reject at a much greater rate than the
conventional weak efficiency tests, in which the variance-covariance matrix incor-
porates the Newey-West-Bartlett correction for heteroscedasticity and serial corre-
lation. Finally, unlike Ito (1990), we find no strong pattern between horizon length
and number of rejections.
30
This is a general test, not only because it allows for an alternative hypothesis of higher-order
serial correlation of specified order but also because it allows for serial correlation to be generated
by AR, MA, or ARMA processes.
31
We use the F-statistic because the w2 test statistics tend to over-reject, while the F-tests have
more appropriate significance levels (see Kiviet 1987).
43 Rationality and Heterogeneity of Survey Forecasts 1231
the same variables. For the univariate unbiasedness regressions, the null of micro-
homogeneity is given by H0: aih ¼ ajh, for all i, j 6¼ i. Before testing for homoge-
neous intercepts in Eq. 43.7, we must specify the form for our GMM system
variance-covariance matrix. Keane and Runkle (1990) first accounted for cross-
sectional correlation (in price level forecasts) using a GMM estimator on pooled
data. Bonham and Cohen (2001) tested the pooling specification by replacing
Zellner’s (1962) SUR variance-covariance matrix with a GMM counterpart
that incorporates the Newey-West single-equation corrections (used in our
individual equation tests above) plus allowances for corresponding
cross-covariances, both contemporaneous and lagged. Bonham and Cohen (2001)
1232 R. Cohen et al.
32
Elliott and Ito (1999) used single-equation estimation that incorporated a White correction for
heteroscedasticity and a Newey-West correction for serial correlation. (See the discussion below
of Ito’s tests of forecaster heterogeneity.)
43 Rationality and Heterogeneity of Survey Forecasts 1233
Under the null hypothesis of no cross-correlation, CD a N ð0; 1Þ.33 See Table 43.21
for CD test results. We tested for cross-correlation in forecast errors from lag zero up
to lags four and eight for the 1 and 3-month forecast horizons, respectively.
(The nonstationarity of the 6-month forecast error precludes using the CD test at
that horizon.) At the 1-month horizon, cross-correlations from lags zero to four are
each significant at the 5 % level. Since rational forecasts allow for (individual)
serial correlation of forecast errors at lags of h-1 or less, and h ¼ 2 for the 1-month
horizon, the cross-correlations at lags two through four indicate violations of weak
efficiency. Similarly, at the 3-month horizon, where h-1 ¼ 5, there is significant
cross-correlation at lag six.34 However, it should be noted that, for many
lags shorter than h, one cannot reject the null hypothesis that there are no
cross-correlated forecast errors.
33
Unlike Breusch and Pagan’s (1980) LM test for cross-sectional dependence, Pesaran’s (2004)
CD test is robust to multiple breaks in slope coefficients and error variances, as long as the
unconditional means of the variables are stationary and the residuals are symmetrically distributed.
34
There are three instances of statistically significant negative test statistics for lags greater than
h-1, none for lags less than or equal to h-1. Thus, some industries produce relatively high forecast
errors several periods after others produce relative low forecast errors, and this information is not
fully incorporated in some current forecasts.
43 Rationality and Heterogeneity of Survey Forecasts 1235
current forecast), but only for the exporters at the 3- and 6-month horizons.
However, there are no rejections of micro-homogeneity for any horizon.35
The micro-homogeneity test results are very different for both the 1- and 3-month
systems of univariate unbiasedness regressions in Tables 43.7 and 43.8. (Recall that
35
The nonrejection of micro-homogeneity in bivariate regressions does not, however, mean that one
can avoid aggregation bias by using the mean forecast. Even if the bivariate regressions were
correctly interpreted as joint tests of unbiasedness and weak efficiency with respect to the current
forecast, and even if the regressions had sufficient power to reject a false null, the micro-homogeneity
tests would be subject to additional econometric problems. According to the Figlewski-Wachtel
(1983) critique, successfully passing a pretest for micro-homogeneity does not ensure that estimated
coefficients from such consensus regressions will be consistent. See Sect. 43.2.1.
43 Rationality and Heterogeneity of Survey Forecasts 1237
unbiasedness was rejected for all groups at the 6-month horizon due to the
nonstationarity of the forecast error.) Despite having only one failure of unbiasedness
at the 5 % level for the two shorter horizons, micro-homogeneity is rejected at a level
of virtually zero for both horizons. The rejection of micro-homogeneity at the
1-month horizon occurs despite the failure to reject unbiasedness for any of the
industry groups. We hypothesize that the consistent rejection of micro-homogeneity
regardless of the results of individual unbiasedness tests is the result of sufficient
variation in individual bias estimates as well as precision in these estimates.
According to these tests, aggregation of individual forecasts into a mean forecast is
invalid at all horizons.
1238 R. Cohen et al.
In Table 43.23, we replicate Ito’s (1990) and Elliott and Ito’s (1999) test for
forecaster “heterogeneity.” This specification regresses the deviation of the indi-
vidual forecast from the cross-sectional average forecast on a constant. Algebrai-
cally, Ito’s regression can be derived from the individual forecast error regression
by subtracting the mean forecast error regression. Thus, because it simply replaces
the forecast error with the individual deviation from the mean forecast, it does not
suffer from aggregation bias (c.f. Figlewski and Wachtel (1983)) or pooling bias
(c.f. Zarnowitz 1985) (Table 43.24).36, 37
sei, t, h sem, t, h ¼ ai, h am þ ei, t, h em, t (43.16)
36
Recall that our group results are not entirely comparable to Ito’s (1990), since our data set, unlike
his, combines insurance companies and trading companies into one group and life insurance
companies and import-oriented companies into another group.
37
Chionis and MacDonald (1997) performed an Ito-type test on individual expectations data from
Consensus Forecasts of London.
1240 R. Cohen et al.
banks and brokers at the 1-month horizon and life insurance and import companies
at the 6-month horizon.38 Since Ito’s regressions have a similar form (though not
a similar economic interpretation) to the tests for univariate unbiasedness in
Tables 43.7 and 43.8, it is not surprising that micro-homogeneity tests on the
four-equation system of Ito equations produce rejections at a level of virtually
zero for all three horizons.
43.6 Conclusions
38
Elliott and Ito (1999), who have access to forecasts for the 42 individual firms in the survey, find
that, for virtually the same sample period as ours, the null hypothesis of a zero deviation from the
mean forecast is rejected at the 5 % level by 17 firms at the 1-month horizon, 13 firms for the
3-month horizon, and 12 firms for the 6-month horizon. These authors do not report results by
industry group.
1242 R. Cohen et al.
Using only 2 years of data, Ito (1990) found the same result for exporters,
which he described as a type of “wishful thinking.”
The unbiasedness results are almost entirely reversed when we test the
hypothesis using the conventional bivariate specification. That is, the joint
hypothesis of zero intercept and unit slope is rejected for all groups at the
1-month horizon, but only for exporters and the 3- and 6-month horizons. Thus,
in stark contrast to the univariate unbiasedness tests, as well as Ito’s (1990)
bivariate tests, forecast performance does not deteriorate with increases in
the horizon.
Also, since Engle and Granger (1987) have showed that cointegrated variables
have an error correction representation, we impose joint “unbiasedness” restrictions
first used by Hakkio and Rush (1989) on the ECM. However, we show that these
restrictions also represent sufficient, not necessary, conditions, so these tests could
tend to over-reject. We then develop and test restrictions which are both necessary
and sufficient conditions for unbiasedness. The test results confirm that the greater
rate of rejections of the joint “unbiasedness” restrictions in the ECM is caused by
the failure of the implicit restriction of weak efficiency with respect to the lagged
forecast. When we impose the restriction that the coefficient of the forecast equals
one, the ECM unbiasedness test results mimic those of the simple univariate
unbiasedness tests. For this data set, at least, it does not appear that an ECM
provides any value added over the simple unbiasedness test. Furthermore, since
the error correction term is not statistically significant in any regressions, it is
unclear whether the ECM provides any additional insight into the long-run adjust-
ment mechanism of exchange rate changes.
The failure of more general forms of weak efficiency is borne out by two types of
explicit tests for weak efficiency. In the first type, we regress the forecast error on
single and cumulative lags of mean forecast error, mean forecasted depreciation, and
actual depreciation. We find many rejections of weak efficiency. In the second type,
we use the Godfrey (1978) LM test for serial correlation of order h through h + 6 in the
residuals of the forecast error regression. Remarkably, all but one LM test at a specified
lag length produces a rejection at less than a 10 % level, with most at less than a 5 %
level. (As in the case of the univariate unbiasedness test, all weak efficiency tests at the
6-month horizon fail due to the nonstationarity of the forecast error.)
Whereas Ito (1990) and Elliott and Ito (1999) measured diversity as
a statistically significant deviation of an individual’s forecast from the cross-
sectional average forecast, we perform a separate test of micro-homogeneity for
each type of rationality test – unbiasedness as well as weak efficiency – that we first
conducted at the industry level. In order to conduct the systems estimation and
testing required for the micro-homogeneity test, our GMM estimation and inference
make use of an innovative variance-covariance matrix that extends the Keane and
Runkle (1990) counterpart from a pooled to an SUR-type structure. Our variance-
covariance matrix takes into account not only serial correlation and heterosce-
dasticity at the individual level (via a Newey-West-Bartlett correction) but also
forecaster cross-correlation up to h-1 lags. We document the statistical significance
of the cross-sectional correlation using Pesaran’s (2004) CD test.
43 Rationality and Heterogeneity of Survey Forecasts 1243
39
We put less weight on the results of the weaker tests for micro-homogeneity in the bivariate
regressions.
40
He also included regressors for adaptive expectations and the forward premium.
1244 R. Cohen et al.
The null hypothesis of micro-homogeneity is that the slope and intercept coeffi-
cients in the equation of interest are equal across individuals. This chapter considers
the case of individual unbiasedness regressions such as Eq. 43.2 in the text, repeated
here for convenience,
stþh st ¼ ai, h þ bi, h sei, t, h st þ ei, t, h (43.17)
S ¼ Fy þ e (43.18)
References
Banerjee, A., Dolado, J. J., Galbraith, J. W., & Hendry, D. F. (1993). Co-integration, error correction
and the econometric analysis of non-stationary data. Oxford: Oxford University Press.
Batchelor, R., & Dua, P. (1990a). Forecaster ideology, forecasting technique, and the accuracy of
economic forecasts. International Journal of Forecasting, 6, 3–10.
1246 R. Cohen et al.
Batchelor, R., & Dua, P. (1990b). Product differentiation in the economic forecasting industry.
International Journal of Forecasting, 6, 311–316.
Batchelor, R., & Dua, P. (1992). Conservatism and consensus-seeking among economic fore-
casters. Journal of Forecasting, 11, 169–181.
Batchelor, R., & Peel, D. A. (1998). Rationality testing under asymmetric loss. Economics Letters,
61, 49–54.
Bonham, C. S., & Cohen, R. H. (1995). Testing the rationality of price forecasts: Comment.
American Economic Review, 85(1), 284–289.
Bonham, C. S., & Cohen, R. H. (2001). To aggregate, pool, or neither: Testing the rational-
expectations hypothesis using survey data. Journal of Business & Economic Statistics, 19,
278–291.
Boothe, P., & Glassman, D. (1987). Comparing exchange rate forecasting models. International
Journal of Forecasting, 3, 65–79.
Breusch, T. S. (1978). Testing for autocorrelation in dynamic linear models. Australian Economic
Papers, 17, 334–355.
Breusch, T. S., & Pagan, A. R. (1980). The Lagrange multiplier test and its application to model
specifications in econometrics. Review of Economic Studies, 47, 239–253.
Bryant, R. C. (1995). The “exchange risk premium,” uncovered interest parity, and the treatment
of exchange rates in multicountry macroeconomic models. Technical report, The Brookings
Institution.
Cavaglia, S. M. F. G., Verschoor, W. F. C., & Wolff, C. C. P. (1994). On the biasedness of forward
foreign exchange rates: Irrationality or risk premia? Journal of Business, 67, 321–343.
Chionis, D., & MacDonald, R. (1997). Some tests of market microstructure hypotheses in the
foreign exchange market. Journal of Multinational Financial Management, 7, 203–229.
Clements, M. P., & Hendry, D. F. (1998). Forecasting economic time series. Cambridge:
Cambridge University Press.
Cohen, R. H., & Bonham, C. S. (2006). Specifying the forecast generating process for exchange
rate survey forecasts. Working paper, Department of Economics, University of Hawaii Manoa.
Davies, A., & Lahiri, K. (1995). A new framework for testing rationality and measuring aggregate
shocks using panel data. Journal of Econometrics, 68, 205–227.
Dickey, D. A., & Fuller, W. A. (1979). Distribution of the estimators for autoregressive time series
with a unit root. Journal of the American Statistical Association, 74, 427–431.
Ehrbeck, T., & Waldmann, R. (1996). Why are professional forecasters biased? Agency versus
behavioral explanations. The Quarterly Journal of Economics, 111, 21–40.
Elliott, G., & Ito, T. (1999). Heterogeneous expectations and tests of efficiency in the yen/dollar
forward exchange rate market. Journal of Monetary Economics, 43, 435–456.
Engle, R., & Granger, C. W. S. (1987). Cointegration and error correction representation,
estimation and testing. Econometrica, 55, 251–276.
Figlewski, S. (1978). Market “efficiency” in a market with heterogeneous information. Journal of
Political Economy, 86, 581–597.
Figlewski, S. (1982). Information diversity and market behavior. Journal of Finance, 37, 87–102.
Figlewski, S. (1984). Information diversity and market behavior: A reply. Journal of Finance, 39,
299–302.
Figlewski, S., & Wachtel, P. (1983). Rational expectations, informational efficiency, and tests
using survey data: A reply. Review of Economics and Statistics, 65, 529–531.
Frankel, J. A., & Froot, K. A. (1987). Using survey data to test standard propositions regarding
exchange rate expectations. American Economic Review, 77, 133–153.
Frankel, J. A., & Froot, K. A. (1990). Chartists, fundamentalists, and trading in the foreign
exchange market. American Economic Review, 80, 181–185.
Froot, K. A., & Frankel, J. A. (1989). Forward discount bias: Is it an exchange risk premium?
Quarterly Journal of Economics, 104, 139–161.
Gavin, W. T. (2003) FOMC forecasts: Is all the information in the central tendency? Economic
Review, Federal Reserve Bank of St. Louis, May/June, 27–46.
43 Rationality and Heterogeneity of Survey Forecasts 1247
Godfrey, L. G. (1978). Testing for higher order serial correlation in regression equations when the
regressors include lagged dependent variables. Econometrica, 46, 1303–1310.
Goldberg, M. D., & Frydman, R. (1996). Imperfect knowledge and behaviour in the foreign
exchange market. Economic Journal, 106, 869–893.
Granger, C. W. J. (1991). Developments in the study of cointegrated economic variables. In
R. F. Engle & C. W. J. Granger (Eds.), Long run economic relationships: Readings in
Cointegration (pp. 65–80). Oxford: Oxford University Press.
Hakkio, C. S., & Rush, M. (1989). Market efficiency and cointegration: An application to the
sterling deutschemark exchange markets. Journal of International Money and Finance, 8,
75–88.
Haltiwanger, J. C., & Waldman, M. (1989). Rational expectations in the aggregate. Economic
Inquiry, 27, 619–636.
Hamilton, J. (1994). Time series analysis. Princeton: Princeton University Press.
Hansen, L. P., & Hodrick, R. J. (1980). Forward exchange rates as optimal predictors of future spot
rates: An econometric analysis. Journal of Political Economy, 88, 829–853.
Holden, K., & Peel, D. A. (1990). On testing for unbiasedness and efficiency of forecasts.
Manchester School, 58, 120–127.
Ito, T. (1990). Foreign exchange rate expectations: Micro survey data. American Economic
Review, 80, 434–449.
Ito, T. (1994). Short-run and long-run expectations of the yen/dollar exchange rate. Journal of the
Japanese and International Economies, 8, 119–143.
Keane, M. P., & Runkle, D. E. (1990). Testing the rationality of price forecasts: New evidence
from panel data. American Economic Review, 80, 714–735.
Kirman, A. P. (1992). Whom or what does the representative individual represent? Journal of
Economic Perspectives, 6, 117–136.
Kiviet, J. F. (1987). Testing linear econometric models. Ph.D. Thesis, University of Amsterdam.
Lai, K. (1990). An evaluation of survey exchange rate forecasts. Economics Letters, 32, 61–65.
Lamont, O. (2002). Macroeconomic forecast and microeconomic forecasters. Journal of Economic
Behavior and Organization, 48, 265–280.
Laster, D., Bennett, P., & Geoum, I. S. (1999). Rational bias in macroeconomic forecasts.
Quarterly Journal of Economics, 293–318.
LeBaron, B. (2000) Technical trading profitability in foreign exchange markets in the 1990’s.
Technical report, Brandeis U.
Leitch, G., & Tanner, J. E. (1991). Economic forecast evaluation: Profits versus the conventional
error measures. The American Economic Review, 81, 580–590.
Liu, P., & Maddala, G. S. (1992). Rationality of survey data and tests for market efficiency in the
foreign exchange markets. Journal of International Money and Finance, 11, 366–381.
Lyons, R. K. (2002). Foreign exchange: Macro puzzles, micro tools. Economic Review, Federal
Reserve Bank of San Francisco, 51–69.
MacDonald, R. (1992). Exchange rate survey data: A disaggregated G-7 perspective. The Man-
chester School, LX(Supplement), 47–62.
Maynard, A., & Phillips, P. C. B. (2001). Rethinking and old empirical puzzle: Econometric
evidence on the forward discount anomaly. Journal of Applied Econometrics, 16, 671–708.
Mincer, J., & Zarnowitz, V. (1969). The evaluation of economic forecasts. In Economic forecasts
and expectations. New York: National Bureau of Economic Research.
Muth, J. (1961). Rational expectations and the theory of price movements. Econometrica, 29,
315–335.
Newey, W., & West, K. (1987). A simple positive semi-definite heteroscedasticity and autocor-
relation consistent covariance matrix. Econometrica, 55, 703–708.
Osterberg, W. P. (2000). New results on the rationality of survey measures of exchange-rate
expectations. Federal Reserve Bank of Cleveland Economic Review, 36, 14–21.
Pesaran, H. M. (2004). General diagnostic tests for cross section dependence in panels. CESifo
working paper series 1229, CE-Sifo Group Munich.
1248 R. Cohen et al.
Contents
44.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1250
44.2 Stochastic Volatility and GARCH . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1252
44.3 Serial Correlation in Returns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1255
44.4 Persistence, Co-Persistence, and Non-Normality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1258
44.4.1 Case 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1260
44.4.2 Case 2 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1260
44.5 Weighted GARCH . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1260
44.6 Empirical Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1261
44.6.1 Index Futures, Market Index, and Stock Price Data . . . . . . . . . . . . . . . . . . . . . . . . . 1261
44.6.2 Estimates of the Autoregressive Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1263
44.6.3 Conditional Variance Estimates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1264
44.6.4 Weighted GARCH Estimates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1267
44.6.5 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1270
44.7 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1273
Appendix . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1274
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1275
Abstract
The behavior of financial asset price data when observed intraday is quite
different from these same processes observed from day to day and longer
sampling intervals. Volatility estimates obtained from intraday observed data
can be badly distorted if anomalies and intraday trading patterns are not
accounted for in the estimation process.
In this paper I consider conditional volatility estimators as special cases of
a general stochastic volatility structure. The theoretical asymptotic distribution
of the measurement error process for these estimators is considered for particular
G.L. Gannon
Deakin University, Burwood, VIC, Australia
e-mail: gerard@deakin.edu.au; gleonon@yahoo.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1249
DOI 10.1007/978-1-4614-7750-1_44,
# Springer Science+Business Media New York 2015
1250 G.L. Gannon
Keywords
ARCH • Asymptotic distribution • Autoregressive parameters • Conditional var-
iance estimates • Constant/dynamic hedging • Excess kurtosis • Index futures •
Intraday returns • Market anomalies • Maximum likelihood estimates • Mis-
specification • Mis-specified returns • Persistence • Serial correlation •
Stochastic volatility • Stock/futures • Unweighted GARCH • Volatility
co-persistence
44.1 Introduction
Nelson and Foster (1994) derive and discuss properties for the ARCH process as the
observation interval reduces to daily records (h!0) when the underlying process is
driven by an assumed continuous diffusion process. Nelson and Foster (1991)
generalized a Markov process with two state variables, hXt and hs2t , only one of
which hXt is ever directly observable. The conditional variance hs2t is defined
conditional on the increments in hXt per unit time and conditional on an information
set hBt. Modifying the notation from h to d (to account for intraday discretely
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1253
observed data), and employing the notation d!0 to indicate reduction in the
observation interval from above, when their assumptions 2, 3, and 10 hold, when
d is small, dXt, j(ds2t )) is referred to as a near diffusion if for any T, 0 T 1,
(dXt, j (ds2t ))0tT ) (Xt, j(s2t ))0tT.
If we assume these data-generating processes are near diffusions, then the
general discrete time stochastic volatility structure, defined in Nelson and Foster
(1991), may be described using the following modified notation:
2 3 2 3 2 3
d Xðkþ1Þd d Xkd mðd Xkd ; skd Þ
6 7 6 7 6 7
4 5 ¼ 4
5 þ d :4 5
j d sð2kþ1Þd j d s2kd lðd Xkd ; d skd Þ
2 2 31 2 3
d skd Lj, x ðd Xkd ; d skd Þ 2 d Z1, kd
6 76 7
þ d1=2 4 54 5
Lj, x ðd Xkd ; d skd Þ L2 ðd Xkd ; d skd Þ Z
d 2, kd
(44:1)
where (dZ1,kd,d Z2,kd)k¼0,1 is i.i.d. with mean zero and identity covariance matrix.
In Eq. 44.1 d is the size of the observation interval, X may describe the asset price
return, and s2 the volatility of the process. It is not necessary to assume the data-
generating processes are stationary or ergodic, but the crucial assumption is that the
data-generating processes are near diffusions.
In the ARCH specification, dZ2,kd is a function of dZ1,kd so that ds2kd(dhkd) can be
inferred from past values of the one observable process dXkd. This is not true for
a general stochastic volatility structure where there are two driving noise terms.
For the first-order Markov ARCH structure, a strictly increasing function of
∧2
estimates d st d h∧
t of the conditional variance process dst (dht) is defined as f(s ),
2 2
∧ mean
and estimates of the conditional
∧ ∧
per∧unit
of time of the increments in X and
f(s2) are defined as m x, s and k x, s . Estimates of ds2kd are updated by the
recursion:
∧2 ∧2 ∧ ∧
f d sðkþ1Þd ¼ s d skd þ d k d Xkd ; d skd
∧
∧2
(44:2)
∧
þ d a d Xkd ; d skd g d Z1, kd ; d Xkd , d skd
1=2
∧ ∧
where k ð:Þ, a ð:Þ, m ð:Þ, and g(.) are continuous on bounded (j(s2), x) sets and
g(z1, x.s2) assumed continuous everywhere with the first three derivatives of g with
respect to z1 well defined and bounded. The function g(dZ1,kd,.) is normalized
to
have mean zero and unit conditional variance. Nonzero drifts in f d s2kd are
∧
allowed for in the k ð:Þ term and non-unit conditional variances accounted for
in the a(.) term. The second term on the right measures the change in f d s2kd
forecast by the ARCH structure while the last term measures the surprise change.
1254 G.L. Gannon
∧
The GARCH(1,1)
2 structure
2 is obtained by setting fðs2 Þ ¼ s2 , k ðx;sÞ ¼ o ys2 ,
gðz1 Þ ¼ z1 1 =SD z1 , and a(x,s) ¼ a . s2 . SD(z21). The parameters o, y, and a
index a family of GARCH(1,1) structures. In a similar manner the EGARCH volatility
∧
structure can be defined. That is, by selectingf ð:Þ, k ð:Þ,gð:Þ, and a(.), various “filters”
can be defined within this framework.
Other conditions in Nelson and Foster (1991) define the rate at which the
normalized measurement error process dQt mean reverts relative to (dXt,ds2t ). It
becomes white noise as d!0 on a standard time scale but operates on a faster
time scale mean-reverting more rapidly. Asymptotically optimal choice of a(.)
and f(.) given g(.) can be considered with respect to minimizing the asymptotic
variance of the measurement error. This is considered on a faster time scale
(T + t) than T. The asymptotically optimal choice of g(.) depends upon the
assumed relationship between Z1 and Z2. In the ARCH structure, Z2 is a function
of Z1 so that the level driving ds2t can be recovered from shocks driving dX2t .
Without further structure in the equation specified for s2t , we are unable to
recover information about changes in s2t . Their discussion is strictly in terms
of constructing a sequence of optimal ARCH filters which minimize the asymp-
totic variance of the asymptotic distribution of the measurement errors. This
approach is not the same as choosing an ARCH structure that minimizes the
measurement error variance for each d.
The asymptotic distribution of the measurement error process, for large t and
small d,
∧2
s
d Tþtd 1=2 d s2
1=2
Q ; s2
;
d T d T d TX ¼ q; s2 ; x
Tþtd
with derivatives evaluated as f0 d s2T , j0 (ds2T), etc. and the notation simplified as
f0 and j0 is approximately normal with mean
h∧ 00
i h ∧
i
ð2s2 j0 Þ k d =j0 a2 j =2ðj0 Þ3 ld =f0 þ L2 f0 3 þ 2 as: md m d :E½gz
d1=2
a : E½ Z1 : g z
(44:3)
and variance
General results in Nelson and Foster (1991, 1994) for the GARCH(1,1)
structure are that GARCH(1,1) can be more accurately measured firstly
the less variable and the smaller is ds2t , second the thinner the tails of Z1, and
third the more the true data-generating mechanism resembles an ARCH
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1255
Consider a first-order autoregressive process for the spot asset price return
(an AR(D) representation):
st ¼ r1 st1 þ et (44:5)
where st is the return on the asset (the difference in the natural logarithm of the spot
asset price levels or the difference in the spot asset price levels) between time t and
t1, r1 is the first-order serial correlation coefficient between s at time t and t1 and
et is an assumed homoskedastic disturbance term, s2e and 1 < r1 < 1. This equation
provides an approximation to the time-series behavior of the spot asset price return.
An alternative specification to Eq. 44.5 is a simple autoregressive equation for
the level of the spot asset price S (or natural logarithm of the levels) at time t and
t1, (an AR(L) representation)
St ¼ r1 St1 þt at (44:6)
f t ¼ at þ y1 at1 , (44:7)
where ft is the index futures price change and at is an assumed mean zero, serially
uncorrelated shock variable with a homoskedastic variance, s2a , and 1 < y1 < 0.
The basis change is defined as
bt ¼ f t i t , (44:8)
where f and i are the index futures return and index portfolio return, respectively.
Whether shocks generated from Eqs. 44.5 or 44.6 generate differences in
parameter estimates and measures of persistence obtained from conditional
second-moment equations is an issue. Measures of persistence for the basis change
may be badly distorted from employing index futures and index portfolio changes.
Measures of persistence for index futures and index portfolios may be badly
distorted by employing observed price changes.
The simplest GARCH structure derived from Eq. 44.1 for the conditional
variance is the GARCH(1,1):
where ht(s2t ) is the conditional variance at time t and e2t1 are squared unconditional
shocks generated from any assumed first-moment equation and 0 a1, b1 1 and
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1257
ht ¼ o 1 þ b1 þ þ bJ1 þ a1 e2t1 þ b1 e2t2 þ þ bJ1 e2tJ1 þ Remainder:
(44:10)
E s2t ¼ E a2t : (44:11)
When Eq. 44.7 is the moving average (MA) representation, the unconditional
variance relative to shocks generated via Eq. 44.11 is
E s2t MA ¼ E 1 þ y2 a2t , (44:12)
and when Eq. 44.5 is the autoregressive (AR) representation, the unconditional
variance relative to shocks generated via Eq. 44.11 is
1 r1 2
E s2t AR ¼ E at : (44:13)
1 þ r1
The conditional variance from a GARCH(1,1) structure for Eq. 44.11 can be
rewritten as
ht ðsÞ ¼ o 1 þ b1 þ þ bJ1 þ a1 a2t1 þ b1 a2t2 þ þ bJ1 a2tJ1
þ Remainder: (44:14)
and if Eq. 44.5 is the representation, then, relative to the conditional variance
equation from Eq. 44.11,
ht ðsÞAR ¼ o 1 þ b1 þ þ bJ1
1 r1 2 1 r1 2 1 r1 2
þ a1 at1 þ b1 at2 þ bJ1 atJ1 þ Remainder:
1 þ r1 1 þ r1 1 þ r1
(44:16)
If o, a1, and b1 were equivalent in Eqs. 44.14, 44.15, and 44.16, when the
conditional variance is driven by Eq. 44.7 with y1 negative, then ht(s)MA > ht(s),
and when the conditional variance is driven by Eq. 44.5 with r1 negative, then
ht(s)AR > ht(s) and with r1 positive then ht(s)AR < ht(s). However, given the scaling
factor in Eq. 44.15 relative to Eq. 44.16, the potential for distortions to GARCH
parameter estimates is greater when the underlying process is driven by Eq. 44.5
relative to Eq. 44.7.
it ¼ g1 xt þ it
(44:18)
f t ¼ g2 xt þ ft ,
the common “news” factor xt is IGARCH, in the co-persistence structure, while the
idiosyncratic parts are assumed jointly independent and independent of xt and not
IGARCH. The individual processes have infinite unconditional variance. If a linear
combination is not IGARCH, then the unconditional variance of the linear combi-
nation is finite and a constant hedge ratio (defined below) leads to substantial
reduction in portfolio risk.
A time-varying hedge ratio can lead to greater reduction in portfolio risk under
conditions discussed in Ghose and Kroner (1994) when the processes are
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1259
it ¼ g1 xi t þ i t
(44:19)
f t ¼ g 2 xf t þ f t
then the estimated constant hedge ratio which is short g units of the spot for every
1 unit long in the futures is
" # 2 2 qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi3 3
∧ ∧ ∧ ∧
∧ cov ðf; iÞ ∧ p var ½xf var ½xi
g¼ ∧
¼ 4g 2 4 ∧
5 þ var Z i 5 (44:20)
var ðiÞ jt var ½xi
jt
∧
where r is the correlation between xft and xit.
When both xf and xi follow IGARCH processes and no common factor structure
exists, then the estimated constant hedge diverges. Ghose and Kroner (1994)
investigate this case.
When xft follows an IGARCH process but xit is weak GARCH, then the
estimated constant hedge ratio cannot be evaluated. There are two problems:
(a) The estimated sample variance of xf in Eq. 44.20 is infinite as T ! 1.
∧ pffiffiffi ∧ ∧
(b) p ¼ c ov xf = v a r xf var ½xi so that there is no linear combination
jt
of xf and xi which can provide a stationary unconditional variance.
This last observation has a direct parallel from the literature for cointegration in
the means of two series.
If xf is an approximate I(1) process and xi is I(0), then there is no definable linear
combination of xf and xi.
When observing spot index and futures prices over successively finer intervals,
the co-persistence structure may not hold for at least two further reasons. This
argument relates directly to the horizon t+kd for the hedging strategy. This argu-
ment also relates directly to distortions possibly induced onto a dynamic hedging
strategy, as d!0.
Perverse behavior can be observed in spot index level changes as oversampling
becomes severe. The smoothing effect due to a large proportion of the portfolio
entering the non- and thin-trading group generates a smoothly evolving process
with short-lived shocks generated by irregular news effects.
1260 G.L. Gannon
General results assume that the z0ts are drawn from a continuous distribution.
When sampling futures price data at high frequency, then the discrete nature of the
price recording mechanism guarantees that there are discontinuities in return-
generating processes. The distribution of the z0ts can become extremely peaked due
to multiple small price changes and can have very long thin tails due to abrupt shifts
in the distribution. As d ! 0, 1 < E[1n(b1 + a1zti 2
)] < + 1 for a large range of
values for 0 a1, b1 1 and a1 + b1 < 1, and this depends on the distribution
induced by oversampling and resultant reduction in (a1 + b1). In the limit E[ (zt)4]/[E
(z2t )]2 ! 1. Then even-numbered higher moments of zt are unbounded as d!0. This
oversampling can lead to two extreme perverse effects generated by bid/ask bounce
or zero price changes. The effect depends upon liquidity in the respective markets.
44.4.1 Case 1
a1 ! 1, a1 þ b1 > 1 and E ln b1 þ a1 z2ti > 0:
44.4.2 Case 2
Oversampling can produce many zero price changes in thin markets. In this latter
case as d!0 then a1 + b1 ! 0.
This explanation can apply to relatively illiquid futures (and spot asset) price
changes. That is, conditional heteroskedasticity disappears as oversampling
becomes severe.
The effect on the basis change when there is a relatively illiquid futures market
and oversampling may be badly distorted. As well, failure to account for anomalies
in conditional variance equations can severely distort estimates.
Foster (1991, 1994), this diffusion limit is l/j0 + L2j00 /2 (j0 )3. The effect on the
expression for the bias in the asymptotic distribution of the measurement error
00 00
process can be explosive if derivatives in the terms a2 f =2ðf0 Þ3 L2 j =2ðj0 Þ3
cannot be evaluated because of discontinuities in the process. This can happen when
important intraday effects are neglected in the conditional variance equation speci-
fication. As well, the bias can diverge as d!0 if the dZ1,kd terms are badly distorted.
Occasional large jumps in the underlying distribution contribute large Op(1) move-
ments while the near diffusion components contribute small Op(d1/2) increments. When
sampling intraday financial data, there are often many small price changes which tend
to be dominated by occasional large shifts in the underlying distribution.
Failure to account for intraday effects (large shocks to the underlying distribu-
tion) can lead to a mixture of Op(1) and Op(d1/2) effects in the process. One
approach is to specify this mixed process as a jump diffusion. An alternative is to
account for these effects by incorporating activity measures in the specification of
conditional variance equations.
It follows that failure
to account
for these Op(1) effects can lead to an explosive
∧
measurement error d h t d ht . However, in empirical applications this measure-
ment error is unobservable since dht is unobservable. Failure to account for these
jumps in the underlying distribution imply that the unweighted GARCH(1,1)
structure cannot satisfy the necessary assumptions required to approximate
a diffusion limit.
The data has been analyzed for the common trading hours for 1992 from the ASX
and SFE, i.e., from 10.00 a.m. to 12.30 p.m. and from 2.00 p.m. to 4.00 p.m.
1262 G.L. Gannon
This dataset was first employed in examples reported in Gannon (1994) and
a similar analysis undertaken in Gannon (2010) for the SVL models as is
undertaken in this paper for the GARCH and GARCH-W models. Further details
of the sampling and institutional rules operating in these markets are
briefly reported in the Appendix of this paper. Transactions for the Share Price
Index futures (SPI) were sampled from the nearest contract 3 months to expira-
tion. The last traded price on the SPI levels and stock price levels were then
recorded for each observation interval. The All Ordinaries Index (AOI) is the
recorded level at the end of each observation interval. During these common
trading hours, the daily average number of SPI futures contracts traded 3 months
to expiration was 816. As well, block trades were extremely rare in this series.
Transactions on the heavily capitalized stock prices were extremely dense within
the trading day.
These seven stocks were chosen from the four largest market capitalized group-
ings according to the ASX classification code in November 1991, i.e., general
industrial, banking, manufacturing, and mining. The largest capitalized stocks
were chosen from the first three categories as well as the four largest capitalized
mining stocks. This selection provides for a diversified portfolio of very actively
traded stocks which comprised 32.06 % of total company weights from the
300 stocks comprising the AOI.
All datasets were carefully edited in order to exclude periods where the
transaction capturing broke down. The incidence of this was rare. As well, lags
were generated and therefore the effects of overnight records removed. A natural
logarithmic transformation of the SPI and AOI prices is undertaken prior to
analysis.
Opening market activity for the SPI is heaviest during the first 40 min of
trading. Trade in the SPI commences at 9.50 a.m. but from 10.30 a.m. onwards
volume of trade tapers off until the lunchtime close. During the afternoon
session, there is a gradual increase in volume of trade towards daily market
close at 4.10 p.m. Excluding the market opening provides the familiar U-shaped
pattern of intraday trading volume observed on other futures markets. SPI price
volatility is highest during the market opening period with two apparent reverse
J-shaped patterns for the two daily trading sessions (small J effect in afternoon
session). The first and last 10 min of trade in the SPI are excluded from this
dataset.
Special features govern the sequence at which stocks open for trade on the
ASX. Individual stocks are allocated a random opening time to within plus or
minus 30 s of a fixed opening time. Four fixed opening times, separated by 3-min
intervals starting at 10.00 a.m., operated throughout 1992. Four alphabetically
ordered groups then separately opened within the first 10 min of trading time.
The last minute of trading on the ASX is also subject to a random closing time
between 3.59 p.m. and 4.00 p.m. The effect of these institutional procedures on
observed data series can be potentially severe.
Both of these activity effects in market opening prices of trading on the SFE, and
ASX should be accounted for in the estimation process.
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1263
Table 44.1 Autoregressive parameter estimates for SPI, AOI, and basis change
Interval ft it bt
Full day
30 min 0.0324 0.0909 0.1699
(1.54) (4.33) (8.19)
15 min 0.0497 0.0953 0.1348
(3.45) (6.42) (9.15)
05 min 0.0441 0.3317 0.0137
(5.14) (41.2) (1.60)
Excluding market open
30 min 0.0299 0.1213 0.1947
(1.34) (5.48) (8.89)
15 min 0.0301 0.2531 0.2145
(1.91) (16.6) (13.9)
05 min 0.0532 0.3284 0.0968
(5.84) (38.2) (10.7)
Asymptotic t-statistics in brackets
ft ¼ Ft – Ft is the difference in the observed log level of the SPI
it ¼ It – I is the difference in the observed log level of the AOI
Equation 44.5, i.e., an autoregressive specification for the differences, an [AR(D)], is estimated for
both data series with one lag only for the autoregressive parameter
Dummy variables are included, for the first data series, in order to account for market opening
effects for the SPI, institutional features governing market opening on the ASX and therefore
effects transmitted to the basis change. Two separate dummy variables are included for the first
pair of 5-min intervals
This form for the basis change is bt ¼ ft – it
In the lower panel, results are reported from synchronized trading from 10.30 a.m. That is, the first
40 min and first 30 min of normal trade in the SPI and AOI, respectively, is excluded
In Table 44.1 the first-order autoregressive parameter estimate is reported for the
observed differenced series for the SPI (f), AOI (i), and basis change (b). In the top
panel, these parameter estimates are from observations for the full (synchronized)
trading day. These equations include dummy variables to account for institutional
market opening effects. In the following tables of results, SING refers to singular-
ities in the estimation process.
For the SPI futures price process, low first-order negative serial correlation is
initially detected in the log of the price change. As the sampling interval is reduced,
low first-order positive serial correlation can be detected in the series. This feature
of the data accords with order splitting and non-trading-induced effects.
Low positive first-order serial correlation can be detected in differences of the
log of the market index and serial correlation increases. Positive serial correlation is
high at 15-min intervals for the opening excluded set. When sampling the market
index at 5-min intervals, substantial positive first-order serial correlation is detected
1264 G.L. Gannon
in the log of the price change process. Miller et al. (1994) demonstrate that thin
trading and non-trading in individual stocks induce a positive serial correlation in
the observed spot index price change process. However, the smoothing effect from
non-trading in individual stocks, averaging bid/ask bounce effects in heavily traded
stocks, and under-differencing the aggregate process also contribute.
Of interest is the reduction in serial correlation of the basis spread as d ! 0. As
the log futures price change moves into the order splitting/non-price change region,
positive serial correlation is induced in the log of the futures price change. This, in
part, helps offset the increasing positive serial correlation induced in the log of the
spot index.
Table 44.2 GARCH estimates for the SPI (Ft), AOI (It), and basis change (Bt) for AR(L) and
AR(D) specifications
Ft It Bt
Full day
30 min
a1 0.0527 0.0517 0.1019 0.1002 0.0904 0.0741
(7.84) (7.58) (7.23) (6.91) (9.04) (7.70)
b1 0.9266 0.9284 0.0196 0.0227 0.0204 0.0251
(115) (111) (2.09) (2.24) (2.06) (2.31)
15 min
a1 0.0602 0.0592 0.1922 0.1514 0.1050 0.1000
(14.7) (14.5) (13.2) (11.4) (13.6) (13.6)
b1 0.9151 0.9167 0.0893 0.1100 0.0852 0.1186
(224) (225) (9.89) (10.9) (5.45) (7.17)
05 min
a1 0.0358 0.0362 0.3540 0.2338 0.1705 0.1685
(34.0) (34.6) (35.8) (27.9) (39.6) (38.5)
b1 0.9535 0.9530 0.2074 0.2428 0.4345 0.4469
(1075) (1087) (26.8) (29.3) (43.3) (45.6)
Excluding market open
30 min
a1 0.0448 0.0444 0.0668 0.0693 0.0684 0.0631
(8.31) (8.12) (8.71) (9.22) (11.5) (10.5)
b1 0.9473 0.9478 0.9074 0.9050 0.9114 0.9172
(155) (150) (92.9) (93.3) (114) (109)
15 min
a1 0.0385 0.0394 0.0867 0.0835 0.0310 0.0403
(14.8) (14.5) (12.9) (13.2) (12.9) (18.8)
b1 0.9551 0.9564 0.8832 0.8872 0.9499 0.9537
(382) (369) (113) (115) (232) (374)
05 min
a1 5.8E-5 0.0329 0.3326 0.2229 4.6E-5 0.0236
(33.7) (30.3) (37.5) (33.6) (27.4) (35.9)
b1 1.000 0.9638 0.4470 0.5795 1.000 0.9742
SING (906) (35.2) (46.4) SING (1353)
Asymptotis t-statistics in brackets
These preliminary results have been obtained from observations sampled for all
four futures contracts and a continuous series constructed for 1992. The AOI is not
affected by contract expiration. In Table 44.3, GARCH(1,1) estimates for the
separate SPI futures contracts, 3 months to expiration, and synchronously sampled
observations on the AOI are recorded. The full data series corresponding to
synchronized trading on the SFE and ASX is employed.
The same form of dummy variable set was imposed in both first and second-
moment equations. As well, a post-lunchtime dummy is included to account for the
1266 G.L. Gannon
Table 44.3 GARCH estimates for the SPI and AOI 3 months to expiration
MAR JUN SEP DEC
Ft It Ft It Ft It Ft It
30 min
a1 0.042 0.026 0.145 0.015 0.019 0.089 0.060 0.147
(2.37) (0.94) (3.95) (0.76) (2.96) (3.65) (3.24) (3.70)
b1 0.933 0.054 0.116 0.004 0.977 0.021 0.918 0.037
(35.6) (2.05) (0.56) (0.21) (118) (1.12) (35.0) (1.30)
15 min
a1 0.028 0.186 SING 0.107 0.037 0.183 0.036 0.195
(3.11) (5.85) (4.75) (4.96) (6.84) (6.08) (5.80)
b1 0.958 0.108 0.998 0.035 0.956 0.055 0.958 0.102
(72.6) (6.65) (3235) (2.11) (120) (3.75) (114) (3.67)
05 min
a1 0.051 0.240 SING 0.307 0.174 0.378 0.039 0.409
(7.86) (12.4) (14.8) (22.2) (17.0) (11.0) (19.3)
b1 0.881 0.231 0.999 0.118 0.649 0.161 0.956 0.259
(52.9) (12.1) (21170) (7.16) (38.5) (10.6) (247) (16.9)
GARCH(1,1) parameter estimates for the 3 months corresponding to expiration of the March,
June, September, and December contracts for 1992. Full-day data series are employed with
opening and post-lunchtime dummy variables in both first and conditional second-moment
equations. An AR(L) specification of the (log) mean equation is employed for these results
break in daily market trade at the SFE during 1992. The log of levels is specified for
the first-moment equations.
There is some instability within this set of parameter estimates. However,
a similar pattern emerges within the set of futures and the set of market index
estimates as was observed for the full-day series for 1992. The futures conditional
variance parameter estimates are close to the IGARCH boundary while the index
conditional variance estimates are not. This again implies that these processes
cannot be co-persistent in variance for these samples and observation intervals.
In order to obtain further insight into the, seemingly, perverse results for the
market index, a similar analysis was undertaken on seven of the largest market
capitalized stocks which comprised the AOI during 1992.
Relevant market opening and closing dummy variables were included in both
first and conditional second-moment equations accordingly. These effects were not
systematic in the first-moment equations and are not reported. The stock price
processes have not been transformed to natural logarithmic form for these estima-
tions. This is because the weighted levels of the stock prices are employed in
construction of the market index.
As the observation interval is reduced for these stock prices:
(i) The autoregressive parameter estimates for the price levels equation
converges to a unit root.
(ii) The first-order serial correlation coefficient for the price difference equation
moves progressively into the negative region.
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1267
If these were the only stocks comprising the construction of the index, then we
might expect to see increasing negative serial correlation in the index. But this
observation would only apply if these processes were sampled from a continuous
process which was generated by a particular form of ARMA specification. Only in
the case of data observed from a continuous process could the results on temporal
aggregation of ARMA processes be applied.
These results should not be surprising as the combination of infrequent price
change and “price bounce” from bid/ask effects starts to dominate the time series.
As well, these bid/ask boundaries can shift up and down. When these processes are
aggregated, the effects of price bounce can cancel out. As well, the smoothing
effect of thinly and zero traded stocks within the observation intervals dampens and
offsets individual negative serial correlation observed in these heavily traded stocks
(Table 44.4).
Some of these autoregressive parameter estimates are quite high for the
AR(D) specifications. However, it is apparent that there is almost no difference in
GARCH(1,1) estimates from either the AR(L) or AR(D) specifications at each
observation interval for any stock price. In some instances estimation breaks
down at 5 min intervals.
If the conditional variances of these stock price movements contain common
news and announcement effects, then it should not be surprising that the
weighted aggregated process is not persistent in variance. This can happen
when news affects all stocks in the same market. As well, smoothing effects
from thin-traded stocks help dampen volatility shocks observed in heavily traded
stocks. These news and announcement effects may be irrelevant when observing
these same processes at daily market open to open or close to close. These news
and announcement effects may be due to private information filtering onto the
market prior to and following market open. It is during this period that overnight
information effects can be observed in both price volatility and volume. As well,
day traders and noise traders are setting positions. However, the ad hoc applica-
tion of dummy variables is not sufficient to capture the interaction between
volatility and volume. In the absence of specific measures of these news “vari-
ables,” the effects cannot be directly incorporated into a structural model.
However, these effects are often captured in the price volatility and reflected in
increased trading activity.
Weighted GARCH estimates for the futures (log) price process with the accumu-
lated number of futures contracts traded within the interval t to t1 are reported in
Table 44.5. This choice ensures that volume measures are recorded within the
interval that actual prices define.
The weighting variable employed in the index (log) price process is the squared
shock from the futures price mean equation within the interval t to t1. There is no
natural “volume” of trade variable available for the market index. The form of mean
1268 G.L. Gannon
Table 44.4 Unconditional mean and GARCH(1,1) estimates: Australian stock prices
AR(L) AR(D) AR(L) AR(D)
Interval p1 r1 a1 b1 a1 b1
NAB
60 min 0.999 0.160 0.125 0.762 0.122 0.774
(17.2) (39.2) (16.4) (40.1)
30 min 0.999 0.135 0.086 0.862 0.081 0.876
(25.3) (123) (24.4) (135)
15 min 0.999 0.199 0.226 0.627 0.220 0.629
(22.5) (70.0) (22.2) (61.1)
05 min 1.00 0.287 2.4E-5 0.042 2.3E-5 0.049
(0.94) (0.28) (0.53) (0.24)
BHP
60 min 1.00 0.034 0.096 0.843 0.098 0.839
(9.26) (45.6) (9.14) (44.2)
30 min 1.00 0.087 0.067 0.898 0.066 0.899
(14.2) (129) (13.7) (128)
15 min 1.00 0.104 0.054 0.923 0.054 0.924
(20.5) (281) (20.2) (279)
05 min 1.00 0.123 0.132 0.795 0.130 0.799
(66.0) (361) (61.9) (363)
BTR
60 min 0.995 0.035 0.194 SING 0.196 SING
(9.68) (SING) (9.68) (SING)
30 min 0.998 0.109 0.232 0.294 0.231 0.297
(14.3) (7.36) (13.9) (7.55)
15 min 0.999 0.109 0.136 0.615 0.131 0.626
(17.6) (31.5) (17.5) (32.8)
05 min 1.00 0.074 0.076 0.831 0.074 0.835
(45.9) (231) (45.5) (233)
WMC
60 min 1.00 0.052 0.015 0.981 0.015 0.981
(6.58) (368) (6.54) (363)
30 min 1.00 0.016 0.272 0.234 0.265 0.251
(13.7) (5.39) (13.70 (5.85)
15 min 1.00 0.008 0.185 0.594 0.186 0.597
(19.6) (33.7) (19.3) (34.0)
05 min 1.00 0.065 0.013 0.003 SING SING
CRA
60 min 0.999 0.031 0.332 0.058 0.333 0.074
(13.0) (1.52) (12.5) (1.88)
30 min 0.999 0.008 0.231 0.456 0.231 0.458
(10.2) (5.89) (9.98) (5.69)
15 min 1.00 0.003 0.132 0.687 0.132 0.688
(24.0) (58.8) (23.8) (58.5)
(continued)
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1269
equation is the same as the generated corresponding results for Table 44.3, i.e., an
AR(L). However, the results are almost identical when alternative forms for the
mean equation are employed, i.e., AR(L) and AR(D).
The specifications generating the reported stock price estimates are augmented
to include a measure of trade activity within the observation interval. These
measures are the accumulated number of stocks traded in each individual stock
within the interval t to t1. The estimates are reported in Table 44.6. The condi-
tional variance parameter estimates are almost identical from the AR(L) and
AR(D) specifications of the mean equation. The logarithmic transformation has
not been taken for these stock prices.
Direct comparison of these GARCH parameter estimates (a1 and b1) with those
from the unweighted GARCH(1,1) estimates demonstrates the importance of this
measure of activity. The change in GARCH parameter estimates is striking.
1270 G.L. Gannon
Table 44.5 Weighted GARCH estimates for the SPI and AOI 3 months to expiration
March June September December
Ft It Ft It Ft It Ft It
30 min
a1 0.093 0.040 0.134 0.027 SING 0.079 0.030 0.007
(2.07) (1.51) (4.58) (1.10) (2.73) (1.01) (0.40)
b1 0.014 0.039 SING 0.068 SING 0.033 SING 0.025
(0.24) (1.62) (2.74) (1.91) (1.24)
15 min
a1 0.052 0.185 0.175 0.022 0.039 0.042 0.076 0.058
(1.84) (3.97) (6.67) (1.50) (2.47) (2.35) (2.89) (3.04)
b1 0.004 0.086 0.000 0.163 0.000 0.106 0.001 0.136
(0.19) (3.82) (0.00) (7.61) (0.00) (4.99) (0.01) (5.67)
05 min
a1 0.000 0.156 0.095 0.170 0.004 0.162 0.030 0.128
(4.04) (8.34) (7.79) (9.55) (32.1) (9.58) (8.32) (7.30)
b1 0.000 0.238 SING 0.232 0.000 0.270 0.000 0.327
(5.06) (11.9) (11.8) (0.12) (16.0) (7.79) (17.8)
The measures of persistence from these weighted estimates are never near the
IGARCH boundary. These effects are summarized in Table 44.7.
These results are generated from an AR(L) specification of the first-moment
equation. By adequately accounting for contemporaneous intraday market activity,
the time persistence of volatility shocks becomes less relevant. It follows that
deviations of the estimated conditional variance from the true (unobservable)
conditional variance are reduced (Table 44.8).
44.6.5 Discussion
3. The basis change between the log futures price change and the log index level
change displayed two surprising effects:
(i) The autoregressive parameter for the basis change is initially negative, but
the strength of this effect weakens. This can be attributed to the log of the
Table 44.7 Measures of persistence from GARCH(1,1) and weighted GARCH equations for the
SPI 3 months to expiration
G G-W G G-W G G-W G G-W
Interval March June September December
30 min 0.975 0.107 0.261 SING 0.996 SING 0.978 SING
15 min 0.986 0.056 SING 0.175 0.993 0.039 0.994 0.077
05 min 0.932 0.000 SING SING 0.823 0.004 0.995 0.030
Table 44.8 Measures of persistence from GARCH(1,1) and weighted GARCH equations for
Australian stock price processes
NAB BHP BTR WMC
60 min 0.887 0.175 0.939 0.178 SING 0.069 0.996 0.150
30 min 0.948 0.311 0.965 0.241 0.526 0.167 0.506 0.115
15 min 0.853 0.275 0.977 0.242 0.751 0.093 0.779 0.131
CRA MIM CML
60 min 0.390 0.136 0.995 SING 0.995 SING
30 min 0.687 0.079 0.445 0.145 0.990 0.137
15 min 0.819 0.000 0.763 0.108 0.843 0.056
An AR(L) specification has been employed to generate these results
The measure of persistence is the calculated as a+b from the conditional variance equations
G represents persistence obtained from a GARCH(1,1) specification
G-W represents persistence obtained from a weighted GARCH specification
SING indicates that one or more of these GARCH parameters could not be evaluated due to
singularities
The GARCH(1,1) parameter estimates for the AOI at 5-min intervals were
different when alternative forms of first-moment equations were specified. There
was no perceivable difference in the other processes at this sampling frequency. It
would then appear that increasing positive serial correlation (smoothing) in the
observed returns process has a greater distorting effect on GARCH(1,1) parameter
estimates than increasing negative serial correlation (oscillation) in observed
returns processes. The most important effect is mis-specification of the conditional
variance equations from failure to adequately account for the interaction between
market activity and conditional variance (volatility).
Some implications of these results are:
7. When aggregating stock prices, which may be close to the IGARCH boundary,
persistence in variance can be low for the market index because (i) there is
common persistence present in heavily traded stock prices which when aggre-
gated do not display persistence in variance because heavily traded stock prices
react to market specific news instantaneously, (ii) the smoothing effect of less
heavily traded stocks dampens the volatility clustering which is often observed
in other financial assets such as exchange rates, and (iii) high volatility and
volume of trade effects within these markets following opening is better mea-
sured by employing relevant measures of activity than an ad hoc approach.
8. There is a strong and quantifiable relationship between activity in these markets
and volatility.
44.7 Conclusion
The behavior of financial asset price data observed intraday is quite different from
these data observed at longer sampling intervals such as day to day. Market
anomalies which distort intraday observed data mean that volatility estimates
1274 G.L. Gannon
obtained from data observed from day-to-day trades will provide different volatility
estimates. This latter feature then depends upon when the data is sampled within the
trading day. If these anomalies and intraday trading patterns are accounted for in the
estimation process, then better estimates of volatility are obtainable by employing
intraday observed data. However, this is dependent on a sampling interval that is
not so fine that these estimators break down.
The specific results indicate that serial correlation in returns processes can
distort parameter estimates obtainable from GARCH estimators. However,
induced excess kurtosis may be a more important factor in distortions to esti-
mates. The most important factor is mis-specification of the conditional variance
(GARCH) equation from omission of relevant variables which explain the
anomalies and trading patterns observed in intraday data. Measures of activity
do help explain systematic shifts in the underlying returns distribution and in
this way help explain “jumps” in the volatility process. This effect can be
observed in the likelihood function and in asymptotic standard errors of
weighting (mixing) variables. One feature is that the measure of volatility
persistence observed in unweighted univariate volatility estimators is reduced
substantially with inclusion of weighting variables.
Appendix
At the time the ASX data were collected, the exchange had just previously moved
from floor to screen trading with the six main capital city exchanges linked via
satellite and trade data streamed to trading houses and brokers instantaneously via
a signal G feed. The SFE maintained Pit trading for all futures and options on
futures contracts at the time.
Legal restrictions on third party use and development of interfaces meant the
ASX had a moratorium on such usage and development. The author was required to
obtain special permission from the ASX to capture trade data from a live feed from
broking house Burdett, Buckeridge, and Young (BBY). There was a further delay in
reporting results of research following the legal agreement obtained from the ASX.
Trade data for stock prices and volume of trade were then sampled into 5-min files
and subsequently into longer sampling interval files. The market index was refreshed
at 1-min intervals and the above sampling scheme repeated. Futures price trades were
supplied in two formats via feed: Pit (voice recorded) data and Chit data. Although
the Pit data provides an instantaneous record of trade data during the trading day,
some trades are lost during frantic periods of activity. The Chit records are of every
trade (price, volume, buyer, seller, time stamped to the nearest second, etc.).
The recorded Chits are placed in a wire basket on a carriageway and transferred up
the catwalk where recorders on computers enter details via a set of simplified
keystrokes. The average delay from trade to recording is around 30 s for the Chit
trades. These are then fed online to trading houses and brokers. At the end of the
trading day, these recorded trades are supplemented with a smaller set of records that
44 Stochastic Volatility Structures and Intraday Asset Price Dynamics 1275
were submitted to the catwalk late, e.g., morning trades that may have gone to lunch
in a brokers pocket and submitted during the afternoon session and also some late
submitted trades.
We created the intraday sampled files from both the Pit and Chit records.
However, we employed the Chit trades for analysis in this paper so as to have the
correct volume of trade details for each trading interval. All trades were reallocated
using the time stamps to the relevant time of trade, including trades not submitted
on time but supplied as an appendix to the trading day data. In this study the average
number of late Chits were not a high proportion of daily trades. These futures price
records were then sampled into relevant 5-min records and longer sampling frames
generated in the same manner as was employed for the stock prices.
For all series the first price and last price closest to the opening and closing nodes
for each sampling interval were recorded with volume of trade the accumulated
volume of trade within the interval defined by first and last trade.
References
Brown, S. (1990). Estimating volatility. In S. Figlewski et al. (Eds.), Financial options: From
theory to practice (pp. 516–537). Homewood: Business One Irwin.
Duffee, G., Kupiec, P., & White, A. P. (1990). A primer on program trading and stock price
volatility: A survey of the issues and evidence (Working Paper No. 109). FEDS Board of
Governors of the Federal Reserve System.
Edwards, F. R. (1988). Futures trading and cash market volatility: Stock index and interest rate
futures. Journal of Futures Markets, 8, 421–439.
Gannon, G. L. (1994). Simultaneous volatility effects in index futures. Review of Futures Markets,
13, 1027–1066.
Gannon, G. L. (2010). Simultaneous volatility transmissions and spillovers: Theory and evidence.
Review of Pacific Basin Financial Markets and Policies, 13, 127–156.
Ghose, D., & Kroner, K. F. (1994). Common persistence in conditional variances: Implications for
optimal hedging. Paper presented at the 1994 Meeting of the Australasian Econometric
Society.
McCurdy, T., & Morgan, I. G. (1987). Tests of the martingale hypothesis for foreign currency
futures with time varying volatility. International Journal of Forecasting, 3, 131–148.
Miller, M. H., Muthuswamy, J., & Whaley, R. E. (1994). Mean reversion of standard and poor’s
500 index basis changes: Arbitrage-induced or statistical illusion? Journal of Finance, 49,
479–513.
Nelson, D. B. (1990a). ARCH models as diffusion approximations. Journal of Econometrics, 45, 7–38.
Nelson, D. B. (1990b). Stationarity and persistence in the GARCH(1,1) model. Econometric
Reviews, 6, 318–334.
Nelson, D. B., & Foster, D. P. (1991). Estimating conditional variances with misspecified ARCH
models: Asymptotic theory. Graduate School of Business, University of Chicago, mimeo.
Nelson, D. B., & Foster, D. P. (1994). Asymptotic filtering theory for univariate ARCH models.
Econometrica, 62, 1–41.
Optimal Asset Allocation Under VaR
Criterion: Taiwan Stock Market 45
Ken Hung and Suresh Srivastava
Contents
45.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1278
45.2 Value at Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1279
45.3 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1281
45.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1288
Appendix 1: Value at Risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1288
Appendix 2: Optimal Portfolio Under a VaR Constraint . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1289
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1290
Abstract
Value at risk (VaR) measures the worst expected loss over a given time horizon
under normal market conditions at a specific level of confidence. These days,
VaR is the benchmark for measuring, monitoring, and controlling downside
financial risk. VaR is determined by the left tail of the cumulative probability
distribution of expected returns. Expected probability distribution can be
generated assuming normal distribution, historical simulation, or Monte
Carlo simulation. Further, a VaR-efficient frontier is constructed, and an
asset allocation model subject to a target VaR constraint is examined.
This paper examines the riskiness of the Taiwan stock market by determining
the VaR from the expected return distribution generated by historical simulation.
Our result indicates the cumulative probability distribution has a fatter left tail,
compared with the left tail of a normal distribution. This implies a riskier market.
K. Hung (*)
Division of International Banking & Finance Studies, Texas A&M International University,
Laredo, TX, USA
e-mail: ken.hung@tamiu.edu
S. Srivastava
University of Alaska Anchorage, Anchorage, AK, USA
e-mail: afscs@uaa.alaska.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1277
DOI 10.1007/978-1-4614-7750-1_45,
# Springer Science+Business Media New York 2015
1278 K. Hung and S. Srivastava
Keywords
Value at risk • Asset allocation • Cumulative probability distribution • Normal
distribution • VaR-efficient frontier • Historical simulation • Expected return
distribution • Two-sector asset allocation model • Delta • Gamma • Corner
portfolio • TAIEX
45.1 Introduction
1
Extensive discussion of value at risk can be found in Basak and Shapiro (2001), Beder (1995),
Dowd (1998), Fong and Vasicek (1997), Hendricks (1996), Hoppe (1999), Jorion (1997, 1997),
Schachter (1998), Smithson and Minton (1996a, b), and Talmor (1996).
45 Optimal Asset Allocation Under VaR Criterion: Taiwan Stock Market 1279
2
Institutional use of VaR can be found in Basel (1995, 1998a, b, c, 1999), Danielsson et al. (1998),
and Danielsson Hartmann and de Vries (1998).
1280 K. Hung and S. Srivastava
simulation is its sensitivity to sample data. Many scholars pointed out that if
October 1987 is included into the observation period, then it would make great
difference to the estimation of VaR. Another problem with historical simulation is
that the left tail of actual return distribution is at zero stock prices. In other words, it
would not be accurate to assume a zero probability of loss that is greater than the
past loss. Lastly, the estimation of historical simulation is more complicated than
that of variance-covariance analysis. The VaR needs to be reestimated every time
the level of reliability or holding period changes.
Monte Carlo simulation can be used to generate future return distribution
for a wide range of financial products. It is done in two steps. First, a stochastic
process is specified for each financial variable along with appropriate parameters.
Second, simulated prices are determined for each variable, and portfolio loss is
calculated. This process is repeated 1,000 times to produce a probability distribu-
tion of losses. Monte Carlo simulation is the most powerful tool for generating
the entire probability distribution function and can be used to calculate VaR
for a wide range of financial products. However, it is time consuming and
expensive to implement.
Modern investment portfolio theories try to achieve optimal asset allocation via
maximizing the risk premium per unit risk, also known as the Sharpe ratio (Elton
and Gruber 1995). Within the framework of mean-variance, market risk is defined
as the expected probable variance of investment portfolio. To estimate risk with
standard deviation implies investors pay the same attention to the probabilities of
negative and positive returns. Yet investors have different aversion to investment’s
downside risk than to capital appreciation. Some investors may use semi-variance
to estimate the downside risk of investment. However, semi-variance has not
become popular.
Campbell et al. (2001) have developed an asset allocation model that takes VaR
as one of its constraints. This model takes the maximum expected loss preset by
risk managers (VaR) as a constraint to maximize expected return. In other words,
the optimal investment portfolio deduced from this model meets the constraint of
VaR. This model is similar to the mean-variance model that generates
the Sharpe index. If the expected return is a normal distribution, then this model
is identical with mean-variance model. Details of this model are presented in
the Appendix.
Other researchers have examined four models to introduce VaR for ex-ante
asset allocation of optimal investment portfolio: mean-variance (MV) model,
mini-max (MM) model, scenario-based stochastic programming (SP) model, and
a model that combines stochastic programming and aggregation/convergence
(SP-A). The investment portfolio constructed using the SP-A model has a higher
return in all empirical and simulation tests. Robustness test indicates that VaR
strategy results in higher risk tolerance than risk assessment that takes severe loss
into consideration. Basak and Shapiro (2001) pointed out that the drawback of risk
management lies in its focus on loss probability instead of loss severity. Although
loss probability is a constant, when severe loss occurs, it has greater negative
consequence than non-VaR risk management.
45 Optimal Asset Allocation Under VaR Criterion: Taiwan Stock Market 1281
Lucas and Klaassen (1998) pointed out the importance of correctly assessing the
fat-tailed nature of return distribution. If the real return is not in normal distribution,
the asset allocation under the hypothesis of normal distribution will result in
non-efficiency or non-feasibility. An excellent discussion of VaR and risk
measurements is presented by Jorion (2001).
In July 1997, financial crisis broke out in Southeast Asian nations and then prolifer-
ated to other Asian regions and led to a series of economic problems. Taiwan had also
been attacked by financial crisis in late 1998. Domestic financial markets fluctuated.
Corporations and individuals greatly suffered. The proliferation of financial crisis
within or among countries makes it impossible for corporations and individuals to
ignore market risk. Risk measurement is the first thing to do before investment.
The Taiwan market general weighted stock index, individual weighted stock price
indexes, and interbank short loan interest rate used in this research paper are obtained
from the data base of AREMOS. We divide the historical period into two groups,
from 1980 to 1999 and from 1991 to 1999, so as to analyze the impact of violent stock
market fluctuation on VaR estimation, such as the New York stock market collapse in
October 1987 and Taiwan stock market dramatic uprising from 1988 to 1990. The
first period is rather long and can indicate the nature of dramatic stock fluctuation.
The second period is rather short and can reflect the change of stock market tendency.
This paper employs historical simulation to reproduce the daily fluctuations of
returns for electrical machinery, cement, food, pulp and paper, plastics and
petroleum, and textile and fiber stocks trading in the Taiwan stock market during
the periods of 1980–1999 and 1991–1999. We estimate their VaRs under reliability
levels of 95 %, 97.5 %, and 99 %. The expected return of investment portfolio in
1999 is the sum of annual mean returns of various stocks multiplied with their
respective weights. We use this expected return to estimate year 1999 optimal stock
holding proportion and analyze the impact of different historical simulation periods
on optimal asset allocation.
Table 45.1 presents the summary of TSE general weighted stock index
(daily data) and estimated VaR for periods 1980–1999 and 1991–1999. Table 45.2
presents cumulative probability distributions of TSE daily index return for the
1980–1999 period and daily returns under the assumption of normality. It shows
that at confidence level lower than 95.8 % (e.g., 90 %), the left-tail probability for
historical distribution is higher than the normal return probability (4.2 %). Hence,
under the normal distribution assumption, the VaR is overestimated, and this leads
to an overcautious investment decision. At confidence level higher than 95.8 %
(e.g., 97.5 %), the left-tail probability for historical distribution is lower than the
normal return probability (4.2 %). Hence, under the normal distribution assump-
tion, the VaR is underestimated, and this leads to an overactive investment decision.
Figure 45.1 is the graphical presentation of the data in Table 45.2. The solid blue
line represents cumulative probability distributions of TSE daily index return, and the
dashed red line represents the normal distribution. The bottom panel is the enlarged
view of the left tail. This graph also indicates that VaR estimated using extreme
values of historical distribution will lead to an overactive investment decision.
Table 45.3 reports annualized returns and standard deviations for TSE daily
index and six selected industries: cement, electrical machinery, food, pulp and
paper, plastics and petroleum, and textile and fiber. For the 1980–1999 period,
the food industry had the greatest risk with a standard deviation of 53.47 % and
15.13 % annual return, whereas the overall market had a standard deviation of
50.80 % with 23.22 % annual return. Textile and fiber was the least risky industry
with a standard deviation of 41.21 % and 12.78 % annual return. For the 1991–1999
period, the electrical machinery industry had the greatest risk with a standard
45 Optimal Asset Allocation Under VaR Criterion: Taiwan Stock Market 1283
1
Cumulative Probability Empirical
0.8
Normal
0.6
0.4
0.2
0
−8% −6% −4% −2% 0% 2% 4% 6% 8%
return
0.3
0.25 Empirical
0.2 Normal
Probability
0.15
0.1
0.05
0
−8% −7% −6% −5% −4% −3% −2% −1%
return
Fig. 45.1 Cumulative probability distribution of Taiwan daily stock index. Period: 1980–1999.
Left tail of the cumulative probability of Taiwan weighted daily stock index. Period 1980–1999.
Lower panel shows a fat left tail
Table 45.3 TSE index and selected industry’s returns and standard deviations
0.250
r(p, 1980–1998)
0.200
expected return
0.150
0.100
0.050
0.000
72.500 73.000 73.500 74.000 74.500 75.000 75.500 76.000 76.500
risk
0.250
r(p, 1980–1998)
0.200
expected return
0.150
0.100
0.050
0.000
76.500 77.000 77.500 78.000 78.500 79.000 79.500
risk
Fig. 45.2 VaR-efficient frontier. The upper figure refers to investment portfolio of electrical
machinery and plastics and petroleum stocks, and the lower figure refers to investment portfolio of
cement and food stocks. VaR is set at reliability level of 95 %. Expected return and VaR are
estimated from TSE industry indexes from 1980 to 1998
deviation of 42.267 % and 26.12 % annual return, whereas the overall market had
a standard deviation of 36.58 % with 9.40 % annual return. The cement industry
was the least risky industry with a standard deviation of 24.92 % and 0.45 % annual
return. Next we constructed a two-industry optimal portfolio subject to VaR
constraint. The optimal asset allocation for the two-industry portfolio is obtained
by maximizing S(p) (derivation discussed in Appendix). The resulting
VaR-efficient frontiers are plotted in Fig. 45.2. The upper panel in Fig. 45.2 refers
45 Optimal Asset Allocation Under VaR Criterion: Taiwan Stock Market 1285
Table 45.4 Optimal asset allocation for the two-industry portfolio obtained by maximizing S(p)
at different level of confidence
Confidence level
Portfolio choices 95 % 97.5 % 99 %
Electrical machinery cement {1,0}; {1,0}a {1,0}; {1,0} {1,0}; {1,0}
Electrical machinery food {1,0}; {1,0} {1,0}; {1,0} {1,0}; {1,0}
Electrical machinery pulp and paper {1,0}; {1,0} {1,0}; {1,0} {1,0}; {1,0}
Electrical machinery plastics and {1,0}; {1,0} {1,0}; {1,0} {1,0}; {1,0}
petroleum
Electrical machinery textile and fiber {1,0}; {1,0} {1,0}; {1,0} {1,0}; {1,0}
Cement food {0,1}; {0,1} {0,1}; {0,1} {0,1}; {0,1}
Cement pulp and paper {0,1}; {0,1} {0,1}; {0,1} {0,1}; {0,1}
Cement plastics and petroleum {0.12, 0.88}; {0.04, 0.96}; {0.17, 0.83};
{0,1} {0,1} {0,1}
Cement textile and fiber {0.67, 0.33}; {0.6, 0.4}; {0,1} {0.15, 0.85};
{0,1} {0,1}
Food pulp and paper {1,0}; {1,0} {1,0}; {1,0} {0.98, 0.02};
{0,1}
Food plastics and petroleum {1,0}; {1,0} {1,0}; {1,0} {1,0}; {1,0}
Food textile and fiber {1,0}; {1,0} {1,0}; {1,0} {1,0}; {1,0}
Pulp and paper plastics and petroleum {0,1}; {0,1} {0,1}; {0,1} {0,1}; {0,1}
Pulp and paper textile and fiber {0,1}; {0,1} {0,1}; {0,1} {0,1}; {0,1}
Plastics and petroleum textile and fiber {0.99, 0.01}; {0.89, 0.11}; {1,0}; {1,0}
{1,0} {1,0}
First set {x, y} refers to the historical simulation for period 1980–1998, and second set {x, y} refers
to the historical simulation for period 1991–1998
a
{1, 0} represents 100 % investment in electrical machinery industry and 0 % investment in
cement industry
Table 45.5 Optimal allocation for two-industry portfolio (historical simulation, period
1980–1998)
Confidence Cement Food Portfolio VaR Lending, Cement Food Cash
level (%) (%) (%) (c, p*) B yuan VaR* (%) (%) (%)
95 100 0 31.12 0 31.12 100 0 0
97.5 100 0 42.57 121 31.12 87.9 0 12.1
99 100 0 53.92 216 31.12 78.4 0 21.6
95 0 100 27.49 0 27.49 0 100 0
97.5 0 100 40.22 139 27.49 0 86.1 13.9
99 0 100 56.27 267 27.49 0 73.3 26.7
95 12 88 27.52 0 27.52 12 88 0
97.5 4 96 41.99 154 27.52 3.4 81.2 15.4
99 17 83 53.42 246 27.52 12.8 62.6 24.6
Pulp and Textile Portfolio Pulp and Textile
Confidence paper and fiber VaR(c, Lending, paper and fiber Cash
level (%) (%) (%) p*) B yuan VaR* (%) (%) (%)
95 0 100 29.58 0 29.58 0 100 0
97.5 0 100 41.97 132 29.58 0 86.8 13.2
99 0 100 53.85 230 29.58 0 77 23
Textile Textile
Confidence Cement and fiber Portfolio Lending, Cement and fiber Cash
level (%) (%) (%) VaR(c, p*) B yuan VaR* (%) (%) (%)
95 67 33 25.86 0 25.86 12 88 0
97.5 60 40 41.97 172 25.86 49.7 33.1 17.2
99 15 85 52.51 256 25.86 11.2 63.2 25.6
Plastics Plastics
and Textile Portfolio and Textile
Confidence petroleum and fiber VaR Lending, petroleum and fiber Cash
level (%) (%) (%) (c, p*) B yuan VaR* (%) (%) (%)
95 99 1 28.43 0 28.43 99 1 0
97.5 89 11 41.35 139 28.43 76.6 9.5 13.9
99 100 0 55.50 253 28.43 74.7 0 25.3
Confidence Food Textile and Portfolio Lending, Food Textile and Cash
level (%) (%) fiber (%) VaR(c, p*) B yuan VaR* (%) fiber (%) (%)
95 100 0 27.49 0 27.49 100 0 0
97.5 100 0 40.22 121 27.49 87.9 0 12.1
99 100 0 56.27 267 27.49 73.3 0 26.7
Two-industry portfolio with initial investment of 1,000 yuan
Allocations for other industry combinations are available to interested readers
constraint. In opposite case, the VaR* constraint is specified at a higher level than
the portfolio VaR(c, p*); then investors will borrow money to invest in risky assets
(B < 0). Tables 45.5 and 45.6 list examples of investment in two-industry stocks
and T-bills for periods 1980–1999 and 1991–1999, respectively. In each case, the
target VaR* in the preset constraint is at a 95 % level of confidence, and 1,000 yuan
is invested in the portfolio. In the first panel of Table 45.5, 1,000 yuan is invested in
45 Optimal Asset Allocation Under VaR Criterion: Taiwan Stock Market 1287
Table 45.6 Optimal allocation for two-industry portfolio (historical simulation, period
1991–1998)
Confidence Cement Food Portfolio VaR Lending, Cement Food Cash
level (%) (%) (%) (c, p*) B yuan VaR* (%) (%) (%)
95 100 0 30.55 0 30.55 100 0 0
97.5 100 0 42.57 128 30.55 87.2 0 12.8
99 100 0 53.92 221 30.55 77.9 0 22.1
95 0 100 24.57 0 24.57 0 100 0
97.5 0 100 34.68 117 24.57 0 88.3 11.7
99 0 100 48.41 238 24.57 0 76.2 23.8
95 0 100 27.19 0 27.19 0 100 0
97.5 0 100 35.95 100 27.19 0 90 10
99 0 100 48.59 213 27.19 0 78.7 21.3
Pulp and Textile Portfolio Pulp and Textile
Confidence paper and fiber VaR(c, Lending, paper and fiber Cash
level (%) (%) (%) p*) B yuan VaR* (%) (%) (%)
95 0 100 27.75 0 27.75 0 100 0%
97.5 0 100 38.98 124 27.75 0 87.6 12.4
99 0 100 49.06 212 27.75 0 78.8 21.2
Textile Textile
Confidence Cement and fiber Portfolio Lending, Cement and fiber Cash
level (%) (%) (%) VaR(c, p*) B yuan VaR* (%) (%) (%)
95 0 100 27.75 0 27.75 0 100 0
97.5 0 100 38.98 124 27.75 0 87.6 12.4
99 0 100 49.06 212 27.75 0 78.8 21.2
Plastics Textile Plastics Textile
and and and and
Confidence petroleum fiber Portfolio Lending, petroleum fiber Cash
level (%) (%) (%) VaR(c, p*) B yuan VaR* (%) (%) (%)
95 100 0 27.19 0 27.19 0 100 0
97.5 100 0 35.95 100 27.19 0 90 10
99 100 0 48.59 213 27.19 0 78.7 21.3
Confidence Food Textile and Portfolio Lending, Food Textile and Cash
level (%) (%) fiber (%) VaR(c, p*) B yuan VaR* (%) fiber (%) (%)
95 100 0 24.57 0 24.57 100 0 0
97.5 100 0 34.68 117 24.57 88.3 0 11.7
99 100 0 48.41 238 24.57 76.2 0 23.8
Two-industry portfolio with initial investment of 1,000 yuan
Allocations for other industry combinations are available to interested readers
electrical machinery stocks and 0 in cement stocks. These allocations are from
Table 45.4. Portfolio VaR(c, p*) is 31.12, 42.57, and 53.92 yuan at 95 %, 97.5 %,
and 99 % respectively. This leads to the lending of 0, 121, and 216 yuan to meet the
target VaR of 31.12. Thus, a 97.5 % portfolio consists of 87.9 % in electrical
machinery stocks, 0 in cement stocks, and 21.1 % in T-bills.
1288 K. Hung and S. Srivastava
45.4 Conclusion
TSE daily index was found to be riskier than the market risk under the assumption of
normal distribution for market returns. This resulted in the left tail of cumulative return
distribution being fatter and a higher value at risk, indicating an overactive investment
activity. For asset allocation model under a constrained VaR framework, most of the
optimal portfolios have predominant investment in stocks from one industry. Hence, it
will be inappropriate to comment on the optimal allocation of future investment
portfolio based on the past stock performance of this unique period studied.
Let W0 be the initial investment and R be the rate of return of a portfolio. The value
of the portfolio at the end of the target horizon will be W ¼ W0(1 + R). Let m and s
be the expected return and standard deviation of R. The lowest portfolio value at the
confidence level c is defined as W* ¼ W0 (1 + R*). The relative VaR is the dollar
loss relative to the mean:
VaRðzeroÞ ¼ W 0 W ¼ W 0 R (45.2)
W* and R* are minimum value and cutoff return, respectively. In this paper we are
discussing absolute VaR. The general form of VaR can be derived from the probability
distribution of the future portfolio value f(w). For a given confidence level c, the worst
possible portfolio value W* is such that probability of exceeding W* is c:
ð1
c¼ f ðwÞdw (45.3)
W
ð W
p ¼ Pðw W Þ ¼ 1 c ¼ f ðwÞdw (45.4)
1
ð 1c
VaR ¼ Nð1 cÞ ¼ ’ðeÞde (45.5)
1
Let rp be the total portfolio return at the end of the holding period and T then the
expected wealth at the end of holding period; T can be written as
EðW T Þ ¼ ðW 0 þ BÞ 1 þ rp Bð1 þ rf Þ (45.9)
Max: EðW T Þ
(45.10)
s:t: PrfW T ðW 0 VaR Þg ð1 cÞ
Performance measure S(p) and borrowed amount can be deduced from Eq. 45.5.
Let p* be the maximizing portfolio and q(c, p) defines the quantile that corresponds
to probability (1 c) which can be obtained from portfolio return’s cumulative
density function. Maximizing portfolio p* is defined as
1290 K. Hung and S. Srivastava
p : max SðpÞ ¼ rp rf
(45.11)
p
W 0 r f W 0 qðc; pÞ
Initial wealth in the denominator of Eq. 45.6 is a scale constant and does not
affect the asset allocation. Let VaR(c, p) denote portfolio p’s VaR, and then the
denominator of Eq. 45.6 can be written as
rp rf
p : max SðpÞ ¼
p Fðc; pÞ
W 0 ðVaR VaRðc; p ÞÞ
B¼
F0 ðc, p0 Þ
References
Basak, S., & Shapiro, A. (2001). Value-at-risk based risk management: Optimal policies and asset
prices. The Review of Financial Studies, 14(2), 371–405.
Beder, T. S. (1995). VAR: Seductive but dangerous. Financial Analysts Journal, 51(5), 12–24.
Campbell, R., Huisman, R., & Koedijk, K. (2001). Optimal portfolio selection in a value-at-risk
framework. Journal of Banking and Finance, 25, 1789–1804.
Dowd, K. (1998). Beyond value at risk: The new science of risk management. New York: Wiley.
45 Optimal Asset Allocation Under VaR Criterion: Taiwan Stock Market 1291
Elton, E. J., & Gruber, M. J. (1995). Modern portfolio theory and investment analysis (5th ed.).
New York: Wiley.
Fong, G., & Vasicek, O. A. (1997). A multidimensional framework for risk analysis. Financial
Analysts Journal, 53, 51–57.
Hendricks, D. (1996). Evaluation of value-at-risk models using historical data. FRBNY Economic
Policy Review, 2, 39–69.
Hoppe, R. (1999). It’s time we buried value-at-risk. Risk Professional, 1. October page 18.
London.
Hull, J., & White, A. (1998). Incorporating volatility updating into the historical simulation
method for value-at-risk. Journal of Risk, 1, 5–19.
Johnson, R., & Wichern, D. (2007). Applied multivariate statistical analysis (6th ed.). New York:
Pearson Education.
Jorion, P. (1997). In defense of VAR. Derivatives Strategy, 2, 20–23.
Jorion, P. (2001). Value at risk (2nd ed.). New York: McGraw Hill.
Lucas, A., & Klaassen, P. (1998). Extreme returns, downside risk, and optimal asset allocation.
Journal of Portfolio Management, 25(1), 71–79.
Morgan, J. P. (1996). RiskMetricsTM – technical document. New York: Morgan Guaranty Trust
Company.
Roy, A. D. (1952). Safety first and the holding of assets. Econometrica, 20, 431–449.
Schachter, B. (1998). An irreverent guide to value at risk. Risks and Rewards, (1), 17–18.
Smithson, C., & Minton, L. (1996a). Value-at-risk (1): The debate on the use of VaR. Risk, 9,
25–27.
Smithson, C., & Minton, L. (1996b). Value-at-risk (2): The debate on the use of VaR. Risk, 9,
38–39.
Talmor, S. (1996). History repeat itself. The Banker, (1), 75–76.
Alternative Methods for Estimating Firm’s
Growth Rate 46
Ivan E. Brick, Hong-Yi Chen, and Cheng-Few Lee
Contents
46.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1294
46.2 The Discounted Cash Flow Model and the Gordon Growth Model . . . . . . . . . . . . . . . . . . 1295
46.3 Internal Growth Rate and Sustainable Growth Rate Models . . . . . . . . . . . . . . . . . . . . . . . . . . 1300
46.4 Statistical Methods . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1303
46.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1309
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1309
Abstract
The most common valuation model is the dividend growth model. The growth rate
is found by taking the product of the retention rate and the return on equity. What
is less well understood are the basic assumptions of this model. In this paper, we
demonstrate that the model makes strong assumptions regarding the financing mix
of the firm. In addition, we discuss several methods suggested in the literature on
estimating growth rates and analyze whether these approaches are consistent with
the use of using a constant discount rate to evaluate the firm’s assets and equity.
This chapter is a slightly revised version of Chapter 64 of Encyclopedia of Finance, 2nd Edition
and Brick et al. (2014).
I.E. Brick (*)
Department of Finance and Economics, Rutgers, The State University of New Jersey,
Newark/New Brunswick, NJ, USA
e-mail: ibrick@andromeda.rutgers.edu
H.-Y. Chen
Department of Finance, National Central University, Taoyuan, Taiwan
e-mail: fnhchen@ncu.edu.tw
C.-F. Lee
Department of Finance and Economics, Rutgers Business School, Rutgers, The State University of
New Jersey, Piscataway, NJ, USA
Graduate Institute of Finance, National Chiao Tung University, Hsinchu, Taiwan
e-mail: lee@business.rutgers.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1293
DOI 10.1007/978-1-4614-7750-1_46,
# Springer Science+Business Media New York 2015
1294 I.E. Brick et al.
The literature has also suggested estimating growth rate by using the average
percentage change method, compound-sum method, and/or regression methods.
We demonstrate that the average percentage change is very sensitive to extreme
observations. Moreover, on average, the regression method yields similar but
somewhat smaller estimates of the growth rate compared to the compound-sum
method. We also discussed the inferred method suggested by Gordon and Gordon
(1997) to estimate the growth rate. Advantages, disadvantages, and the interrela-
tionship among these estimation methods are also discussed in detail.
Keywords
Growth rate • Discount cash flow model • Internal growth rate • Sustainable
growth rate • Compound sum method
46.1 Introduction
One of the more highly used valuation models is that developed by Gordon and
Shapiro (1956) and Gordon (1962) known as the dividend growth model. In security
analysis and portfolio management, growth rate estimates of earnings, dividends, and
price per share are important factors in determining the value of an investment or
a firm. These publications demonstrate that the growth rate is found by taking the
product of the retention rate and the return on equity. What is less well understood are
the basic assumptions of this model. In this paper, we demonstrate that the model
makes strong assumptions regarding the financing mix of the firm.
In addition, we will also discuss several methods suggested in the literature on
estimating growth rates. We will analyze whether these approaches are consistent
with the use of using a constant discount rate to evaluate the firm’s assets and
equity. In particular, we will demonstrate that the underlying assumptions of the
internal growth rate model (whereby no external funds are used to finance growth)
is incompatible with the constant discount rate model of valuation.
The literature has also suggested estimating growth rate by taking the average of
percentage change of dividends over a sample period, taking the geometric average
of the change in dividends or using regression analysis to estimate the growth rate
(e.g., Lee et al. 2009; Lee et al. 2012; Lee et al. 2000; and Ross et al. 2010). Gordon
and Gordon (1997) suggest first using the Capital Asset Pricing Model (CAPM) to
determine the cost of equity of the firm and then using the dividend growth model to
infer the growth rate. Advantages, disadvantages, and the interrelationship among
these estimation methods are also discussed in detail.
This paper is organized as follows. In Sect. 46.2 we present the Gordon and
Shapiro model (1956). We discuss the inherent assumptions of the model and its
implied method to estimate the growth rate. Section 46.3 analyzes the internal
growth rate and sustainable growth rate models. Section 46.4 describes leading
statistical methods for estimating firm’s growth rates. We will also present the
46 Alternative Methods for Estimating Firm’s Growth Rate 1295
inferred method suggested by Gordon and Gordon (1997) to estimate the growth
rate. Concluding remarks appear in Sect. 46.5.
The traditional academic approach to evaluate a firm’s equity is based upon the
constant discount rate method. One approach uses the after-tax weighted average
cost of capital as a discount rate. This model is expressed as:
X
1
CFut
Value of Equity ¼ Debtt , (46:1)
t¼1 ð1 þ ATWACOCÞt
where CFut is the expected unlevered cash flow of the firm at time t and Debtt is the
market value of debt outstanding. ATWACOC equals L(1 t)Rd + (1 L)r where
L is the market value proportion of debt, t is the corporate tax rate, Rd is the cost of
debt and r is the cost of equity. The first term on the right hand side of Eq. 46.1 is the
value of the assets. Subtracting out the value of debt yields the value of equity. The
price per share is therefore the value of equity divided by the number of shares
outstanding. Alternatively, the value of equity can be directly found by discounting
the dividends per share by the cost of equity, or more formally:
X
1
dt
Value of Common Stock ðP0 Þ ¼ , (46:2)
t¼1 ð1 þ r Þt
where dt is the dividend per share at time t. Boudreaux and Long (1979), and
Chambers et al. (1982) demonstrate the equivalence of these two approaches
assuming that the level of that the level of debt is a constant percentage of the
value of the firm.1 Accordingly:
X
1
Xt
Debtt
t¼1 ð1 þ ATWACOCÞt X
1
dt
¼ (46:3)
#of Shares Outstaning t¼1 ð1 þ r Þt
If we assume that dividends per share grow at a constant rate g, then Eq. 46.2 is
reduced to the basic dividend growth model2:
d1
P0 ¼ : (46:4)
ð r gÞ
1
See Brick and Weaver (1984, 1997) concerning the magnitude of error in the valuation using
a constant discount rate when the firm does not maintain a constant market based leverage ratio.
2
Gordon and Shapiro’s (1956) model assume that dividends were paid continuously and hence
P0 ¼ d1/(r g).
1296 I.E. Brick et al.
where Et is the earnings in period t.3 If the firm’s retention rate is constant and used
in new investment, then the earnings at time t is
Ptþ1 Pt
gP ¼ : (46:8)
Pt
Recognizing that Pt and Pt+1 can be defined by Eq. 46.4, while noting that dt+2 is
equal to dt+1(1 + br) then:
d tþ2 dtþ1
d tþ2 d tþ1 dtþ1 ð1 þ br Þ d tþ1
gP ¼ k rb k rb ¼ ¼ ¼ br: (46:9)
d tþ1 dtþ1 dtþ1
k rb
Thus, under the assumption of a constant retention rate, for a one-period model,
dividends, earnings, and prices are all expected to grow at the same rate.
The relationship between the growth rate, g, the retention rate, b, and the return
on equity, r, can be expanded to a multi-period setting as the following numerical
example illustrates. In this example, we assume that the book value of the firm’s
assets equal the market value of the firm. We will assume that the growth rate of the
firm sales and assets is 4 % and the tax rate is equal to 40 %. The book value of the
3
Earnings in this model are defined using the cash-basis of accounting and not on an accrual basis.
46 Alternative Methods for Estimating Firm’s Growth Rate 1297
assets at time 0 is $50 and we assume a depreciation rate of 10 % per annum. The
amount of debt outstanding is $12.50 and amount of equity outstanding is $37.50.
We assume that the cost of debt, Rd, is 12 % and the cost of equity, r, is 25 %,
implying an ATWACOC of 20.55 %. The expected dividend at t ¼ 1, d1, must
satisfy Eq. 46.4. That is, 37.50 ¼ d1/(0.25 0.04).
The unlevered cash flow is defined as Sales less Costs (excluding the depreci-
ation expense) less Investment less the tax paid. Tax paid is defined as the tax rate
(which we assume to be 40 %) times Sales minus Costs minus the Depreciation
Expense. Recognizing that the value of the firm is given by CFu1/(ATWACOC g),
if firm value is $50, g ¼ 4 % and ATWACOC is 20.55 %, then the expected
unlevered cash flow is at time 1 is $8.28. We assume that the asset turnover ratio
is 1.7. Hence, if assets at time 0 is $50, the expected sales at time 1 is $85. To obtain
the level of investment, note that the depreciation expense at time 1 is $5. If the
book value of assets equals $52, then the firm must invest $7. To obtain an expected
unlevered cash flow at t ¼ 1 of $8.28, the Gross Profit Margin is assumed to be
approximately 26.03 %, resulting in expected costs at time 1 of $62.88. The interest
expense at time 1, is the cost of debt times the amount of debt outstanding at time
zero, or $1.50. The Earnings Before Taxes (EBT) is defined as Sales – Costs –
Interest Expense – Depreciation Expense, which equals $15.63 at time 1. 40 % of
EBT is the taxes paid or $6.25 resulting in a net income (NI) of $9.38. ROE, which
equals Net Income/Book Value of Equity at the beginning of the period is 25 %.
Since the aggregate level of dividends at time 1 is $7.88, then the dividend payout
ratio (1 b) is 84 %. Note that b is therefore equal to 16 % and b ROE ¼ 4%.4
Further note that the firm will increase its book value of equity via retention of
NI by $1.50 (RE in the table). In order to maintain a leverage ratio of 25 %, the firm
must increase the level of debt from time 0 to time 1 by $0.50. The entries for time
periods 2–5 follow the logical extension of the above discussion, and as shown in
the table, the retention rate b is 16 % and ROE ¼ 25 % for each period. Again the
product of b and ROE results in the expected growth rate of 4 %. Further note, that
g ¼ 4 % imply that sales, costs, book value of asset, depreciation, unlevered cash
flow, cash flow to stockholders, value of debt and value of equity to increase by 4 %
per annum.
Investors may use a one-period model in selecting stocks, but future profitability
of investment opportunities plays an important role in determining the value of the
firm and its EPS and dividend per share. The rate of return on new investments can
be expressed as a fraction, c (perhaps larger than 1), of the rate of return security
holders require (r):
k ¼ cr: (46:10)
4
Generally, practioners define ROE as the ratio of the Net Income to the end of year Stockholders
Equity. Here we are defining ROE as the ratio of the Net Income to the beginning of the year
Stockholders Equity. Brick et al. (2012) demonstrate that the practitioner’s definition is one of the
sources for the Bowman Paradox reported in the Organization Management literature.
1298 I.E. Brick et al.
Table 46.1 The book value of the firm’s assets equal the market value of the firm (growth rate
is 4 %)
0 1 2 3 4 5
Assets $50.00 $52.00 $54.08 $56.24 $58.49 $60.83
Debt $12.50 $13.00 $13.52 $14.06 $14.62 $15.21
Equity $37.50 $39.00 $40.56 $42.18 $43.87 $45.62
Rd 0.12 0.12 0.12 0.12 0.12 0.12
r 0.25 0.25 0.25 0.25 0.25 0.25
ATWACOC 0.2055 0.2055 0.2055 0.2055 0.2055 0.2055
Asset turnover 1.7 1.7 1.7 1.7 1.7
GPM 0.26029 0.26029 0.26029 0.26029 0.26029
Sales $85.00 $88.40 $91.94 $95.61 $99.44
Cost $62.88 $65.39 $68.01 $70.73 $73.55
Depreciation $5.00 $5.20 $5.41 $5.62 $5.85
Interest exp. $1.50 $1.56 $1.62 $1.69 $1.75
EBT $15.63 $16.25 $16.90 $17.58 $18.28
Tax $6.25 $6.50 $6.76 $7.03 $7.31
NI $9.38 $9.75 $10.14 $10.55 $10.97
DIV $7.88 $8.19 $8.52 $8.86 $9.21
New debt $.50 $0.52 $0.54 $0.56 $0.59
CFu $8.28 $8.61 $8.95 $9.31 $9.68
Firm value $50.00 $52.00 $54.08 $56.24 $58.49 $60.83
Investment $7.00 $7.28 $7.57 $7.87 $8.19
Vequity $37.50 $39.00 $40.56 $42.18 $43.87 $45.62
RE $1.50 $1.56 $1.62 $1.69 $1.75
ROE 0.25 0.25 0.25 0.25 0.25
1-b 0.84 0.84 0.84 0.84 0.84
g 0.04 0.04 0.04 0.04 0.04
ð1 bÞE1
k¼ : (46:11)
ð1 cbÞP0
Table 46.2 The market value of the firm and equity is greater than its book value
0 1 2 3 4 5
Assets $50.00 $52.00 $54.08 $56.24 $58.49 $60.83
Firm value $60.00 $62.40 $64.90 $67.49 $70.19 $73.00
Debt $12.50 $13.00 $13.52 $14.06 $14.62 $15.21
Equity $47.50 $49.40 $51.38 $53.43 $55.57 $57.79
Rd 0.12 0.12 0.12 0.12 0.12 0.12
r 0.25 0.25 0.25 0.25 0.25 0.25
ATWACOC 0.2129 0.2129 0.2129 0.2129 0.2129 0.2129
Asset turnover 1.7 1.7 1.7 1.7 1.7
GPM 0.3093 0.3093 0.3093 0.3093 0.3093
Sales $85.00 $88.40 $91.94 $95.61 $99.44
Cost $58.71 $61.06 $63.50 $66.04 $68.68
Depreciation $5.00 $5.20 $5.41 $5.62 $5.85
Interest exp. $1.50 $1.56 $1.62 $1.69 $1.75
EBT $19.79 $20.58 $21.41 $22.26 $23.15
Tax $7.92 $8.23 $8.56 $8.91 $9.26
NI $11.88 $12.35 $12.84 $13.36 $13.89
DIV $9.98 $10.37 $10.79 $11.22 $11.67
New debt $.50 $0.52 $0.54 $0.56 $0.59
CFu $10.38 $10.79 $11.22 $11.67 $12.14
Firm value $60.00 $62.40 $64.90 $67.49 $70.19 $73.00
Investment $7.40 $7.70 $8.00 $8.32 $8.66
Vequity $47.50 $49.40 $51.38 $53.43 $55.57 $57.79
RE $1.90 $1.98 $2.06 $2.14 $2.22
Market based ROE 0.25 0.25 0.25 0.25 0.25
1-b 0.84 0.84 0.84 0.84 0.84
g 0.04 0.04 0.04 0.04 0.04
firm and equity is greater than its book value. This example is depicted in
Table 46.2. The basic assumptions of the model is as follows: We will assume
that the growth rate of the firm sales and book value of the assets is 4 %. The book
value of the assets at time 0 is again $50 and we assume a depreciation rate of 10 %
per annum. However, note that the market value of the firm is $60. The entries for
Debt and Equity represent market values. The amount of debt outstanding is $12.50
and amount of equity outstanding is now $47.50. We assume that the cost of debt,
Rd, is 12 % and the cost of equity, r, is 25 %, implying an ATWACOC of 21.29 %.
For the valuation of the firm to be internally consistent, the unlevered cash flow at
time 1 is $10.38. Similarly, the value of equity to be internally consistent, the
expected dividends at t ¼ 1 is $9.98. Note that net income is $11.88 implying
a dividend payout ratio of 84 % and a retention rate of 16 %. The book value based
ROE, k, is found by taking the net income divided by the book value of equity. In
our example, implied book value of equity is $37.50. Hence, k ¼ 31.68 %, implying
1300 I.E. Brick et al.
that the book value ROE is greater than the cost of equity which is the required rate
of return. But g is given by the market value based ROE which is defined as Net
Income over market value of equity. That is r ¼ 25 %. Note again, br is 4 %.
An investor could predict next year’s dividends, the firm’s long-term growth
rate, and the rate of return stockholders require (perhaps using the CAPM to
estimate r) for holding the stock. Equation 46.4 could then be solved for the
theoretical price of the stock that could be compared with its present price. Stocks
that have theoretical prices above actual price are candidates for purchase; those
with theoretical prices below their actual price are candidates for sale or for
short sale.
The internal growth rate model assumes that the firm can only finance its growth by
its internal funds. Consequently, the cash to finance growth must come from only
retained earnings. Therefore, retained earnings can be expressed as
where
p ¼ the profit margin on all sales;
S ¼ annual sales; and
DS ¼ the increase in sales during the year.
Because retained earnings is the only source of new funds, the use of cash
represented by the increase in assets must equal the retained earnings:
DS pb
¼ , (46:13)
S T pb
where T ¼ the ratio of total assets to sales. If we divide both numerator and
denominator of Eq. 46.13 by T and make rearrange the terms, then we can show
that the internal growth rate is:
46 Alternative Methods for Estimating Firm’s Growth Rate 1301
DS pb=T b ROA
g¼ ¼ ¼ , (46:14)
S 1 pb=T 1 b ROA
where ROA is the return on assets. The internal growth rate is the maximum growth
rate that can be achieved without debt or equity kind of external financing. But note
this assumption of not issuing new debt or common stock to finance growth is
inconsistent with the basic assumption of the constant discount rate models that the
firm maintains a constant market based leverage ratio. Hence, this model cannot be
used to estimate the growth rate and be employed by the Gordon Growth Model.
Higgins (1977, 1981, 2008) has developed a sustainable growth rate under
assumption that firms can generate new funds by using retained earnings or issuing
debt, but not issuing new shares of common stock. Growth and its management
present special problems in financial planning. From a financial perspective, growth
is not always a blessing. Rapid growth can put considerable strain on a company’s
resources, and unless management is aware of this effect and takes active steps to
control it, rapid growth can lead to bankruptcy. Assuming a company is not raising
new equity, the cash to finance growth must come from retained earnings and new
borrowings. Further, because the company wants to maintain a target debt-to-equity
ratio equal to L, each dollar added to the owners’ equity enables it to increase its
indebtedness by $L. Since the owners’ equity will rise by an amount equal to
retained earnings, the new borrowing can be written as:
The use of cash represented by the increase in assets must equal the two sources
of cash (retained earnings and new borrowings)5:
DS pbð1 þ LÞ
g¼ ¼ : (46:15)
S T pbð1 þ LÞ
5
Increased in Assets is the net increase in assets. The total investment should also include the
depreciation expense as can be seen in our examples delineated in Tables 46.1 and 46.2. But
depreciation expense is also a source of funding. Hence, it is netted out in the relationship between
increases in assets and retained earnings and new borrowings.
1302 I.E. Brick et al.
In Eq. 46.15 the DS/S or g is the firm’s sustainable growth rate assuming no
infusion of new equity. Therefore, a company’s growth rate in sales must equal the
indicated combination of four ratios, p, b, L, and T. In addition, if the company’s
growth rate differs from g, one or more of the ratios must change. For example,
suppose a company grows at a rate in excess of g, then it must either use its assets
more efficiently, or it must alter its financial policies. Efficiency is represented by the
profit margin and asset-to-sales ratio. It therefore would need to increase its profit
margin (p) or decrease its asset-to-sales ratio (T) in order to increase efficiency. Financial
policies are represented by payout or leverage ratios. In this case, a decrease in its payout
ratio (1-b) or an increase in its leverage (L) would be necessary to alter its financial
policies to accommodate a different growth rate. It should be noted that increasing
efficiency is not always possible and altering financial policies are not always wise.
If we divide both numerator and denominator of Eq. 46.15 by T and rearrange
the terms, then we can show that the sustainable growth rate can be shown as
Please note that, in the framework of internal growth rate and sustainable growth
rate presented above, the source of cash are taken from the end of period values of
assets and assumed that the required financing occurs at the end of the period.
However, Ross et al. (2010) show that if the source of cash is from the beginning of
the period, the relationship between the use and the source of cash can be expressed
for the internal growth rate model as DST ¼ pSb and for the sustainable growth rate
model, DST ¼ pbS + pbSL . Such relationship will result an internal growth rate of
b ROA and a sustainable growth rate of b ROE. For example, Table 46.3
assumes identical assumptions to that of Table 46.1, but now we will assume
a growth rate of 4.1667 % and use total asset, total equity, and total debt from the
beginning of the period balance sheet to calculate the net income. Recall that ROE
is the net income divided by stockholders’ equity at the beginning of the period.
Note that the product of ROE and b will yield 4.1667 %.
Note that the intent of the Higgins’ sustainable growth rate allows only internal
source and external debt financing. Chen et al. (2013) incorporate Higgins (1977)
and Lee et al. (2011) frameworks, allowing company use both external debt and
equity, and derive a generalized sustainable growth rate as
b ROE l Dn P=E
gðtÞ ¼ þ , (46:17)
1 b ROE 1 b ROE
where
l ¼ degree of market imperfection;
Dn ¼ number of shares of new equity issued;
P ¼ price per share of new equity; and
E ¼ total equity:
46 Alternative Methods for Estimating Firm’s Growth Rate 1303
Table 46.3 The book value of the firm’s assets equal the market value of the firm (sustainable
growth rate is 4.1667 %)
0 1 2 3 4 5
Assets $50.00 $52.08 $54.25 $56.51 $58.87 $61.32
Value $50.00 $52.08 $54.25 $56.51 $58.87 $61.32
Debt $12.50 $13.02 $13.56 $14.13 $14.72 $15.33
Equity $37.50 $39.06 $40.69 $42.39 $44.15 $45.99
R 0.12 0.12 0.12 0.12 0.12 0.12
Re 0.25 0.25 0.25 0.25 0.25 0.25
ATWACOC 0.2055 0.2055 0.2055 0.2055 0.2055 0.2055
Asset turnover 1.7 1.7 1.7 1.7 1.7
GPM 0.26029 0.26029 0.26029 0.26029 0.26029
Sales $85.00 $88.54 $92.23 $96.07 $100.08
Cost $62.88 $65.49 $68.22 $71.07 $74.03
Depreciation $5.00 $5.21 $5.43 $5.65 $5.89
Interest exp. $1.50 $1.56 $1.63 $1.70 $1.77
EBT $15.63 $16.28 $16.95 $17.66 $18.40
Tax $6.25 $6.51 $6.78 $7.06 $7.36
NI $9.38 $9.77 $10.17 $10.60 $11.04
DIV $7.81 $8.14 $8.48 $8.83 $9.20
New debt $7.60 $7.92 $8.25 $8.59 $8.95
CFu $8.19 $8.53 $8.89 $9.26 $9.64
Value $50.00 $52.08 $54.25 $56.51 $58.87 $61.32
Investment $7.08 $7.38 $7.69 $8.01 $8.34
RE $1.56 $1.63 $1.70 $1.77 $1.84
ROE 0.25 0.25 0.25 0.25 0.25
(1-b) 0.833333 0.833333 0.833333 0.833333 0.833333
g 0.041667 0.041667 0.041667 0.041667 0.041667
Comparing Eq. 46.17, the generalized sustainable growth rate has an additional
positive term, 1lDnp=E
ð1DÞROE , when the new equity issue is taken into account.
Therefore, Chen et al. (2013) show that Higgins’ (1977) sustainable growth rate
is underestimated because of the omission of the source of the growth related to new
equity issue.
Instead of relying on financial ratios to estimate firm’s growth rates, one may use
statistical methods to determine firm’s growth rates. A simple growth rate can be
estimated by calculating the percentage change in earnings over a time period, and
taking the arithmetic average. For instance, the growth rate in earnings over one
period can be expressed as:
1304 I.E. Brick et al.
Et Et1
gt ¼ : (46:18)
Et1
1X n
g¼ g: (46:19)
n t¼1 t
X t ¼ X 0 ð 1 þ gÞ t , (46:20)
or
1=t
Xt
g¼ 1, (46:21)
X0
where
X0 ¼ measure in the current period (measure can be sales, earnings, or
dividends); and
Xt ¼ measure in period t.
This method is called the discrete compound sum method of growth-rate esti-
mation. For this approach to be consistent with the dividend growth model, the
duration of each period (e.g., quarterly or yearly) must be consistent with the
compounding period used in the dividend growth model.
Another method of estimating the growth rate uses the continuous compounding
process. The concept of continuous compounding process can be expressed math-
ematically as
Xt ¼ X0 egt : (46:22)
g mt
X t ¼ X0 1 þ , (46:23)
m
1 m
lim 1 þ ¼ e: (46:24)
m!1 m
m
g mt 1 ð g Þgt
lim Xt ¼ lim X0 1 þ ¼ X0 lim 1 þ ¼ X0 egt : (46:25)
m!1 m!1 m m!1 m=g
1 Xt
g ¼ ln : (46:26)
t X0
If you estimate the growth rate via Eq. 46.26, you are implicitly assuming the
dividends are growing continuously and therefore the dividend growth model. In
this case, according to Gordon and Shapiro’s (1956) model, P0 ¼ d0/(r g).
To use all the information available to the security analysts, two regression
equations can be employed. These equations can be derived from Eqs. 46.20 and
46.22 by taking the logarithm (ln) on both sides of equation:
If Eq. 46.27 can be used to estimate the growth rate, then the antilog of the
regression slope estimate would equal the growth rate. For the continuous
compounding process,
Both Eqs. 46.27 and 46.28 indicate that Xn is linearly related to t; and the growth
rate can be estimated by the ordinary least square (OLS) regression. For example,
growth rates for EPS and DPS can be obtained from an OLS regression by using
EPSt
ln ¼ a0 þ a1 T þ e1t , (46:29)
EPS0
and
DPSt
ln ¼ b0 þ b1 T þ e2t , (46:30)
DPS0
where EPSt and DPSt are earnings per share and dividends per share, respectively,
in period t, and T is the time indicators (i.e., T ¼ 1, 2, . . ., n). We denote a^1 and b^1 as
1306 I.E. Brick et al.
Table 46.4 Dividend behavior of firms Pepsico and Wal-Mart in dividends per share (DPS)
Year T PEP WMT Year T PEP WMT
1981 1 3.61 1.73 1996 16 0.72 1.33
1982 2 2.4 2.5 1997 17 0.98 1.56
1983 3 3.01 1.82 1998 18 1.35 1.98
1984 4 2.19 1.4 1999 19 1.4 1.25
1985 5 4.51 1.91 2000 20 1.51 1.41
1986 6 1.75 1.16 2001 21 1.51 1.49
1987 7 2.30 1.59 2002 22 1.89 1.81
1988 8 2.90 1.11 2003 23 2.07 2.03
1989 9 3.40 1.48 2004 24 2.45 2.41
1990 10 1.37 1.9 2005 25 2.43 2.68
1991 11 1.35 1.14 2006 26 3.42 2.92
1992 12 1.61 1.4 2007 27 3.48 3.17
1993 13 1.96 1.74 2008 28 3.26 3.36
1994 14 2.22 1.02 2009 29 3.81 3.73
1995 15 2.00 1.17 2010 30 3.97 4.2
6
If the earnings (or dividend) process follows Eq. 46.27, we can get same results from the
non-restricted model as Eqs. 46.29 and 46.30.
46 Alternative Methods for Estimating Firm’s Growth Rate 1307
Table 46.5 Estimated dividend growth rates for Pepsico and Wal-Mart
Pepsico (%) Wal-Mart (%)
Arithmetic average 4.64 8.99
Geometric average 0.99 5.45
Compound-sum method 0.99 5.30
Regression method (continuous) 0.56 7.04
Regression method (discrete) 0.56 7.29
Pepsico
2
1.8
1.6
1.4
1.2
DPS
1
0.8
0.6 True DPS
0.4 Predicted DPS
0.2
0
0 5 10 15 20 25 30
Time
DPSt
ln = − 0.6236 + 0.1947 T + εt
DPS0 (0.0056) (0.0113)
Wal-Mart
1.4
1.2
1
0.8
DPS
Consider the following situation. Dividends in years 1, 2 and 3 are $2, $4 and $2.
The arithmetic average of growth rate is 25 % but the true growth rate is 0 %. The
difference in the two average techniques will be greater when the variability of the
data is larger. Therefore, it is not surprising that we find differences in the estimated
growth rates using arithmetic average and geometric average methods for Pepsico
and Wal-Mart in Table 46.5.
1308 I.E. Brick et al.
Table 46.6 Estimated dividend growth rates for 50 randomly selected companies
Firms with positive growth Firms with negative growth
50 Firms (%) (35 firms) (%) (15 firms) (%)
Arithmetic average 4.95 7.27 0.47
Geometric average 0.93 3.00 3.88
Compound-sum 0.83 2.91 4.02
method
Regression method 0.66 2.32 3.22
(continuous)
Regression method 0.71 2.37 3.15
(discrete)
The regression method uses more available information than the geometric
average, discrete compounding and continuous compounding methods in that it
takes into account the observed growth rates between the first and last period of the
sample. A null hypothesis test can be used to determine whether the growth rate
obtained from the regression method is statistically significantly different from zero
or not. However, logarithms cannot be taken with zero or negative numbers. Under
this circumstance the arithmetic average will be a better alternative.
We further randomly select 50 companies from S&P 500 index firms, which paid
dividends during 1981–2010, to estimate their dividend growth rates by arithmetic
average method, geometric average method, compound-sum method, and the
regression method in terms of discrete and continuous compounding processes.
Table 46.6 shows averages of estimated dividend growth rates for 50 random
companies by different methods. As we discussed before, the arithmetic average
is sensitive to extreme values and has an upward bias. We, therefore, find a larger
average of the estimated dividend growth rate using the arithmetic average method.
We also find that on average, the geometric, and compound sum methods yield
relatively smaller growth rate estimates as compared to the estimates obtained using
the regression methods to estimate growth rate. However, it appears that estimates
obtained using the geometric, compound sum and regression methods are very
similar.
Finally, Gordon and Gordon (1997) suggest that one can infer the growth rate
using the dividend growth model. In particular, the practitioner can use regression
analysis to calculate the beta of the stock and use the CAPM to estimate the cost of
equity. Since
d 0 ð 1 þ gÞ
P0 ¼ (46:31)
ðr gÞ
and the price of the stock is given by the market, the cost of equity is obtained using
the CAPM, and d0 and the current dividend is known, one can infer the growth rate
using Eq. 46.31. If the inferred growth rate is less than the practitioner’s estimate,
then the recommendation will be to buy the stock. On the other hand, if the inferred
46 Alternative Methods for Estimating Firm’s Growth Rate 1309
46.5 Conclusion
The most common valuation model is the dividend growth model. The growth rate
is found by taking the product of the retention rate and the return on equity. What is
less well understood are the basic assumptions of this model. In this paper, we
demonstrate that the model makes strong assumptions regarding the financing mix
of the firm. In addition, we discuss several methods suggested in the literature on
estimating growth rates and analyze whether these approaches are consistent with
the use of using a constant discount rate to evaluate the firm’s assets and equity. In
particular, we demonstrate that the underlying assumptions of the internal growth
rate model (whereby no external funds are used to finance growth) are incompatible
with the constant discount rate model of valuation. The literature has also suggested
estimating growth rate by using the average percentage change method, compound-
sum method, and/or regression methods. We demonstrate that the average percent-
age change is very sensitive to extreme observations. Moreover, on average, the
regression method yields similar but somewhat smaller estimates of the growth rate
compared to the compound-sum method. We also discussed the inferred method
suggested by Gordon and Gordon (1997) to estimate the growth rate. Advantages,
disadvantages, and the interrelationship among these estimation methods are also
discussed in detail. Choosing an appropriate method to estimate firm’s growth rate
can yield a more precise estimation and be helpful for the security analysis and
valuation. However, all of these methods use historical information to obtain
growth estimates. To the extent that the future may differ from the past, will
ultimately determine the efficacy of any of these methods.
References
Boudreaux, K. J., & Long, R. W. (1979). The weighted average cost of capital as a cutoff rate:
A further analysis. Financial Management, 8, 7–14.
Brick, I. E., & Weaver, D. G. (1984). A comparison of capital budgeting techniques in identifying
profitable investments. Financial Management, 13, 29–39.
Brick, I. E., & Weaver, D. G. (1997). Calculating the cost of capital of an unlevered firm for use in
project evaluation. Review of Quantitative Finance and Accounting, 9, 111–129.
Brick, I. E., Palmon, O., & Venezia, I. (2012). The risk-return (Bowman) paradox and accounting
measurements. In I. Venezia & Z. Wiener (Eds.), Bridging the GAAP: Recent advances in
finance and accounting. London: World Scientific Publishing Company.
Brick, I. E., Chen H. Y., Hsieh C. H., & Lee C. F. (2014, forthcoming). A comparison of
alternative models for estimating firm’s growth rate. Review of Quantitative Finance and
Accounting.
1310 I.E. Brick et al.
Chambers, D. R., Harris, R. S., & Pringle, J. J. (1982). Treatment of financing mix in analyzing
investment opportunities. Financial Management, 11, 24–41.
Chen, H. Y., Gupta, M. C., Lee, A. C., & Lee, C. F. (2013). Sustainable growth rate, optimal
growth rate, and optimal payout ratio: A joint optimization approach. Journal of Banking and
Finance, 37, 1205–1222.
Fama, E. F., & French, K. R. (1997). The cross-section of expected stock returns. The Journal of
Finance, 47, 427–465.
Gordon, J., & Gordon, M. (1997). The finite horizon expected return model. The Financial
Analysts Journal, 53, 52–61.
Gordon, M., & Shaprio, E. (1956). Capital equipment analysis: The required rate of profit.
Management Science, 3, 102–110.
Higgins, R. C. (1977). How much growth can a firm afford? Financial Management, 6, 7–16.
Higgins, R. C. (1981). Sustainable growth under inflation. Financial Management, 10, 36–40.
Higgins, R. C. (2008). Analysis for financial management (9th ed.). New York: McGraw-Hill.
Lee, C. F., & Lee, A. C. (2013). Encyclopedia of finance (2nd ed.). New York: Springer.
Lee, C. F., Lee, J. C., & Lee, A. C. (2000). Statistics for business and financial economics
(2nd ed.). Singapore: World Scientific Publishing Company.
Lee, A. C., Lee, J. C., & Lee, C. F. (2009). Financial analysis, planning and forecasting: Theory
and application (2nd ed.). Singapore: World Scientific Publishing Company.
Lee, C. F., Finnerty, J. E., Lee, A. C., Lee, J., & Wort, D. H. (2012a). Security analysis, portfolio
management, and financial derivatives (2nd ed.). New York: Springer.
Lee, C. F., Gupta, M. C., Chen, H. Y., & Lee, A. C. (2012b). Optimal payout ratio under
uncertainty and the flexibility hypothesis: Theory and empirical evidence. Journal of Corpo-
rate Finance, 17, 483–501.
Ross, S. A., Westerfield, R. W., & Jaffe, J. (2010). Corporate finance (9th ed.). New York:
McGraw-Hill/Irwin.
Econometric Measures of Liquidity
47
Jieun Lee
Contents
47.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1312
47.2 Low-Frequency Liquidity Proxies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1313
47.2.1 The Roll Measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1313
47.2.2 Effective Tick . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1314
47.2.3 Amihud (2002) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1315
47.2.4 Lesmond, Ogden, and Trzcinka (LOT 1999) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1316
47.3 High-Frequency Liquidity Proxies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1318
47.3.1 Percent Quoted Spread . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1318
47.3.2 Percent Effective Spread . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1318
47.4 Empirical Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1319
47.4.1 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1319
47.4.2 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1319
47.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1321
Appendix 1: Solution to LOT (1990) Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1321
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1322
Abstract
A security is liquid to the extent that an investor can trade significant quantities of
the security quickly, at or near the current market price, and bearing low transac-
tion costs. As such, liquidity is a multidimensional concept. In this chapter,
I review several widely used econometrics or statistics-based measures that
researchers have developed to capture one or more dimensions of a security’s
liquidity (i.e., limited dependent variable model (Lesmond, D. A. et al. Review of
Financial Studies, 12(5), 1113–1141, 1999) and autocovariance of price changes
(Roll, R., Journal of Finance, 39, 1127–1139, 1984). These alternative proxies
have been designed to be estimated using either low-frequency or high-frequency
J. Lee
Economic Research Institute, Bank of Korea, Seoul, South Korea
e-mail: jelee@bok.or.kr
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1311
DOI 10.1007/978-1-4614-7750-1_99,
# Springer Science+Business Media New York 2015
1312 J. Lee
data, so I discuss four liquidity proxies that are estimated using low-frequency data
and two proxies that require high-frequency data. Low-frequency measures permit
the study of liquidity over relatively long time horizons; however, they do not
reflect actual trading processes. To overcome this limitation, high-frequency
liquidity proxies are often used as benchmarks to determine the best
low-frequency proxy. In this chapter, I find that estimates from the effective tick
measure perform best among the four low-frequency measures tested.
Keywords
Liquidity • Transaction costs • Bid-ask spread • Price impact • Percent effective
spread • Market model • Limited dependent variable model • Tobin’s model •
Log likelihood function • Autocovariance • Correlation analysis
47.1 Introduction
A security is liquid to the extent that an investor can trade significant quantities of
the security quickly, at or near the current market price, and bearing low transaction
costs. A security’s liquidity is an important characteristic variable, relevant in asset
pricing studies, studies of market efficiency, and even corporate finance. In the asset
pricing literature, researchers have considered whether liquidity is a priced risk
factor (e.g., Amihud and Mendelson 1986; Brennan and Subrahmanyam 1996;
Amihud 2002; Pastor and Stambaugh 2003). In corporate finance, researchers
have found that liquidity is related to capital structure, mergers and acquisitions,
and corporate governance (e.g., Lipson 2003; Lipson and Mortal 2007, 2009;
Bharath 2009; Chung et al. 2010).
In these and many other studies, researchers have chosen from a variety of
liquidity measures that have been developed. In turn, the variety of available
liquidity measures reflects the multidimensional aspect of liquidity. Note that the
definition of liquidity given above features four dimensions of liquidity: trading
quantity, trading speed, price impact, and trading cost. Some extant measures focus
on a single dimension of liquidity, while others encompass several dimensions.
For instance, the bid-ask spread measure in Amihud and Mendelson (1986), the
estimator of the effective spread in Roll (1984), and the effective tick estimator in
Goyenko et al. (2009) relate to the trading cost dimension. The turnover measure of
Datar et al. (1998) captures the trading quantity dimension. The measures in
Amihud (2002) and Pastor and Stambaugh (2003) are relevant to price impact.
The number of zero trading volume days in Liu (2006) emphasizes trading speed.
Finally, and different from the others, the measure in Lesmond et al. (1999) encom-
passes several dimensions of liquidity.
Among the available measures, this chapter focuses on six liquidity proxies,
including four that are commonly estimated using low-frequency data (i.e., daily
closing prices) and two that are commonly estimated using high-frequency
data (i.e., intraday trades and quotes). The low-frequency measures are in
Roll (1984), Goyenko et al. (2009), Lesmond et al. (1999), and Amihud (2002).
47 Econometric Measures of Liquidity 1313
The high-frequency measures are the percent quoted spread and the percent
effective spread. The low-frequency proxies are advantageous because they are
more amenable to the study of liquidity over relatively long time horizons and
across countries. However, they are limited because they do not directly reflect
actual trading processes, while the high-frequency measures do. Thus, high-
frequency liquidity proxies are often used as benchmarks to determine the best
low-frequency proxy. This is not a universal criterion, however, because each
measure captures a different dimension of liquidity and may lead to different
results in specific cross-sectional or time-series applications.
The remainder of this chapter is organized as follows. In Sects. 47.2 and 47.3,
I introduce and briefly discuss each of the low-frequency and high-frequency
liquidity measures, respectively. Section 47.4 provides an empirical analysis of
these liquidity measures, including the aforementioned test of the best
low-frequency measure. Section 47.5 concludes.
Below I describe four widely used measures of liquidity: the Roll (1984) measure;
effective tick; the Amihud (2002) measure; and the Lesmond et al. (1999) measure.
Roll (1984) develops a measure of the effective bid-ask spread. He assumes that the
true value of a stock follows a random walk and that Pt, the observed closing price
on day t, is equal to the stock’s true value plus or minus half of the effective spread.
He also assumes that a security trades at either the bid price or the ask price, with
equal frequency. This relationship can be expressed as follows:
s
Pt ¼ Pt þ Qt
2
þ1 with probability 1=2 ðbuyer initiatedÞ
Qt IID
1 with probability 1=2 ðseller initiatedÞ
Under the assumption that Qt is IID, the variance and covariance of DPt can be
easily calculated:
s2
var½DPt ¼ :
2
s2
cov½DPt ; DPt1 ¼ :
hs 4 i
s
cov½DPt ; DPt1 ¼ cov ðQt Qt1 Þ, ðQt1 Qt2 Þ
2 2
s2
¼ ½covðQt Qt1 Þ, covðQt1 Qt2 Þ
4
s2
¼ ½covðQt ; Qt1 Þ covðQt1 ; Qt1 Þ
4
þcovðQt1 ; Qt2 Þ covðQt ; Qt2 Þ
s2 s2 1 2 1 2 s2
¼ ½varðQt1 Þ ¼ ð1 0Þ þ ð1 0Þ ¼
4 4 2 2 4
Roll’s measure is simple and intuitive: If P* is fixed so that prices take only two
values, bid or ask, and if the current price is the bid, then the change between
current price and previous price must be either 0 or –s and the price change between
current price and next price must be either 0 or s. Analogous possible price changes
apply when the current price is the ask.
The Roll measure S is generally estimated using daily data on price changes.
Roll (1984) and others have found that for some individual stocks, the
autocovariance that defines S is positive, rather than negative, so that S is undefined.
In this case, researchers generally choose one of three solutions: (1) treat the
observation as missing, (2) set the Roll spread estimate to zero, or (3) multiply
the covariance by negative one, calculate S, and multiply this estimate by negative
one to produce a negative spread estimate. In my empirical analysis to follow, I find
that results are insensitive to the alternative solutions, so I only report results of
setting S to zero when the observed autocovariance is positive.
Goyenko et al. (2009) and Holden (2009) develop an effective tick measure that is
based on price clustering and changes in tick size. Below I describe the effective
tick measure in Goyenko et al. (2009), which is elegant in its simplicity.
47 Econometric Measures of Liquidity 1315
Consider four possible bid-ask spreads for a stock: $1/8, $1/4, $1/2, and $1. If the
spread is $1/4, the authors assume that bid and ask prices are associated with only
even quarters. Thus, if an odd-eighth transaction price shows up, it is instead
inferred that the spread is $1/8. The number of quotes that occur at $1/8 spread is
given by N1. The number of quotes at odd-quarter fractions ($1/8 and $1/4) is N2.
The number of quotes at odd-half ($1/8, $1/4, and $1/2) is N3. Finally, the number
of whole dollar quotes is given by N4. The following is the proportion of each price
fraction observed during the day:
Ni
Fi ¼ XI for i ¼ 1, . . . , I:
i¼1
Ni
Next, suppose that the unconstrained probability of the effective ith estimated
spread is
2Fi i ¼ 1
U ¼ 2Fi Fi1 i ¼ 2, . . . , i
Fi Fi1 i ¼ I:
The measures in Amihud (2002) and Pastor and Stambaugh (2003) both purport to
capture the price impact dimension of liquidity. Goyenko et al. (2009) show that the
Amihud (2002) measure performs well in measuring price impact while the Pastor
and Stambaugh (2003) measure is dominated by other measures. Pastor and
Stambaugh (2003, p. 679) also caution against their measure as a liquidity measure
for individual stocks, reporting large sampling errors in individual estimates.
1316 J. Lee
Referring also to Hasbrouck (2009), I do not discuss the Pastor and Stambaugh
(2003) measure. The Amihud (2002) measure is a representative proxy for price
impact, i.e., the daily price response associated with one dollar of trading volume:
jRetit j
Amiit ¼ ,
Volit
where Retit is the stock i’s return on day t and Volit is the stock i’s dollar volume on
day t. The average is calculated over all positive-volume days, since the ratio is
undefined for zero volume days.
The Lesmond et al. (LOT 1999) liquidity measure is based on the idea that an
informed trader observing a mispriced stock will execute a trade only if the
difference between the current market price and the true price exceeds the
trader’s transaction costs; otherwise, no trade will occur. Therefore, they
argue that a stock with high transaction costs will have less frequent price
movements and more zero returns than a stock with low transaction costs.
Based on this relationship, they develop a measure of the marginal trader’s
effective transaction costs for an individual stock. Their measure utilizes the
limited dependent variable regression model of Tobin (1958) and Rosett (1959)
applied to the “market model.”
The market model implies that a security’s return reflects the effect of new
information on the value of the stock, which can be divided into two components:
contemporaneous market-wide information (biRmt) and firm-specific information eit.
In an ideal market without frictions such as transaction costs, new information
will be immediately reflected into the security’s price, so Rit is the true return on
security i.
Region3
Fig. 47.1 This figure illustrates the relationship between the observed return on a stock in the
presence of transaction costs that inhibit trading, Rit, and its true return in the absence of transaction
costs, R*it, where the latter reflects the true effects of new market-wide or firm-specific information.
The relationship can be divided into three regions: (1) Region 1, where the value of new information
is negative and exceeds transaction costs; (2) Region 2, where the transaction costs exceed the value
of new information regardless of the direction of the value of information; and (3) Region 3, where
the value of new information is positive and exceeds transaction costs
returns. In the presence of transaction costs, Lesmond et al. (1999) propose the
following relationship between observed and true returns:
where a1i < 0 and a2i > 0. a1i is the transaction costs for the marginal investor when
information has a negative shock (selling), and a2i is the transaction costs for the
marginal investor when information has a positive shock (buying). Consequently,
the difference between a1i and a2i is a measure of round-trip transaction costs.
If the true return exceeds transaction costs, the marginal investor will continu-
ously trade, and the market price will respond until, for the next trade, marginal
profit is equal to marginal transaction costs. If the transaction costs are greater than
true returns, then the marginal investors will not trade, price will not move, and
consequently the zero returns will occur. Therefore, in this model the frequency of
zero returns is a simple alternative measure of transaction costs. The relationship
between observed returns and true returns is illustrated in Fig. 47.1.
1318 J. Lee
Next I describe two well-known spread proxies that can be estimated using high-
frequency data. These are the percent quoted spread and percent effective spread.
The ask (bid) quotation is the price at which shares can be purchased (sold) with
immediacy. The difference, known as the percent quoted spread, is the cost of
a round-trip transaction and is generally expressed as a proportion of the average of
the bid and ask prices:
Askit Bidit
Percent quoted spreadit ¼ :
Mit
In the empirical analysis in the next section, I estimate percent quoted spreads
using high-frequency data. Following convention, for each stock and trading day,
I find the highest bid and lowest ask prices over all venues at every point during the
day, denoting these “inside” ask and bid prices as Askit and Bidit, respectively. Mit
is then the average of, or midpoint between, Askit and Bidit. I then calculate the
average percent quoted spread for a stock and day as the time-weighted average of
all spreads observed for that stock during the day. Finally, percent quoted spread
for each stock is calculated by averaging the daily estimates across all trading days
within a given month.
Some trades occur within the range of inside bid and ask quotes, as when simulta-
neous buy and sell market orders are simply crossed. Thus, the inside bid-ask spread
may overestimate the realized amount of this component of transaction costs.
Hasbrouck’s (2009) measure of percent effective spread attempts to adjust for
this bias. For a given stock, percent effective spread is computed for all trades
relative to the prevailing quote midpoint:
Pit Mit
Percent effective spreadit ¼ 2Dit ,
Mit
where, for stock i, Dit is the buy-sell indicator variable which takes a value of 1 (–1)
for buyer-initiated (seller-initiated) trades, Pit is the transaction price, and Mit is the
midpoint of the most recently posted bid and ask quotes. The average percent
effective spread for each day is a trade-weighted average across all trades during
the day. The monthly percent effective spread for each security is calculated by
averaging across all trading days within a given month.
47 Econometric Measures of Liquidity 1319
47.4.1 Data
I estimate liquidity measures for NYSE, AMEX, and NASDAQ common stocks
over the years 1993–2008. I estimate the low-frequency measures using daily data
from the Center for Research in Security Prices (CRSP) database. I estimate the
high-frequency measures using the New York Stock Exchange Trades and Auto-
mated Quotes (TAQ) database. TAQ data is available only since 1993, which is
therefore the binding constraint in terms of staring year. In order to be included in
the sample, a stock must have at least 60 days of past return data. I discard
certificates, American Depositary Receipts (ADRs), shares of beneficial interest,
units, companies incorporated outside United States, American Trust components
closed-end funds, preferred stocks, and Real Estate Investment Trusts (REITs).
Regarding estimating the high-frequency measures, I determine the highest bid
and lowest ask across all quoting venues at every point during the day (NBBO
quotes) and then follow filters referring to Huang and Stoll (1997) and Brownless
and Gallo (2006). To reduce errors and outliers, I remove (1) quotes if either the bid
or ask price is negative; (2) quotes if either the bid or ask size is negative; (3) quotes
if bid-ask spread is greater than $5 or negative; (4) the quotes if transaction price is
negative; (5) quotes before-the-open and after-the-close trades and quotes;
(6) quotes if the bid, ask, or trade price differ by more than 20 % from the previous
quote or trade price; (7) quotes originating in market other than the primary
exchange because regional quotes tend to closely follow the quotes posted by the
primary exchange; and (8) %effective spread/%quoted spread>4.0.
Table 47.1 reports correlations among the various liquidity estimates. In this table,
observations are pooled across all stocks and all months. All correlations are
reliably positive and substantial in magnitude, ranging from 0.382 to 0.971. The
two high-frequency measures, percent effective spread and percent quoted spread,
are very highly correlated (0.971). Using percent effective spread as our high-
frequency “benchmark,” its correlations with the low-frequency measures are 0.742
(Roll), 0.757 (effective tick), 0.621 (Amihud), and 0.586 (LOT). Based on the
aforementioned criterion, these results indicate that the effective tick and Roll
measures are the “best” low-frequency measures, as they have the highest correla-
tion with percent effective spread.
Table 47.2 presents the time-series means of monthly correlations of percent
effective spread with each of the low-frequency measures for the full sample period
as well as subperiods 1993–2000 (pre-decimalization) and 2001–2008 (post-
decimalization). For three of the four low frequencies measured, the correlation
with percent effective spread is higher in the first subperiod than the second
subperiod, which may reflect differential effects of decimalization on the various
1320 J. Lee
Table 47.2 Average cross-sectional correlations with percent effective spread, monthly
estimates
Roll Eff. tick Ami LOT
1993–2008 0.662 0.685 0.684 0.561
1993–2000 0.748 0.754 0.662 0.670
2001–2008 0.576 0.615 0.706 0.452
For each month, I estimate the cross-sectional correlation between the liquidity proxies from the
low-frequency data and percent effective spread from TAQ. This table presents the average cross-
sectional correlations across all months. A stock is excluded only if it trades for less than 60 days
prior to an observation or if liquidity estimates are missing
dimensions of liquidity. For the full period as well as the first subperiod, effective
tick has the correlation with percent effective spread, while for the second period
the Amihud measure has the highest correlation with percent effective spread.
Table 47.3 shows stock-by-stock time-series correlations between the high-
frequency measure percent effective spread and each of the low-frequency
measures, using the full-period data but also breaking the observations down by
the exchange on which a stock trades. For the full sample as well as every exchange,
the Amihud and effective tick estimates have relatively high correlations with
percent effective spread, while the correlations are relatively low for the Roll and
LOT measures.
47 Econometric Measures of Liquidity 1321
47.5 Conclusions
This chapter discusses several popular measures of liquidity that are based on
econometric approaches and compares them via correlation analysis. Among the
four low-frequency liquidity proxies, I find that the effective tick measure is
generally more highly correlated with the high-frequency measure (i.e., percent
effective spread). Thus, by this criterion the effective tick measure is the
“best” low-frequency measure of liquidity. However, since each liquidity measure
captures only part of the multidimensional nature of liquidity, it is difficult to judge
which measure is best. Consequently, from among the available measures of
liquidity, a researcher should choose the measure that is consistent with their
research purpose or perhaps consider employing several of them.
where
1
s i
Y Rit þ a2i b Rmt
Rit þ a1i bi Rmt
F i
F
2
si si
Y Rit þ a2i b Rmt
Ø i
,
3
s i
(47.5)
where Ø refers to the standard normal density function and F refers to the
cumulative normal distribution. The product is over the Region 1, 2, and 3 of
observations for which R*it < a1i, a1i < R*it < a2i, and R*it > a2i, respectively. The
log likelihood function is
2 3
6 1 7 1
log L ¼ S1 log4 5 S1 ðRit þ a1i bi Rmt Þ2 þ S2 log½F2 F1
1 2si 2
ð2psi Þ
2
2 2 3
6 1 7
þ S3 log6 7 1 S ðR þ a2i bi Rmt Þ2 :
4 1 5 2si 2 3 it
ð2psi 2 Þ2
(47.6)
Given Eq. 47.6, a1i, a2i, bi, and si can be estimated. The difference between a2i
and a1i is the proxy of a round-trip transaction cost in the LOT model.
References
Amihud, Y. (2002). Illiquidity and stock returns: Cross-section and time-series effects. Journal of
Financial Markets, 5, 31–56.
Amihud, Y., & Mendelson, H. (1986). Asset pricing and the bid-ask spread. Journal of Financial
Economics, 17(2), 223–249.
Bharath, S. T., Pasquariello, P., et al. (2009). Does asymmetric information drive capital structure
decisions? Review of Financial Studies, 22(8), 3211–3243.
Brennan, M. J., & Subrahmanyam, A. (1996). Market microstructure and asset pricing: on the
compensation for illiquidity in stock returns. Journal of Financial Economics, 41, 441–464.
Brennan, M., Chordia, T., et al. (1998). Alternative factor specifications, security characteristics, and
the cross-section of expected stock returns. Journal of Financial Economics, 49(3), 345–373.
Brownlees, C. T., & Gallo, G. M. (2006). Financial econometric analysis at ultra-high frequency:
Data handling concerns. Computational Statistics & Data Analysis, 51, 2232–2245.
47 Econometric Measures of Liquidity 1323
Chung, K. H., Elder, J., & Kim, J. (2010). Corporate governance and liquidity. Journal of
Financial and Quantitative Analysis, 45(02), 265–291.
Datar, V. T., Naik, N. Y., & Radcliffe, R. (1998). Liquidity and stock returns: An alternative test.
Journal of Financial Markets, 1, 203–219.
Goyenko, R. Y., Holden, C. W., et al. (2009). Do liquidity measures measure liquidity? Journal of
Financial Economics, 92(2), 153–181.
Hasbrouck, J. (2009). Trading costs and returns for US equities: Estimating effective costs from
daily data. Journal of Finance, 64(3), 1445–1477.
Holden, C. W. (2009). New low-frequency spread measures. Journal of Financial Markets, 12(4),
778–813.
Huang, R. D., & Stoll, H. R. (1997). The components of the bid-ask spread: A general approach.
Review of Financial Studies, 10(4), 995–1034.
Lesmond, D. A., Ogden, J. P., et al. (1999). A new estimate of transaction costs. Review of
Financial Studies, 12(5), 1113–1141.
Lipson, M. L. (2003). Market microstructure and corporate finance. Journal of Corporate Finance,
9(4), 377–384.
Lipson, M. L., & Mortal, S. (2007). Liquidity and firm characteristics: Evidence from mergers and
acquisitions. Journal of Financial Markets, 10(4), 342–361.
Lipson, M. L., & Mortal, S. (2009). Liquidity and capital structure. Journal of Financial Markets,
12(4), 611–644.
Liu, W. (2006). A liquidity-augmented capital asset pricing model. Journal of Financial Econom-
ics, 82, 631–671.
Maddala, G. S. (1986). Limited-dependent and qualitative variables in econometrics. Cambridge:
Cambridge University Press.
Pástor, L., & Stambaugh, R. (2003). Liquidity risk and expected stock returns. Journal of Political
Economy, 111, 642–685.
Roll, R. (1984). A simple implicit measure of the effective bid-ask spread in an efficient market.
Journal of Finance, 39, 1127–1139.
Rosett, R. N. (1959). A statistical model of friction in economics. Econometrica: Journal of the
Econometric Society, 27, 263–267.
Tobin, J. (1958). Estimation of relationships for limited dependent variables. Econometrica:
Journal of the Econometric Society, 26, 24–36.
A Quasi-Maximum Likelihood
Estimation Strategy for Value-at-Risk 48
Forecasting: Application to Equity
Index Futures Markets
Contents
48.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1326
48.2 ARMA-GARCH Model with Normal and Tempered Stable Innovations . . . . . . . . . . . . 1328
48.3 VaR for the ARMA-GARCH Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1329
48.3.1 VaR and Backtesting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1329
48.4 Introduction of Trading Volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1332
48.4.1 Different Variants of Trading Volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1332
48.4.2 Lagged Relative Change of Trading Volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1334
48.4.3 Lagged Trading Volume or Forecasting Contemporaneous
Trading Volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1334
48.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1337
Appendix: VaR on the CTS Random Variable . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1337
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1339
O. Carchano
Department of Financial Economics, University of Valencia, Valencia, Spain
e-mail: oscar.carchano@uv.es
Y.S.A. Kim
College of Business, Stony Brook University, Stony Brook, NY, USA
e-mail: aaron.kim@stonybrook.edu
E.W. Sun
KEDGE Business School and BEM Management School, Bordeaux, France
e-mail: edward.sun@kedgebs.com
S.T. Rachev
Department of Applied Mathematics and Statistics, College of Business, Stony Brook University,
SUNY, Stony Brook, NY, USA
FinAnalytica, Inc, New York, NY, USA
e-mail: svetlozar.rachev@stonybrook.edu
F.J. Fabozzi (*)
EDHEC Business School, EDHEC Risk Institute, Nice, France
e-mail: frank.fabozzi@edhec.edu; fabozzi321@aol.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1325
DOI 10.1007/978-1-4614-7750-1_48,
# Springer Science+Business Media New York 2015
1326 O. Carchano et al.
Abstract
We present the first empirical evidence for the validity of the ARMA-GARCH
model with tempered stable innovations to estimate 1-day-ahead value at risk
in futures markets for the S&P 500, DAX, and Nikkei. We also provide
empirical support that GARCH models based on normal innovations
appear not to be as well suited as infinitely divisible models for predicting
financial crashes. The results are compared with the predictions based on
data in the cash market. We also provide the first empirical evidence on
how adding trading volume to the GARCH model improves its forecasting
ability.
In our empirical analysis, we forecast 1 % value at risk in both spot and
futures markets using normal and tempered stable GARCH models following
a quasi-maximum likelihood estimation strategy. In order to determine
the accuracy of forecasting for each specific model, backtesting using Kupiec’s
proportion of failures test is applied. For each market, the model with
a lower number of violations is preferred. Our empirical result indicates the
usefulness of classical tempered stable distributions for market risk management
and asset pricing.
Keywords
Infinitely divisible models • Tempered stable distribution • GARCH models •
Value at risk • Kupiec’s proportion of failures test • Quasi-maximum likelihood
estimation strategy
48.1 Introduction
1
For a description of ARCH and GARCH modeling, see Chap. 8 in Rachev et al. (2007). The
chapter of the same reference describes ARCH and GARCH modeling with infinite variance
innovations. Engle et al. (2008) provide the basics of ARCH and GARCH modeling with
applications to finance.
48 A Quasi-Maximum Likelihood Estimation Strategy for Value-at-Risk Forecasting 1327
Despite the considerable amount of research in this area, there are no studies that
use trading volume in an effort to improve the capability of models to forecast
1-day-ahead VaR. Typically, in a VaR context, trading volume is only employed as
a proxy for “liquidity risk” – the risk associated with trying to close out a position.
In this paper, in contrast to prior studies, we analyze the impact of introducing
trading volume on the ability to enhance performance in forecasting VaR 1 day
ahead. We empirically test whether the introduction of trading volume will
reduce the number of violations (i.e., the number of times when the observed loss
exceeds the estimated one) in the spot and futures equity markets in the USA,
Germany, and Japan.
The remainder of this paper is organized as follows. ARMA-GARCH models
with normal and tempered stable innovations are reviewed in Sect. 48.2. In
Sect. 48.3, we discuss parameter estimation of the ARMA-GARCH models and
forecasting daily return distributions. VaR values and backtesting of the ARMA-
GARCH models are also reported in Sect. 48.2, along with a comparison of the
results for (1) the spot and futures markets and (2) the normal and tempered stable
innovations. Trading volume is introduced into the ARMA-GARCH model with
tempered stable innovations in Sect. 48.4. VaR and backtesting of the ARMA-
GARCH with different variants of trading volume are presented and compared to
the results for models with and without trading volume. We summarize our
principal findings in Sect. 48.5.
yt ¼ ayt1 þ bst1 et1 þ st et þ ct
: (48.1)
s2t ¼ a0 þ a1 s2 t1 e2 t1 þ b1 s2 t1
model, et stdCTS(a, l+, l_). This distribution does not have a closed-form
solution for its probability density function. Instead, it is defined by its character-
istic function as follows: Let a 2 (0,2)\{1},C, l+, l_ > 0, and m ∈ ℝ. Then
a random variable X is said to follow the classical tempered stable (CTS) distribu-
tion if the characteristic function of X is given by
fx ðuÞ ¼ fCTS u : a, C, lþ , l, m
¼ exp ium iuCT ð1 aÞ la1 þ l
a1
(48.2)
a a a a
þ CGðaÞ ðlþ iuÞ lþ þ ðl iuÞ l ,
1 ∂n
Cn ðXÞ ¼ log E eiuX ju ¼ 0, n ¼ 1, 2, 3, . . . : :
i ∂u
n n
For the tempered stable distribution, we have E[X] ¼ c1(X) ¼ m. The cumulants
of the tempered stable distribution for n ¼ 2, 3, . . . are
n an
cn ðXÞ ¼ CGðn aÞ lanþ þ ð1Þ l :
By substituting the appropriate value for the two parameters m and C into the
three tempered stable distributions, we can obtain tempered stable distributions
with zero mean and unit variance. That is, X CTS(a, C, l+, l, 0) has zero mean
and unit variance by substituting
a2 1
C ¼ Gð2 aÞ la2 þ þl : (48.3)
In this section, we discuss VaR for the ARMA-GARCH model with normal and
tempered stable innovations.
where Pt(A) is the conditional probability of a given event A for the information
until time t.
Two models are considered: normal-ARMA(1,1)-GARCH(1,1) and
stdCTS-ARMA(1,1)-GARCH(1,1). For both models, the parameters have been
estimated for the time series between December 14, 2004 and December
31, 2008. For each daily estimation, we worked with 10 years of historical daily
performance for the S&P 500, DAX 30, and Nikkei 225 spot and futures indices.
More specifically, we used daily returns calculated based on the closing price of
those indices. In the case of futures indices, we constructed a unique continuous
time series using the different maturities of each futures index following the
methodology proposed by Carchano and Pardo (2009).3 Then, we computed
VaRs for both models.
The maximum likelihood estimation method (MLE) is employed to
estimate parameters of the normal-ARMA(1,1)-GARCH(1,1) model. For the CTS
distribution, the parameters are estimated as follows4:
1. Estimate parameters a0, a1, b1, a, b, c with normal innovations by the
MLE. Volatility clustering is captured by the GARCH model.
2. Extract residuals using those parameters. The residual distribution still presents
fat tail and skewness.
3. Fit the parameters of the innovation distribution (CTS) to the extracted residuals
using MLE. The fat tailed and skewed features of the residual distribution are
captured.
In order to determine the accuracy of VaR for the two models, backtesting using
Kupiec’s proportion of failures test (Kupiec 1995) is applied. We first calculate the
number of violations. Then, we compare the number of violations with the
conventional number of exceedances at a given significance level. In Table 48.1 the
number of violations and p-values for Kupiec’s backtest for the three stock indices
over the 41-year periods are reported. Finally, we sum up the number of violations
and their related p-values for 1 % VaRs for the normal and CTS-ARMA-GARCH
models.
2
VaR on the CTS distribution is described in the Appendix.
3
Thus, the last trading day of the front contract is chosen as the rollover date. Then, the return of
the day after the rollover date is calculated as the quotient between the closing price of the
following contract and the previous closing price of such contract. By doing so, all the returns
are taken from the same maturity.
4
A quasi-MLE strategy is followed because the ARMA-GARCH CTS model has too many
parameters. If all the parameters are estimated at once, then the GARCH parameters go to zero.
This strategy is also followed in Kim et al. (2009, 2010, 2011). For a discussion of the quasi- MLE
methodology, see Rachev et al. (2007, pp. 292–293) or Verbeek (2004, pp. 182–184).
48 A Quasi-Maximum Likelihood Estimation Strategy for Value-at-Risk Forecasting 1331
Based on Table 48.1, we conclude the following for the three stock indices. First,
a comparison of the normal and tempered stable models indicates that there are no
cases using the tempered stable model at the 5 % significance level, whereas the normal
model is rejected five times. This evidence is consistent with the findings of Kim
et al. (2011). Second, a comparison of the spot and futures indices indicates that spot
data provide less than or the same number of violations than futures data. One potential
explanation is that futures markets are more volatile, particularly, when the market
falls.5 This overreaction to bad news could cause the larger number of violations.
5
We compared the spot and futures series when the markets discount bad news (negative returns).
We find that for the three stock indices, futures volatility is significantly greater than spot volatility
at a 5 % significance level. Moreover, for all three stock indices, the minimum return and the 1 %
percentile return are also lower for futures data than spot data.
1332 O. Carchano et al.
In the previous section, we showed the usefulness of the tempered stable model for
stock index futures. Motivated by the vast literature linking trading volume and
volatility, for the first time we investigate whether the introduction of trading
volume in the CTS model could improve its ability to forecast 1-day-ahead VaR.
Let (St)t0 be the asset price process and (yt)t0 be the return process of (St)t0
defined by yt ¼ log StS1
t
. We propose the following ARMA(1,1)-GARCH(1,1) with
trading volume model:
yt ¼ ayt1 þ bst1 et1 þ st et þ c
(48.4)
s2t ¼ a0 þ a1 s2 t1 e2 t1 þ b1 s2 t1 þ g1 Volt1 ,
where e0 ¼ 0 and a sequence (et)tcN ¼ 0 of iid real random variables. The innovation
et is assumed to be the tempered stable innovation. We will consider the standard
classical tempered stable distributions. This new ARMA(1,1)-GARCH(1,1)-V
model is defined as follows:
For the S&P 500 cash and futures markets, we test the following versions of trading
volume in order to determine which one would be the most appropriate:
• Lagged trading volume in levels: V(t 1)
• Logarithm of lagged trading volume: log [V(t 1)]
• Relative change of lagged trading volume: log [V(t 1)/V(t 2)]
48 A Quasi-Maximum Likelihood Estimation Strategy for Value-at-Risk Forecasting 1333
The spot series trading volume is in dollars; for the futures series, the trading value
is in number of contracts. We can calculate the volume of the futures market in dollars
too. The tick value of the S&P 500 futures contract is 0.1 index points or $25.
Multiplying the number of contracts by the price and finally by $250 (the contract’s
multiple), we obtain the trading volume series for the futures contract in dollars. Thus,
for the futures contract we get three new versions of trading volume to test:
• Lagged trading volume in dollars: V$(t 1)
• Logarithm of trading volume in dollars: log [V$(t 1)]
• Relative change of lagged trading volume in dollars: log [V$(t 1)/V$(t 2)]
By doing that, we can determine which series (in dollars or in contracts) seems to
be more useful for the futures index.
In Table 48.2 we report the number of violations and p-values of Kupiec’s backtest
for the different versions of the CTS-ARMA-GARCH-V model for the S&P 500 spot
1334 O. Carchano et al.
and futures indices. We count the number of violations and the corresponding
p-values for 1 %-VaRs of both markets. From Table 48.2, we conclude the following:
• The model with the lagged trading volume in level is rejected at the 1 %
significance level in all 4 years for the S&P 500 spot, and for the second period
(2005–2006) for the S&P 500 futures.
• The logarithm of trading volume in the model is rejected at the 5 % significance
level for the spot market for the third period (2006–2007), but it is not rejected in
any period for the futures market.
• The relative change of the lagged volume is not rejected at the 5 % significance
level in any period in either market. Of the three versions of trading volume tests,
this version seems to be the most useful for both spot and futures markets.
• The results for trading volume in contracts and the trading volume in dollars in the
futures market indicate that the former is rejected at the 1 % significance level only
for the lagged trading volume in level in the second period (2005–2006). Trading
volume in dollars is rejected three times, for the lagged trading volume in levels
for the third period (2006–2007) and for the lagged relative trading volume change
in the last two periods (2006–2007 and 2007–2008). These findings suggest that
the trading volume in contracts is the preferred measure.
As we have just seen, the variant of trading volume that seems more useful
for forecasting 1-day-ahead VaR using CTS-ARMA-GARCH is the relative change
of trading volume. Next, we compare the original CTS-ARMA-GARCH model
with the new CTS-ARMA-GARCH-V model where V is the lagged relative change
of trading volume. Table 48.3 shows the number of violations and p-values of
Kupiec’s backtest for the two models for the three stock indices and both markets.
We sum up the number of violations and the corresponding p-values for 1 % VaRs for
each case.
Our conclusions from Table 48.3 are as follows. For the spot markets, the
introduction of trading volume does not mean a reduction in the number of
violations in any period for any index. However, for the futures markets, the
numbers of violations are the same or lower for the model with trading
volume than with the original model. Thus, by introducing trading volume, we
get a slightly more conservative model, increasing the VaR forecasted for futures
equity markets.
Although there is some evidence which supports the relationship between lagged
trading volume and volatility, the literature is not as extensive as the studies
that establish a strong link between volatility and contemporaneous trading
48 A Quasi-Maximum Likelihood Estimation Strategy for Value-at-Risk Forecasting 1335
volume. As there are countless ways to try to forecast trading volume, we begin by
introducing contemporaneous trading volume relative change in the model as
a benchmark to assess whether it is worthwhile to forecast trading volume.
In Table 48.4 we show the number of violations and p-values of Kupiec’s
backtest for the CTS-ARMA-GARCH with contemporaneous and lagged
relative change of trading volume for the three stock indices for both markets.
We count the number of violations and the corresponding p-values for 1 % VaRs for
the six indices.
Our conclusions based on the results reported in Table 48.4 are as follows.
First, with the exception of the S&P 500 futures, the introduction of the contem-
poraneous relative change of trading volume in the model is rejected at the 1 %
significance level for the last period analyzed (2007–2008). In the case of the S&P
1336 O. Carchano et al.
500 futures, it is rejected at the significance level of 5 % for the third period
(2006–2007). Second, the model with lagged relative change of trading volume is
not rejected for any stock index or market. It seems to be more robust than
contemporaneous trading volume (although, in general, there are fewer violations
when using it).
48 A Quasi-Maximum Likelihood Estimation Strategy for Value-at-Risk Forecasting 1337
Our results suggest that it is not worth making an effort to predict contemporane-
ous trading volume because the forecasts will be flawed and two variables
would have to be predicted (VaR and contemporaneous trading volume). Equiva-
lently, the lagged trading volume relative change appears to be more robust because it
is not rejected in any case, although it provides a poor improvement to the model.
48.5 Conclusions
Based on an empirical analysis of spot and futures trading for the S&P 500, DAX
30, and Nikkei 225 stock indices, in this paper we provide empirical evidence
about the usefulness of using classical tempered stable distributions for predicting
1-day-ahead VaR. Unlike prior studies that investigated CTS models in the
cash equity markets, we analyzed their suitability for both spot markets and
futures markets. We find in both markets the CTS models perform better in forecast-
ing 1-day-ahead VaR than models that assume innovations follow the normal law.
Second, we introduced trading volume into the CTS model. Our empirical
evidence suggests that lagged trading volume relative change provides a slightly
more conservative model (i.e., reduces the number of violations) to predict
1-day-ahead VaR for stock index futures contracts. We cannot state the same
for the cash market because the results are mixed depending on the index. After
that, we introduced contemporaneous trading volume to try to improve the forecast-
ing ability of the model, but in the end, it did not seem to be worth the effort. That is,
trading volume appeared not to offer enough information to improve forecasts.
Finally, we compared the number of violations of the estimated VaR in the spot
and futures equity markets. For the CTS model without volume, in general, we find
fewer violations in the spot indices than in the equivalent futures contracts.
In contrast, our results suggest that the number of violations in futures markets
is less in the case of the CTS model with trading volume in comparison to the
CTS model that ignores trading volume. But if we contrast spot and futures equity
markets, violations are still greater for futures than in spot markets. A possible
reason is that futures markets demonstrate extra volatility or an overreaction when
the market falls with respect to their corresponding spot markets.
Let X be a CTS random variable. Since the CTS random variable is continuous and
infinitely divisible, we obtain VaR(X) ¼ FX(), where the cumulative distribu-
tion function FX of X is provided by the following proposition.
ð /
exr ixu fX ðu þ irÞ
FX ð x Þ ¼ ℜ e du , for x 2 ℝ (48.5)
p 0 r ui
where fX(x) is the density function of X. The probability density function fX(t) can be
obtained from the characteristic function fX by the complex inverse formula (see
Doetsch 1970); that is,
ð 1þir
1
f X ðt Þ ¼ eitz fX ðzÞdz,
2p 1þir
and we have
ðx ð 1þip
1
FX ð x Þ ¼ eitz fX ðzÞdzdt
1 2p1þip
ð 1þip ð x
1
¼ eitz dtfX ðzÞdz:
2p 1þip 1
it
lim eitðaþirÞ ¼ lim e ðaþirÞ ¼ lim ert ¼ 0, a 2 ℝ,
t!1 t!1 t!1
and hence
ðx
1 itz x 1
eitz dt ¼ e 1
¼ eixz
1 iz iz
Let
fX ðu þ irÞ
gr ð uÞ ¼ :
r iu
Then we can show that gr ðuÞ ¼ gr ðuÞ with u ∈ ℝ and hence we have
ð1
ð 1
eixu gr ðuÞdu ¼ 2ℜ eixu gr ðuÞdu :
1 0
References
Asai, M., & McAleer, M. (2009). The structure of dynamic correlations in multivariate stochastic
volatility models. Journal of Econometrics, 150, 182–192.
Belsley, D., Kuh, E., & Welsch, R. (1980). Regression diagnostics: Identifying influential data and
sources of collinearity. New York: Wiley.
Bollerslev, T. (1986). Generalized autoregressive conditional heteroscedasticity. Journal of
Econometrics, 31, 307–327.
Carchano, O., & Pardo, A. (2009). Rolling over stock index futures contracts. Journal of Futures
Markets, 29, 684–694.
Clark, P. K. (1973). A subordinated stochastic process model with finite variance for speculative
prices. Econometrica, 41, 135–155.
Conrad, J. S., Hameed, A., & Niden, C. (1994). Volume and autocovariances in short-horizon
individual security returns. Journal of Finance, 49, 1305–1329.
Copeland, T. E. (1976). A model of asset trading under the assumption of sequential information
arrival. Journal of Finance, 34, 1149–1168.
Doetsch, G. (1970). Introduction to the theory and application of the Laplace transformation.
New York: Springer.
Engle, R. (1982). Autoregressive conditional heteroskedasticity with estimates of the variance
with estimates of the variance of United Kingdom inflation. Econometrica, 50, 987–1007.
Engle, R., Focardi, S. M., & Fabozzi, F. J. (2008). ARCH/GARCH Models in applied financial
econometrics. In F. J. Fabozzi (Ed.), Handbook of finance. Hoboken: Wiley. vol III.
Epps, T. W., & Epps, L. M. (1976). The stochastic dependence of security price changes and
transaction volumes: Implications for the mixture-of-distribution hypothesis. Econometrica,
44, 305–321.
Foster, F. J. (1995). Volume-volatility relationships for crude oil futures markets. Journal of
Futures Markets, 15, 929–951.
Fung, H. G., & Patterson, G. A. (1999). The dynamic relationship of volatility, volume, and market
depth in currency futures markets. Journal of International Financial Markets Institutions and
Money, 17, 33–59.
Garcia, P., Leuthold, M. R., & Zapata, H. (1986). Lead-lag relationships between trading volume
and price variability: New evidence. Journal of Futures Markets, 6, 1–10.
Gwilym, O., MacMillan, D., & Speight, A. (1999). The intraday relationship between volume and
volatility in LIFFE futures markets. Applied Financial Economics, 9, 593–604.
Huang, C. Y., & Lin, J. B. (2004). Value-at-risk analysis for Taiwan stock index futures: Fat tails
and conditional asymmetries in return innovations. Review of Quantitative Finance and
Accounting, 22, 79–95.
1340 O. Carchano et al.
Karpoff, J. M. (1987). The relation between price changes and trading volume: A survey. Journal
of Financial and Quantitative Analysis, 22, 109–126.
Kim, Y. S., Rachev, S. T., Chung, D. M., & Bianchi, M. L. (2009). The modified tempered stable
distribution, GARCH-models and option pricing. Probability and Mathematical Statistics, 29,
91–117.
Kim, Y. S., Rachev, S. T., Bianchi, M. L., & Fabozzi, F. J. (2010). Tempered stable and tempered
infinitely divisible GARCH models. Journal of Banking and Finance, 34, 2096–2109.
Kim, Y. S., Rachev, S. T., Bianchi, M. L., Mitov, I., & Fabozzi, F. J. (2011). Time series analysis
for financial market meltdowns. Journal of Banking and Finance, 35, 1879–1891.
Kupiec, P. (1995). Techniques for verifying the accuracy of risk measurement models. Journal of
Derivatives, 6, 6–24.
Ragunathan, V., & Peker, A. (1997). Price variability, trading volume and market depth: Evidence
from the Australian futures market. Applied Financial Economics, 7, 447–454.
Schwert, G. W. (1989). Why does stock market volatility change over time? Journal of Finance,
44, 1115–1153.
Smirlock, M., & Starks, L. (1985). A further examination of stock price changes and transactions
volume. Journal of Financial Research, 8, 217–225.
Tang, L. T., & Shieh, J. S. (2006). Long memory in stock index futures markets: A value-at-risk
approach. Physica A: Statistical Mechanics and its Applications, 366, 437–448.
Tauchen, G., & Pitts, M. (1983). The price variability-volume relationship on speculative markets.
Econometrica, 51, 485–505.
Verbeek, M. (2004). A guide to modern econometrics. Hoboken: Wiley.
Wang, G., & Yau, J. (2000). Trading volume, bid-ask spread, and price volatility in futures
markets. Journal of Futures Markets, 20, 943–970.
Computer Technology for Financial
Service 49
Fang-Pang Lin, Cheng-Few Lee, and Huimin Chung
Contents
49.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1342
49.1.1 Information Technology (IT) for Financial Services . . . . . . . . . . . . . . . . . . . . . . . . 1342
49.1.2 Competitiveness Through IT Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1343
49.2 Performance Enhancement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1346
49.2.1 High-End Computing Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1346
49.2.2 Compute Intensive IT Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1349
49.2.3 Data-Intensive IT Systems . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1352
49.3 Distributed and Parallel Financial Simulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1354
49.3.1 Financial Simulation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1355
49.3.2 Monte Carlo Simulations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1357
49.3.3 Distribution and Parallelism Based on Random Number Generation . . . . . . . 1360
49.4 Case Study and Discussions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1367
49.4.1 Case Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1367
49.4.2 Grid Platforms Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1369
49.5 Conclusions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1375
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1377
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1341
DOI 10.1007/978-1-4614-7750-1_49,
# Springer Science+Business Media New York 2015
1342 F.-P. Lin et al.
Abstract
Securities trading is one of the few business activities where a few seconds processing
delay can cost a company big fortune. The growing competition in the market
exacerbates the situation and pushes further towards instantaneous trading even in
split second. The key lies on the performance of the underlying information system.
Following the computing evolution in financial services, it was a centralized process
to begin with and gradually decentralized into a distribution of actual application logic
across service networks. Financial services have tradition of doing most of its heavy-
duty financial analysis in overnight batch cycles. However, in securities trading it
cannot satisfy the need due to its ad hoc nature and requirement of fast response. New
computing paradigms, such grid and cloud computing, aiming at scalable and
virtually standardized distributed computing resources, are well suited to the chal-
lenge posed by the capital market practices. Both consolidate computing resources by
introducing a layer of middleware to orchestrate the use of geographically distributed
powerful computers and large storages via fast networks. It is nontrivial to harvest the
most of the resources from this kind of architecture. Wiener process plays a central
role in modern financial modeling. Its scaled random walk feature, in essence, allows
millions of financial simulation to be conducted simultaneously. The sheer scale can
only be tackled via grid or cloud computing. In this study the core computing
competence for financial services is examined. Grid and cloud computing will be
briefly described. How the underlying algorithm for financial analysis can take
advantage of grid environment is chosen and presented. One of the most popular
practiced algorithms Monte Carlo simulation is used in our case study for option
pricing and risk management. The various distributed computational platforms are
carefully chosen to demonstrate the performance issue for financial services.
Keywords
Financial service • Grid and cloud computing • Monte Carlo simulation • Option
pricing • Risk management • Cyberinfrastructure • Random number generation •
High-end comptuing • Financial simulation • Information technology
49.1 Introduction
The finance services industry involves a broad range of organizations such as banks,
credit card companies, insurance companies, consumer finance companies, stock
brokerages, investment funds, and some government-sponsored enterprises. The
industry represents a significant share of the global market. Information technology
(IT) in the financial service industry is considered as an indispensable tool for
productivity as well as competitiveness in the market. The IT spending in financial
service industry grows constantly across different industry verticals (banking,
insurance, and securities and investments). The impact directly from the use of
advanced IT brings on financial services industry on the rise.
49 Computer Technology for Financial Service 1343
The structure of the industry has changed significantly in the last two decades as
companies, which are not traditionally viewed as financial service providers, have
taken advantage of opportunities created by technology to enter the market. New
technology-based services keep emerging. These changes are direct result of the
interaction of technology with the industrial environment, such as economic atmo-
sphere, societal pressures, and the legal/regulatory environment in which the
financial service industry operates. The effects of IT on the internal operations,
the structure, and the types of services offered by the financial service industry have
been particularly profound (Phillips et al. 1984; Hauswald and Marquez 2003;
Griffiths and Remenyi 2003). IT technology has been and continues to be both
a motivator and facilitator of change in the financial service industry, which
ultimately leads to competitiveness of the industry. The change is in particular
radical after 1991 when the World Wide Web was invented by Tim Berners-Lee
and his group for information sharing in the community of high energy physics. It
was later introduced to the rest of the world, which subsequently changed the face
of how people doing business today.
Informational considerations have long been recognized to determine not only the
degree of competition but also the pricing and profitability of financial services and
instruments. Recent technological progress has dramatically affected the production
and availability of information, thereby changing the nature of competition in such
informationally sensitive markets. Hauswald and Marquez (2003) investigate how
advances in information technology (IT) affect competition in the financial services
industry, particularly credit, insurance, and securities markets. Two aspects of
improvement in IT are focused: better processing and easier dissemination of infor-
mation. In other words, two dimensions of technology progress that affect competi-
tion in financial services can be defined as advances in the ability to process and
evaluate information and in the ease of obtaining information generated by compet-
itors. While better technology may result in improved information processing, it
might also lead to low cost or even free access to information through, for example,
informational spillovers. They show that in the context of credit screening, better
access to information decreases interest rates and the returns from screening. On the
other hand, an improved ability to process information increases interest rates and
bank profits. Hence predictions regarding financial claims’ pricing hinge on the
overall effect ascribed to technological progress. Their results conclude that in
general financial market informational asymmetries drive profitability.
The viewpoint of Hauswald and Marquez is adopted in this work. Assuming
competitors in the dynamics of financial market possess similar capacity, the infor-
mational asymmetries can be created sometimes only between seconds and now are
possible to be achieved through the outperformance of underlying IT platforms.
0
2004 2005 2006 2007 2008 2009 2010 2011 2012
News reference volume
Fig. 49.1 The trend history from Google Trend according to global Search Volume and global
News Reference Volume, in which the alphabetic letters represent the specific events that relate to
each curve
trading application logic across service networks. Financial services have tradition
of doing most of its heavy lifting financial analysis in overnight batch cycles.
However, in securities trading it cannot satisfy the need due to its ad hoc nature
and requirement of fast response.
New computing paradigms, grid computing and cloud computing were subse-
quently emerged in the last decade. The grid computing was initially incorporated
into the core context of a well-referenced Atkins’ report of National Science Board
of the United States, namely, “Revolutionizing Science and Engineering Through
Cyberinfrastructure” (Atkins et al. 2003), which lays down a visionary path for
future IT platform development of the world. One may observe this trend from
statistics from Google Trend regarding the global Search Volume and global News
Reference Volume of key phrases of “cluster computing,” “grid computing,”
“cloud computing,” and “Big Data” (Fig. 49.1), which represents four main stream
computing paradigms in high-end quantitative analysis.
Cluster computing is a group of coupled computers that work closely together so
that in many respects they can be viewed as though they are a single computer.
They are connected with high-speed local area networks and the purpose is usually
to gain more compute cycles with better cost performance and higher availability.
The grid computing aims at virtualizing scalable geographically distributed com-
puting and observatory resources to maximize compute cycles and data transaction
rates with minimum cost. Cloud computing is more of recent development owing to
the similar technology used in global information services providers, such as
Google and Amazon. The cloud is referred to as a subset of Internet if to be
explained in a simplest fashion. Within the cloud the computers also talk with
servers instead of communicating with each other similarly to that of peer-to-peer
computing (Milojicic et al. 2002). There are no definitive definitions for the above
49 Computer Technology for Financial Service 1345
In this section two types of grid systems, compute intensive and data intensive,
respectively, are introduced. The classification of the types is based on various grid
applications. Traditionally, the grid systems provide a general platform to harvest or to
scavenge, if used only in idle status, compute cycles for a collection of resources
across boundaries of institutional administration. In real world most applications are in
fact data centric. For example, in a trading center, it collects tick-by-tick volume data
from all related financial markets and is driven by informational flows, hence typical
data centric. However, as noted in Sect. 49.3.2.1, the core competence still lies on the
performance enhancement of the IT system. The following two subsections will give
more details of compute intensive as well as data-intensive grid systems by a survey of
current development of grids specifically for financial services. In some cases, e.g.,
high-frequency data with real-time analysis, two systems have to work together to get
better performance. Our emphasis will be more on compute intensive grid system.
ProActive
Server Sub- ProActive Worker
Client
Server
ProActive
JavaSpace
virtual shared
DB memory (to v3)
Fig. 49.2 Architecture of PicsouGrid for option pricing based on Monte Carlo simulation
(Stokes-Rees et al. 2007)
2.Distribute Slave 0
Main CoG Java SSL
Cluster the tasks
1.Send service Data Provider
at location X
request Text Slave 1 (Reuters News
Processing Globus Triarch SSL
Client Source)
Service
4.Notify the user Slave 2
about the results Time Series “Machines” “Machines”
Service 3.Receive Data Provider (FinGrid) (Reuters)
Slave 3 at location Y
results (Numerical
Data)
facilitates the execution of parallel tasks in the main cluster and is distributed to
a set of slave machines (nodes), and the third tier comprises the connection of the
slave machines to the data providers. This work focuses on small scale and
dedicated grid system. It pumps in real and live numerical and textual data from
say Reuters and performs real-time sophisticated data mining analysis. This is
a good prototype for financial grid. However, it will encounter similar problem
as that of PicsouGrid if it is to scale up. The model is more successful in
automatically combining real data and the analysis.
3. IBM Japan collaborates with life insurance company and adopts PC grids
concept to scavenge more compute cycles (Tanaka 2003):
In this work an integrated risk management system (see Fig. 49.4) is modi-
fied, in which the future scenarios of red circle of Fig. 49.4 are send via grid
middleware to a cluster of PCs. According to the size of the given PCs, the
49 Computer Technology for Financial Service 1351
PV of each asset
Basic equations
Future scenarios
(Stochastic
differential eqs.)
Scenarios
Default-free
interest rates
(domestic. foreign)
1 2 3 N
Exchange rate
Distribution of future value of
portfolio (individual assets)
Return valuation
Risk valuation
number scenarios are then divided in a work balanced manner for each PC. This
is the most typical use of compute intensive grid systems and a good practice for
production system. However, the key issues that discussed in the above two
cases cannot be answered in this study. Similar architecture can also be found in
EGrid (Leto et al. 2005).
4. UK e-Science developed a grid service discovery in the financial market sector
focusing on integration of different knowledge flows (Bell and Ludwig 2005).
From application’s viewpoint, business and technical architecture of financial
service applications may be segmented by product, process, or geographic con-
cerns. Segmented inventories make inter silo reuse difficult. The service integration
model is adopted and a loosely coupled inventory – containing differing explicit
1352 F.-P. Lin et al.
SEDI4G
COMPONENT
REPOSTORY
1
SCR
(Apache JNDI or
MDS)
SEDI4G
SEDI4G CLIENT DISCOVERY SEDI4G MATCHING
4 3
VIEW SCV CONTROL SERVICE SMAS
SERVICE SDCS
5
Ontology
6
Described in
Fig. 49.5 The semantic discovery for Grid Services Architecture (SEDI4G) (Bell and Ludwig
2005)
capability knowledge. Three use cases were specifically chosen in this work to
explore the use of semantic searching:
Use case 1 – Searching for trades executed with a particular counterparty
Use case 2 – Valuing a portfolio of interest rate derivative products
Use case 3 – Valuing an option-based product
The use cases were chosen to provide examples of three distinct patterns of use –
aggregation, standard selection, and multiple selection. The architecture (see
Fig. 49.5) is bound specifically with the user cases. The advantage for grid in this
case is that it can be easily tailored into specific user need to integrate different
applications, which is a crucial strength of using grid.
Grid in financial services from the perspective of web services towards financial
services industry. The perspective is more on transactional side. Once the bottle-
neck of compute cycle is solved, the data-centric nature will play the key role again.
49 Computer Technology for Financial Service 1353
The knowledge flows back to the customized business logic should provide the best
path for users to access the live data of interest. There is no strong focus of
development on this data-intensive grid system. Even in FinGrid, which claims in
streaming live data for real-time analysis, the data issue remains part of compute
grids. However, the need for dynamic data management is obvious as mentioned in
Amad et al. (2004). Hereby, we like to introduce and implement a dynamic data
management software Ring Buffer Network Bus (RBNB) DataTurbine to serve
such a purpose.
RBNB DataTurbine was used recently to support global environmental obser-
vatory network, which involves linking with ten of thousand of sensors and is able
to obtain the observed data online. It meets grid/cyberinfrastructure
(CI) requirements with regard to data acquisition, instrument management, and
state-of-health monitoring including reliable data capture and transport, persistent
monitoring of numerous data channels, automated processing, event detection and
analysis, integration across heterogeneous resources and systems, real-time tasking
and remote operations, and secure access to system resources. To that end, stream-
ing data middleware provides the framework for application development and
integration.
Use cases of RBNB DataTurbine include adaptive sampling rates, failure detec-
tion and correction, quality assurance, and simple observation (see Tilak
et al. 2007). Real-time data access can be used to generate interest and buy-in
from various stakeholders. Real-time streaming data is a natural model for many
applications in observing systems, in particular event detection and pattern recog-
nition. Many of these applications involve filters over data values, or more gener-
ally, functions over sliding temporal windows. The RBNB DataTurbine
middleware provides a modular, scalable, robust environment while providing
security, configuration management, routing, and data archival services. The
RBNB DataTurbine system acts as an intermediary between dissimilar data mon-
itoring and analysis devices and applications. As shown in Fig. 49.6, a modular
architecture is used, in which a source or “feeder” program is a Java application that
acquires data from an external live data sources and feeds it into the RBNB server.
Additional modules display and manipulate data fetched from the RBNB server.
This allows flexible configuration where RBNB serves as a coupling between
relatively simple and “single purpose” suppliers of data and consumers of data,
both of which are presented a logical grouping of physical data sources. RBNB
supports the modular addition of new sources and sinks with a clear separation of
design, coding, and testing (ref. Fig. 49.6). From the perspective of distributed
systems, the RBNB DataTurbine is a “black box” from which applications and
devices send data and receive data. RBNB DataTurbine handles all data manage-
ment operations between data sources and sinks, including reliable transport,
routing, scheduling, and security. RBNB accomplishes this through the innovative
use of memory and file-based ring buffers combined with flexible network objects.
Ring buffers are a programmer-configurable mixture of memory and disk, allowing
system tuning to meet application-dependent data management requirements. Net-
work bus elements perform data stream multiplexing and routing. These elements
1354 F.-P. Lin et al.
Client
Live Data
DAV
Source
TCP/IP
Client
HTTP
Client
TCP/IP
Live Data
Source
API
RBNB
Client
Feeder
Client
Plugin
Plugin
combine to support seamless real-time data archiving and distribution over existing
local and wide area networks. Ring buffers also connect directly to client applica-
tions to provide streaming-related services including data stream subscription,
capture, rewind, and replay. This presents clients with a simple, uniform interface
to real-time and historical (playback) data.
There are wide variety of sophisticated financial models developed, to name a few,
ranging from analysis in time series, fractals, nonlinear dynamics, and agent-based
modeling to applications in optional pricing, portfolio management, and market risk
measure, etc. (Schmidt 2005), in which option pricing and VaR calculations of
market risk measure can be considered crucial and one of the most practiced
activities in market trading.
a European-style Put Option (or European Put). If it can be exercised any date
before its maturity, the option is referred to as an American-style Put Option
(or American Put).
To price options in computational finance, we use the following notation: K is
the strike price; T is the time to maturity of the option; St is the stock price at time t;
r is the risk-free interest rate; m is the drift rate of the underlying asset (a measure of
the average rate of growth of the asset price); s is the volatility of the stock; and
V denotes the option value. Here is an example to illustrate the concept of option
pricing. Suppose an investor enters into a call option contract to buy a stock at price
K after 3 months. After 3 months, the stock price is St. If St > K then one
can exercise one’s option by buying the stock at price K and by immediately
selling in the market to make a profit of ST K. On the other hand, if he
ST K to be buy the stock. Hence, we see that a call option to buy the stock at
time T at price K will get payoff (ST K)+, where (ST K)+ max(ST K, 0)
(Schmidt 2005; Hull 2003).
shortfall (ES), also called expected tail loss (ETL) or conditional VaR, is the
expected value of the losses in excess of VaR:
XN
^I ¼ 1 f ðxi Þ (49.4)
N i¼1
where B is a subcube of [0, 1]d containing the origin, A(B, qn) is the number of
points in qn that fall into B, and Vd(B) is the d-dimensional Lebesgue measure of B.
The elements of qn are said uniformly distributed if its discrepancy D*N ! 0 as
N ! 1. From the theory of uniform distribution sequences (Kuipers and
Niederreiter 1974), the estimate of the integral using a uniformly distributed
XN
sequence {qn} is ^I ¼ N1 f ðqn Þ, as N ! 1 then ^I ! I . The integration error
n¼1
bound is given by the Koksman-Hlawka inequality:
1X N
I f ðqn Þ V ðf ÞDN ðqn Þ (49.7)
N n¼1
where V(f) is the variation of the function in the sense of Hardy and Krause (see
Kuipers and Niederreiter 1974), which is assumed to be finite.
The inequality suggests a smaller error can be obtained by using sequences
with smaller discrepancy. The discrepancy of many uniformly distributed
sequences satisfies O((log N)d/N). These sequences are called low-discrepancy
(LD) sequences (Chen et al. 2006). Inequality (49.7) shows that the estimates
using a LD sequence satisfy the deterministic error bound O((log N)d/N).
where W is a standard Wiener process, also known as Brownian motion. Under the
risk-neutral measure, the drift m is set to m ¼ r.
To simulate the path followed by S, suppose the life of the option has been
divided into n short intervals of length Dt(Dt ¼ T/n), the updating of the stock price
at t + Dt from t is (Hull 2003):
pffiffiffiffiffi
StþDt St ¼ rSt Dt þ sSt Z Dt (49.9)
49 Computer Technology for Financial Service 1359
where Z is a standard random variable, i.e., Z(0,1). This enables the value of SDt to
be calculated from initial value St at time Dt, the value at time 2Dt to be calculated
from SDt, and so on. Hence, a completed path for S has been constructed.
In practice, in order to avoid discretization errors, it is usual to simulate lnS
rather than S. From It ^o’s lemma, the process followed by of Eq. 49.9 is (Bratley and
Fox 1988)
s2
dlnS ¼ r dt þ sdz (49.10)
2
so that
s2 pffiffiffiffiffi
lnStþDt lnSt ¼ r dt þ sZ Dt (49.11)
2
or equivalently
s2 pffiffiffiffiffi
StþDt ¼ St exp r dt þ sZ Dt (49.12)
2
1X n
1X n h i
V¼ Vi ¼ erT max ST ðiÞ K, 0 (49.13)
n i¼1 n i¼1
The QMC simulations follow the same steps as the MC simulations, except that
the pseudorandom numbers are replaced by LD sequences. The basic LD sequences
known in literature are Halton (1960), Sobol (1967), and Faure (1982). Niederreiter
(1992) proposed a general principle of generating LD sequences. In finance, several
examples have shown that the Sobol sequence is superior to others. For example,
Galanti and Jung (1997) observed that the Sobol sequence outperforms the Faure
sequence, and the Faure marginally outperforms the Halton sequence. In this
research, we use Sobol sequence in our experiments. The generator used for
generating the Sobol sequence comes from the modified algorithm 659 of Joe and
Kuo (2003).
The bootstrap (Efon 1981; Efron and Tibshirani 1986) is a simple and straight-
forward method for calculating approximated biases, standard deviations, confi-
dence intervals, and so forth, in almost any nonparametric estimation problem.
Method is a keyword here, since little is known about the bootstrap’s theoretical
basis, except that (a) it is closely related to the jackknife in statistic inferring;
(b) under reasonable condition, it gives asymptotically correct results; and (c) for
some simple problems which can be analyzed completely, for example, ordinary
linear regression, the bootstrap automatically produces standard solutions.
The bootstrap method is straightforward. Suppose we observe returns Xi ¼ xi,
i ¼ 1, 2, , n, where the Xi are independent and identically distributed (iid) according
to some unknown probability distribution F. The Xi may be real valued and
two-dimensional or take values in a more complicated space. A given parameter y(F),
perhaps the mean, median, correlation, and so forth, is to be estimated, and we agree
to use the estimate ^y ¼ y F ^ , where F^ is the empirical distribution function putting
mass 1/n at each observed value xi. We wish to assign some measure of accuracy to ^y.
Let s(F) be some measure
of accuracy that we would use if F were known, for
example, sðFÞ ¼ SDF ^y , the standard deviation ^
of y when X1, X2, , Xn F .
(idd)
other words, X*1, X*2, , X*n is an independent sample of size n drawn with replace-
ment from the set of observations {x1,x2, , xn}. (b) This gives a bootstrap empirical
distribution function F ^ , the empirical distribution
of the n values X*1, X*2, , X*n,
and a corresponding bootstrap value ^y ¼ y F ^ . (c) Steps (a) and (b) are
repeated, independently, in a large number of times, say N, giving bootstrap values
^ 1 , y
y ^ 2 , , y
^ N . (d) The value of s ^ is approximated, in the case where s(F) is the
standard deviation by the sample standard deviation of the ^y values, where
Xn j
^¼
m ^y
j¼1
(49.14)
N
and
Xn
j 2
^y m
^
j¼1
^2 ¼
s (49.15)
N1
Eqs. 49.8 and 49.9 for approximating the underlying prices St+Dt, or the boot-
strap samples of return X*i , the solution methods involve basic market parame-
ters, drift m, volatility s, and risk-free interest rate r, current underlying price
S or return X, strike price K, and Wiener process, which isprelatedffiffiffiffiffi to time to
maturity Dt and standard random variable Z, i.e., DW ¼ Z Dt . Monte Carlo
methods simulate this nature of the Brownian motion directly. It follows Wiener
process and approximates the standard random variable Z by introducing pseudo
iid random number into each Wiener process. When the simulation number is
large enough, e.g., if n in Eq. 49.13 is large enough, the mean value will
approach the exact solution. The large number for n also implies the perfor-
mance problems are the key problems for Monte Carlo methods. One the other
hand, the iid property of the random number Z shows possible solution to tackle
the performance problem through mass distribution and/or parallelism. The
solution method centers on the random number generation.
The techniques of random number generation can be developed in a simple form
through the approximation of a d-dimensional integral, e.g., (Eq. 49.3). Mass
distribution and parallelism required solutions of for large dimension. However,
most modern techniques in random number generation have limitations. In this
study, both tradition pseudorandom number generation and high-dimensional
low-discrepancy random number generator are considered.
Following Sect. 49.3.2.1 better solution can be achieved by making use of Sobol
sequences, which were proposed by Sobol (1967). A computer implementation
in Fortran 77 was subsequently given by Bratley and Fox (1988) as Algorithm 659.
Other implementations are available as C, Fortran 77, or Fortran 90 routines in the
popular Numerical Recipes collection of software. However, as given, all these
implementations have a fairly heavy restriction on the maximum value of
d allowed. For Algorithm 659, Sobol sequences may be generated to approximate
integrals in up to 40 dimensions, while the Numerical Recipes routines allow the
generation of Sobol sequences to approximate integrals in up to six dimensions
only. The FinDer software of Paskov and Traub (1995) provides an implementation
of Sobol sequences up to 370 dimensions, but it is licensed software. As computers
become more powerful, there is an expectation that it should be possible to
approximate integrals in higher and higher dimensions. Integrals in hundreds of
variables arise in applications such as mathematical finance (e.g., see Paskov and
Traub (1995)). Also, as new methods become available for these integrals, one
might wish to compare these new methods with Sobol sequences. Thus, it would be
desirable to extend these existing implementations such as Algorithm 659 so they
may be used for higher-dimensional integrals. We remark that Sobol sequences are
now considered to be examples of (t, d)-sequences in base 2. The general theory of
these low-discrepancy (t, d)-sequences in base b is discussed in detail in
Niederreiter (1992). The generation of Sobol sequences is clearly explained in
Bratley and Fox (1988). We review the main points so as to show what extra data
would be required to allow Algorithm 659 to generate Sobol sequences to approx-
imate integrals in more than 40 dimensions. To generate the j th component of the
1362 F.-P. Lin et al.
for k
sj + 1, where is the bit-by-bit exclusive-OR operator. The initial values
m1, j , m2, j , , msj , j can be chosen freely provided that each Mk,j, 1 k sj is odd
and less than 2k. The “direction numbers” {v1,j,v2,j, } are defined by
mk, j
v1, j (49.18)
2k
Then xi,j, the j th component of the ith point in a Sobol sequence, is given by
x i , j ¼ b1 v 1 , j b2 v 2 , j (49.19)
Where bl is the lth bit from the right when i is written in binary, that is, ( b2b1)2
is the binary representation of i. In practice, a more efficient Gray code implemen-
tation proposed by Antonov and Saleev (1979) is used; see Bratley and Fox (1988)
for details. We then see that the implementation in Bratley and Fox (1988) may be
used to generate Sobol sequences to approximate integrals in more than 40 dimen-
sions by providing more data in the form of primitive polynomials and direction
numbers (or equivalently, values of m1, j , m2, j , , msj , j ). When generating such
Sobol sequences, we need to ensure that the primitive polynomials used to generate
each component are different and that the initial values of the mk,j’s are chosen
differently for any two primitive polynomials of the same degree. The error bounds
for Sobol sequences given in Sobol (1967) indicate we should use primitive poly-
nomials of as low a degree as possible. We discuss how additional primitive
polynomials may be obtained in the next section. After these primitive polynomials
have been found, we need to decide upon the initial values of the mk,j for 1ksj.
As explained above, all we require is that they be odd and that mk,j <2k. Thus, we
could just choose them randomly, subject to these two constraints. However, Sobol
and Levitan (1976) introduced an extra uniformity condition known as Property A.
Geometrically, if the cube [0, 1]d is divided up by the planes xj ¼ 1/2 into 2d equally
sized subcubes, then a sequence of points belonging to [0, 1]d possesses Property
A if, after dividing the sequence into consecutive blocks of 2d points, each one of
49 Computer Technology for Financial Service 1363
the points in any block belongs to a different subcube. Property A is not that useful
to have for large d because of the computational time required to approximate an
integral using 2d points. Also, Property A is not enough to ensure that there are no
bad correlations between pairs of dimensions. Nevertheless, Property A would
seem a reasonable criterion to use in deciding upon a choice of the initial mk,j.
The numerical results for Sobol sequences given in Sect. 49.4 suggest that the
direction numbers obtained here are indeed reasonable. Other ways of obtaining the
direction numbers are also possible. For example, in Cheng and Druzdzel (2000),
the initial direction numbers are obtained by an interesting technique of minimizing
a measure of uniformity in two dimensions. This technique may alleviate the
problem of bad correlations between pairs of dimensions that was mentioned
above. Sobol (1967) showed that a Sobol sequence used to approximate
a d-dimensional integral possesses Property A if and only if
2 3
v1, 1, 1 v2, 1, 1 vd , 1, 1
6 v1, 2, 1 v2, 2, 1
vd , 2, 1 7
Vd ¼ 6
4
7 (49.21)
⋮ ⋱ ⋮ 5
v1, d, 1 v2, d, 1 vd, d, 1
With vk,j,1 denoting the first bit after the binary point of vk,j,. The primitive
polynomials and direction numbers used in Algorithm 659 are taken from Sobol
and Levitan (1976), and a subset of this data may be found in Sobol (1967). Though
it is mentioned in Sobol (1967) that Property A is satisfied for d 16, that is,
det(Vd) ¼ 1 (mod 2) for all d 16, our calculations showed that Property A is
actually satisfied for d 16. As a result, we change the values of the mk,j for 21
j 40, but keep the primitive polynomials. For j
41, we obtain additional
primitive polynomials. The number of primitive polynomials of degree
s is f(2s 1)/s, where f is Euler’s totient function. Including the special case for
j ¼ 1 when all the Mk,j are 1, this allows us to approximate integrals in up to
dimension d ¼ 1,111 if we use all the primitive polynomials of degree 13 or less.
We then choose values of the Mk,j so that we can generate Sobol sequences
satisfying Property A in dimensions d up to 1,111. This is done by generating
some values randomly, but these are subsequently modified so that the condition
det(Vd) ¼ 1 (mod 2) is satisfied for all d up to 1,111. This process involves
evaluating values of the vk,j,1’s to obtain the matrix Vd and then evaluating the
determinant of Vd. A more detailed discussion of this strategy is given in the next
section. It is not difficult to produce values to generate Sobol’s points for approx-
imating integrals in even higher dimensions.
The following figures are the two-dimensional plots of high-dimensional Sobol
sequences of Joe and Kuo with d ¼ 1,000. It is compared with pseudorandom number
1364 F.-P. Lin et al.
Comparison of Randomness
Pseudo Random Number
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Comparison of Randomness
Sobol Seq dimension 1:2
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Fig. 49.7 Pseudorandom number plot comparing with quasi-random number of Sobol for
dimensions 1 and 2
Comparison of Randomness
Sobol Seq dimension 498:501
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Comparison of Randomness
Sobol Seq dimension 990-997
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Fig. 49.8 Comparison of adjacent dimensions in quasi-random number Sobol sequence. The
dimensions are chosen according to prime numbers. There is a high discrepancy found in higher
dimensions of Sobol sequence modified by Joe and Kuo (2003)
and Kuo. The results are plotted in Figs. 49.8 and 49.9. It is found that there are
stronger correlations between Sobol sequences of nonadjacent dimensions in the
fashion of the dimensional comparison of their randomness. Larger numbers of
sampling points, e.g., 10,000, are also tested and the patterns persist. It implied the
1366 F.-P. Lin et al.
Comparison of Randomness
Sobol Seq dimension 2:498
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Comparison of Randomness
Sobol Seq dimension 2:990
1
0.8
0.6
0.4
0.2
0
0 0.2 0.4 0.6 0.8 1
Fig. 49.9 Comparison of nonadjacent dimensions. High discrepancy is found in their correlations
and forms clusters of islands in the distribution
violation of idd assumption and may incur problems in mass distribution and paral-
lelism, in which each process the random number is generated independently without
knowing what other processes are doing. The dependency may deteriorate the quality
of randomness. Nevertheless, in our numerical experiments, there are no significant
differences found thus far (Figs. 49.10 and 49.11).
49 Computer Technology for Financial Service 1367
1.2
Distribution in One Vector
0.8
Probability
0.6
0.4
0.2
0
0 100 200 300 400 500 600 700 800 900 1000
Number of Coordinates
Fig. 49.10 The distribution of probability in directional vector vi,j of Sobol sequences at
i ¼ 3,000 with j ∈ {1, ,1000}. The mean of the distribution is 0.491357, which approaches
the mean of the normal distribution 0.5
13.5
Mean of Asian Option Pricing
13
12.5
12
11.5
11
10.5
10
0 200 400 600 800 1000 1200 1400 1600
Number of Simulations
13.5
Mean of Asian Option Pricing
13
12.5
12
11.5
11
10.5
10
0 200 400 600 800 1000 1200 1400 1600
Number of Simulations
Fig. 49.11 The convergence history of the mean value of Asian option pricing with risk-free
interest rate r ¼ 0.1, underlying asset spot price S ¼ 100, strike price X ¼ 100, duration to maturity
T ¼ 1, and volatility s ¼ 0.3: The comparison is based on a single dimension of the extended high-
dimensional Sobol sequences. The quasi-random number generator (QRNG) outperforms pseu-
dorandom number generator. The test also is conducted to compare the convergence history
between different dimensions in Sobol sequences and found that all perform consistently as
shown in the right figure, in which the low dimension and high dimension are chosen for the
comparison
49 Computer Technology for Financial Service 1369
The various grid platforms are carefully chosen to demonstrate the performance
issue in finance services, which include a small diskless remote boot Linux (DBRL)
PC clusters, large-scale and geographically widely distributed test-bed the Pacific
Rim Applications and Grid Middleware Assembly (PRAGMA) compute grid, and
a densely distributed at-home style PC grid, which resembles the clouding
computing.
OS Image
00101010
01101011
111000
Duplicate OS Images
OS Image OS Image OS Image
OS Image OS Image
00101010 00101010 00101010
00101010 00101010
01101011 01101011 01101011
01101011 01101011
111000 111000 111000
111000 111000
Fig. 49.12 The schematic of DRBL system: DRBL duplicates image files of operational system
via network to the clients, in which the clients’ original operational systems are untouched.
Therefore, the clients are temporarily turned into dedicated compute resources, which also provide
additional security to financial data
Table 49.2 Comparison of speedup ratios based on the calculations in Table 49.1
DRBL cluster FORMOSA II IBM cluster 1350
Asian option pricing (AOP) 28.68 29.46 30.16
Rainbow option pricing (ROP) 27.31 28.59 29.34
VaR calculation (based on ROP 27.20 28.10 29.10
with bootstrap)
Unit: speedup ratio: CPU(nonparallel single node)/CPU(parallel single node)
the clients. The clients’ original operational systems are not used. The clients are
therefore temporarily turned into dedicated compute resources, which also provide
additional security to financial data.
The test case here involves Monte Carlo simulations on Asian option pricing,
on rainbow option pricing and on a bootstrap VaR calculation, respectively
(see Sect. 49.3.2). The market parameters are given as risk-free interest rate
r ¼ 0.1, underlying asset spot price S ¼ 100, strike price X ¼ 100, duration
to maturity T ¼ 125, and volatility s ¼ 0.3. For the rainbow options a linear
weighted combination of 4 underlying assets is assumed with underlying prices of
Si ¼ 100, 110, 120, and 130 and the corresponding weightings of volatility si ¼ 0.3,
0.4, 0.5, and 0.6. The correlation matrix is taken to be
0 1
0:5 0:4 0:5
rij ¼ @ 0:4 0:3 0:4 A
0:5 0:4 0:6
The calculation of the VaR uses the same 4-dimensional rainbow options with
additional expectation of return 0.07, 0.08, 0.09, and 0.10, respectively. They are
calculated in DRBL cluster as well as benchmarked with two cluster-based super-
computers in NCHC. The results are shown in Tables 49.1 and 49.2.
In Table 49.1, instead of giving a total wall clock time of the calculation with
some given numbers of Monte Carlo simulations or paths, a more useful averaged
single Monte Carlo simulation based on 1,000,000 simulations is used to demon-
strate the performance when different system architectures are used. The results
show that the traditional big irons, i.e., supercomputers, still outperform the cluster.
1372 F.-P. Lin et al.
Yet, considering there is no extra cost invested in the computing resources and still
obtains compute cycle in a sizable manner, the approach is appealing to be further
developed in to a fully operational production system.
Applications
TDDFT QM/MD CCAM mpiBLAST iGAP ...
Globus (required)
Table 49.3 Comparison of performance between Group A, which consists of 13 nodes from
NCHC and 15 from UCSD, and Group B, which consists of 122 nodes collectively from UCSD,
AIST, NCHC, and Osaka University. The details of resources are referred to (http://pragma-goc.
rocksclusters.org/pragma-doc/resources.html)
Group A Group B
Asian option pricing (AOP) 68 56
Rainbow option pricing (ROP) 278 234
VaR calculation (based on ROP with bootstrap) 271 229
Unit: seconds/per Monte Carlo path. Averaging from 1000,000 Monte Carlo paths
Table 49.4 Comparison of speedup ratios based on the calculations in Table 49.3
Group A Group B
Asian option pricing (AOP) 25.24 107.58
Rainbow option pricing (ROP) 25.81 110.34
VaR calculation (based on ROP with bootstrap) 25.63 110.09
Unit: seconds/per Monte Carlo path. CPU(nonparallel single node)/CPU(parallel single node)
and allocated to individual computers on the grid-like network. Even with increas-
ingly faster CPUs, the power of PCs is not comparable to those of supercomputers,
but in a networked environment where individual PCs simply process complex task
in parallel, PCs can deliver surprisingly high performance. This is the core concept
of CP grid computing, and it came into a reality several years ago (Table 49.5).
The commoditization and the increased processing speed of PCs lead the growth
of idle CPU power. This will facilitate the construction of PC grid computing
system along with improvement of communication environment by broadband
connectivity (Chen et al. 2006). In practice, a PC grid platform Korea@Home
(or K@H) is used in our study. Its architecture is shown in Fig. 49.14. It is based on
MS Windows. Asian option pricing is used to demonstrate the performance of the
1374 F.-P. Lin et al.
Secure/Intelligent
P2P Agent Technology Agent Agent
Cooperation
Cooperation Cooperation
Distribution
Distribution
Internet
Subscription
Distribution
job request
result
transmission [6T Application Filed]
[Resource
[Large-scale Application] Application Provider
Management] Korea@Home Server
Fig. 49.14 The architecture of Korea@Home, a specific @Home style PC grid used in our case
study (Jun-Weon Yoon 2008)
49.5 Conclusions
Securities trading is one of the few business activities where a few seconds
processing delay can cost a company big fortune. The growing competitive in the
market exacerbates the situation and pushes further towards instantaneous trading
even in split second. The key lies on the performance of the underlying information
system. Following the computing evolution in financial services, it was
a centralized process to begin with and gradually decentralized into a distribution
of actual application logic across service networks. Financial services have tradi-
tion of doing most of its heavy lifting financial analysis in overnight batch cycles.
However, in securities trading it cannot satisfy the need due to its ad hoc nature and
requirement of immediate response. A new computing paradigm, grid computing,
aiming at virtualizing scale-up distributed computing resources, is well suited to the
challenge posed by the capital market practices.
In this study we revisit the theoretical background of how performance will
affect the market competition. The core concept lies on information asymmetry.
Due to the advance of IT, even in split second, it will be a matter of win or lose in
real market practice. After establishing the motivation, we review recent grid
development specifically used for finance service. Monte Carlo simulations are
chosen not only because of its popularity in real world but also because of its nature
so-called “fine grain” or mass parallelism approach. The success of Monte Carlo
simulations lies on better random number generators. The well-recognized Sobol
sequences as a quasi-random number generator are carefully studied to ensure the
quality of Monte Carlo simulations when employed for mass parallelism. Then
some popular basic option pricing models, collectively Asian option pricing,
rainbow option pricing, and VaR calculation with constant market parameters,
are introduced as drivers to introduce more details of grids for better finance
service. Finally, we test various grid platforms, based on the methodology of
mass parallelism and mass distribution, with the drivers. The real market data are
also used, but at this stage they are only used to demonstrate the dynamic data
management, in which grids can offer better.
During this study, we encountered system architect Koschnick from Z€urcher
Kantonalbank of Switzerland (Koschnick 2008). Coincidentally, the system they
plan to migrate from big irons is the similar system to that of DRBL with additional
1376 F.-P. Lin et al.
Fig. 49.15 RBNB DataTurbine streaming open for data channels of iShares MSCI Taiwan Index
(ETF) and TSEC weighted index from May 31, 2005, to May 31, 2008, and 30 days tick-by-tick
trading data from Taiwan Futures Exchange Center (TAIFEX) (Real data plot in collaboration
with Strandell et al. 2007)
49 Computer Technology for Financial Service 1377
virtual local area networks (VLAN) for security. The system is used for overnight
batch job as well as real-time trading practice. It is confirmed that for the years to
come, financial services providers will adopt more grid or grid-based technology to
enhance their competitiveness.
Our future work will be following the current work, continuously using the
current grid platforms and extending them to the use high-frequency real market
data. Along the track of this development, we will also develop sophisticated Monte
Carlo-based option pricing and risk management based on tick-by-tick daily market
information.
References
Abramson, D., et al. (2006). Deploying scientific applications to the PRAGMA grid testbed:
strategies and lessons. In Sixth IEEE international symposium on cluster computing and the
grid, Singapore, pp. 241–248.
Ahmad, K., Gillam, L., & Cheng, D. (2005). Textual and quantitative analysis: Towards a new,
e-mediated social science. In Proceedings of the 1st international conference on e-Social
science, Manchester, pp. 22–24.
Almond, J., & Snelling, D. (1999). UNICORE: Uniform access to supercomputing as an element
of electronic commerce. Future Generation Computer Systems, 15(5), 539–548.
Amad, K., et al. (2004). Financial information grid –an ESRC e-Social social science pilot.
In Proceedings of the third UK e-Science programme all-hands meeting, Nottingham.
Antonov, I. A., & Saleev, V.M. (1979). An economic method of computing LP[tau]-sequences.
USSR Computational Mathematics and Mathematical Physics, 19(1), 252–256.
Atkins, D.E., et al. (2003). Revolutionizing science and engineering through cyberinfrastructure.
Report of the National Science Foundation Blue-Ribbon Advisory Panel on Cyberinfras-
tructure. National Science Foundation.
Bell, D., & Ludwig, A. S. (2005). Grid service discovery in the financial markets sector. In 7th
international conference on information technology interfaces, Cavtat, Croatia.
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of
Political Economy, 81, 637–654.
Bratley, P., & Fox, B. L. (1988). Algorithm 659: Implementing Sobol’s quasirandom sequence
generator. ACM Transactions on Mathematical Software, 14, 88–100.
Bughin, J., Chui, M., & Manyika, J. (2010). Cloud, big data and smart assets: Ten tech-enabled
business trends to watch. McKinsey Quarterly August, 7–8.
Chang, F., Dean, J., Ghemawat, S., Hsieh, W.C., Wallach, D.A., Burrows, M., Chandra, M., Fikes, A.,
Gruber, R.E. (2006). Big table: A distributed storage system for structured data. In OSDI 06:
Proceedings of the 7th USENIX Symposium on Operating Systems Design and Implementation,
Seattle, WA, p. 15-15.
Cheng, J., & Druzdzel, M. J. (2000). AIS-BN: An Adaptive importance sampling algorithm for
evidential reasoning in large bayesian networks. Journal of Artificial Intelligence Research,
13, 155–188.
Chen, G., Thulasiraman, P., & Thulasiram, R.K. (2006). Distributed quasi-monte carlo algorithm
for option pricing on HNOWs using mpC. In Proceedings of the 39th annual simulation
symposium (ANSS’06), Huntsville.
Cizek, P., Härdle, W., & Weron, R. (2011). Statistical tools for finance and insurance. Berlin:
Springer.
Dean, J., & Ghemawat, S. (2004). Map reduce: Simpli_ed data processing on large clusters.
In OSDI 04, San Francisco, CA, pp. 137–150.
Dowd, K. (2002). Measuring market risk. Chichester/England/New York: Wiley.
1378 F.-P. Lin et al.
Efon, B. (1981). Censored data and the bootstrap. Journal of the American Statistical Association,
76(374), 312–319.
Efron, B., & Tibshirani, R. (1986). Bootstrap methods for standard errors, confidence intervals,
and other measures of statistical accuracy. Statistical Science, 1(1), 54–75.
Faure, H. (1982). Discrepance de Suites Associees a un Systeme de Numeration (en Dimension s).
Acta Arithmetica, 41, 337–351.
Foster, I., & Kessleman, C. (2004). The grid: Blueprint for a new computing infrastructure
(2nd ed.). Amsterdam: Morgan Kaufmann.
Foster, I., Zhao, Y., Raicu, I., & Lu, S. (2008). Cloud computing and grid computing 360-degree
compared. In Grid computing environments workshop, Austin, TX, pp. 1–10.
Galanti, S., & Jung, A. (1997). Low-discrepancy sequences: Monte Carlo simulation of option
prices. Journal of Derivatives, 5(1), 63–83.
Ghoshal, D., Canon, R. S., & Ramakrishnan, L. (2011). I/O performance of virtualized cloud
environments. In The second international workshop on data intensive computing in the
clouds, Seattle, WA, pp 71–80.
Gillam, L., Ahmad, K., & Dear, G. (2005). Grid-enabling social scientists: The Fingrid infra-
structure. In Proceedings of the 1st international conference on e-social science,
Manchester, pp. 22–24.
Griffiths, P., & Remenyi, D. (2003). Information technology in financial services:
A model for value creation. Electronic Journal of Information Systems Evaluation,
6(2), 107–116.
Grimshaw, A. S., & Wulf, W. A. (1997). The legion vision of a worldwide virtual computer.
Communications of the ACM, 40(1), 39–45.
Halton, J. H. (1960). On the efficiency of certain quasirandom sequences of points in evaluating
multidimensional integrals. Numerische Mathematik, 2, 84–90.
Hauswald, R., & Marquez, R. (2003). Information technology and financial services competition.
The Review of Financial Studies, 16(3), 921–948.
Hull, J. C. (2003). Options, futures and other derivatives (5th ed.). Upper Saddle River:
Prentice Hall.
Joe, S., & Kuo, F. Y. (2003). Remark on algorithm 659: Implementing Sobol’s quasirandom
sequence generator. ACM Transactions on Mathematical Software, 29(1), 49–57.
Korpela, E., et al. (2001). SETI@home-massively distributed computing for SETI. Computing in
Science and Engineering, 3(1), 78–83.
Koschnick, T. (2008). Cluster & security: Separating application environments. In Int’l
Supercomputing conference and exhibition, Dresden, pp. 17–20.
Kuipers, L., & Niederreiter, H. (1974). Uniform distribution of sequence. New York: Wiley.
Lauret, J., Walker, M., Goasguen, M., & Hajdu, L. (2010). From grid to cloud, the STAR
experience. In SciDAC 2010 proceedings, Chattanooga, TN.
Leto, A., et al. (2005). EGRID Project: Experience report on the implementation of a grid
infrastructure for the analysis of financial data. In Proceedings of the international conference
on information technology: Coding and computing, Las Vegas, NV.
Litzkow, M. J., Livny, M., & Mutka, M. W. (1988). Condor-a hunter of idle workstations. In 8th
international conference on distributed computing systems, proceeding, San Jose, CA,
pp. 104–111.
Mell, P., Grance, T. (2011). The NIST definition of cloud computing (pp. 800-145). Special
Publication.
Milojicic, S. D., et al. (2002). Peer-to-peer computing (HL laboratories research report).
Niederreiter, H. (1992). Random number generation and Quasi-Monte Carlo methods. In CBMS-
NSF regional conference series in appl. math. Philadelphia, PA.
Paskov, S. H., & Traub, J. F. (1995). Faster valuation of financial derivatives. Journal of Portfolio
Management, 22(1), 113–120.
Phillips, A., et al. (1984). Effects of information technology on financial services systems.
Washington, D. C.: U.S. Congress, Office of Technology Assessment, OTA-CIT-202.
49 Computer Technology for Financial Service 1379
Ramakrishnan, L., Canon, R. S., Muriki, K., Sakrejda, I., & Wright, N. J. (2011). Evaluating
interconnect and virtualization performance for high performance computing. In Proceedings
of 2nd international workshop on performance modeling, benchmarking and simulation of high
performance computing systems, Seattle, WA, pp. 1–2.
Schmidt, A. B. (2005). Quantitative finance for physicists: An introduction. Academic Press
Advanced Finance Series, Elsevier, Oxford.
Smarr, L., & Catlett, C. (1992). Metacomputing. Communications of the ACM archive,
35(6), 44–52.
Snir, M., et al. (1996). MPI: The complete reference. The MIT Press, Cambridge, Massachusetts.
Sobol, I. M. (1967). On the distribution of points in a cube and the approximate evaluation of
integers. U.S.S.R. Computational Mathematics and Mathematical Physics, 7(4), 86–112.
Sobol, I. M., & Levitan, Y. L. (1976). The production of points uniformly distributed in a
multidimensional cube. Tech. Rep. 40, Institute of Applied Mathematics, USSR Academy of
Sciences (in Russian).
Stokes-Rees, I., et al. (2007). Multi-cluster parallel job submission: Experiences with Monte Carlo
simulations for computational finance on Grid5000 (IRIS PicsouGrid Project Report).
Strandell, E., Tilak, S., Chou, H. M., Wang, Y. T., Lin, F. P., Arzberger, P. Fountain, T. (2007).
Data management at kenting’s underwater ecological observatory. In Proceedings of the third
International Conference on Intelligent Sensors, Sensor Networks and Information Processing,
Melbourne, Australia.
Tanaka, S. (2003). Joint development project applies grid computing technology to financial risk
management. Nikkei Computer, special section on grid computing.
Tilak, S., Hubbard, P. Miller, M., & Fountain, T. (2007). The ring buffer network bus (RBNB) data
turbine streaming data middleware for environmental observing systems. In Proceedings of
third IEEE international conference on e-science and grid computing, Bangalore, India,
pp. 125–133.
Weron, R. (2004). Handbook of computational statistics. Berlin: Springer.
Yelick, K., Coghlan, S., Draney, B., Canon, R. S. (2011). The magellan report on cloud computing
for science. ASCR, DOE. Washington, DC.
Yoon, J. (2008) Korea@Home (PC Grid project in Korea). GridAsia 2008: Singapore.
Long-Run Stock Return and the Statistical
Inference 50
Yanzhi Wang
Contents
50.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1382
50.2 Long-Run Return Estimation in Event-Time Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1383
50.2.1 Return Estimations: CAR, BHAR, and Earnings
Announcement Returns . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1384
50.2.2 Benchmark Problem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1386
50.2.3 Statistical Inference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1387
50.3 Long-Run Return Estimation in Calendar-Time Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . 1392
50.3.1 Return Estimations: Mean Monthly Return and Factor Models . . . . . . . . . . . . . 1393
50.3.2 Conditional Market Model and Ibbotson’s RATS . . . . . . . . . . . . . . . . . . . . . . . . . . . 1394
50.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1396
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1396
Abstract
This article introduces the long-run stock return methodologies and their statis-
tical inference. The long-run stock return is usually computed by using a holding
strategy more than 1 year but up to 5 years. Two categories of long-run return
methods are illustrated in this article: the event-time approach and calendar-time
approach. The event-time approach includes cumulative abnormal return,
buy-and-hold abnormal return, and abnormal returns around earnings announce-
ments. In former two methods, it is recommended to apply the empirical
distribution (from the bootstrapping method) to examine the statistical inference,
whereas the last one uses classical t-test. In addition, the benchmark selections in
the long-run return literature are introduced. Moreover, the calendar-time
approach contains mean monthly abnormal return, factor models, and Ibbotson’s
Y. Wang
Yuan Ze University, Taiwan
Department of Finance, College of Management, National Taiwan University, Taipei, Taiwan
e-mail: yzwang@ntu.edu.tw
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1381
DOI 10.1007/978-1-4614-7750-1_50,
# Springer Science+Business Media New York 2015
1382 Y. Wang
Keywords
Long-run stock return • Buy-and-hold return • Factor model • Event time •
Calendar time • Cumulative abnormal • Return • Ibbotson’s RATS • Conditional
market model • Bootstrap • Zero-investment portfolio
50.1 Introduction
The long-run stock return has been an important facet of the stock performance
of firms since the 1990s. Ritter (1991) starts the line of long-run return studies
by investigating the initial public offering (IPO) cases. He finds that the post-IPO
stock performance is poor in the long run. His paper is then followed by many
scholars who study other important corporate events and asset pricing anomalies.1
Fama (1998) reviews these papers which engage in long-run event studies. Hence,
the long-run return method is nowadays a standard way for the stock performance
of firms.
I review the methodologies about the long-run stock abnormal return in this
article. Since the mid-1990s, some papers started to be aware of the properties of
long-term stock performance and suggested various ways to calculate the long-run
abnormal return. Therefore, Barber and Lyon (1997), Kothari and Warner (1997),
and Lyon et al. (1999) review and compare long-run return methods in the late
1990s. Based on these methodology papers, I update recent developments about
the new methods that are not mentioned in these two papers, such as the earnings
announcement abnormal return applications (La Porta et al. 1997), conditional
market model method (Eberhart et al. 2004; Petkova and Zhang 2005), Ibbotson’s
RATS (Ibbotson 1975; Agrawal et al. 1992; Peyer and Vermaelen 2009), and zero-
investment portfolio method (Daniel and Titman 1997; Eberhart et al. 2004). I also
make a clearer categorization on these long-run stock return methodologies given
that the long-run return methodologies nowadays are much more mature than in the
late 1990s.
Two main categories of the long-run stock performance illustrated in this paper
are event-time and calendar-time approaches. The event-time approach includes
cumulative abnormal return (CAR), buy-and-hold abnormal return (BHAR),
rebalanced buy-and-hold abnormal return (RBHAR), and abnormal returns around
1
For corporate events, papers have studied seasoned equity offerings, mergers, dividend initiations
and omissions, quarterly earnings announcements, share repurchases, proxy flights, stock splits
and spinoffs, and other corporate events for their long-run stock performance. For asset pricing
anomalies, papers investigate value premium, momentum profit, research and development profit,
accrual effect, asset growth, net share issuance, and other anomalies in terms of the long-term
impact.
50 Long-Run Stock Return and the Statistical Inference 1383
To estimate the long-run stock return, it is nature to identify an event to track the
stock performance of a firm. Figure 50.1a shows a general time line for the long-run
return estimation. For any specific event and its event day noted by day 0, we can
start a holding strategy by purchasing the stock from event day 0 to 251th day (for
1-year return), to 503th day (for 2-year return), 755th day (for 3-year return), to
1,007th day (for 4-year return), or to 1,259th day (for 5-year return). Papers usually
compute the long-run stock return up to 5 years (Ritter 1991). As indicated in
Fig. 50.1a, an accounting reporting lag is required between the event date and the
previous fiscal year-end with at least 4 months.
1384 Y. Wang
Return
b Fiscal Formation Fiscal
Fiscal
Year Year Day Year
End -1 End 0 (July 1st ) End +1
Fig. 50.1 (a) time line of general event study. (b) time line of asset pricing study
For the asset pricing study, we usually do not have a specific event day for
tracking the compounding daily returns. As suggested in Fama and French (1992,
1993), a general investment strategy starts the return formation from 1 July of each
year by exerting the accounting information in the previous fiscal year-end with at
least 6 months reporting lag. Figure 50.1b shows the time line for asset pricing
studies. Because asset pricing papers start from 1st July, the long-run returns are
estimated based on compounded monthly returns. Basically, 1-year return includes
12 monthly returns, while 5-year return is computed upon 60 monthly returns.
The cumulative abnormal return (CAR) is estimated as follows. For a given bench-
mark E(Ri,t), the abnormal return ARi,t ¼ Ri,t – E(Ri,t). The average of estimations
for CAR is
X
N
ARi, t
X
T
i¼1
CAR ¼ , (50.1)
t¼1
N
50 Long-Run Stock Return and the Statistical Inference 1385
in which the Nt is for number of observations at time t and T for the holding period
(252 for a year if the return is computed based on daily returns; 12 for a year if the
return is computed based on monthly returns).
The buy-and-hold abnormal return is defined as
" # " #
Y
T Y
T
1 þ Ri, t 1 1 þ E Ri , t 1
X
N
t¼1
X
N
t¼1
BHAR ¼ : (50.2)
i¼1
N i¼1
N
Thus, we can use classical t-test or other methods (will be illustrated in later
section) to carry out the statistical inference for CAR and BHAR. One modification
on BHAR is the rebalanced buy-and-hold abnormal return (RBHAR), which comes
from the combination of rebalanced return (Rreb) and BHAR. Given the rebalanced
return as
0 1
XN
T B
Ri , t C
Y B C
R ¼ B i¼1 C
B1 þ N C 1,
reb
(50.3)
t¼1 @ A
the RBHAR is to calculate buy-and-hold return for a certain period (e.g., a year) and
rebalance the portfolio equally for every specific period. Taking the 1-year
rebalanced BHAR as the example, we should compute 1-year BHAR for each
event year (i.e., first to fifth event year) and obtain the average BHAR for every
event year. Finally, we obtain compounding return for this average rebalanced
BHAR with yearly rebalancing. Because BHAR has inflated compounding return,
which results in many outliers in the long-run return estimation, RBHAR with
rebalancing every year is able to reduce the impact from extreme values (Ikenberry
et al. 1995; Chan et al. 2010). In general, RBHAR could be described as
8 " # 9
> Y
T >
>
> 1 þ Ri, j 1>
>
Y>
year5 < X
N
j¼1
>
=
RBHAR ¼ 1þ
>
year1 >
N >
>
>> i¼1 >
>
: ;
8 " # 9
> Y
T >
>
> 1 þ E Ri , j 1>
>
Y>
year5 < X
N
j¼1
>
=
1þ (50.4)
>
year1 >
N >
>
>
> i¼1 >
>
: ;
Note that T stands for 252 days as an event year, and year 1 to year 5 represent
the first event year to the fifth event year.
1386 Y. Wang
Distinct from CAR and BHAR, the quarterly earnings announcement return is
the estimate of the long-term stock performance by successive short-term quarterly
announcement returns (La Porta et al. 1997; Denis and Sarin 2001; Chan
et al. 2004). In other words, we can use the 2-day or 3-day abnormal returns
around quarterly announcement dates, and the successive earnings surprises should
be aligned with the long-run stock abnormal return. The benchmark selection
determines the long-run abnormal return estimation because of its statistical
property (will be discussed in next section), so the results may be changed
when benchmark settings are altered. This sensitive outcome driven from the
benchmark problem leads to an inconclusive long-run abnormal return. Yet, the
short-run return is not sensitive to the selection of benchmarks. For example, given
10 % and 20 % of market index returns, the 3-day market index returns are
expected to be 0.12 % and 0.24 % only. When an event occurs, the 3-day
announcement return could be generally above 1 %, which significantly exceeds
any selected benchmark returns. Therefore, if a corporate event is followed by
profitability improvements that are not observed by investors, then there should be
successive earnings surprises following the event date, and the short-term
announcement abnormal returns around the earnings announcement dates
should be positive. To capture the long-run abnormal return, papers usually study
12–20 quarters (for 3–5 years) for the quarterly earnings announcement abnormal
return. Generally, the quarterly earnings announcement abnormal returns capture
about 25–40 % of the long-run stock return of a firm (Bernard and Thomson 1989;
Sloan 1996).
Long-run stock return has some issues regarding the calculation and testing when
we select the benchmarks for the expected return of the firm. Namely, the conven-
tional methodologies for the long-run stock return might be biased if the chosen
matching procedure is inappropriate. First, some long-term return measures have
the rebalancing and new-listing biases problems. Second, long-run stock return is
positive skewed, thus the traditional t-test is inappropriate for the long-run return.
Although I will review the skewness-adjusted t-statistics and empirical p-value in
the next section, matching firm method is another way to alleviate the skewness
concern in the long-run return studies.
The first benchmark for the expected return is the CRSP equal-weighted
index return, which includes whole stocks in CRSP database and computes the
simple average of stock returns. However, this approach involves rebalancing
problem, which ignores the transaction costs from broker’s fee and tax to
the government. To maintain the weights on stocks equally, investors must sell
profitable stocks and buy stocks with loss. This rebalancing leads to huge transac-
tion costs that are not considered in the CRSP equal-weighted index return, making
the abnormal return underestimated. In fact, CAR per se also has this rebalancing
problem because of its average in cross section.
50 Long-Run Stock Return and the Statistical Inference 1387
The second benchmark is the CRSP value-weighted index return, which also
comes from CRSP database. This index return uses firm value (equal to the price
timing shares outstanding) as the weighted scheme to compute the weighted
average of stock returns. Most importantly, it is a rebalance-free benchmark and
accordingly has no transaction cost during the holding period (except the beginning
and end). Therefore, recent papers tend to use CRSP value-weighted index return
instead of CRSP equal-weighted index return as the benchmark return.
The third benchmark is the reference portfolio return. Before constructing the
reference portfolio, we may need matching criterions. As the important pricing
factors, size and book-to-market (BM) ratio are the two most important in deter-
mining the expected return (Fama and French 1992, 1993, 1996; Lakonishok
et al. 1994; Daniel and Titman 1997). In particular, we have to determine
the matching pool, which includes stocks that are irrelevant to the sample firm.
We form 50 size and book-to-market portfolios (ten size portfolios and five book-
to-market portfolios in each size decile) where the size and book-to-market cutoff
points are obtained from stock in NYSE exchange. We then are able to compute
either equal-weighted or value-weighted portfolio return as the expected return.
The last one is the matching firm method. Because the long-run stock return is
positive skewed, we may take the long-run return of the matching firm as the
expected return. The skewed return of the sample firm and skewed return of the
matching firm offset each other and make the abnormal return symmetric (Barber
and Lyon,1997). In addition, matching firm method avoids the new-listing bias and
rebalancing problem. In general, the matching criterions of the matching firm are
similar to the reference portfolio. Within each reference portfolio, a matching firm
could be selected by minimizing the book-to-market difference between the sample
firm and the matching firm (Ikenberry et al. 1995). Sometimes, papers use few
matching firms but not single matching firm as the benchmark to avoid few outlier
impacts (e.g., Lee 1997) or use different matching variable (e.g., Ikenberry and
Ramnath 2002; Eberhart et al. 2004). Generally, various matching methods under
size and book-to-market effect controls do not largely change the results.
The most important statistical problem for the long-run abnormal return is the
skewness. The minimum loss of a long-term stock investment is 100 % while
the maximum potential gain approaches infinite. Thus, the distribution of the long-
run stock return is positive skewed. If we test the long-run stock return by a standard
normal distribution, then we tend to reject the null hypothesis (that suggests
no abnormal return) for negative returns and accept the null for positive returns.
This misspecification leads to a type I error in the distributional left tail but causes
a type II error in the distributional right tail.
To solve the skewness problem, Barber and Lyon (1997) suggest the matching
firm method because abnormal stock return is the return difference between the
sample firm and matching firm, making the skewness from matching firm and
1388 Y. Wang
sample firm offset by each other. In addition, there are two ways to alleviate the
testing problem under the skewness, one is the skewness-adjusted t-statistic and the
other is the empirical p-value, suggested by Ikenberry et al. (1995) and Lyon
et al. (1999).
Given an abnormal return ARi, the skewness-adjusted t-statistic illustrated by
Lyon et al. (1999) is tested as follows:
pffiffiffiffi 1 1
tsa ¼ N ðSÞ þ ^g S2 þ ^g ,
3 6N
where
X
N 3
ARi AR
AR i¼1
S¼ , and ^g ¼ : (50.5)
sðARi Þ NsðARi Þ3
Note that the S is the conventional t-statistic and ^g is the coefficient for the
skewness adjustment.
The second suggested statistical inference method is the empirical p-value. This
approach uses bootstrapping method to construct an empirical distribution with
general long-term return features. We use bootstrapping method to select the
pseudo-sample firm. Thus, we are able to compare sample firm and pseudo-sample
firm as the base of the empirical p-value. In fact, it is possible that we may face
moment conditions (higher than third moment condition) in the long-term return
estimation, and the skewness-adjusted t-statistic is not enough to capture the return
characteristic. Also, the strong cross-sectional correlations among sample observa-
tions in BHARs can lead to poorly specified test statistics (Fama 1998; Lyon
et al. 1999; Brav 2000). Under the empirical distribution, we can examine the
statistical inference without a parametric distribution but are able to capture more
unknown statistical features.
As mentioned above, the empirical p-value is generated from the empirical
distribution from bootstrapping, and the empirical distribution well controls the
skewness and time-dependent properties of the long-run stock returns (Ikenberry,
et al. 1995; and Lyon et al. 1999; Chan et al. 2004). In addition, this empirical
p-value also solves the statistical inference problem in RBHAR because we may
have too few observations in times series for computing standard deviation.2
To construct the empirical distribution, we need to find pseudo-sample firms that
share similar firm characteristics as the sample firm but do not have the interested
corporate events. Next, we construct 25 size and book-to-market portfolios (five
size portfolios and five book-to-market portfolios in each size quintile) from all
2
For example, if we compute a 5-year RBHAR, then we have five averages of BHARs in five event
years only.
50 Long-Run Stock Return and the Statistical Inference 1389
nonevent firms. The nonevent firm selection criterions are similar to the matching
firm selection. Then, we randomly sample one pseudo-sample firm out of the
corresponding portfolio for each sample firm. For example, if an event firm is
with the second size quintile and third book-to-market quintile, then we
randomly choose a pseudo firm that is also with the second size quintile and third
book-to-market quintile. Hence, we are able to obtain N pseudo firms for N sample
firms. Upon these pseudo-sample firms, we calculate the long-run return by CAR,
BHAR, or RBHAR method. Finally, we repeat the sampling and long-run return
estimation for 1,000 times and obtain 1,000 averages of long-run returns for pseudo-
sample firms. The empirical distribution is plotted according to the frequency
distribution diagram of those 1,000 average returns of pseudo-sample firms.
If the long-run return is larger than the (1a) percentile of the 1,000 average
pseudo-sample firm returns, then we obtain the p-value as a for testing the
positive average abnormal return. Similarly, if the long-run return is smaller than
a percentile of the 1,000 average pseudo-sample firm returns, then we obtain the
p-value as a for testing the negative average abnormal return.
Figures 50.2a–50.2d are empirical distributions for 1-year to 4-year long-run
returns upon size/book-to-market controlled pseudo-sample firms of US repurchase
firms during 1980–2008. I collect the repurchase data from SDC database as the
example for the empirical distribution construction. For those empirical distribu-
tions, it is obvious that 4-year return has more outliers than 1-year return. Moreover,
the 4-year return figure has lower kurtosis and is more positive skewed. Obviously,
the shape of long-run returns does not obey normal distribution, and the empirical
p-value is more relevant to the long-run abnormal return testing.
I also show the specification (statistical size) for different long-run return methods
in Table 50.1, which is obtained from Table 5 of Barber and Lyon (1997) and Table 3
of Lyon et al. (1999). They show the percentage of 1,000 random samplings from
200 firms rejecting the null hypothesis that suggests no abnormal return in terms of
CAR, BHAR, and RBHAR with different benchmark and testing methods. First, the
CAR with matching portfolio as the benchmark has type I error in left tail when
measuring the abnormal return in 5 years. Second, the matching firm method yields
good specification no matter how we focus on size control, BM control, or
a combination control for both size and BM ratio. Third, skewness-adjusted t-statistics
has type I error, implying that skewness is not the only statistical feature that we
should address. Forth, empirical p-value method performs well even when adopting
the reference portfolio as the benchmark, at least for 10 % significance level.
Figure 50.3 shows the testing power of alternative tests by using BHAR as the
primary method, and this figure is originally plotted in Fig. 1 of Lyon et al. (1999).
The empirical distribution performs better in testing power than classical t-test,
no matter what the standard empirical distribution or the bootstrapped skewness-
adjusted t-statistic is employed.
In sum, in the event-time approach, the BHAR is suggested. Matching firm
is a better matching method than other approaches. In statistical testing, the
empirical p-value could be the best way due to its well statistical size control and
testing power.
1390 Y. Wang
a 70
60
50
40
30
20
10
0
0.08
0.09
0.1
0.1
0.11
0.12
0.13
0.14
0.15
0.16
0.17
0.18
0.19
0.19
0.2
0.21
0.22
0.23
0.24
0.25
0.26
b 40
35
30
25
20
15
10
5
0
0.29
0.3
0.31
0.32
0.33
0.33
0.34
0.35
0.36
0.37
0.37
0.38
0.39
0.4
0.41
0.41
0.42
0.43
0.44
0.45
c 35
30
25
20
15
10
5
0
0.54
0.54
0.55
0.56
0.57
0.57
0.58
0.59
0.59
0.6
0.61
0.61
0.62
0.63
0.64
0.64
0.65
0.66
0.66
0.67
0.68
0.68
d 30
25
20
15
10
5
0
0.77
0.78
0.79
0.79
0.80
0.81
0.82
0.83
0.83
0.84
0.85
0.86
0.87
0.87
0.88
0.89
0.90
0.91
0.91
Fig. 50.2 (a) One-year return distribution of size/BM controlled pseudo-sample firms. (b)
Two-year return distribution of size/BM controlled pseudo-sample firms. (c) Three-year return
distribution of size/BM controlled pseudo-sample firms. (d) Four-year return distribution of size/
BM controlled pseudo-sample firms
50 Long-Run Stock Return and the Statistical Inference 1391
100
empirical
Percentage of Samples Rejecting Null
p value
80
bootstrapped
skewness-adj.
60 t statistic
40
t statistic,
control firm
20
0
−20 −15 −10 −5 0 5 10 15 20
Induced Level of Abnormal Return (%)
Fig. 50.3 Power of alternative tests in random ample. This figure is from Fig. 1 of Lyon
et al. (1999, p. 180). The percentage of 1,000 random samples of 200 firms rejecting the null
hypothesis of no annual buy-and-hold abnormal return at various induced levels of abnormal
return (horizontal axis) based on control firm method, bootstrapped skewness-adjusted t-statistic,
and empirical p-values
The first method in the calendar-time approach is the mean monthly abnormal
return (MMAR). We can start from a 5-year holding strategy on a corporate event.
For any given calendar month (e.g., June 2002), we need to include a stock if it had
the corporate event in the past 60 months (e.g., looking backward at a period of May
2002–June 1997). We then need to form a monthly portfolio for this specific calendar
month, and the portfolio return could be computed upon equal-weighted, value-
weighted or log-value-weighted scheme. Next, we repeat the abovementioned step
for all calendar months throughout the sample period, and then the mean monthly
return is the time-series average of monthly portfolio returns.
Similar to the benchmark problem in event-time approach, we need to select the
CRSP market index return, reference portfolio, or the matching firm as the bench-
mark for the MMAR. For any benchmark E(Ri), the MMAR is
X
T
RPt RM
t
t¼1
MMAR ¼ ,
T
where
X
Nt X
Nt
wi, t Ri, t wi, t E Ri, t
i¼1 i¼1
RPt ¼ t ¼
, and RM : (50.6)
Nt Nt
T is for calendar month in this setting; RPt is the monthly portfolio return of
sample firms; RM t is the monthly portfolio return of benchmarks; and Nt is the
number of observations in each calendar month. The statistical inference can be
either classical t-statistic or the Newey and West (1987) estimation.
As suggested by Fama (1998), Mitchell and Stafford (2000), and Schultz (2003),
factor model is a robust method in estimating the long-run stock abnormal return.
The standard Fama and French three-factor and Carhart four-factor models could be
described as
RPt r f , t ¼ a þ b r m, t r f , t þ sSMBt þ hHMLt þ et , (50.7)
RPt r f , t ¼ a þ b r m, t r f , t þ sSMBt þ hHMLt þ mMOMENTUMt þ et ,
(50.8)
1394 Y. Wang
where RPt is the sample firm portfolio return for each calendar month, and could
be obtained from the equal-weighted, value-weighted, or log-value-weighted
average. This average return in calendar is similar to what we compute in the
MMAR. rf is the risk-free rate, usually the short-term Treasury bill rate; rm is
usually computed as CRSP value-weighted index return; SMB is small firm
portfolio return minus big-firm portfolio return; HML is the high book-to-market
portfolio return minus low book-to-market portfolio return; MOMENTUM is
the winner portfolio return minus loser portfolio return where winner and
loser portfolios are identified by past 1-year return. SMB, HML, and MOMENTUM
are applied to control size and book-to-market and momentum effects,
respectively (Fama and French 1992, 1993; Jegadeesh and Titman 1993;
Lakonishok et al. 1994; Carhart 1997). The abnormal return is the regression
intercept and can be tested based on the t-values or Newey and West (1987)
estimator.
Next, I introduce a modification of the factor model analysis: the zero-
investment portfolio method. Daniel and Titman (1997) and Eberhart et al.
(2004) study the long-run return by using the zero-investment portfolio approach
to control for both risk and firm characteristic effects. To form the factor
model under a zero-investment portfolio strategy, we have to buy sample stocks
and short-sell matching stocks. Taking Carhart (1997) four-factor model as the
example, we have
P
RPt RM
t ¼ a a
M
þ ðbP bM Þ r m, t r f , t þ ðsP sM ÞSMBt
þ ðhP hM ÞHMLt þ ðmP mM ÞMOMENTUMt þ et , (50.9)
and we use (aP–aM) as the abnormal return controlling for both risks and
firm characteristics. It is also the hedging portfolio return controlled for the
common risk factors. The matching firm selection criterions can apply the steps
in benchmark problem section. For other modifications of the factor model analysis,
Eberhart et al. (2004) provide more examples in their Table 3.
One major challenge to the standard market model is that the risk loadings
are assumed to be unchanged. To estimate the factor loadings, we usually need
long time series to obtain the estimated risk loadings, and the fixed risk
loading over time is naturally assumed in the OLS analysis. Yet, the magnitude
of the risk of a firm could be changed; in particular many corporate events
change the risk of the firm (e.g., R&D increases could be followed by risk increases,
and share repurchase could be followed by risk decreases). Accordingly it is
needed to introduce the conditional market model to address the time-varying
market model.
50 Long-Run Stock Return and the Statistical Inference 1395
There are at least two ways to address the time-varying risks in the regression
model as the conditional market model. To simplify the problem, I use the CAPM
as the first example. First, the systematic risk could change along with some
firm characteristics and macroeconomic variables. Petkova and Zhang (2005) use
the following regression analysis to estimate abnormal return:
RPt r f , t ¼ a þ ðb0 þ b1 DIV t þ b2 DEFt þ b3 TERMt þ b4 TBt Þ r m, t r f , t þ et :
(50.10)
(b0 + b1DIVt + b2DEFt + b3TERMt + b4TBt) is the bt that accommodates
to the time-varying risk loading. They assume that the risk changes with the
dividend yield (DIV), the default spread (DEF), the term spread (TERM), and the
short-term Treasury bill rate (TB). Again, the abnormal return is the intercept a.
The second method is the rolling regression, as suggested by Petkova and Zhang
(2005) and Eberhart et al. (2004). If we have a sample period from January 1990,
then we can use the portfolio returns in the first 60 months (i.e., January
1990–December 1994) to carry out the Carhart (1997) four-factor regression.
We substitute the estimated factor loadings from these 60 monthly returns into
the equity premiums in 61th month (i.e., January 1995) and then obtain the
expected portfolio return for 61th month. Thus, the abnormal return for 61th
month is from the portfolio return of the sample firm minus the expected portfolio
return. Next, we need to repeat the abovementioned steps for every month by
rolling return windows. Finally, we estimate the abnormal return as the average
of abnormal returns across time and use the time-series volatility to test the
statistical significance.
The final method relating to the time-varying risk is the Ibbotson (1975) RATS
though it is not under the family of the factor model analysis. The original setting of
Ibbotson RATS is designed for the long-run return estimation, yet recent papers
use this method combining the factor model analysis to measure the long-run
abnormal return (e.g., Peyer and Vermaelen 2009). Based on Carhart (1997)
four-factor model, we regress the security excess return on the Carhart (1997)
four factors for each month in the event time. Given a 60-month holding
strategy, we have to carry out this regression for 1st month to 60th month following
the corporate event date. Then, the abnormal return for month t is the intercept of
this four-factor regression:
Ri, t r f , t ¼ at þ bt r m, t r f , t þ st SMBt þ ht HMLt þ mt MOMENTUMt þ et :
(50.11)
The regression analysis is similar to what I introduce in Eq. 50.8; however,
the regression is examined every event month t. For every event month or event
year, we can obtain the average abnormal return as the average of the intercepts
(at), which is obtained from a model allowing time-varying risks.
1396 Y. Wang
50.4 Conclusion
The long-run return studies have been investigated for many corporate events and
asset pricing studies in the past two decades. I introduce the long-run stock return
methodologies and their statistical inference adopted in recent papers. Two cate-
gories of long-run return methods are illustrated: the event-time approach and
calendar-time approach. Under the event-time category, we have methods includ-
ing cumulative abnormal return, buy-and-hold abnormal return, and abnormal
returns around earnings announcements. Although the event-time approach is
able to be implemented as an investment strategy in real world, it also raises
more benchmark and statistical inference problems. Under the calendar-time cate-
gory, we have mean monthly abnormal return, factor models, and Ibbotson’s
RATS. Generally, calendar-time approach is more popular due to its robustness
and variety. For any long-run return study, I may suggest that combining works on
those methodologies could be necessary.
References
Agrawal, A., Jaffe, J., & Mandelker, G. (1992). The post-merge performance of acquiring firms:
A re-examination of an anomaly. Journal of Finance, 47, 1605–1621.
Barber, B., & Lyon, J. (1997). Detecting long-run abnormal stock returns: The empirical power
and specification of test statistics. Journal of Financial Economics, 43, 341–372.
Bernard, V., & Thomas, J. (1989). Post-earnings-announcement drift: Delayed price response or
risk premium. Journal of Accounting Research, 27, 1–36.
Brav, A. (2000). Inference in long-horizon event studies: A Bayesian approach with application to
initial public offerings. Journal of Finance, 55, 1979–2016.
Carhart, M. (1997). On persistence in mutual fund performance. Journal of Finance, 52, 57–82.
Chan, K., Ikenberry, D., & Lee, I. (2004). Economic sources of gain in share repurchases. Journal
of Financial and Quantitative Analysis, 39, 461–479.
Chan, K., Ikenberry, D., Lee, I., & Wang, Y. (2010). Share repurchases as a potential tool to
mislead investors. Journal of Corporate Finance, 16, 137–158.
Daniel, K., & Titman, S. (1997). Evidence on the characteristics of cross sectional variation in
stock returns. Journal of Finance, 52, 1–33.
Denis, D., & Sarin, A. (2001). Is the market surprised by poor earnings realizations following
seasoned equity offerings? Journal of Financial and Quantitative Analysis, 36(2), 169–193.
Eberhart, A., Maxwell, W., & Siddique, A. (2004). An examination of long-term abnormal stock
returns and operating performance following R&D increases. Journal of Finance, 59(2),
623–650.
Fama, E., & French, K. (1992). The cross-section of expected returns. Journal of Finance, 47,
427–466.
Fama, E., & French, K. (1993). Common risk factors in the returns on stocks and bonds. Journal of
Financial Economics, 33, 3–56.
Fama, E., & French, K. (1996). Multifactor explanations of asset pricing anomalies. Journal of
Finance, 51, 55–84.
Fama, E. F., (1998). Market efficient, long-term returns, and behavioral finance. Journal of
Financial Economics, 49, 283–306.
Ibbotson, R. (1975). Price performance of common stock new issues. Journal of Financial
Economics, 2, 235–272.
50 Long-Run Stock Return and the Statistical Inference 1397
Ikenberry, D., & Ramnath, S. (2002). Underreaction to self-selected news events: The case of
stock splits. Review of Financial Studies, 15, 489–526.
Ikenberry, D., Lakonishok, J., & Vermaelen, T. (1995). Market underreaction to open market share
repurchases. Journal of Financial Economics, 39, 181–208.
Ikenberry, D., Lakonishok, J., & Vermaelen, T. (2000). Stock repurchases in Canada: Performance
and strategic trading. Journal of Finance, 55, 2373–2397.
Jegadeesh, N., & Titman, S. (1993). Returns to buying winners and selling losers: Implications for
stock market efficiency. Journal of Finance, 48, 65–91.
Kothari, S. P., & Warner, J. (1997). Measuring long-horizon security price performance. Journal
of Financial Economics, 43, 301–339.
La Porta, R., Lakonishok, J., Shleifer, A., & Vishny, R. (1997). Good news for value stocks:
Further evidence on market efficiency. Journal of Finance, 52(2), 859–874.
Lakonishok, J., Shleifer, A., & Vishny, R. (1994). Contrarian investment, extrapolation, and risk.
Journal of Finance, 49, 1541–1578.
Lee, I. (1997). Do firms knowingly sell overvalued equity? Journal of Finance, 52, 1439–1466.
Loughran, T., & Ritter, J. (2000). Uniformly least powerful tests of market efficiency. Journal of
Financial Economics, 55, 361–389.
Lyon, J., Barber, B., & Tsai, C. L. (1999). Improved methods for tests of long-run abnormal stock
returns. Journal of Finance, 54, 165–201.
Mitchell, M., & Stafford, E. (2000). Managerial decisions and long-term stock price performance.
Journal of Business, 73, 287–329.
Newey, W., & West, K. (1987). A simple, positive semi-definite, heteroskedasticity and autocor-
relation consistent covariance matrix. Econometrica, 55, 703–708.
Petkova, R., & Zhang, L. (2005). Is value riskier than growth. Journal of Financial Economics, 78,
187–202.
Peyer, U., & Vermaelen, T. (2009). The nature and persistence of buyback anomalies. Review of
Financial Studies, 22, 1693–1745.
Ritter, J. (1991). The long-run performance of initial public offerings. Journal of Finance, 46,
1–27.
Schultz, P. (2003). Pseudo market timing and the long-run underperformance of IPOs. Journal of
Finance, 58, 483–517.
Sloan, R. (1996). Do stock prices fully reflect information in accruals and cash flows about future
earnings? Accounting Review, 71, 289–315.
Value-at-Risk Estimation via
a Semi-parametric Approach: Evidence 51
from the Stock Markets
Contents
51.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1400
51.2 Empirical Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1403
51.2.1 Parametric Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1403
51.2.2 Semi-parametric Method . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1406
51.3 Evaluation Methods of Model-Based VaR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1406
51.3.1 Log-Likelihood Ratio Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1407
51.3.2 Binary Loss Function or Failure Rate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1408
51.3.3 Quadratic Loss Function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1408
51.3.4 The Unconditional Coverage Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1408
51.3.5 Unexpected Loss . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1409
51.4 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1409
51.4.1 Data Preliminary Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1409
51.4.2 Estimation Results for Alternate VaR Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1410
51.4.3 The Results of VaR Performance Assessment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1413
51.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1422
Appendix 1: The Left-Tailed Quantiles of the Standardized SGT . . . . . . . . . . . . . . . . . . . . . . . . . . . 1423
Appendix 2: The Procedure of Parametric VaR Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1423
Appendix 3: The Procedure of Semi-parametric VaR Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1428
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1429
C.-F. Lee
Department of Finance and Economics, Rutgers Business School, Rutgers, The State University of
New Jersey, Piscataway, NJ, USA
Graduate Institute of Finance, National Chiao Tung University, Hsinchu, Taiwan
e-mail: lee@business.rutgers.edu; cflee@business.rutgers.edu
J.-B. Su (*)
Department of Finance, China University of Science and Technology, Nankang, Taipei, Taiwan
e-mail: jungbinsu@cc.cust.edu.tw; jungbinsu@gmail.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1399
DOI 10.1007/978-1-4614-7750-1_51,
# Springer Science+Business Media New York 2015
1400 C.-F. Lee and J.-B. Su
Abstract
This study utilizes the parametric approach (GARCH-based models) and the
semi-parametric approach of Hull and White (Journal of Risk 1: 5–19, 1998)
(HW-based models) to estimate the Value-at-Risk (VaR) through the accuracy
evaluation of accuracy for the eight stock indices in Europe and Asia stock markets.
The measure of accuracy includes the unconditional coverage test by Kupiec
(Journal of Derivatives 3: 73–84, 1995) as well as two loss functions, quadratic
loss function, and unexpected loss. As to the parametric approach, the parameters of
generalized autoregressive conditional heteroskedasticity (GARCH) model are
estimated by the method of maximum likelihood and the quantiles of asymmetric
distribution like skewed generalized student’s t (SGT) can be solved by composite
trapezoid rule. Sequentially, the VaR is evaluated by the framework proposed by
Jorion (Value at Risk: the new benchmark for managing financial risk. New York:
McGraw-Hill, 2000). Turning to the semi-parametric approach of Hull and White
(Journal of Risk 1: 5–19, 1998), before performing the traditional historical
simulation, the raw return series is scaled by a volatility ratio where the volatility
is estimated by the same procedure of parametric approach. Empirical results show
that the kind of VaR approaches is more influential than that of return distribution
settings on VaR estimate. Moreover, under the same return distributional setting,
the HW-based models have the better VaR forecasting performance as compared
with the GARCH-based models. Furthermore, irrespective of whether the GARCH-
based model or HW-based model is employed, the SGT has the best VaR forecast-
ing performance followed by student’s t, while the normal owns the worst VaR
forecasting performance. In addition, all models tend to underestimate the real
market risk in most cases, but the non-normal distributions (student’s t and SGT)
and the semi-parametric approach try to reverse the trend of underestimating.
Keywords
Value-at-Risk • Semi-parametric approach • Parametric approach • Generalized
autoregressive conditional heteroskedasticity • Skewed generalized student’s t •
Composite trapezoid rule • Method of maximum likelihood • Unconditional
coverage test • Loss function
51.1 Introduction
Over the last two decades, a number of global and national financial disasters have
occurred due to failures in risk management procedures. For instance, US Savings and
Loan crisis of 1989–1991, Japanese asset price bubble collapse of 1990, Black
Wednesday of 1992–1993, 1994 economic crisis in Mexico, 1997 Asian Financial
Crisis, 1998 Russian financial crisis, financial crisis of 2007–2010, followed by the
late 2000s recession, and the 2010 European sovereign debt crisis. The crises caused
many enterprises to be liquidated and many countries to face near depressions in
their economies. These painful experiences once again underline the importance
of accurately measuring financial risks and implementing sound risk management
51 Valuet-isk Estimation via a Semi-rametric Approach 1401
policies. Hence, Value-at-Risk (VaR) is a widely used risk measure of the risk of loss
on a specific portfolio of financial assets because it is an attempt to summarize the
total risk with a single number. For example, if a portfolio of stocks has a 1-day 99 %
VaR of US$1,000, there is a 1 % probability that the portfolio will fall in value by
more than US$1,000 over a 1-day period. In other words, we are 99 % certain that we
will not lose more than US$1,000 in the next 1 day, where 1 day is the time horizon,
99 % is the confidence level, and the US$1,000 is the VaR of the portfolio.
VaR estimates are currently based on either of three main approaches: the
historical simulation, the parametric method, and the Monte Carlo simulation.
The Monte Carlo simulation is a class of computational algorithms that rely on
repeated random sampling to compute their results. That is, this approach allows for
an infinite number of possible scenarios you are exposing yourself to huge model
risks in determining the likelihood of any given path. In addition, as you had more
and more variables that could possibly alter your return paths, model complexity
and model risks also increase in scale. Like historical simulation, however, this
methodology removes any assumption of normality and thus, if modeled accu-
rately, probably would give the most accurate measure of the portfolio’s true
VaR. Besides, little research such as Vlaar (2000) had applied this approach to
estimate the VaR. The parametric method is also known as variance/covariance
approach. This method is popular because the only variables you need to do the
calculation are the mean and standard deviation of the portfolio, indicating the
simplicity of the calculations. The parametric method assumes that the returns of
the portfolios are normally distributed and serially independent. In practice, this
assumption of return normality has proven to be extremely risky. Indeed, this was
the biggest mistake that LTCM made gravely underestimating their portfolio
risks. Another weakness with this method is the stability of the standard deviation
through time as well as the stability of the variance/covariance matrix in your
portfolio. However, it is easy to depict how correlations have changed over
time particularly in emerging markets and through contagion in times of
financial crisis. Additionally, numerous studies focused on the parametric approach
of generalized autoregressive conditional heteroskedasticity (GARCH) family
variance specifications (i.e., risk metrics, asymmetric power ARCH (APARCH),
exponential GARCH (EGARCH), threshold GARCH (TGARCH), integrated
GARCH (IGARCH), and fractional IGARCH (FIGARCH)) to estimate the VaR
(see Vlaar (2000), Giot and Laurent (2003a, b), Gencay et al. (2003), Cabedo and
Moya (2003), Angelidis et al. (2004), Huang and Lin (2004), Hartz et al. (2006), So
and Yu (2006), Sadeghi and Shavvalpour (2006), Bali and Theodossiou (2007),
Bhattacharyya et al. (2008), Lee et al. (2008), Lu et al. (2009), Lee and Su (2011),
and so on). Lately, in the empirical study of parametric VaR approach, several
researches have utilized the other type of volatility specifications besides GARCH
family such as the ARJI-GARCH-based model (hereafter ARJI) of Chan and
Maheu (2002) which combines the GARCH specification of volatility and
autoregressive jump intensity (ARJI) in jump intensity (see Su and Hung (2011)
Chang et al. (2011), and so on). Moreover, the other types of long memory volatility
specifications such as fractional integrated APARCH (FIAPARCH) and hyperbolic
1402 C.-F. Lee and J.-B. Su
GARCH (HYGARCH) besides FIGARCH mentioned above are also used to estimate
the VaR (see Aloui and Mabrouk (2010), Degiannakis et al. (2012)). As to distribution
setting, some special distributions like the Weibull distribution (see Gebizlioglu
et al. (2011)), the asymmetric Laplace distribution (see Chen et al. (2012)), and the
Pearson type-IV distribution (see Stavroyiannis et al. (2012)) are also employed to
estimate VaR.
The historical simulation assumes that the past will exactly replicate the future.
The VaR calculation of this approach is literally ranking all of your past historical
returns in terms of lowest to highest and computing with a predetermined confi-
dence rate what your lowest return historically has been.1 In addition, several
studies such as Vlaar (2000), Gencay et al. (2003), Cabedo and Moya (2003), and
Lu et al. (2009) had applied this approach to estimate the VaR. Even though it is
relatively easy to implement, there is a couple of shortcomings of this approach, and
first of all is that it imposes a restriction on the estimation assuming asset returns are
independent and identically distributed (iid) which is not the case. From empirical
evidence, it is known that asset returns are clearly not independent as they exhibit
volatility clustering.2 Therefore, it can be unrealistic to assume iid asset returns.
Second restriction relates to time. Historical simulation applies equal weight to all
returns of the whole period, and this is inconsistent with the nature where there is
diminishing predictability of data that are further away from the present.
These two shortcomings of historical simulation lead this paper to use the
approach proposed by Hull and White (1998) (hereafter, HW method) as
a representative of the semi-parametric approach. This semi-parametric approach
combines the abovementioned parametric approach of GARCH-based variance
specification with the weighted historical simulation. The weighted historical
simulation applies decreasing weights to returns that are further away from
the present, which overcomes the inconsistency of historical simulation with
diminishing predictability of data that are further away from the present. Hence,
this study utilizes the parametric approach (GARCH-N, GARCH-T, and GARCH-
SGT models) and the semi-parametric approach of Hull and White (1998) (HW-N,
HW-T, and HW-SGT models), totaling six models, to estimate the VaR for the
eight stock indices in Europe and Asia stock markets, then uses three accuracy
measures: one likelihood ratio test (the unconditional coverage test (LRuc)
of Kupiec (1995)) and two loss functions (the average quadratic loss function
(AQLF) of Lopez (1999) and the unexpected loss (UL)) to compare the forecasting
ability of the aforementioned models in terms of VaR.
Our results show that the kind of VaR approaches is more influential than that
of return distribution settings on VaR estimate. Moreover, under the same return
distributional setting, the HW-based models have the better VaR forecasting
1
This means if you had 200 past returns and you wanted to know with 99 % confidence what’s the
worst you can do, you would go to the 2nd data point on your ranked series and know that 99 % of
the time you will do no worse than this amount.
2
Large changes tend to be followed by large changes, of either sign, and small changes tend to be
followed by small changes.
51 Valuet-isk Estimation via a Semi-rametric Approach 1403
In this paper, there are two approaches of calculating VaR to be introduced, that is, the
parametric method and the semi-parametric approach. Here, we use the GARCH(1,1)
model with three conditional distributions, namely, the normal, student’s t, and SGT
distributions, to estimate the corresponding volatility in terms of different stock
indices then employ the framework of Jorion (2000) to evaluate the VaR of paramet-
ric approach whereas utilizing the weighting scheme of volatility proposed by Hull
and White (1998) (hereafter, HW method) which is a straightforward extension of
traditional historical simulation to calculate the VaR of semi-parametric VaR.
Many time series data of financial assets appear to exhibit autocorrelated and vola-
tility clustering. Bollerslev et al. (1992) showed that the GARCH(1,1) specification
works well in most applied situations. Furthermore, the unconditional distribution of
those returns displays leptokurtosis and a moderate amount of skewness. Hence,
this study thus considers the applicability of the GARCH(1,1) model with three
conditional distributions, namely, the normal, student’s t, and SGT distributions, to
estimate the corresponding volatility in terms of different stock indices and use the
GARCH model as an official delegate of the VaR model.
where et is the current error and m and st2 are the conditional mean and
variance of return, respectively. Moreover, the variance parameters o, a, and b
are the parameters to be estimated and obey the constraints o, a, b > 0 and
a + b < 1. IID denotes that the standardized errors et are independent and
identically distributed. Since et is drawn from the standard normal distribution,
the probability density function for et is
2
1 e
f ðet Þ ¼ pffiffiffiffiffiffi exp t , (51.3)
2p 2
VaRN ^ tþ1jt
tþ1jt ¼ m þ Fc ðet Þ s (51.5)
where Fc(et) is the left-tailed quantile at c% for the standardized normal distribu-
^ tþ1jt is the one-step-ahead forecasts of the standard deviation of the returns
tion. s
conditional on all information upon the time t.
nþ1
Gð0:5ðn þ 1ÞÞ e2t 2
f ð et Þ ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi 1 þ (51.6)
Gð0:5 nÞ pðn 2Þ n2
where G(•) is the gamma function and n is the shape parameter. Hence, the
log-likelihood function of the GARCH-T model can be expressed as
51 Valuet-isk Estimation via a Semi-rametric Approach 1405
" #
Gð0:5ðn þ 1ÞÞ
LðcÞ ¼ ln f ðrt jOt1 ; cÞ ¼ ln pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
Gð0:5nÞ pðn 2Þ
" # (51.7)
nþ1 rt m 2 1
ln st ln 1 þ ð n 2Þ
2 st
^ tþ1jt
VaRTtþ1jt ¼ m þ Fc ðet ; nÞ s (51.8)
nþ1
j e t þ dj k k
f ð et Þ ¼ C 1 þ (51.9)
½1 þ sign ðet þ dÞlk yk
1
12 pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
where y ¼ Sð1lÞ B k1 ; kn 2 B k3 ; n2
k , SðlÞ ¼ 1 þ 3l2 4A2 l2 ,
1 n
0:5 3 n2
0:5 k
1 n
1
A ¼ B k2 ; n1
k B k;k B k; k , d ¼ S2lA
ðlÞ, C ¼ 2y B k ; k
where k, n, and l are scaling parameters and C and y are normalizing constants
ensuring that f(•) is a proper p.d.f. The parameters k and n control the height and tails
of density with constraints k > 0 and n > 2, respectively. The skewness parameter l
controls the rate of descent of the density around the mode of et with 1 < l < 1. In
the case of positive (resp. negative) skewness, the density function skews toward the
right (resp. left). Sign is the sign function, and B(•) is the beta function. The
parameter n has the degrees of freedom interpretation in case l ¼ 0 and
k ¼ 2. The log-likelihood function of the GARCH-SGT model thus can be written as
k
nþ1 rt m
LðcÞ ¼ ln f ðrt jOt1 ; cÞ ¼ ln C lnst
ln 1 þ þ d
k st
k (51.10)
rt m k
1 þ sign þd l y
st
1406 C.-F. Lee and J.-B. Su
VaRSGT ^ tþ1jt
tþ1jt ¼ m þ Fc ðet ; k; l; nÞ s (51.11)
In this paper, we use the approach proposed by Hull and White (1998) (hereafter, HW
method) as a representative of the semi-parametric approach. The HW method is
a straightforward extension of traditional historical simulation. Instead of using the
actual historical percentage changes in market variables for the purposes of calculat-
ing VaR, we use historical changes that have been adjusted to reflect the ratio of the
current daily volatility to the daily volatility at the time of the observation and assume
that the variance of each market variable during the period covered by the historical
data is monitored using a GARCH model. The methodology is explained in the
following three steps: First, use a raw return series, {r1, r2, r3,......, rt¼T}, to fit the
GARCH(1,1) models with alternative distributions expressed as in Sect. 51.2.1. Thus,
a series of daily volatility estimates, {s1, s2, s3,......, st¼T}, are obtained where T is
the number of estimated samples. Second, the modified return series are obtained by
the raw return series multiplied by the ratio of the current daily volatility to the daily
volatility at the time of the observation, sT/si. That is, the modified return series are
expressed as {r1*, r2*, r3*,......, rt¼T*}, where ri* ¼ ri(sT/si). Finally, sort the returns
ascendingly to achieve the empirical distribution. Thus, VaR is the percentile that
corresponds to the specified confidence level.
The HW-GARCH-SGT model (simply called HW-SGT) implies that the
standardized residual return of the GARCH-SGT model is applied by the HW
approach to estimate the VaR so are HW-N and HW-T models.
Many financial institutions have been required to hold capital against their market risk
exposure, while the market risk capital requirements are based on the VaR estimates
3
See Faires and Burden (2003) for more details.
51 Valuet-isk Estimation via a Semi-rametric Approach 1407
generated by the financial institutions’ own risk management models. Explicitly, the
accuracy of these VaR estimates is of concern to both financial institutions and their
regulators. Hence, model accuracy is important to all VaR model users. To compare
the forecasting ability of the aforementioned models in terms of VaR, this study
considers three accuracy measures: the unconditional coverage test of Kupiec (1995)
which are quite standard in the literatures. Moreover, the quadratic loss function and
the unexpected loss are introduced and used for determining the accuracy of model-
based VaR measurements.
where LRr and LRu are, respectively, the maximum value of the log-likelihood
values under the null hypothesis of the restricted model and the alternative
hypothesis of the unrestricted model and m is the number of the restricted
parameters in the restricted model. For example, LRN for GARCH-SGT model
could be used to test the null hypothesis that log-returns are normally distributed
against the alternative hypothesis that they are SGT distributed. The null hypoth-
esis for testing normality is H0:k ¼ 2, l ¼ 0 and n!1, and the alternative
hypothesis is H1:k∊R+, n > 2 and |l| < 1. Restate, LRN ¼ 2(LRr LRu)
w2(3) where LRr and LRu are, respectively, the maximum value of the
log-likelihood values under the null hypothesis of restricted model (GARCH-N
model) and the alternative hypothesis of unrestricted model (GARCH-SGT
model) and m is the number of the restricted parameters in the restricted model
(k ¼ 2, l ¼ 0 and n!1) and equal to 3 in this case. At the same inference, LRN
for GARCH-T model follows the w2(1) distribution with one degree of freedom.
Moreover, LRT for GARCH-SGT model follows the w2(2) distribution with two
degrees of freedom.
1408 C.-F. Lee and J.-B. Su
If the predicted VaR is not able to cover the realized loss, this is termed a violation.
A binary loss function (BLF) is merely the reflection of the LR test of unconditional
coverage test and gives a penalty of one to each exception of the VaR. The BLF for
long position can be defined as follows:
1 if rtþ1 < VaRtþ1jt ,
BLtþ1 ¼ (51.13)
0 if rtþ1 VaRtþ1jt :
where BLt+1 represents the 1-day-ahead BLF for long position. If a VaR model truly
provides the level of coverage defined by its confidence level, then the average
binary loss function (ABLF) or the failure rate over the full sample will equal c for
the (1 c)th percentile VaR.
The quadratic loss function (QLF) of Lopez (1999) penalizes violations differently
from the binary loss function and pays attention to the magnitude of the violation.
The QLF for long position can be expressed as
2
1 þ rtþ1 VaRtþ1jt if rtþ1 < VaRtþ1jt ,
QLtþ1 ¼ (51.14)
0 if rtþ1 VaRtþ1jt :
where QLt+1 represents the 1-day-ahead QLF for long position. The quadratic term
in Eq. 51.14 ensures that large violations are penalized more than the small
violations which provides a more powerful measure of model accuracy than the
binary loss function.
represents the full sample size. The LRuc test can be employed to test whether the
sample point estimate is statistically consistent with the VaR model’s prescribed
confidence level or not.
The unexpected loss (UL) will equal the average magnitude of the violation over
the full sample. The magnitude of the violation for long position is given by
rtþ1 VaRtþ1jt if rtþ1 < VaRtþ1jt ,
Ltþ1 ¼ (51.16)
0 if rtþ1 VaRtþ1jt :
where Lt+1 is the 1-day-ahead magnitude of the violation for long position.
The study data comprises daily prices of the following eight stock indices: the
Austria ATX (6/29/1999–8/10/2009), the Belgium Brussels (10/19/1999–
8/10/2009), the France CAC40 (10/22/1999–8/10/2009) and the Switzerland
Swiss (9/8/1999–8/10/2009) in Europe, the India Bombay (7/8/1999–8/10/2009),
the Malaysia KLSE (6/23/1999–8/10/2009), the South Korea KOSPI (6/21/1999–
8/10/2009), and the Singapore STRAITS (8/24/1999–8/10/2009) in Asia, where the
numbers in parentheses are the start and end dates for our sample. Daily closing
spot prices for the study period, totaling 2,500 observations, were obtained from
http://finance.yahoo.com. The stock returns are defined as the first difference in the
logarithms of daily stock prices then multiplied by 100.
Table 51.1 summarizes the basic statistical characteristics of return series for both
the estimation and forecast periods. Notably, the average daily returns are all
negative (resp. positive) for forecast (resp. estimation) period and very small
compared with the variable standard deviation, indicating high volatility. Except
the Brussels of estimation period and the CAC40, Swiss and Bombay of forecast
period, all returns series almost exhibit negative skewness for both the estimation
and forecast periods. The excess kurtosis all significantly exceeds zero at the 1 %
level, indicating a leptokurtic characteristic. Furthermore, J-B normality test statis-
tics are all significant at the 1 % level and thus reject the hypothesis of normality
and confirm that neither return series is normally distributed. Moreover, the
Ljung-Box Q2(20) statistics for the squared returns are all significant at the 1 %
1410 C.-F. Lee and J.-B. Su
level and thus indicate that the return series exhibit linear dependence and strong
ARCH effects. Therefore, the preliminary analysis of the data suggests the use of
a GARCH model to capture the fat-tails and time-varying volatility found in these
stock indices return series.
Descriptive graphs (levels of spot prices, density of the daily returns against
normal distribution) for each stock index are illustrated in Fig. 51.1a–h. As shown
in Fig. 51.1, all stock indices have experienced a severe slide in price levels and
display pictures of volatile bear markets for forecast period. Moreover, comparing
density graphs against the normal distribution shows that each return distribution of
data employed exhibits non-normal characteristics. This provides evidence in favor
of some of skewed, leptokurtic, and fat-tailed return distributions. These results are
in line with those of Table 51.1.
Date
c CAC40 stock index : overall-sample daily return density for overall
period period(CAC40)
7000 1.00 0.45
CLOSE Mean −0.01090
0.75 0.40
6000 Std Error 1.58137
0.50 0.35
Skewness 0.02402
0.25 0.30 Exc Kurtosis 5.01027
5000 0.25
0.00
0.20
4000 −0.25 0.15
−0.50 0.10
3000
−0.75 0.05
2000 −1.00 0.00
−10 −5 0 5 10 15
00
01
02
03
04
05
06
07
08
09
20
20
20
20
20
20
20
20
20
20
Date
d Swiss stock index : overall-sample
daily return density for overall
period
10000 1.00 period(Swiss)
CLOSE 0.6
9000 0.75 Mean −0.00717
0.50 0.5 Std Error 1.31938
8000 Skewness 0.06587
0.25 0.4
7000 Exc Kurtosis 5.91237
0.00 0.3
6000
−0.25
0.2
5000 −0.50
0.1
4000 −0.75
−1.00 0.0
3000
−10 −5 0 5 10 15
00
01
02
03
04
05
06
07
08
09
20
20
20
20
20
20
20
20
20
20
Date
20 1
02
20 3
04
20 5
20 6
07
20 8
09
9
0
0
0
19
20
20
20
20
Date
f KLSE stock index : overall-sample daily return density for overall
period period(KLSE)
1600 1.00 0.7
CLOSE 0.75 Mean 0.01615
1400 0.6
0.50 Std Error 1.03179
1200 0.5 Skewness −0.62305
0.25 Exc Kurtosis 7.87111
0.4
1000 0.00
0.3
−0.25
800 0.2
−0.50
600 −0.75 0.1
400 −1.00 0.0
−15 −10 −5 0 5 10
99
00
01
02
03
04
05
06
07
08
09
19
20
20
20
20
20
20
20
20
20
20
Date
g KOSPI stock index : overall-sample daily return density for overall
period period(KOSPI)
2200 1.00 0.35
2000 CLOSE 0.75 Mean 0.02528
0.30 Std Error 1.93219
1800 0.50
0.25 Skewness −0.46118
1600 0.25 Exc Kurtosis 3.94058
1400 0.00 0.20
1200
−0.25 0.15
1000
800 −0.50 0.10
600 −0.75 0.05
400 −1.00 0.00
99
00
01
02
03
04
05
06
07
08
09
−20 −10 0 10 20 30
19
20
20
20
20
20
20
20
20
20
20
Date
h STRAITS stock index : overall-sample daily return density for overall
period period(STRAITS)
4000 1.00 0.6
CLOSE 0.75 Mean 0.00730
3500 0.5 Std Error 1.35780
0.50
3000 Skewness −0.39368
0.25 0.4
Exc Kurtosis 5.34473
2500 0.00 0.3
−0.25
2000 0.2
−0.50
1500 −0.75 0.1
1000 −1.00
0.0
99
00
01
02
03
04
05
06
07
08
09
−10 −5 0 5 10 15
19
20
20
20
20
20
20
20
20
20
20
Date
Fig. 51.1 The stock index in level and daily return density (versus normal) for whole sample (a) ATX,
(b) Brussels, (c) CAC40, (d) Swiss, (e) Bombay, (f) KLSE, (g) KOSPI, (h) STRAITS stock indices
51 Valuet-isk Estimation via a Semi-rametric Approach 1413
most distant day. In this procedure, according to the theory of Sect. 51.2, the
out-of-sample VaR is computed for the next 500 days.
Table 51.2 lists the estimation results4 of the GARCH-N, GARCH-T, and
GARCH-SGT models for the ATX, Brussels, CAC40, and Swiss stock indices in
Europe and the Bombay, KLSE, STRAITS, and KOSPI stock indices in Asia during
the first in-sample period. The variance coefficients o, a, and b are all positive and
significant almost at the 1 % level. Furthermore, the sums of parameters a and b for
these three models are less than one thus ensuring that the conditions for stationary
covariance hold. As to the fat-tail parameters in student’s t distribution, the fat-tail
parameter (n) ranges from 4.9906 (KLSE) to 14.9758 (CAC40) for GARCH-T
model. All these shape parameters are all significant at 1 % level and obey the
constraint n > 2 and thereby implying that the distribution of returns has larger,
thicker tails than the normal distribution. Turning to the shape parameters in SGT
distribution, the fat-tail parameter (n) ranges from 4.9846 (KLSE) to 21.4744
(KOSPI), and the fat-tail parameter (k) is between 1.5399 (KOSPI) and 2.3917
(Bombay). The skewness parameter (l) ranges from 0.1560 (Bombay) to
0.0044(KLSE). Moreover, these three coefficients are almost significant at the
1 % level and thereby these negative skewness parameters imply that the distribu-
tion of returns has a leftward tail. Therefore, both fat-tails and skewness cannot be
ignored in modeling these stock indices returns. The Ljung-Box Q2(20) statistics
for the squared returns are all not significant at the 10 % level and thus indicate that
serial correlation does not exist in standard residuals, confirming that the GARCH
(1,1) specification in these models is sufficient to correct the serial correlation of
these eight return series in the conditional variance equation.
Moreover, as shown in Table 51.2, the LRN statistics for both GARCH-T and
GARCH-SGT models are all significant at the 1 % level, indicating that the null
hypothesis of normality for either stock index is rejected. These results thus imply that
both the student’s t and SGT distributions closely approximate the empirical return
series as compared with the normal distribution. Furthermore, except for ATX and
KLSE stock indices, the LRT statistics of GARCH-SGT model are all significant,
implying that the SGT distribution more closely approximates the empirical return
series than the student’s t does. To sum up, the SGT distribution closely approximates
the empirical return series followed by student’s t and normal distributions.
4
The parameters are estimated by QMLE (quasi-maximum likelihood estimation; QMLE) and the
BFGS optimization algorithm, using the econometric package of WinRATS 6.1.
1414 C.-F. Lee and J.-B. Su
then uses three accuracy measures: one likelihood ratio test (the unconditional
coverage test (LRuc) of Kupiec (1995)) and two loss functions (the average qua-
dratic loss function (AQLF) of Lopez (1999) and the unexpected loss (UL)) to
compare the forecasting ability of the aforementioned models in terms of VaR.
Figure 51.2 graphically illustrates the long VaR forecasts of the GARCH-N,
GARCH-T, and GARCH-SGT models at alternate levels (95 %, 99 %, and 99.5 %)
for all stock indices. Tables 51.3, 51.4 and 51.5 provide the failure rates and the results
of the prior three accuracy evaluation tests (LRuc, AQLF, and UL) for the aforemen-
tioned six models at the 95 %, 99 %, and 99.5 % confidence levels, respectively. As
observed in Tables 51.3, 51.4 and 51.5, we find that, except for a few cases at the 99 %
and 99.5 % confidence levels, all models tend to underestimate real market risk
because the empirical failure rate is higher than the theoretical failure rate in most
cases. The abovementioned exceptional cases emerge at the GARCH-SGT model
of 99 % level (CAC40); both the GARCH-T and GARCH-SGT models of 99.5 %
level (KLSE and STRAITS); the HW-N (KLSE), HW-T (KLSE), and HW-SGT
(KLSE and STRAITS) models of 99 % level; and the HW-N (STRAITS), HW-T
(ATX and STRAITS), and HW-SGT (ATX, KLSE, and STRAITS) models of 99.5 %
level, where the stock indices in parentheses behind the models are the exceptional
cases. Moreover, the empirical failure rate of the above exceptional cases is
lower than the theoretical failure rate, indicating that the non-normal distributions
(student’s t and SGT) and the semi-parametric approach try to reverse the trend of
underestimating real market risk, especially at the 99.5 % level.
1416 C.-F. Lee and J.-B. Su
10 10 10 10
5 5 5 5
0 0 0 0
Return
Return
Return
Return
-5 −5 −5 −5
10 10 10 10
5 5 5 5
0 0
Return
Return
0 0
Return
Return
−5 −5 −5 −5
−15 −15
−15 −15
−20 −20
−20 −20
2008 2009
2008 2009
Date
Date
f KLSE index: VaR at
e Bombay index: VaR at alternate levels
alternate levels 5.0 5.0
15 15
2.5 2.5
10 10
0.0 0.0
5 5 −2.5 −2.5
−5.0 −5.0
Return
Return
0 0
Return
Return
−7.5 −7.5
−5 −5
−10.0 −10.0
−10 −10
−12.5 −12.5
−15 −15 −15.0 −15.0
10 10 10 10
5 5 5 5
0 0 0 0
Return
Return
Return
Return
−5 −5 −5 −5
Fig. 51.2 Long VaR forecasts at alternative level for the normal, student’s t, and SGT distribu-
tion. (a) ATX, (b) Brussels, (c) CAC40, (d) Swiss, (e) Bombay, (f) KLSE, (g) KOSPI,
(h) STRAITS stock indices
Panel F. KLSE
N 0.0560(0.36*) 0.20084 –0.04595 0.0740(5.31) 0.23255 –0.05623
T 0.0600(0.99*) 0.21827 –0.05152 0.0700(3.76*) 0.23146 –0.05276
SGT 0.0580(0.64*) 0.21375 –0.05007 0.0640(1.90*) 0.22378 –0.05306
Panel G. STRAITS
N 0.0580(0.64*) 0.21980 –0.06233 0.0640(1.90*) 0.24194 –0.06675
T 0.0640(1.90*) 0.25978 –0.07222 0.0620(1.41*) 0.23921 –0.06511
SGT 0.0620(1.41*) 0.24670 –0.06736 0.0560(0.36*) 0.23035 –0.06428
Panel H. KOSPI
N 0.0740(5.31) 0.27949 –0.08826 0.0600(0.99*) 0.24267 –0.07816
T 0.0760(6.18) 0.32360 –0.09624 0.0620(1.41*) 0.25234 –0.07522
SGT 0.0672(2.82*) 0.27616 –0.08236 0.0620(1.41*) 0.24475 –0.07531
Notes: 1. *Indicates that the model passes the unconditional coverage test at the 5 % significance
level and the critical value of the LRuc test statistics at the 5 % significance level is 3.84. 2. The red
(resp. blue) font represents the lowest (resp. highest) AQLF and unexpected loss when the
predictive accuracies of three different innovations with the same VaR method are compared.
3. The delete-line font represents the lowest AQLF and unexpected loss when the predictive
accuracies of two different VaR methods with the same innovation are compared. 4. The model
acronyms stand for the following methods: HW-GARCH non-parametric method proposed by Hull
and White (1998), GARCH parametric method of GARCH model, N the standard normal distri-
bution, T the standardized student’s t distribution, SGT the standardized SGT distribution proposed
by Theodossiou (1998)
51 Valuet-isk Estimation via a Semi-rametric Approach 1419
Panel B. Brussels
N 0.0260(8.97) 0.13491 –0.02538 0.0180(2.61*) 0.03685 –0.01726
T 0.0180(2.61*) 0.11962 –0.01947 0.0160(1.53*) 0.03465 –0.01559
SGT 0.0160(1.53*) 0.11033 –0.01711 0.0160(1.53*) 0.03536 –0.01596
Panel C. CAC40
N 0.0200(3.91) 0.14794 –0.01913 0.0160(1.53*) 0.07472 –0.01808
T 0.0100(0.00*) 0.14325 –0.01494 0.0100(0.00*) 0.05628 –0.01436
SGT 0.0060(0.94*) 0.13144 –0.01354 0.0140(0.71*) 0.05992 –0.01466
Panel D. Swiss
N 0.0260(8.97) 0.12387 –0.01967 0.0180(2.61*) 0.03938 –0.01353
T 0.0160(1.53*) 0.11943 –0.01516 0.0140(0.71*) 0.03528 –0.01341
SGT 0.0140(0.71*) 0.10151 –0.01244 0.0140(0.71*) 0.03512 –0.01338
Panel E. Bombay
N 0.0300(13.16) 0.23811 –0.03748 0.0120(0.18*) 0.04961 –0.01751
T 0.0220(5.41) 0.22706 –0.02822 0.0120(0.18*) 0.04539 –0.01586
SGT 0.0180(2.61*) 0.19232 –0.02152 0.0120(0.18*) 0.04504 –0.01617
Panel F. KLSE
N 0.0160(1.53*) 0.16522 –0.01892 0.0080(0.21*) 0.07681 –0.01341
T 0.0100(0.00*) 0.16891 –0.01585 0.0060(0.94*) 0.08114 –0.01425
SGT 0.0100(0.00*) 0.16278 –0.01551 0.0060(0.94*) 0.07965 –0.01384
Panel G. STRAITS
N 0.0240(7.11) 0.16096 –0.01848 0.0120(0.18*) 0.09107 –0.01763
T 0.0100(0.00*) 0.17606 –0.01568 0.0100(0.00*) 0.07477 –0.01403
SGT 0.0100(0.00*) 0.16258 –0.01406 0.0080(0.21*) 0.07278 –0.01361
Panel H. KOSPI
N 0.0220(5.41) 0.18050 –0.02675 0.0200(3.91) 0.03799 –0.01639
T 0.0200(3.91) 0.20379 –0.02199 0.0180(2.61*) 0.04722 –0.01942
SGT 0.0163(1.68*) 0.15465 –0.01563 0.0180(2.61*) 0.04487 –0.01941
Note: 1. *indicates that the model passes the unconditional coverage test at the 5 % significance
level and the critical value of the LRuc test statistics at the 5 % significance level is 3.84. 2. The red
(resp. blue) font represents the lowest (resp. highest) AQLF and unexpected loss when the
predictive accuracies of three different innovations with the same VaR method are compared.
3. The delete-line font represents the lowest AQLF and unexpected loss when the predictive
accuracies of two different VaR methods with the same innovation are compared. 4. The model
acronyms stand for the following methods: HW-GARCH non-parametric method proposed by
Hull and White (1998), GARCH parametric method of GARCH model, N the standard normal
distribution, T the standardized student’s t distribution, SGT the standardized SGT distribution
proposed by Theodossiou (1998)
1420 C.-F. Lee and J.-B. Su
Table 51.5 Out-of-sample long VaR performance at the 99.5 % confidence level
GARCH HW-GARCH
Failure rate (LRuc) AQLF UL Failure rate (LRuc) AQLF UL
Panel A. ATX
N 0.0960(17.75) 0.30360 –0.10151 0.0960(17.75) 0.30642 –0.10308
T 0.0960(17.75) 0.3 2058 –0.10865 0.0920(15.04) 0.30342 –0.10384
SGT 0.0900(13.75) 0.29877 –0.10212 0.0980(19.18) 0.32263 –0.10858
Panel B. Brussels
N 0.0780(7.10) 0.20246 –0.07099 0.0680(3.08*) 0.18374 –0.06673
T 0.0720(4.51) 0.19494 –0.07020 0.0620(1.41*) 0.16978 –0.06268
SGT 0.0660(2.45*) 0.18178 –0.06602 0.0620(1.41*) 0.17067 –0.06262
Panel C. CAC40
N 0.0740(5.31) 0.21446 –0.06485 0.0700(3.76*) 0.20928 –0.05844
T 0.0800(8.07) 0.22598 –0.06887 0.0720(4.51) 0.19510 –0.05617
SGT 0.0760(6.18) 0.21146 –0.06281 0.0700(3.76*) 0.19390 –0.05600
Panel D. Swiss
N 0.0760(6.18) 0.18567 –0.06181 0.0620(1.41*) 0.15104 –0.05088
T 0.0740(5.31) 0.18729 –0.06345 0.0600(0.99*) 0.14222 –0.04694
SGT 0.0740(5.31) 0.17506 –0.05698 0.0560(0.36*) 0.13928 –0.04697
Panel E. Bombay
N 0.0780(7.10) 0.34428 –0.10214 0.0800(8.07) 0.33643 –0.09967
T 0.0820(9.11) 0.35999 –0.10617 0.0780(7.10) 0.33182 –0.09639
SGT 0.0700(3.76*) 0.31488 –0.09366 0.0800(8.07) 0.33629 –0.09824
Panel F. KLSE
N 0.0560(0.36*) 0.20084 –0.04595 0.0740(5.31) 0.23255 –0.05623
T 0.0600(0.99*) 0.21827 –0.05152 0.0700(3.76*) 0.23146 –0.05276
SGT 0.0580(0.64*) 0.21375 –0.05007 0.0640(1.90*) 0.22378 –0.05306
Panel G. STRAITS
N 0.0580(0.64*) 0.21980 –0.06233 0.0640(1.90*) 0.24194 –0.06675
T 0.0640(1.90*) 0.25978 –0.07222 0.0620(1.41*) 0.23921 –0.06511
SGT 0.0620(1.41*) 0.24670 –0.06736 0.0560(0.36*) 0.23035 –0.06428
Panel H. KOSPI
N 0.0740(5.31) 0.27949 –0.08826 0.0600(0.99*) 0.24267 –0.07816
T 0.0760(6.18) 0.32360 –0.09624 0.0620(1.41*) 0.25234 –0.07522
SGT 0.0672(2.82*) 0.27616 –0.08236 0.0620(1.41*) 0.24475 –0.07531
Note: 1.*Indicates that the model passes the unconditional coverage test at the 5 % significance
level and the critical value of the LRuc test statistics at the 5 % significance level is 3.84. 2. The red
(resp. blue) font represents the lowest (resp. highest) AQLF and unexpected loss when the
predictive accuracies of three different innovations with the same VaR method are compared.
3. The delete-line font represents the lowest AQLF and unexpected loss when the predictive
accuracies of two different VaR methods with the same innovation are compared. 4. The model
acronyms stand for the following methods: HW-GARCH non-parametric method proposed by Hull
and White (1998), GARCH parametric method of GARCH model, N the standard normal distri-
bution, T the standardized student’s t distribution, SGT the standardized SGT distribution proposed
by Theodossiou (1998)
51 Valuet-isk Estimation via a Semi-rametric Approach 1421
level, Table 51.3 describes that the HW-N, HW-T, and HW-SGT models pass the
LRuc tests with a total of 5, 5, and 6 stock indices, respectively. Moreover, for
99 % confidence level, Table 51.4 depicts that the HW-N, HW-T, and HW-SGT
models pass the LRuc tests with a total of 7, 8, and 8 stock indices, respectively;
for 99.5 % confidence level, Table 51.5 illustrates that the HW-N, HW-T, and
HW-SGT models pass the LRuc tests with a total of 7, 8, and 8 stock indices,
respectively. Hence, under all confidence levels, there is a total of 19, 21, and
22 acceptances for HW-N, HW-T, and HW-SGT models (the semi-parametric
approach), respectively.
From the abovementioned results, we can find the following two important
phenomena: First, under the same return distributional setting, the number of
acceptance of the HW-based models is greater or equal than those of the
GARCH-based models, irrespective of whether the case of individual level
(95 %, 99 %, or 99.5 %) or all levels (95 %, 99 %, and 99.5 %) is considered.
For example, with regard to all levels, the number of acceptance of the HW-N
model (19) is greater than those of the GARCH-N models (6). These results reveal
that the HW-based models (semi-parametric approach) have the better VaR fore-
casting performance as compared with GARCH-based models (parametric
approach). Second, the number of acceptance of the SGT distribution is the greatest
followed by the student’s t and normal distributions, irrespective of whether the
GARCH-based model (parametric) or HW-based model (semi-parametric
approach) is employed. For instance, with regard to all levels, the number of
acceptance of the GARCH-SGT model (20) is the greatest followed by the
GARCH-T model (14) and GARCH-N model (6). These results indicate that
the SGT has the best VaR forecasting performance followed by student’s t while
the normal owns the worst VaR forecasting performance.
Turning to the other two accuracy measures (i.e., AQLF and UL), the two loss
functions (the average quadratic loss function (AQLF) and the unexpected loss
(UL)) reflect the magnitude of the violation which occur as the observed return
exceeds the VaR estimation. The smaller the AQLF and UL are generated,
the better the forecasting performance of the models is. As observed in
Tables 51.3, 51.4 and 51.5, we can also find the following two important phe-
nomena which are similar as those of the back-testing as was mentioned above:
First, under the same return distributional setting, the AQLF and UL generated by
the HW-based models are smaller than those generated by the GARCH-based
models, irrespective of whether the 95 %, 99 %, or 99.5 % level is considered.
These results reveal that the HW-based models (semi-parametric approach)
significantly have the better VaR forecasting performance as compared with
GARCH-based models (parametric approach), which is in line with the results
of the back-testing. Second, for all confidence levels, the GARCH-SGT
model yields the lowest AQLF and UL for most of the stock indices.
Moreover, for most of the stock indices, the GARCH-N model produces the
highest AQLF and UL for both 99 % and 99.5 % levels, while the GARCH-T
model gives the highest AQLF and UL for 95 % level. These results indicate
that the GARCH-SGT model significantly owns the best out-of-sample VaR
1422 C.-F. Lee and J.-B. Su
performance, while the GARCH-N model appears to have the worst out-of-
sample VaR performance. On the contrary, for all confidence levels, the HW-N
model bears the highest AQLF and UL for most of the stock indices, while the
HW-SGT model gives the lowest AQLF and UL for half of the stock indices,
indicating that the HW-N model significantly owns the worst out-of-sample VaR
performance, while the HW-SGT model appears to bear the highest out-of-
sample VaR performance. Consequently, it seems reasonable to conclude that
the SGT has the best VaR forecasting performance followed by student’s t, while
the normal owns the worst VaR forecasting performance, which appears to be
consistent with the results of back-testing.
To sum up, according to the three accuracy measures, the HW-based models
(semi-parametric approach) have the better VaR forecasting performance as com-
pared with GARCH-based models (parametric approach), and the SGT has the best
VaR forecasting performance followed by student’s t, while the normal owns the
worst VaR forecasting performance. In addition, the kind of VaR approach is more
influential than that of return distribution setting on VaR estimate.
51.5 Conclusion
This study utilizes the parametric approach (GARCH-N, GARCH-T, and GARCH-
SGT models) and the semi-parametric approach of Hull and White (1998) (HW-N,
HW-T, and HW-SGT models), totaling six models, to estimate the VaR for the
eight stock indices in Europe and Asia stock markets, then uses three accuracy
measures: one likelihood ratio test (the unconditional coverage test (LRuc) of
Kupiec (1995)) and two loss functions (the average quadratic loss function
(AQLF) of Lopez (1999) and the unexpected loss (UL)) to compare the forecasting
ability of the aforementioned models in terms of VaR.
The empirical findings can be summarized as follows. First, according to the
results of the log-likelihood ratio test, the SGT distribution closely approximates
the empirical return series followed by student’s t and normal distributions.
Second, in terms of the failure rate, all models tend to underestimate the real market
risk in most cases, but the non-normal distributions (student’s t and SGT) and
the semi-parametric approach try to reverse the trend of underestimating real
market risk, especially at the 99.5 % level. Third, the kind of VaR approaches is
more influential than that of return distribution settings on VaR estimate.
Moreover, under the same return distributional setting, the HW-based models
(semi-parametric approach) have the better VaR forecasting performance as com-
pared with the GARCH-based models (parametric approach). Finally, irrespective
of whether the GARCH-based model (parametric) or HW-based model (semi-
parametric approach) is employed, the SGT has the best VaR forecasting perfor-
mance followed by student’s t, while the normal owns the worst VaR forecasting
performance.
51 Valuet-isk Estimation via a Semi-rametric Approach 1423
The standardized SGT distribution was derived by Lee and Su (2011) and expressed
as follows:
nþ1
j e t þ dj k k
f ð et Þ ¼ C 1 þ k k (51.17)
½1 þ signðet þ dÞl y
1
12 pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
where y ¼ Sð1lÞ B k1 ; kn 2 B k3 ; n2
k , SðlÞ ¼ 1 þ 3l2 4A2 l2 ,
1 n
0:5 3 n2
0:5 k
1 n
1
A ¼ B k2 ; n1
k B k;k B k; k , d ¼ S2lA
ðlÞ, C ¼ 2y B k ; k
where k, n, and l are scaling parameters and C and y are normalizing constants
ensuring that f(•) is a proper p.d.f. The parameters k and n control the height and tails
of density with constraints k > 0 and n > 2, respectively. The skewness parameter l
controls the rate of descent of the density around the mode of et with 1 < l < 1. In
the case of positive (resp. negative) skewness, the density function skews toward the
right (resp. left). Sign is the sign function, and B(•) is the beta function. The parameter
n has the degrees of freedom interpretation in case l ¼ 0 and k ¼ 2. Particularly, the
SGT distribution generates the student’s t distribution for l ¼ 0 and k ¼ 2. Moreover,
the SGT distribution generates the normal distribution for l ¼ 0, k ¼ 2, and n ¼ 1.
As observed from Table 51.2, the shape parameters in SGT distribution, the
fat-tail parameter (n) ranges from 4.9846 (KLSE) to 21.4744 (KOSPI), and
the fat-tail parameter (k) is between 1.5399 (KOSPI) and 2.3917 (Bombay). The
skewness parameter (l) ranges from 0.1560 (Bombay) to 0.0044 (KLSE).
Therefore, the left-tailed quantiles of the SGT distribution with various combinations
of shape parameters (0.15 l 0.05; 1.0 k 2.0; n ¼ 10) at alternate levels are
obtained by the composite trapezoid rule and are listed in Table 51.6. Moreover,
Fig. 51.3 depicts the left-tailed quantiles surface of SGT (versus normal) distribution
with various combinations of shape parameters (0.25 l 0.25; 0.8 k 2.0;
n ¼ 10 and 20) at 10 %, 5 %, 1 %, and 0.5 % levels. Notably, Fc(et;k ¼ 2, l ¼ 0, n ¼ 1)
where c ¼ 0.1, 0.05, 0.01, and 0.005 in Fig. 51.3 represents the left-tailed quantiles of
normal distribution at 10 %, 5 %, 1 %, and 0.5 % levels, which is 1.28155,
1.64486, 2.32638, and 2.57613, respectively.
The parametric method is very popular because the only variables you need to do
the calculation are the mean and standard deviation of the portfolio, indicating the
simplicity of the calculations. Moreover, from the literatures’ review mentioned
above, numerous studies focused on the parametric approach of the GARCH family
1424
Table 51.6 The left-tailed quantiles of SGT distribution with n ¼ 10 and various combinations (k, l) at alternate levels
k\l 0.15 0.10 0.05 0.00 0.05
Panel A. 10 % level
1.0 1.6848(1.1536) 1.6443(1.1359) 1.6003(1.1165) 1.5532(1.0958) 1.5033(1.0739)
1.1 1.7005(1.1831) 1.6615(1.1659) 1.6193(1.1472) 1.5742(1.1273) 1.5266(1.1062)
1.2 1.7103(1.2066) 1.6729(1.1901) 1.6324(1.1721) 1.5893(1.1530) 1.5439(1.1329)
1.3 1.7160(1.2257) 1.6800(1.2098) 1.6413(1.1927) 1.6001(1.1744) 1.5568(1.1552)
1.4 1.7189(1.2414) 1.6843(1.2261) 1.6471(1.2097) 1.6078(1.1923) 1.5665(1.1740)
1.5 1.7196(1.2544) 1.6863(1.2398) 1.6508(1.2241) 1.6132(1.2075) 1.5738(1.1901)
1.6 1.7189(1.2652) 1.6869(1.2512) 1.6528(1.2363) 1.6168(1.2204) 1.5792(1.2039)
1.7 1.7171(1.2744) 1.6864(1.2610) 1.6536(1.2467) 1.6191(1.2316) 1.5831(1.2159)
1.8 1.7147(1.2822) 1.6850(1.2694) 1.6536(1.2557) 1.6205(1.2413) 1.5860(1.2264)
1.9 1.7116(1.2888) 1.6831(1.2766) 1.6528(1.2635) 1.6211(1.2498) 1.5881(1.2356)
2.0 1.7083(1.2946) 1.6807(1.2828) 1.6516(1.2704) 1.6211(1.2573) 1.5894(1.2438)
Panel B. 5 % level
1.0 1.6848(1.7252) 1.6443(1.6851) 1.6003(1.6418) 1.5532(1.5955) 1.5033(1.5468)
1.1 1.7005(1.7349) 1.6615(1.6966) 1.6193(1.6553) 1.5742(1.6114) 1.5266(1.5652)
1.2 1.7103(1.7397) 1.6729(1.7031) 1.6324(1.6639) 1.5893(1.6222) 1.5439(1.5785)
1.3 1.7160(1.7413) 1.6800(1.7064) 1.6413(1.6690) 1.6001(1.6294) 1.5568(1.5880)
1.4 1.7189(1.7406) 1.6843(1.7073) 1.6471(1.6716) 1.6078(1.6340) 1.5665(1.5947)
1.5 1.7196(1.7385) 1.6863(1.7065) 1.6508(1.6726) 1.6132(1.6368) 1.5738(1.5995)
1.6 1.7189(1.7353) 1.6869(1.7047) 1.6528(1.6723) 1.6168(1.6382) 1.5792(1.6027)
1.7 1.7171(1.7318) 1.6864(1.7021) 1.6536(1.6711) 1.6191(1.6387) 1.5831(1.6048)
1.8 1.7147(1.7270) 1.6850(1.6990) 1.6536(1.6694) 1.6205(1.6383) 1.5860(1.6061)
1.9 1.7116(1.7224) 1.6831(1.6955) 1.6528(1.6671) 1.6211(1.6375) 1.5881(1.6067)
2.0 1.7083(1.7177) 1.6807(1.6918) 1.6516(1.6646) 1.6211(1.6362) 1.5894(1.6068)
C.-F. Lee and J.-B. Su
51
Panel C. 1 % level
1.0 3.1972(3.1305) 3.0964(3.0358) 2.9871(2.9336) 2.8702(2.8247) 2.7467(2.7100)
1.1 3.1281(3.0522) 3.0323(2.9630) 2.9289(2.8673) 2.8188(2.7657) 2.7029(2.6589)
1.2 3.0623(2.9797) 2.9711(2.8956) 2.8732(2.8057) 2.7693(2.7106) 2.6602(2.6109)
1.3 3.0005(2.9132) 2.9136(2.8336) 2.8207(2.7489) 2.7224(2.6596) 2.6195(2.5661)
1.4 2.9430(2.8522) 2.8600(2.7768) 2.7717(2.6968) 2.6785(2.6126) 2.5811(2.5247)
1.5 2.8896(2.7963) 2.8103(2.7247) 2.7261(2.6489) 2.6376(2.5693) 2.5452(2.4864)
1.6 2.8402(2.7451) 2.7643(2.6769) 2.6839(2.6049) 2.5995(2.5295) 2.5117(2.4511)
1.7 2.7945(2.6980) 2.7216(2.6329) 2.6447(2.5644) 2.5641(2.4927) 2.4804(2.4184)
1.8 2.7522(2.6547) 2.6821(2.5924) 2.6083(2.5270) 2.5312(2.4588) 2.4512(2.3881)
1.9 2.7129(2.6147) 2.6454(2.5550) 2.5745(2.4925) 2.5006(2.4274) 2.4240(2.3600)
2.0 2.6764(2.5777) 2.6113(2.5204) 2.5431(2.4606) 2.4720(2.3983) 2.3986(2.3339)
Panel D. 0.5 % level
1.0 3.9229(3.7705) 3.7941(3.6512) 3.6542(3.5224) 3.5043(3.3851) 3.3459(3.2405)
1.1 3.8012(3.6396) 3.6789(3.5277) 3.5470(3.4077) 3.4064(3.2802) 3.2583(3.1464)
1.2 3.6897(3.5224) 3.5736(3.4173) 3.4490(3.3050) 3.3168(3.1862) 3.1780(3.0618)
Valuet-isk Estimation via a Semi-rametric Approach
−0.8 Fα=0.10(ε;κ,λ,n=20)
−1.2
−1.4
−1.6
Fα=0.10(ε;κ,λ,n=∞)
−1.8
2
1.5 0.4
Fα=0.10(ε;κ,λ,n=10)
0.2
1 0
−0.2
Kappa(κ) 0.5 −0.4 lambda(λ)
−1
−1.2
LHS Critical Value
−1.4
−1.6
−1.8
−2 Fα=0.05(ε;κ,λ,n=∞)
2 Fα=0.05(ε;κ,λ,n=10)
1.5 Fα=0.05(ε;κ,λ,n=20) 0.4
0.2
1 0
−0.2
Kappa(κ) 0.5 −0.4 lambda(λ)
F∝=0.01(ε;κ,λ,n=20)
−2
−2.5
LHS Critical Value
−3
−3.5 Fα=0.01(ε;κ,λ,n=10)
F (ε;κ=2,λ=0,n=∞)
α=0.01
−4
2
1.5 0.4
0.2
1 0
−0.2
Kappa(κ) 0.5 −0.4 lambda(λ)
−2
−2.5
LHS Critical Value
−3
−3.5
F (ε;κ,λ,n=20)
α=0.005 F(ε;κ=2,λ=0,n=∞)
−4 α=0.005
−4.5
2 F(ε;κ,λ,n=10)
α=0.005
1.5 0.4
0.2
1 0
−0.2
Kappa(κ) 0.5 −0.4
lambda(λ)
Fig. 51.3 The left-tailed quantiles of SGT distribution with n ¼ 10, 20, and various combinations
(k, l). (a) 10 %, (b) 5 %, (c) 1 %, (d) 0.5 % confidence levels
parameters is estimated, and then the 1-day-ahead VaR can be calculated for the
next 500 days.
Step 4: For the out-sample period (500 days), via the comparable results between
the one-step-ahead VaR forecasts and the observed returns, the 1-day-ahead
BLF, QLF, and UL can be calculated by using Eqs. 51.13, 51.14 and 51.16. On
the other hand, the unconditional coverage test, LRuc, is evaluated by employing
Eq. 51.15. Thereafter, with regard to the GARCH-based models with alternate
distributions (GARCH-N, GARCH-T, and GARCH-SGT), the unconditional
coverage test (LRuc) and three loss functions (failure rate, AQLF, and UL) are
obtained and are reported in the left panel of Tables 51.3, 51.4 and 51.5 for 95 %,
99 %, and 99.5 % levels.
In this paper, we use the approach proposed by Hull and White (1998) as
a representative of the semi-parametric approach. This method mainly couples
a weighting scheme of volatility with the traditional historical simulation. Hence,
it can be regarded as a straightforward extension of traditional historical simulation.
The weighting scheme of volatility is expressed as follows. Instead of using the
actual historical percentage changes in market variables for the purposes of calcu-
lating VaR, we use historical changes that have been adjusted to reflect the ratio of
the current daily volatility to the daily volatility at the time of the observation and
assume that the variance of each market variable during the period covered by the
historical data is monitored using a GARCH-based models. We take an example of
the HW-SGT model. This methodology is explained in the following five steps:
Step 1: For each data series, using the econometric package of WinRATS 6.1, the
parameters are estimated with a sample of 2,000 daily returns by quasi-
maximum likelihood estimation (QMLE) of log-likelihood function such as
Eq. 51.10 and by the BFGS optimization algorithm. Thus, with c ¼ [m, o, a,
b, k, l, n], the vector of parameters is estimated. This step is the same as the first
step of parametric approach. Consequently, a series of daily volatility estimates,
{s1, s2, s3,......, st¼T}, are obtained where T is the number of estimated samples
and equals 2,000 in this study.
Step 2: The modified return series are obtained by the raw return series multiplied
by the ratio of the current daily volatility to the daily volatility at the time of the
observation, sT/si. That is, the modified return series are expressed as {r1*, r2*,
r3*,......, rt¼T*}, where ri* ¼ ri(sT/si).
Step 3: Resort this modified return series ascendingly to achieve the empirical
distribution. Thus, VaR is the percentile that corresponds to the specified
confidence level. Then the one-step-ahead VaR forecasts are compared with
the observed returns, and the comparative results are recorded for subsequent
evaluation using statistical tests.
Step 4: The estimation period is then rolled forwards by adding one new day and
dropping the most distant day. By replicating steps 1–3, the vector of parameters
51 Valuet-isk Estimation via a Semi-rametric Approach 1429
is estimated, and then the 1-day-ahead VaR can be calculated for the next
500 days. This step is the same as the third step of parametric approach.
Step 5: For the out-sample period (500 days), via the comparable results between
the one-step-ahead VaR forecasts and the observed returns, the 1-day-ahead
BLF, QLF, and UL can be calculated by using Eqs. 51.13, 51.14, and 51.16.
On the other hand, the unconditional coverage test, LRuc, is evaluated by
employing Eq. 51.15. Thereafter, with regard to HW-based models with
alternate distributions (HW-N, HW-T, and HW-SGT), the unconditional
coverage test (LRuc) and three loss functions (failure rate, AQLF, and UL)
are obtained and are reported in the right panel of Tables 51.3, 51.4 and 51.5
for 95 %, 99 %, and 99.5 % levels.
References
Aloui, C., & Mabrouk, S. (2010). Value-at-risk estimations of energy commodities via long-
memory, asymmetry and fat-tailed GARCH models. Energy Policy, 38, 2326–2339.
Angelidis, T., Benos, A., & Degiannakis, S. (2004). The use of GARCH models in VaR estimation.
Statistical Methodology, 1, 105–128.
Bali, T. G., & Theodossiou, P. (2007). A conditional-SGT-VaR approach with alternative GARCH
models. Annals of Operations Research, 151, 241–267.
Bhattacharyya, M., Chaudhary, A., & Yadav, G. (2008). Conditional VaR estimation using
Pearson’s type IV distribution. European Journal of Operational Research, 191, 386–397.
Bollerslev, T. (1986). Generalized autoregressive conditional heteroskedasticity. Journal of
Econometrics, 31, 307–327.
Bollerslev, T., Chou, R. Y., & Kroner, K. F. (1992). ARCH modeling in finance: A review of the
theory and empirical evidence. Journal of Econometrics, 52, 5–59.
Cabedo, J. D., & Moya, I. (2003). Estimating oil price ‘Value at Risk’ using the historical
simulation approach. Energy Economics, 25, 239–253.
Chan, W. H., & Maheu, J. M. (2002). Conditional jump dynamics in stock market returns. Journal
of Business Economics Statistics, 20, 377–389.
Chang, T. H., Su, H. M., & Chiu, C. L. (2011). Value-at-risk estimation with the optimal dynamic
biofuel portfolio. Energy Economics, 33, 264–272.
Chen, Q., Gerlach, R., & Lu, Z. (2012). Bayesian value-at-risk and expected shortfall forecasting via
the asymmetric Laplace distribution. Computational Statistics and Data Analysis, 56, 3498–3516.
Degiannakis, S., Floros, C., & Dent, P. (2012). Forecasting value-at-risk and expected shortfall
using fractionally integrated models of conditional volatility: International evidence. Interna-
tional Review of Financial Analysis, 27, 21–33.
Faires, J. D., & Burden, R. (2003). Numerical methods (3rd ed.). Pacific Grove: Tomson Learning.
Gebizlioglu, O. L., Şenoğlu, B., & Kantar, Y. M. (2011). Comparison of certain value-at-risk
estimation methods for the two-parameter Weibull loss distribution. Journal of Computational
and Applied Mathematics, 235, 3304–3314.
Gençay, R., Selçuk, F., & Ulug€ulyaˇgci, A. (2003). High volatility, thick tails and extreme value
theory in value-at-risk estimation. Insurance: Mathematics and Economics, 33, 337–356.
Giot, P., & Laurent, S. (2003a). Value-at-risk for long and short trading positions. Journal of
Applied Econometrics, 18, 641–664.
Giot, P., & Laurent, S. (2003b). Market risk in commodity markets: A VaR approach. Energy
Economics, 25, 435–457.
Hartz, C., Mittnik, S., & Paolellad, M. (2006). Accurate value-at-risk forecasting based on the
normal-GARCH model. Computational Statistics & Data Analysis, 51, 2295–2312.
1430 C.-F. Lee and J.-B. Su
Huang, Y. C., & Lin, B. J. (2004). Value-at-risk analysis for Taiwan stock index futures: Fat tails
and conditional asymmetries in return innovations. Review of Quantitative Finance and
Accounting, 22, 79–95.
Hull, J., & White, A. (1998). Incorporating volatility updating into the historical simulation
method for value-at-risk. Journal of Risk, 1, 5–19.
Jarque, C. M., & Bera, A. K. (1987). A test for normality of observations and regression residuals.
International Statistics Review, 55, 163–172.
Jorion, P. (2000). Value at risk: The new benchmark for managing financial risk. New York:
McGraw-Hill.
Kupiec, P. (1995). Techniques for verifying the accuracy of risk management models. The Journal
of Derivatives, 3, 73–84.
Lee, C. F., & Su, J. B. (2011). Alternative statistical distributions for estimating value-at-risk:
Theory and evidence. Review of Quantitative Finance and Accounting, 39, 309–331 .
Lee, M. C., Su, J. B., & Liu, H. C. (2008). Value-at-risk in US stock indices with skewed
generalized error distribution. Applied Financial Economics Letters, 4, 425–431.
Lopez, J. A. (1999). Methods for evaluating value-at-risk estimates. Federal Reserve Bank of San
Francisco Economic Review, 2, 3–17.
Lu, C., Wu, S. C., & Ho, L. C. (2009). Applying VaR to REITs: A comparison of alternative
methods. Review of Financial Economics, 18, 97–102.
Sadeghi, M., & Shavvalpour, S. (2006). Energy risk management and value at risk modeling.
Energy Policy, 34, 3367–3373.
So, M. K. P., & Yu, P. L. H. (2006). Empirical analysis of GARCH models in value at risk
estimation. International Financial Markets, Institutions & Money, 16, 180–197.
Stavroyiannis, S., Makris, I., Nikolaidis, V., & Zarangas, L. (2012). Econometric modeling and
value-at-risk using the Pearson type-IV distribution. International Review of Financial Anal-
ysis, 22, 10–17.
Su, J. B., & Hung, J. C. (2011). Empirical analysis of jump dynamics, heavy-tails and skewness on
value-at-risk estimation. Economic Modelling, 28, 1117–1130.
Theodossiou, P. (1998). Financial data and the skewed generalized t distribution. Management
Science, 44, 1650–1661.
Vlaar, P. J. G. (2000). Value at risk models for Dutch bond portfolios. Journal of Banking &
Finance, 24, 1131–1154.
Modeling Multiple Asset Returns by
a Time-Varying t Copula Model 52
Long Kang
Contents
52.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1432
52.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1434
52.3 The Model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1436
52.3.1 Copula . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1436
52.3.2 Modeling Marginal Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1437
52.3.3 Modeling Dependence Structure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1438
52.3.4 Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1440
52.4 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1441
52.5 Empirical Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1442
52.5.1 Marginal Distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1442
52.5.2 Copulas . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1444
52.5.3 Time-Varying Dependence . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1444
52.6 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1448
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1448
Abstract
We illustrate a framework to model joint distributions of multiple asset returns
using a time-varying Student’s t copula model. We model marginal distributions
of individual asset returns by a variant of GARCH models and then use
a Student’s t copula to connect all the margins. To build a time-varying structure
for the correlation matrix of t copula, we employ a dynamic conditional corre-
lation (DCC) specification. We illustrate the two-stage estimation procedures for
the model and apply the model to 45 major US stocks returns selected from nine
L. Kang
Department of Finance, Antai College of Economics and Management, Shanghai Jiao Tong
University, Shanghai, China
The Options Clearing Corporation and Center for Applied Economics and Policy Research,
Indiana University, Bloomington, IN, USA
e-mail: lkang@indiana.edu; kanglong@gmail.com
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1431
DOI 10.1007/978-1-4614-7750-1_52,
# Springer Science+Business Media New York 2015
1432 L. Kang
Keywords
Student’s t copula • GARCH models • Asset returns • US stocks • Maximum
likelihood • Two-stage estimation • Tail dependence • Exceedance correlation •
Dynamic conditional correlation • Asymmetric dependence
52.1 Introduction
each return. Then, we choose a proper copula function to link marginal distributions
of each return to form the joint distribution. As in marginal distributions of each
return, the copula parameters can also be specified as being dependent on previous
observations to make the copula structure time varying for a better fit of data. In this
chapter, we have an AR(1) process for the conditional mean and a GJR-GARCH
(1,1) specification for the conditional volatility for each return. We employ
a Student’s t copula with a time-varying correlation matrix (by a DCC specification)
to link marginal distributions. Usually the specified multivariate model contains
a huge number of parameters, and the estimation by maximum likelihood estimator
(MLE) can be quite challenging. Therefore, we pursue a two-stage procedure,
where all the GARCH models for each return are estimated individually first and
copula parameters are estimated in the second stage with estimated cumulative
distribution functions from the first stage.
We apply our model to modeling log returns of 45 major US stocks selected from
nine sectors with a time span ranging from January 3, 2000 to November 29, 2011.
Our estimation results show that AR(1) and GJR-GARCH(1,1) can reasonably well
capture empirical properties of individual returns. The stock returns possess fat tails
and leverage effects. We plot the estimated conditional volatility on selected stocks
and volatility spikes which happened during the “Internet Bubbles” in the early
2000s and the financial crisis in 2008. We estimate a DCC specification for the time-
varying t copula and also a normal copula for comparison purposes. The parameter
estimates for time-varying t copula are statistically significant, which indicates a
significant time-varying property of the dependence structure. The time-varying
t copula yields significantly higher log-likelihood than normal copula. This improve-
ment of data fitness results from flexibility of t copula (relative to normal copula) and
its time-varying correlation structure.
We plot the time-varying correlation parameter for selected pairs of stocks under
the time-varying t copula model. The correlation parameters fluctuate around
certain averages, and they spike during the 2008 crisis for some pairs. For
45 asset returns, the estimated degree-of-freedom (DoF) parameter of the t copula
is around 25. Together with the estimated correlation matrix of the t copula, this
DoF leads to quite low values of tail dependence coefficients (TDCs). This may
indicate the limitation of t copulas in capturing possibly large tail dependence
behavior for some asset pairs when being used to model a large number of asset
returns. Nevertheless, the time-varying Student’s t copula model has a relatively
flexible parameter structure to account for the dependence among multiple asset
returns and is a very effective tool to model the dynamics of a large number of asset
returns in practice.
This chapter is organized as follows. Section 52.2 gives a short literature review
on recent applications of copulas to modeling financial time series. Section 52.3
introduces our copula model where we introduce copula theory, Copula-GARCH
framework, and estimation procedures. In particular, we elaborate on how to
construct and estimate a time-varying t copula model. Section 52.4 documents
the data source and descriptive statistics for the data set we use. Section 52.5 reports
estimation results and Sect. 52.6 concludes.
1434 L. Kang
1
Alternative approaches are also developed, such as in Ang and Bekaert (2002), Goeij and
Marquering (2004), and Lee and Long (2009), to address non-normal joint distributions of asset
returns.
52 Modeling Multiple Asset Returns by a Time-Varying t Copula Model 1435
2
For a detailed survey on the estimation of Copula-GARCH model, see Chap. 5 of Cherubini
et al. (2004).
3
See Hamilton (1994) and Greene (2003) for more details on maximum likelihood estimation.
1436 L. Kang
52.3.1 Copula
Theorem 52.1 Let F1(x1), . . ., Fn(xn) be given marginal distribution functions and
continuous in x1,. . . , xn, respectively. Let H be the joint distribution of (x1,. . . , xn).
Then there exists a unique copula C such that
n
H ðx1 ; . . . ; xn Þ ¼ CðF1 ðx1 Þ, . . . , Fn ðxn ÞÞ, 8ð x 1 ; . . . ; x n Þ 2 ℝ : (52.1)
Conversely, if we let F1(x1), . . ., Fn(xn) be continuous marginal distribution
functions and C be a copula, then the function H defined by Eq. 52.1 is a joint
distribution function with marginal distributions F1(x1), . . ., Fn(xn).
The above theory allows us to decompose a multivariate distribution function
into marginal distributions of each random variable and the copula form linking the
margins. Conversely, it also implies that to construct a multivariate distribution, we
can first find a proper marginal distribution for each random variable and then
obtain a proper copula form to link the margins. Depending on which dependence
measure used, the copula function mainly, not exclusively, governs the dependence
structure between individual variables. Hence, after specifying marginal distribu-
tions of each variable, the task of building a multivariate distribution solely
becomes to choose a proper copula form which best describes the dependence
structure between variables.
Differentiating Eq. 52.1 with respect to (x1,. . . , xn) leads to the joint density
function of random variables in terms of copula density. It is given as
Y
n
n
hðx1 ; . . . ; xn Þ ¼ cððF1 ðx1 Þ, . . . , Fn ðxn ÞÞ f i ðxi Þ, 8ðx1 ; . . . ; xn Þ 2 ℝ , (52.2)
i¼1
where c(F1(x1), . . ., Fn(xn)) is the copula density and fi(xi) is the density function for
variable i. Equation 52.2 implies that the log-likelihood of the joint density can be
decomposed into components which only involve each marginal density and
a component which involves copula parameters. It provides a convenient structure
for a two-stage estimation, which will be illustrated in details in the following
sections.
4
See Nelsen (1998) and Joe (1997) for a formal treatment of copula theory, and Bouye
et al. (2000), Cherubini et al. (2004), and Embrechts et al. (2002) for applications of copula theory
in finance.
52 Modeling Multiple Asset Returns by a Time-Varying t Copula Model 1437
To better fit the data, we usually assume the moments of distributions of random
variables are time varying and depend on past variables. Therefore, the distribution
of random variables at time t becomes a conditional one, and then the above copula
theory needs to be extended to a conditional case. It is given as follows.5
Theorem 52.2 Let Ot1 be the information set up to time t, and let F1(x1,tjOt1),
. . ., Fn(xn,tjOt1) be continuous marginal distribution functions conditional on
Ot1. Let H be the joint distribution of (x1,. . . , xn) conditional on Ot1. Then
there exists a unique copula C such that
n
H x1 , . . . , xn jOt1 Þ ¼ C F1 x1 jOt1 Þ, . . . , Fn xn jOt1 ÞjOt1 Þ, 8ðx1 ; . . . ; xn Þ 2 ℝ :
(52.3)
Before building a copula model, we need to find a proper specification for marginal
distributions of individual asset returns, as mis-specified marginal distributions
automatically lead to a mis-specified joint distribution. Let xi,t be asset i return at
time t, and its conditional mean and variance are modeled as follows:
5
See Patton (2004).
1438 L. Kang
As shown in Eqs. 52.4, 52.5 and 52.6, we model the conditional mean as an
AR(1) process and the conditional variance as a GJR(1,1) specification.6
We have parameter restrictions as b0,i > 0, b1,i 0, b2,i 0, b2,i + b3,i 0, and
b1, i þ b2, i þ 12 b3, i < 1 . 1(ei,t1 < 0) is an indicator function, which equals one
when ei,t1 < 0 and zero otherwise. We believe that our model specifications can
capture the features of the individual stock returns reasonably well. It is worth
noting that Eqs. 52.4, 52.5 and 52.6 can include more exogenous variables to better
describe the data. Alternative GARCH specifications can be used to describe the
time-varying conditional volatility. We assume i,t is i.i.d. across time and follows
a Student’s t distribution with DoF vi.
Alternatively, to model the conditional higher moments of the series, we can
follow Hansen (1994) and Jondeau and Rockinger (2003) who assume a skewed
t distribution for the innovation terms of GARCH specifications and find that the
skewed t distribution fits financial time series better than normal distribution.
Accordingly, we can assume i,t Skewed T(i,tjvi,t,li,t) with zero mean and unitary
variance where vi,t is DoF parameter and li,t is skewness parameter. The two
parameters are time varying and depend on lagged values of explanatory variables
in a nonlinear form. For illustration purposes, however, we will only use Student’s
t distribution for i,t in this chapter.
Normal copula and Student’s t copula are two copula functions from elliptical
families, which are frequently used in modeling joint distributions of random
variables. In this chapter, we also estimate a normal copula model for comparison
purposes. Let F1 denote the inverse of the standard normal distribution F and F∑,n
be n-dimensional normal distribution with correlation matrix ∑. Hence, the
n-dimensional normal copula is
Cðu; SÞ ¼ FS, N F1 ðu1 Þ, . . . , F1 ðun Þ , (52.7)
where f and f∑,n are the probability density functions (pdfs) of F and F∑,n,
respectively. It can be shown via Sklar’s theorem that normal copula generates
standard joint normal distribution if and only if the margins are standard normal.
6
See Glosten et al. (1993).
52 Modeling Multiple Asset Returns by a Time-Varying t Copula Model 1439
On the other hand, let Tv1 be the inverse of standard Student’s t distribution Tv
with DoF parameter7 v > 2 and TR,v be n-dimensional Student’s t distribution with
correlation matrix R and DoF parameter v. Then n-dimensional Student’s t copula is
Cðu; R; nÞ ¼ T R, n T 1 1
n ðu1 Þ, . . . , T n ðun Þ , (52.9)
qi, j, t
ri, j, t ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi : (52.11)
qi, i, t qj, j, t
7
In contrast to the previous standardized Student’s t distribution, the standard Student’s
t distribution here has variance as v/(v2).
8
Please see Engle and Sheppard (2001) and Engle (2002) for details on the multivariate
DCC-GARCH models.
1440 L. Kang
0
Proof First, (a) and (b) guarantee the system for BtBt is stationary and S exists. With
Q0 ¼ S, (c) guarantees Q0 is positive definite. With (a) to (c), Qt is the sum of
a positive definite matrix, a positive semi-definite matrix, and a positive definite
matrix both with nonnegative coefficients and then is positive definite for all t.
Based on the proposition Eq. 52.1 in Engle and Sheppard (2001), we prove that Rt
is positive definite.
52.3.4 Estimation
X
T
LðYÞ ¼ log f x1, t , . . . , xn, t Xt1 , Y : (52.14)
t¼1
X
T XT X
n
Lðy; g1 ; . . . ; gn Þ ¼ log c u1, t , . . . , un, t Xt1 , y þ f i xi, t Xt1 , gi :
t¼1 t¼1 i¼1
(52.15)
From Eq. 52.15, we observe that the copula and marginal distributions are
additively separate. Therefore, we can estimate the model by a two-stage MLE
procedure. In the first stage, the marginal distribution parameters for each asset are
estimated by MLE, and then with estimated cdf of each asset, we estimate the
copula parameters by MLE. Based on Joe (1997) and Patton (2006b), this two-stage
estimator is consistent and asymptotically normal.
52 Modeling Multiple Asset Returns by a Time-Varying t Copula Model 1441
XT
^¼1
S xt x0t : (52.16)
T t¼1
52.4 Data
We apply our model to modeling log returns of 45 major US stocks from nine
sectors: Consumer Discretionary, Consumer Staples, Energy, Financials, Health
Care, Industrials, Technology, Materials, and Utilities. Table 52.1 shows stock
symbols and company names of the selected 45 companies. We select five major
companies from each sector to form the stock group. The time span ranges from
January 3, 2000 to November 29, 2011 with 2990 observations. We download data
from yahoo finance (http://finance.yahoo.com/). The log returns are calculated from
daily close stock prices adjusted for dividends and splits.
To save space, we only plot and calculate descriptive statistics of nine stocks
with each from one sector. Figure 52.1 plots the log returns of those nine selected
stocks, and there are two periods of volatility clusterings due to “Internet Bubbles”
in the early 2000s and the financial crisis in 2008, respectively. We observe that
during the financial crisis in 2008, major banks, such as Citigroup, incurred huge
negative and positive daily returns. Table 52.2 shows the calculated mean, standard
deviation, skewness, and kurtosis for the nine stocks. The average returns for
the nine stocks are close to zero. Major banks, represented by Citigroup, have
significantly higher volatility. Most of the stocks are slightly positively skewed, and
only two have slight negative skewness. All the stocks have kurtosis greater than
three indicating fat tails, and again major banks have significantly fatter tails.
All the descriptive statistics indicate that the data property of individual returns
needs to be captured by a variant of GARCH specification.
1442 L. Kang
Table 52.1 Symbols and names of 45 selected stocks from nine sectors
Sector Consumer discretionary Consumer Staples Energy
Stock symbol MCD: McDonald’s WMT: Wal-Mart Stores XOM: Exxon Mobil Corp.
Inc.
HD: Home Depot PG: Procter & Gamble CVX: Chevron Corp.
Co.
DIS: Walt Disney Co. KO: Coca-Cola Co. COP:
CONOCOPHILLIPS.
TGT: Target WAG: Walgreen Co. DVN: Devon
Energy Corp.
LOW: Lowe’s MO: Altria Group Inc. SLB: Schlumberger
Limited
Sector Financials Health Care Industrials
Stock symbol C: Citigroup Inc. JNJ: Johnson & GE: General Electric Co.
Johnson
BAC: Bank of PFE: Pfizer Inc. UNP: Union Pacific Corp.
America Corp.
JPM: JPMorgan Chase & ABT: Abbott UTX: United
Co. Laboratories Technologies Corp.
USB: U.S. Bancorp MRK: Merck & MMM: 3 M Co.
Co. Inc.
WFC: Wells Fargo & Co. AMGN: Amgen Inc. BA: Boeing Co.
Sector Technology Materials Utilities
Stock symbol T: AT&T Inc. NEM: Newmont EXC: Exelon Corp.
Mining Corp.
MSFT: Microsoft Corp. DD: E.I. DuPont de FE:
Nemours & Co. FirstEnergy Corp.
IBM: International Business DOW: Dow Chemical Co. PPL: PPL
Machines Corp. Corporation
CSCO: Cisco Systems Inc. FCX: Freeport-McMoRan D: Dominion
Copper & Gold Inc. Resources, Inc.
HPQ: Hewlett-Packard Co. PX: Praxair Inc. DUK: Duke
Energy Corp.
MCD WMT
0.1 0.15
0.05 0.1
Log returns
Log returns
0 0.05
−0.05 0
−0.1 −0.05
−0.15 −0.1
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date (in year) Date (in year)
XOM C
0.2 0.5
0.1
Log returns
Log returns
0 0
−0.1
−0.2 −0.5
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date (in year) Date (in year)
JNJ GE
0.2 0.2
0.1 0.1
Log returns
Log returns
0 0
−0.1 −0.1
−0.2 −0.2
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date (in year) Date (in year)
T NEM
0.2 0.6
0.1 0.4
Log returns
Log returns
0 0.2
−0.1 0
−0.2 −0.2
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date (in year) Date (in year)
EXC
0.2
0.1
Log returns
−0.1
−0.2
00 02 04 06 08 10 12
Date (in year)
Fig. 52.1 The log returns of the nine of our 45 selected stocks with each from one sector have
been plotted
1444 L. Kang
Table 52.2 Descriptive statistics (mean, standard deviation, skewness, and kurtosis) for the nine
of our 45 selected stocks with each from each sector
Stock
symbol MCD WMT XOM C JNJ GE T NEM EXC
Mean 3.73E- 7.21E- 3.15E- 8.05E- 1.96E- 2.89E- 1.23E- 3.65E- 4.48E-
04 06 04 04 04 04 05 04 04
Std. dev. 0.017 0.017 0.017 0.037 0.013 0.022 0.019 0.027 0.018
Skewness 0.21 0.13 0.02 0.48 0.53 0.04 0.12 0.34 0.05
Kurtosis 8.25 7.72 12.52 35.55 17.83 9.99 8.68 8.22 10.58
# obs. 2,990 2,990 2,990 2,990 2,990 2,990 2,990 2,990 2,990
52.5.2 Copulas
We report estimation results for the time-varying t copula parameters in Table 52.4.
All the three parameters a, b, and v are statistically significant. The estimate a is
close to zero and the estimate for b is close to one. The estimate for v is about 25. As
our estimation is carried out on the joint distribution of 45 stock returns, the
estimate for v shed some light on how much Student’s t copula can capture tail
dependence when used to fit a relatively large number of variables. We also report
the log-likelihood for time-varying Student’s t copula and normal copula in
Table 52.4. As the correlation matrix in normal copula is estimated by its sample
correlation, we did not report it here. We find that time-varying t copula has
significantly higher log-likelihood than normal copula, which results from the
more flexible parameter structure of t copula and the time-varying parameter
structure.
Table 52.3 The GARCH estimation results of individual stock returns for the nine of our selected 45 stocks with each from one sector. Values in brackets are
standard errors. The star indicates the statistical significance at a 5 % level
Stock
symbol MCD WMT XOM C JNJ GE T NEM EXC
Conditional mean
a0,i 6.69E-04* 1.77E-05 6.50E-04* 1.03E-06 1.17E-04 1.93E-06 2.73E-04 3.31E-04 7.10E-04*
(2.25E-04) (2.11E-04) (2.41E-04) (2.42E-04) (1.59E-04) (2.32E-04) (2.29E-04) (3.87E-04) (2.36E-04)
a1,i 0.030 0.037* 0.076* 0.003 0.004 0.015 0.009 0.056* 0.019
(0.018) (0.018) (0.019) (0.018) (0.018) (0.018) (0.018) (0.018) (0.019)
Conditional variance
b0,i 1.69E-06* 8.22E-07* 5.97E-06* 1.96E-06* 1.87E-06* 1.26E-06* 1.37E-06* 3.11E-06* 3.91E-06*
(5.21E-07) (3.50E-07) (1.28E-06) (5.53E-07) (4.40E-07) (3.93E-07) (4.47E-07) (1.31E-06) (9.82E-07)
b1,i 0.947* 0.952* 0.901* 0.908* 0.905* 0.948* 0.943* 0.959* 0.897*
(0.007) (0.007) (0.013) (0.009) (0.011) (0.007) (0.007) (0.007) (0.012)
b2,i 0.024* 0.026* 0.025* 0.045* 0.025* 0.017* 0.027* 0.037* 0.066*
(0.008) (0.008) (0.012) (0.011) (0.011) (0.007) (0.008) (0.009) (0.014)
b3,i 0.046* 0.038* 0.092* 0.094* 0.127* 0.067* 0.051* 0.001 0.046*
(0.012) (0.013) (0.018) (0.017) (0.021) (0.012) (0.012) (0.011) (0.018)
Modeling Multiple Asset Returns by a Time-Varying t Copula Model
Degree of freedom
vi 6.187* (0.67) 7.002* (0.92) 9.496* (1.59) 6.505* (0.75) 5.796* (0.62) 6.622* (0.66) 8.403* (1.20) 7.831* (1.09) 9.247* (1.28)
1445
1446 L. Kang
Conditional Volatility
Conditional Volatility
1.5 1.5
1 1
0.5 0.5
0 0
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date (in year) Date (in year)
x 10−3 XOM C
6 0.06
Conditional Volatility
Conditional Volatility
5 0.05
4 0.04
3 0.03
2 0.02
1 0.01
0
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date (in year) Date (in year)
Fig. 52.2 The estimated time-varying conditional volatility for four selected stocks has been
plotted
Table 52.4 The estimates and standard errors for time-varying Student’s t copula. Values in
brackets are standard errors. The star indicates the statistical significance at a 5 % level. We also
report the log-likelihood for time-varying t copula and normal copula
Time-varying t copula Normal copula
Parameter estimates
a 0.0031* (0.0002)
b 0.984* (0.0016)
v 25.81* (1.03)
Log-likelihood of copula component
Log-likelihood 40,445.44 38,096.36
conditional correlation parameters of t copula for four selected pairs of stock returns
in Fig. 52.3. For those four pairs, the conditional correlation parameter fluctuates
around certain positive averages. The two pairs, MCD-WMT and NEM-EXC,
experienced apparent correlation spikes during the 2008 financial crisis. Moreover,
Fig. 52.4 shows the estimated TDCs for the four pairs. We find that with the DoF
around 25, the TDCs for those pairs of stock returns are very low, though some pairs
do exhibit TDC spikes during the 2008 crisis. The low values of TDCs indicate
possible limitations of t copula to account for tail dependence when being used to
model a large number of variables.
52 Modeling Multiple Asset Returns by a Time-Varying t Copula Model 1447
XOM-C MCD-WMT
0.5 0.45
Conditional correlation
Conditional correlation
0.45
0.4
0.4
0.35 0.35
0.3
0.3
0.25
0.2 0.25
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date in year Date in year
C-JNJ NEM-EXC
0.45 0.25
Conditional correlation
Conditional correlation
0.4 0.2
0.35
0.15
0.3
0.25 0.1
0.2 0.05
00 02 04 06 08 10 12 00 02 04 06 08 10 12
Date in year Date in year
Fig. 52.3 The estimated time-varying correlation parameters in t copula for four selected pairs of
stock returns have been plotted
x 10−3 x 10−3
Tail Dependence Coefficient
XOM-C MCD-WMT
4 3.5
3
3
2.5
2 2
1.5
1
1
0 0.5
00 01 02 03 04 05 06 07 08 09 10 11 12 00 01 02 03 04 05 06 07 08 09 10 11 12
Date in year Date in year
x 10−3 x 10−4
Tail Dependence Coefficient
Tail Dependence Coefficient
C-JNJ NEM-EXC
2.5 6
2 5
4
1.5
3
1
2
0.5 1
0 0
00 01 02 03 04 05 06 07 08 09 10 11 12 00 01 02 03 04 05 06 07 08 09 10 11 12
Date in year Date in year
Fig. 52.4 The time-varying tail dependence coefficient (TDC) for the four selected pairs of stock
returns has been plotted
1448 L. Kang
52.6 Conclusion
References
Ang, A., & Bekaert, G. (2002). International asset allocation with regime shifts. Review of
Financial Studies, 15(4), 1137–1187.
Berg, D., & Bakken, H. (2006). A copula goodness-of-fit test based on the probability integral
transform. Working Paper at http://www.danielberg.no/publications/Btest.pdf
Bouyé, E., Valdo D., Ashkan N., Gaël R., & Thierry, R. (2000). Copulas for finance – A reading
guide and some applications. Available at SSRN http://ssrn.com/abstract¼1032533. March.
Chen, X., Yanqin F., & Andrew, J. P. (2004). Simple tests for models of dependence between
multiple financial time series, with applications to U.S. Equity returns and exchange rates
(Discussion Paper 483). London School of Economics, Revised July 2004.
Cherubini, U., Luciano, E., & Vecchiato, W. (2004). Copula methods in finance (Wiley series in
financial engineering). Chichester: Wiley.
De Goeij, P., & Marquering, W. (2004). Modeling the conditional covariance between stock
and bond returns: A multivariate GARCH approach. Journal of Financial Econometrics, 2(4),
531–564.
Deheuvels, P. (1981). A Nonparametric test for independence. Paris: Universite de Paris,
Institut de Statistique.
Engle, R. (2002). Dynamic conditional correlation: A simple class of multivariate generalized
autoregressive conditional heteroskedasticity models. Journal of Business & Economic Statis-
tics, 20(3), 339–350.
52 Modeling Multiple Asset Returns by a Time-Varying t Copula Model 1449
Engle, R. F., & Sheppard, K. (2001). Theoretical and empirical properties of dynamic conditional
correlation multivariate GARCH (NBER Working Paper No. 8554). Cambridge, MA: National
Bureau of Economic Research.
Genest, C., Ghoudi, K., & Rivest, L. (1995). A semiparametric estimation procedure of depen-
dence parameters in multivariate families of distributions. Biometrika, 82, 543–552.
Glosten, L. R., Ravi, J., & Runkle, D. E. (1993). On the relation between the expected value and
the volatility of the nominal excess return on stocks. Journal of Finance, 48(5), 1779–1801.
Greene, W. H. (2003). Econometric analysis. Upper Saddle River: Prentice Hall.
Hamilton, J. D. (1994). Time series analysis. Princeton: Princeton University Press.
Hansen, B. E. (1994). Autoregressive conditional density estimation. International Economic
Review, 35(3), 705–730.
Hu, L. (2006). Dependence patterns across financial markets: A mixed copula approach. Applied
Financial Economics, 16, 717–729.
Joe, H. (1997). Multivariate models and dependence concepts. London: Chapman and Hall.
Joe, H., & Xu, J. J. (1996). The estimation method of inference functions for margins for
multivariate models (Technical Report 166). Vancouver: Department of Statistics, University
of British Columbia.
Jondeau, E., & Rockinger, M. (2002). Conditional dependency of financial series: the copula-
GARCH model (FAME Research Paper 69). Geneva: International Center for Financial Asset
Management and Engineering.
Jondeau, E., & Rockinger, M. (2003). Conditional volatility, skewness, and kurtosis: Existence,
persistence, and comovements. Journal of Economic Dynamics and Control, 27(10),
1699–1737.
Jondeau, E., & Rockinger, M. (2006). The copula-GARCH model of conditional dependencies:
An international stock market application. Journal of International Money and Finance,
25, 827–853.
Lee, T.-H., & Long, X. (2009). Copula-based multivariate GARCH model with uncorrelated
dependent errors. Journal of Econometrics, 150, 207–218.
Nelson, R. B. (1998). An introduction to copula. New York: Springer.
Patton, A. J. (2004). On the out-of-sample importance of skewness and asymmetric dependence
for asset allocation. Journal of Financial Econometrics, 2(1), 130–168.
Patton, A. J. (2006a). Estimation of multivariate models for time series of possibly different
lengths. Journal of Applied Econometrics, 21(2), 147–173.
Patton, A. J. (2006b). Modelling asymmetric exchange rate dependence. International Economic
Review, 47(2), 527–556.
Savu, C., & Trede, M. (2010). Hierarchies of Archimedean copulas. Quantitative Finance,
10, 295–304.
Sklar, A. (1959). Fonctions de repartition an dimensions et leurs marges. Paris: Publications de
l’lnstitut de statistique de l’Universite de Paris.
Trivedi, P. K., & Zimmer, D. M. (2006). Using trivariate copulas to model sample selection
and treatment effects: Application to family health care demand. Journal of Business and
Economic Statistics, 24(1), 63.
Tsafack, G., & Garcia, R. (2011). Dependence structure and extreme comovements in interna-
tional equity and bond markets. Journal of Banking and Finance, 35, 1954–1970.
Xu, Y. (2004). Incorporating estimation risk in copula-based portfolio optimization. Miami:
Department of Economics, University of Miami.
Internet Bubble Examination with
Mean-Variance Ratio 53
Zhidong D. Bai, Yongchang C. Hui, and Wing-Keung Wong
Contents
53.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1452
53.2 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1454
53.3 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1455
53.4 Illustration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1458
53.5 Concluding Remarks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1461
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1463
Abstract
To evaluate the performance of the prospects X and Y, financial professionals are
interested in testing the equality of their Sharpe ratios (SRs), the ratios of the
excess expected returns to their standard deviations. Bai et al. (Statistics and
Probability Letters 81, 1078–1085, 2011d) have developed the mean-variance-
ratio (MVR) statistic to test the equality of their MVRs, the ratios of the excess
expected returns to its variances. They have also provided theoretical reasoning
to use MVR and proved that their proposed statistic is uniformly most powerful
unbiased. Rejecting the null hypothesis infers that X will have either smaller
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1451
DOI 10.1007/978-1-4614-7750-1_53,
# Springer Science+Business Media New York 2015
1452 Z.D. Bai et al.
variance or larger excess mean return or both leading to the conclusion that X is
the better investment. In this paper, we illustrate the superiority of the MVR test
over the traditional SR test by applying both tests to analyze the performance of
the S&P 500 index and the NASDAQ 100 index after the bursting of the Internet
bubble in the 2000s. Our findings show that while the traditional SR test
concludes the two indices being analyzed to be indistinguishable in their per-
formance, the MVR test statistic shows that the NASDAQ 100 index
underperformed the S&P 500 index, which is the real situation after the bursting
of the Internet bubble in the 2000s. This shows the superiority of the MVR test
statistic in revealing short-term performance and, in turn, enables investors to
make better decisions in their investments.
Keywords
Mean-variance ratio • Sharpe ratio • Hypothesis testing • Uniformly most
powerful unbiased test • Internet bubble • Fund management
53.1 Introduction
Internet stocks obtained huge gains in the late 1990s, followed by huge losses from
early 2000. In just 2 years from 1998 to early March 2000, prices of Internet stocks
rose by sixfold and outperformed the S&P 500 by 482 %. Technology stocks
generally showed a similar trend based on the fact that NASDAQ 100 index
quadrupled in value over the same period and outperformed the S&P 500 index
by 268 %. On the other hand, NASDAQ 100 index dropped by 64.28 % in value
during the Internet bubble crash and underperformed the S&P 500 index by
173.87 %.
The spectacular rise and fall of Internet stocks in the late 1990s has stimulated
research into the causes of the Internet stock bubble. Theories had been developed
to explain the Internet bubble. For example, Baker and Stein (2004) develop
a model of market sentiment with irrationally overconfident investors and short-
sale constraints. Ofek and Richardson (2003) provide circumstantial evidence that
Internet stocks attract mostly retail investors who are more prone to be
overconfident about their ability to predict future stock prices than institutional
investors. Perkins and Perkins (1999) suggest that during the Internet boom,
investors were confidently betting on the continued rise of Internet stocks because
they knew that high demand and limited equity float implies substantial upside
returns. Moreover, Ofek and Richardson (2003) provide indirect evidence that
Internet stock prices were supported by a combination of factors such as limited
float, short-sale constraints, and aggressive trend chased by retail investors, whereas
Statman (2002) shows that this asymmetric payoff must have made Internet stocks
appear to be an extremely attractive gamble for risk seekers. On the other hand,
Fong et al. (2008) use stochastic dominance methodology (Fong et al. 2005; Broll
et al. 2006; Chan et al. 2012; Lean et al. 2012) to identify dominant types of risk
preferences in the Internet bull and bear markets. They conclude that investor risk
53 Internet Bubble Examination with Mean-Variance Ratio 1453
preferences (Wong and Li 1999; Wong and Chan 2008) have changed over this
cycle, and the change is related to utility theory (Wong 2007; Sriboonchitta
et al. 2009) and behavioral finance (Lam et al. 2010, 2012).
In this paper, we apply both the mean-variance ratio (MVR) test and the Sharpe
ratio (SR) test to examine the performance of the NASDAQ 100 index and the S&P
500 index during the bursting of the Internet bubble in the 2000s. The tests are
relied on the theory of the mean-variance (MV) portfolio optimization (Markowitz
1952; Bai et al. 2009a, b). The Markowitz efficient frontier also provides the basis
for many important financial economics advances, including the Sharpe-Lintner
capital asset pricing model (CAPM, Sharpe 1964; Lintner 1965) and the well-
known optimal one-fund theorem (Tobin 1958). Originally motivated by the MV
analysis, the optimal one-fund theorem, and the CAPM model, the Sharpe ratio,
the ratio of the excess expected return to its volatility or standard deviation, is
one of the most commonly used statistics in the MV framework. The SR is now
widely used in many different areas in Finance and Economics, from the evalua-
tion of portfolio performance to market efficiency tests (see, e.g., Ofek and
Richardson 2003).
Jobson and Korkie (1981) develop a SR statistic to test for the equality of two
SRs. The test statistic has been modified and improved by Cadsby (1986) and
Memmel (2003). Lo (2002) carries out a more thorough study of the statistical
property of the SR estimator. Using standard econometric methods with several
different sets of assumptions imposed on the statistical behavior of the returns
series, Lo derives the asymptotic statistical distribution for the SR estimator and
shows that confidence intervals, standard errors, and hypothesis tests can be com-
puted for the estimated SRs in much the same way as regression coefficients such as
portfolio alphas and betas are computed.
The SR test statistic developed by Jobson and Korkie (1981) and others provides
a formal statistical comparison of performance among portfolios. One deficiency of
the SR statistic is that it has only an asymptotic distribution. Hence, the SR test has
its statistical properties only for large samples, but not for small samples. Never-
theless, the performance of assets is often compared by using small samples,
especially when markets undergo substantial changes resulting from changes in
short-term factors and momentum. Under these circumstances, it is more meaning-
ful to use limited data to predict the assets’ future performance. In addition, it is not
meaningful to measure SRs for extended periods when the means and standard
deviations of the underlying assets are found empirically to be nonstationary and/or
to possess structural breaks. For small samples, the main difficulty in developing
the SR test is that it is impossible to obtain a uniformly most powerful unbiased
(UMPU) test to check for the equality of SRs. To circumvent this problem, Bai
et al. (2011d) propose to use an alternative statistic, the MVR tests to compare
performance of assets. They also discuss the evaluation of the performance of assets
for small samples by providing a theoretical framework and then invoking both
one-sided and two-sided UMPU MVR tests. Moreover, Bai et al. (2012) further
extend the MVR statistics to compare the performance of prospects after the effect
of the background risk has been mitigated.
1454 Z.D. Bai et al.
Applying the traditional SR test, we fail to reject the possibility of having any
significant difference between the performance of the S&P 500 index and the
NASDAQ 100 index during the bursting of the Internet bubble in the 2000s. This
finding implies that the two indices being analyzed could be indistinguishable in
their performance during the period under the study. However, we conjecture that
this conclusion is most likely to be inaccurate as the lack of sensitivity of the SR test
in analyzing small samples. Thus, we propose to use the MVR test in the analysis.
As expected, the MVR test shows that the MVR of the weekly return on S&P
500 index is different from that on the NASDAQ 100 index. We conclude that the
NASDAQ 100 index underperformed the S&P 500 index during the period under
the study. The proposed MVR test can discern the performance of the two indices
and hence is more informative than tests using the SR statistics for investors to
decide on their investments.
The rest of the paper is organized as follows: Section 53.2 discusses the
data while Sect. 53.3 provides the theoretical framework and discusses the
theory for both one-sided and two-sided MVR tests. In Sect. 53.4, we demonstrate
the superiority of the MVR tests over the traditional SR tests by applying both tests
to analyze the performance of the S&P 500 index and the NASDAQ 100 index
during the bursting of the Internet bubble in the 2000s. This is followed by
Sect. 53.4 which summarizes our conclusions and shares our insights.
53.2 Data
The data used in this study consists of weekly returns on two stock indices: the S&P
500 and the NASDAQ 100 index. We use the S&P 500 index to represent
non-technology or “old economy” firms. Our proxy for the Internet and technology
sectors is the NASDAQ 100 index. Firms represented in the NASDAQ 100 include
those in the computer hardware and software, telecommunications, and biotech-
nology sectors. The NASDAQ 100 index is value weighted.
Our sample period is from January 1, 2000 to December 31, 2002, to study
the effect of the crash in the Internet bubble. Before 2000, there is a clear upward
trend in technology stock prices emerging from around that period and this
period spans a period of intense IPO and secondary market activities for Internet
stocks. Schultz and Zaman (2001) report that 321 Internet firms went public
between January 1999 and March 2000, accounting for 76 % of all new Internet
issues since the first wave of Internet IPOs began in 1996. Ofek and Richardson
(2003) find that the extraordinary high valuations of Internet stocks between the
early 1998 and February 2000 were accompanied by very high trading volume and
liquidity. The unusually high volatility of technology stocks is only partially
explained by the rise in the overall market volatility. Our interest centers on the
bear market from January 1, 2000 to December 31, 2002. All data for this study
are from datastream.
53 Internet Bubble Examination with Mean-Variance Ratio 1455
53.3 Methodology
mX 1 X 2 X 1 X 2
pðx; yÞ ¼ k exp xi 2 xi þ 2 yi 2 yi (53.1)
s 2 2s t 2t
n=2 n=2 2
n2
Where k ¼ ð2ps2 Þ ð2pt2 Þ exp nm
2s 2 exp 2t 2
m m
H 0 : versus H 1 : > (53.2)
s t s t
to compare the performance of their corresponding SRs, ms and t , the ratios of the
excess expected returns to their standard deviations.
If the hypothesis H0 is rejected, it infers that X is the better investment
prospect with larger SR because X has either larger excess mean return or
smaller standard deviation or both. Jobson and Korkie (1981) and Memmel
(2003) develop test statistics to test the hypotheses in Eq. 53.2 for large
samples but their tests would not be appropriate for testing small samples as the
distribution of their test statistics is only valid asymptotically but not valid for
small samples. However, it is especially relevant in investment decisions to test
the hypotheses in Eq. 53.2 for small samples to provide useful investment infor-
mation to investors. Furthermore, as it is impossible to obtain any UMPU test
statistic to test the inequality of the SRs in Eq. 53.2 for small samples, Bai
et al. (2011d) propose to use the following hypothesis to test for the inequality of
the MVRs:
m m
H 01 : versus H11 : > : (53.3)
s2 t 2 s2 t 2
In addition, they develop the UMPU test statistic to test the above hypotheses.
Rejecting the hypothesis H0 infers that X will have either smaller variance or larger
excess mean return or both leading to the conclusion that X is the better investment.
As sometimes investors conduct the two-sided test to compare the MVRs, the
following hypotheses are included in our study:
1456 Z.D. Bai et al.
m m
H 02 : ¼ 2 versus H12 : 6¼ 2 : (53.4)
s 2 t s 2 t
One may argue that the MVR test is that SR test is scale invariant, whereas the
MV ratio test is not. To support the MVR test to be an acceptable alternative test
statistic, Bai et al. (2011d) show the theoretical justification for the use of the MVR
test statistic in the following remark:
Remark 53.1 One may think that the MVR can be less favorable than the SR as the
former is not scale invariant while the latter is. However, in some financial
processes, the mean change in a short period of time is proportional to its variance
change. For example, many financial processes can be characterized by the fol-
lowing diffusion process for stock prices formulated as
where s20 is a given risk level. We call R satisfying Eq. 53.5 the optimal return and
c be its corresponding allocation plan. One could easily extend the separation
theorem and the mutual fund theorem to obtain the solution of Eq. 53.51 from the
following lemma:
1
We note that Bai et al. (2009a, b, 2011c) have also used the same framework as in 53.5.
53 Internet Bubble Examination with Mean-Variance Ratio 1457
Lemma 53.1 For the optimization setting displayed in Eq. 53.5, the optimal return,
R, and its corresponding investment plan, c, are obtained as follows:
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
R ¼ s0 mT S1 m
and
s0
c ¼ pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi S1 m: (53.6)
m S1 m
T
From Lemma 53.1, the investment plan, c, is proportional to the MVR when S
is a diagonal matrix. Hence, when the asset is concluded as superior in performance
utilizing the MVR test, its corresponding weight could then be computed based on
the corresponding MVR test value. Thus, another advantage of using the MVR test
over the SR test is that it not only allows investors to compare the performance of
different assets, but it also provides investors with information of the assets weight.
The MVR test enables investors to compute the corresponding allocation for the
assets. On the other hand, as the SR is not proportional to the weight of the
corresponding asset, an asset with the highest SR would not infer that one should
put highest weight on this asset as compared with our MVR. In this sense, the test
proposed by Bai et al. (2011d) is superior to the SR test.
Bai et al. (2011d) have also developed both one-sided UMPU test and two-sided
UMPU test of equality of the MVRs in comparing the performances of different
prospects with hypotheses stated in Eqs. 53.3 and 53.4, respectively. We first state
the one-sided UMPU test for the MVRs as follows:
Theorem 53.1 Let Xi and Yi (i ¼ 1, 2, , n) be independent random variables with
joint distribution function defined in Eq. 53.1. For the hypotheses setup in Eq. 53.3,
there exists a UMPU level-a test with the critical function f(u, t) such that
1, when u C0 ðtÞ
fðu; tÞ (53.7)
0, when u < C0 ðtÞ
where C0 is determined by
ð1
f n, t ðuÞ du ¼ K 1 ; (53.8)
C0
with
n1
1 !n 1 1
2
u 2
2 ðt1 uÞ 2
f n, t ðuÞ ¼ t2 t3 ,
n n
ð
K 1 ¼ a f n, t ðuÞ du;
O
in which
1458 Z.D. Bai et al.
X
n X
n X
n
U¼ Xi , T1 ¼ Xi þ Yi,
i¼1 i¼1 i¼1
Xn Xn
T2 ¼ X2i , T3 ¼ Y 2i , T ¼ ðT 1 ; T 2 ; T 3 Þ;
i¼1 i¼1
where
ð
K 2 ¼ ð 1 aÞ f n, t ðuÞ du,
ðO
K 3 ¼ ð 1 aÞ u f n, t ðuÞ du:
O
53.4 Illustration
In this section, we demonstrate the superiority of the MVR tests over the traditional
SR tests by illustrating the applicability of the MVR tests to examine the Internet
bubble during January 2000 and December 2002. For simplicity, we only
53 Internet Bubble Examination with Mean-Variance Ratio 1459
demonstrate the two-sided UMPU test.2 The data for this study consists of weekly
returns on two stock indices: the S&P 500 and the NASDAQ 100 index. The sample
period covers from January 2000 to December 2002 in which the data from the first
week of November 2000 to the last week of January 2001 (3 months) are used to
compute the MVR in January 2001, while the data from the first week of December
2000 to the last week of February 2001 are used to compute the MVR in February
2001, and so on. However, if the period used to compute the SRs is too short, the
result would not be meaningful as discussed in our previous sections. Thus, we
utilize a longer period from the first week of February 2000 to the last week of
January 2001 (12 months) to compute the SR ratio in January 2001, from the first
week of March 2000 to the last week of February 2001 to compute the SR ratio in
February 2001, and so on.
Let X with mean mX and variance s2X be the weekly return on S&P 500 while
Y with mean mY and variance s2Y be the weekly return on the NASDAQ 100 index.
We test the following hypotheses:
mX mY mX mY
H0 : ¼ versus H1 : 6¼ : (53.11)
s2X s2Y s2X s2Y
To test the hypotheses in Eq. 53.11, we first compute the values of the test
function U for the MVR statistic shown in Eq. 53.9, then compute the critical values
C1 and C2 under the test level of 5 % for the pair of indices and display the values in
Table 53.1.
For comparison, we also compute the corresponding SR statistic developed by
Jobson and Korkie (1981) and Memmel (2003) such that
^Y m
s ^X s^X m
^Y
z¼ pffiffiffi , (53.12)
^y
which follows standard normal distribution asymptotically with
1 1 2 2 1 2 2 mX mY 2
y¼ 2sX sY 2sX sY sX, Y þ mX sY þ mY sX
2 2
s
T 2 2 sX sY X, Y
to test for the equality of the SRs for the funds by setting the following hypotheses
such that
mX mY mX mY
H 0 : ¼ versus H1 : 6¼ : (53.13)
sX sY sX sY
2
The results of the one-sided test which draw a similar conclusion are available on request.
1460 Z.D. Bai et al.
Table 53.1 The results of the mean-variance ratio test and Sharpe ratio test for NASDAQ and
S&P 500, from January 2001 to December 2002
MVR test SR test
Date month/year U C1 C2 Z
01/2001 0.0556 0.1812 0.1267 1.0906
02/2001 0.0636 0.1843 0.1216 1.8765
03/2001 0.1291 0.2291 0.0643 1.1787
04/2001 0.0633 0.2465 0.1633 0.9590
05/2001 0.0212 0.1937 0.2049 0.8313
06/2001 0.0537 0.1478 0.1983 0.8075
07/2001 0.0421 0.1399 0.1132 0.6422
08/2001 0.1062 0.1815 0.0886 0.6816
09/2001 0.1623* 0.1665 0.2728 1.0125
10/2001 0.1106 0.3507 0.1742 0.5931
11/2001 0.0051 0.2386 0.2825 0.1898
12/2001 0.1190 0.0165 0.2041 0.1573
01/2002 0.0316 0.0744 0.1389 0.0157
02/2002 0.0067 0.1389 0.1013 0.0512
03/2002 0.0216 0.1349 0.0853 0.1219
04/2002 0.0444 0.1739 0.0848 0.1885
05/2002 0.0588 0.1766 0.1094 0.0446
06/2002 0.1477 0.2246 0.0267 0.3408
07/2002 0.2167* 0.0101 0.0578 0.0984
08/2002 0.1526* 0.0452 0.1242 0.1024
09/2002 0.2121* 0.0218 0.0551 0.6304
10/2002 0.0416 0.1249 0.2344 0.0361
11/2002 0.0218 0.1056 0.2150 0.0008
12/2002 0.1265 0.0015 0.2417 0.3908
Note: The MVR test statistic U is defined in Eq. 53.9 and its critical values C1 and C2 are defined in
Eqs. 53.10, respectively. The SR test statistic Z is defined in Eq. 53.12. The level is a ¼ 0.05, and
“*” means significant at levels 5 %. Here, the sample size of the MVR test is 3 months, while the
sample size of the SR test 12 months. Recall that z0.025 1.96
Instead of using a 2-month data to compute the values of our proposed statistic,
we use the overlapping 12-month data to compute the SR statistic. The results are
also reported in Table 53.1.
The limitation of applying the SR test is that it would usually conclude indis-
tinguishable performances between the indices, which may not be the situation in
reality. In this aspect, looking for a statistic to evaluate the difference between
indices for short periods is essential. The situation in reality is that the Internet
stocks registered large gains in the late 1990s, followed by large losses from 2000.
As we mentioned before, the NASDAQ 100 index comprises 100 of the largest
domestic and international technology firms including those in the computer hard-
ware and software, telecommunications, and biotechnology sectors, while the S&P
53 Internet Bubble Examination with Mean-Variance Ratio 1461
5000 NASDAQ
S&P 500
4000
Index
3000
2000
1000
01/03/2000
03/13/2000
05/22/2000
07/31/2000
10/09/2000
12/18/2000
02/26/2001
05/07/2001
07/16/2001
10/01/2001
12/10/2001
02/19/2002
04/29/2002
07/08/2002
09/16/2002
11/25/2002
02/03/2003
04/14/2003
06/23/2003
09/02/2003
11/10/2003
Fig. 53.1 Weekly indices of NASDAQ and S&P 500 from January 3, 2000 to December 31, 2003
500 index represents non-technology or “old economy” firms. After the bursting of
the Internet bubble in the 2000s, as shown in Fig. 53.1, the NASDAQ 100 declined
much more and underperformed the S&P 500. From Table 53.1, we find that the
MVR test statistic does not disappoint us in that it does pick up significant
differences in performances between the S&P 500 and the NASDAQ 100 index
in September 2001, July 2002, August 2002, and September 2002, but SR test does
not conclude any distinguishable performances between the indices. Further to
say, from Table 53.1, we observe that m ^X > m ^ Y in September 2001, July 2002,
August 2002, and September 2002. This infers that the MVR test statistics can
detect the real situation that the NASDAQ 100 index underperformed the S&P
500 index, but the traditional SR test cannot detect any difference. Thus, we
conclude that investors could be able to profiteer from the Internet bubble if they
apply the MVR test.
In this paper, we employ the MVR test statistics developed by Bai et al. (2011d) to
examine the performances between the S&P 500 index and the NASDAQ
100 index during Internet bubble from January 2000 to December 2002. We
illustrate the superiority of the MVR test over the traditional SR test by applying
both tests to analyze the performance of the S&P 500 index and the NASDAQ
100 index after the bursting of the Internet bubble in the 2000s. Our findings show
that while the traditional SR test concludes the two indices being analyzed to be
indistinguishable in their performance, the MVR test statistic shows that the
NASDAQ 100 index underperformed the S&P 500 index, which is the real situation
1462 Z.D. Bai et al.
after the bursting of the Internet bubble in the 2000s. This shows the superiority of
the MVR test statistic in revealing short-term performance and, in turn, enables the
investors to make better decisions about their investments.
There are two basic approaches to the problem of portfolio selection under
uncertainty. One approach is based on the concept of utility theory (Gasbarro
et al. 2007; Wong et al. 2006, 2008). Several stochastic dominance (SD) test statistics
have been developed; see, for example, Bai et al. (2011a) and the references therein
for more information. This approach offers a mathematically rigorous treatment for
portfolio selection, but it is not popular among investors since investors would have
to specify their utility functions and choose a distributional assumption for the returns
before making their investment decisions.
The other approach is the mean-risk (MR) analysis that has been discussed in
this paper. In this approach, the portfolio choice is made with respect to two
measures – the expected portfolio mean return and portfolio risk. A portfolio is
preferred if it has higher expected return and smaller risk. These are convenient
computational recipes and they provide geometric interpretations for the trade-off
between the two measures. A disadvantage of the latter approach is that it is derived
by assuming the Von Neumann-Morgenstern quadratic utility function and that
returns are normally distributed (Hanoch and Levy 1969). Thus, it cannot capture
the richness of the former approach. Among the MR analyses, the most popular
measure is the SR introduced by Sharpe (1966). As the SR requires strong assump-
tions that the returns of assets being analyzed have to be iid, various measures for
MR analysis have been developed to improve the SR, including the Sortino ratio
(Sortino and van der Meer 1991), the conditional SR (Agarwal and Naik 2004), the
modified SR (Gregoriou and Gueyie 2003), value at risk (Ma and Wong 2010),
expected shortfall (Chen 2008), and the mixed Sharpe ratio (Wong et al. 2012).
However, most of the empirical studies, see, for example, Eling and Schuhmacher
(2007), find that the conclusions drawn by using these ratios are basically the same
as that drawn by the SR. Nonetheless, Leung and Wong (2008) have developed
a multiple SR statistic and find that the results drawn from the multiple Sharpe ratio
statistic can be different from its counterpart pair-wise SR statistic comparison,
indicating that there are some relationships among the assets that have not being
revealed using the pair-wise SR statistics. The MVR test could be the right
candidate to reveal these relationships.
One may claim that the limitation of the MVR test statistic is that it can only
draw conclusion for investors with quadratic utility functions and for normal-
distributed assets. Wong (2006), Wong and Ma (2008), and others have shown
that the conclusion drawn from the MR comparison is equivalent to the comparison
of expected utility maximization for any risk-averse investor, not necessarily with
only quadratic utility function, and for assets with any distribution, not necessarily
normal distribution, if the assets being examined belong to the same location-scale
family. In addition, one can also apply the results from Li and Wong (1999) and
Egozcue and Wong (2010) to generalize the result so that it will be valid for any
risk-averse investor and for portfolios with any distribution if the portfolios
being examined belong to the same convex combinations of (same or different)
53 Internet Bubble Examination with Mean-Variance Ratio 1463
Acknowledgment We would like to thank the editor C.-F. Lee for his substantive comments that
have significantly improved this manuscript. The third author would also like to thank Professors
Robert B. Miller and Howard E. Thompson for their continuous guidance and encouragement. The
research is partially supported by grants from North East Normal University, National University
of Singapore, Hong Kong Baptist University and the Research Grants Council of Hong Kong. The
first author thanks the financial support from NSF China grant 11171057, Program for Changjiang
Scholars and Innovative Research Team in University, and the Fundamental Research Funds for
the Central Universities and NUS grant R-155-000-141-112.
References
Agarwal, V., & Naik, N. Y. (2004). Risk and portfolios decisions involving hedge funds. Review of
Financial Studies, 17, 63–98.
Bai, Z. D., Liu, H. X., & Wong, W. K. (2009a). Enhancement of the applicability of Markowitz’s
portfolio optimization by utilizing random matrix theory. Mathematical Finance, 19, 639–667.
Bai, Z. D., Liu, H. X., & Wong, W. K. (2009b). On the markowitz mean-variance analysis of self-
financing portfolios. Risk and Decision Analysis, 1, 35–42.
Bai, Z. D., Wong, W. K., & Zhang, B. Z. (2010). Multivariate linear and non-linear causality tests.
Mathematics and Computers in Simulation, 81, 5–17.
Bai, Z. D., Li, H., Liu, H. X., & Wong, W. K. (2011a). Test statistics for prospect and markowitz
stochastic dominances with applications. Econometrics Journal, 14, 278–303.
Bai, Z. D., Li, H., Wong, W. K., & Zhang, B. Z. (2011b). Multivariate causality tests with
simulation and application. Statistics and Probability Letters, 81, 1063–1071.
Bai, Z. D., Liu, H. X., & Wong, W. K. (2011c). Asymptotic properties of eigenmatrices of a large
sample covariance matrix. Annals of Applied Probability, 21, 1994–2015.
Bai, Z. D., Wang, K. Y., & Wong, W. K. (2011d). Mean-variance ratio test, a complement to
coefficient of variation test and Sharpe ratio test. Statistics and Probability Letters,
81, 1078–1085.
Bai, Z. D., Hui, Y. C., Wong, W. K., & Zitikis, R. (2012). Evaluating prospect performance:
Making a case for a non-asymptotic UMPU test. Journal of Financial Econometrics, 10(4),
703–732.
Baker, M., & Stein, J. C. (2004). Market liquidity as a sentiment indicator. Journal of Financial
Markets, 7, 271–300.
1464 Z.D. Bai et al.
Broll, U., Wahl, J. E., & Wong, W. K. (2006). Elasticity of risk aversion and international trade.
Economics Letters, 91, 126–130.
Broll, U., Egozcue, M., Wong, W. K., & Zitikis, R. (2010). Prospect theory, indifference curves,
and hedging risks. Applied Mathematics Research Express, 2010, 142–153.
Cadsby, C. B. (1986). Performance hypothesis testing with the Sharpe and Treynor measures:
A comment. Journal of Finance, 41, 1175–1176.
Chan, C. Y., de Peretti, C., Qiao, Z., & Wong, W. K. (2012). Empirical test of the efficiency of UK
covered warrants market: Stochastic dominance and likelihood ratio test approach. Journal of
Empirical Finance, 19, 162–174.
Chen, S. X. (2008). Nonparametric estimation of expected shortfall. Journal of Financial Econo-
metrics, 6, 87–107.
Egozcue, M., & Wong, W. K. (2010). Gains from diversification on convex combinations:
A majorization and stochastic dominance approach. European Journal of Operational
Research, 200, 893–900.
Egozcue, M., Fuentes Garcı́a, L., Wong, W. K., & Zitikis, R. (2011). Do investors like to
diversify? A study of Markowitz preferences. European Journal of Operational Research,
215, 188–193.
Eling, M., & Schuhmacher, F. (2007). Does the choice of performance measure influence the
evaluation of hedge funds? Journal of Banking and Finance, 31, 2632–2647.
Fong, W. M., Wong, W. K., & Lean, H. H. (2005). International momentum strategies:
A stochastic dominance approach. Journal of Financial Markets, 8, 89–109.
Fong, W. M., Lean, H. H., & Wong, W. K. (2008). Stochastic dominance and behavior towards
risk: The market for internet stocks. Journal of Economic Behavior and Organization,
68, 194–208.
Gasbarro, D., Wong, W. K., & Zumwalt, J. K. (2007). Stochastic dominance analysis of iShares.
European Journal of Finance, 13, 89–101.
Gregoriou, G. N., & Gueyie, J. P. (2003). Risk-adjusted performance of funds of hedge funds using
a modified Sharpe ratio. Journal of Wealth Management, 6, 77–83.
Hanoch, G., & Levy, H. (1969). The efficiency analysis of choices involving risk. Review of
Economic Studies, 36, 335–346.
Jobson, J. D., & Korkie, B. (1981). Performance hypothesis testing with the Sharpe and Treynor
measures. Journal of Finance, 36, 889–908.
Lam, K., Liu, T., & Wong, W. K. (2010). A pseudo-Bayesian model in financial decision making
with implications to market volatility, under- and overreaction. European Journal of Opera-
tional Research, 203, 166–175.
Lam, K., Liu, T., & Wong, W. K. (2012). A new pseudo Bayesian model with implications to
financial anomalies and investors’ behaviors. Journal of Behavioral Finance, 13(2), 93–107.
Lean, H. H., Phoon, K. F., & Wong, W. K. (2012). Stochastic dominance analysis of CTA funds.
Review of Quantitative Finance and Accounting, doi:10.1007/s11156-012-0284-1.
Leung, P. L., & Wong, W. K. (2008). On testing the equality of the multiple Sharpe ratios, with
application on the evaluation of iShares. Journal of Risk, 10, 1–16.
Li, C. K., & Wong, W. K. (1999). Extension of stochastic dominance theory to random variables.
RAIRO Recherche Opérationnelle, 33, 509–524.
Lintner, J. (1965). The valuation of risky assets and the selection of risky investment in stock
portfolios and capital budgets. Review of Economics and Statistics, 47, 13–37.
Lo, A. (2002). The statistics of Sharpe ratios. Financial Analysis Journal, 58, 36–52.
Ma, C., & Wong, W. K. (2010). Stochastic dominance and risk measure: A decision-theoretic
foundation for VaR and C-VaR. European Journal of Operational Research, 207, 927–935.
Markowitz, H. M. (1952). Portfolio selection. Journal of Finance, 7, 77–91.
Matsumura, E. M., Tsui, K. W., & Wong, W. K. (1990). An extended multinomial-Dirichlet model
for error bounds for dollar-unit sampling. Contemporary Accounting Research, 6, 485–500.
Memmel, C. (2003). Performance hypothesis testing with the Sharpe ratio. Finance Letters,
1, 21–23.
53 Internet Bubble Examination with Mean-Variance Ratio 1465
Ofek, E., & Richardson, M. (2003). A survey of market efficiency in the internet sector. Journal of
Finance, 58, 1113–1138.
Perkins, A. B., & Perkins, M. C. (1999). The internet bubble: Inside the overvalued world of high-
tech stocks. New York: Harper Business.
Schultz, P., & Zaman, M. (2001). Do the individuals closest to internet firms believe they are
overvalued? Journal of Financial Economics, 59, 347–381.
Sharpe, W. F. (1964). Capital asset prices: A theory of market equilibrium under conditions of risk.
Journal of Finance, 19, 425–442.
Sharpe, W. F. (1966). Mutual funds performance. Journal of Business, 39, 119–138.
Sortino, F. A., & van der Meer, R. (1991). Downside risk. Journal of Portfolio Management,
17, 27–31.
Sriboonchitta, S., Wong, W. K., Dhompongsa, D., & Nguyen, H. T. (2009). Stochastic dominance
and applications to finance, risk and economics. Boca Raton: Chapman and Hall/CRC.
Statman, M. (2002). Lottery players/stock traders. Journal of Financial Planning, 14–21.
Tobin, J. (1958). Liquidity preference as behavior towards risk. Review of Economic Studies,
25, 65–86.
Wong, W. K. (2006). Stochastic dominance theory for location-scale family. Journal of Applied
Mathematics and Decision Sciences, 2006, 1–10.
Wong, W. K. (2007). Stochastic dominance and mean-variance measures of profit and loss
for business planning and investment. European Journal of Operational Research,
182, 829–843.
Wong, W. K., & Bian, G. (2000). Robust Bayesian inference in asset pricing estimation. Journal of
Applied Mathematics & Decision Sciences, 4, 65–82.
Wong, W. K., & Chan, R. (2004). The estimation of the cost of capital and its reliability.
Quantitative Finance, 4, 365–372.
Wong, W. K., & Chan, R. (2008). Markowitz and prospect stochastic dominances. Annals of
Finance, 4, 105–129.
Wong, W. K., & Li, C. K. (1999). A note on convex stochastic dominance theory. Economics
Letters, 62, 293–300.
Wong, W. K., & Ma, C. (2008). Preferences over location-scale family. Economic Theory,
37, 119–146.
Wong, W. K., & Miller, R. B. (1990). Analysis of ARIMA-noise models with repeated time series.
Journal of Business and Economic Statistics, 8, 243–250.
Wong, W. K., Chew, B. K., & Sikorski, D. (2001). Can P/E ratio and bond yield be used to beat
stock markets? Multinational Finance Journal, 5, 59–86.
Wong, W. K., Manzur, M., & Chew, B. K. (2003). How rewarding is technical analysis? Evidence
from Singapore stock market. Applied Financial Economics, 13, 543–551.
Wong, W. K., Thompson, H. E., Wei, S., & Chow, Y. F. (2006). Do winners perform better than
losers? A stochastic dominance approach. Advances in Quantitative Analysis of Finance and
Accounting, 4, 219–254.
Wong, W. K., Phoon, K. F., & Lean, H. H. (2008). Stochastic dominance analysis of Asian hedge
funds. Pacific-Basin Finance Journal, 16, 204–223.
Wong, W. K., Wright, J. A., Yam, S. C. P., & Yung, S. P. (2012). A mixed Sharpe ratio. Risk and
Decision Analysis, 3(1–2), 37–65.
Quantile Regression in Risk Calibration
54
Shih-Kang Chao, Wolfgang Karl Härdle, and Weining Wang
Contents
54.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1468
54.2 Methodology . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1472
54.2.1 Constructing Partial Linear Model (PLM) for CoVaR . . . . . . . . . . . . . . . . . . . . . . . 1472
54.2.2 Backtesting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1475
54.2.3 Risk Contribution Measure . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1476
54.3 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1478
54.3.1 CoVaR Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1478
54.3.2 Backtesting . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1480
54.3.3 Global Risk Contribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1482
54.4 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1484
Appendix 1: Local Linear Quantile Regression (LLQR) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1484
Appendix 2: Confidence Band for Nonparametric Quantile Estimator . . . . . . . . . . . . . . . . . . . . . . 1485
Appendix 3: PLM Model Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1487
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1488
Abstract
Financial risk control has always been challenging and becomes now an even
harder problem as joint extreme events occur more frequently. For decision
makers and government regulators, it is therefore important to obtain accurate
€
The financial support from the Deutsche Forschungsgemeinschaft via SFB 649 “Okonomisches
Risiko,” Humboldt-Universität zu Berlin is gratefully acknowledged
S.-K. Chao (*) • W. Wang
Ladislaus von Bortkiewicz Chair of Statistics, C.A.S.E. – Center for Applied Statistics and
Economics, Humboldt–Universität zu Berlin, Berlin, Berlin, Germany
e-mail: shih-kang.chao@cms.hu-berlin.de; wangwein@cms.hu-berlin.de
W.K. Härdle
Ladislaus von Bortkiewicz Chair of Statistics, C.A.S.E. – Center for Applied Statistics and
Economics, Humboldt–Universität zu Berlin, Berlin, Berlin, Germany
Lee Kong Chian School of Business, Singapore Management University, Singapore, Singapore
e-mail: haerdle@wiwi.hu-berlin.de
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1467
DOI 10.1007/978-1-4614-7750-1_54,
# Springer Science+Business Media New York 2015
1468 S.-K. Chao et al.
Keywords
CoVaR • Value-at-Risk • Quantile regression • Locally linear quantile regression •
Partial linear model • Semiparametric model
54.1 Introduction
Sufficiently accurate risk measures are needed not only in crisis times. In the last
two decades, the world has gone through several financial turmoils, and the
financial market is getting riskier and the scale of loss soars. Beside marginal
extremes that can shock even a well-diversified portfolio, the focus of intensified
research in the recent years has been on understanding the interdependence of risk
factors and their conditional structure.
The most popular risk measure is the Value-at-Risk (VaR), which is defined as the
t-quantile of the return distribution at time t + d conditioned on the information set F t:
Here Xt denotes the asset return and t is taking values such as 0.05, 0.01 or 0.001
to reflect negative extreme risk.
Extracting information in economic variables to predict VaR brings quantile
regression into play here, since VaR is the quantile of the conditional asset return
distribution. Engle and Manganelli (2004) propose the nonlinear Conditional
Autoregressive Value-at-Risk (CaViaR) model, which uses (lag) VaR and lag returns.
Chernozhukov and Umantsev (2001) propose linear and quadratic time series models
for VaR prediction. Kuan et al. (2009) propose the Conditional AutoRegressive
Expectile (CARE) model, and argue that expectiles are more sensitive to the scale
of losses. These studies and many others apply quantile regression in a prespecified,
often linear functional form. In a more nonparametric context, Cai and Wang (2008)
estimate the conditioned cdf by a double kernel local linear estimator and find the
54 Quantile Regression in Risk Calibration 1469
quantile by inverting the cdf. Schaumburg (2011) uses the same technique together
with extreme value theory for VaR prediction. Taylor (2008) proposes Exponentially
Weighted Quantile Regression (EWQR) for estimating VaR time series.
The aforementioned studies focus mainly on the VaR estimation for single assets
and do not directly take into account the escalated spillover effect in crisis periods.
This risk of joint tail events of asset returns has been identified and studied. Further,
Brunnermeier and Pedersen (2008) show that the negative feedback effect of a “loss
spiral” and a “margin spiral” leads to the joint depreciation of assets prices. It is
therefore important to develop risk measures which can quantify the contagion effects
of negative extreme event.
Acharya et al. (2010) propose the concept of marginal expected shortfall (MES),
which measures the contribution of individual assets to the portfolio expected shortfall.
Via an equilibrium argument, the MES is shown to be a predictor to a financial
institution’s risk contribution. Brownlees and Engle (2012) demonstrate that the
MES can be written as a function of volatility, correlation, and expectation conditional
on tail events. Huang et al. (2012) propose the distress insurance premium (DIP),
a measure similar to MES but computed under the risk-neutral probability. This
measure can therefore be viewed as the market insurance premium against the event
that the portfolio loss exceeds a low level. Adams et al. (2012) construct financial
indices on return of insurance companies, commercial banks, investment banks, and
hedge funds, and use a linear model for the VaRs of the four financial indices to forecast
the state-dependent sensitivity VaR (SDSVaR). The risk measures proposed above
have some shortcomings though: The computation of DIP is demanding since this
involves the simulation of rare events. MES suffers from the scarcity of data because it
conditions on a rare event.
In Adrian and Brunnermeier (2011) (henceforth AB), the CoVaR concept of
conditional VaR is proposed, which controls the effect of the negative extreme event
of some systemically risky financial institutions. Formally, let C(Xi,t) be some event of
a asset i return Xi,t at time t and take Xj,t as another asset return (e.g., the market index).
The CoVaRtj|i,t is defined as the t-quantile of the conditional probability distribution:
n o
P Xj, t CoVaRtjji, t C Xi, t , Mt ¼ t, (54.2)
d i, t ¼ ^a i þ ^g Τ Mt1 :
VaR (54.4)
i
In the second step, one models the asset j return as a linear function of asset
return i and market variables Mt:
In Eq. 54.5, the variable Xi,t influences the return Xj,t in a linear fashion.
However, the linear parametric model may not be flexible enough to capture the
tail dependence between i and j. The linearity of the conditioned quantile curves of
Xj on Xi is challenged by the confidence bands of the nonparametric quantile curves,
as shown in Fig. 54.1. The left tail quantile from linear parametric quantile
regression (red) lies well outside the confidence band (gray dashed curve) of Hardle
and Song (2010). This motivates empirically that a linear model is not flexible
enough for the CoVaR question at hand.
Nonparametric models can be used to account for the nonlinear structure of the
conditional quantile, but the challenge for using such models is the curse of dimension-
ality, as the quantile regression in CoVaR modeling often involves many variables. Thus,
we resort to semiparametric partial linear model (PLM) which preserves some flexibility
of the nonparametric model while suffers little from the curse of dimensionality.
As an illustration, the VaR/CoVaR of Goldman Sachs (GS) returns are shown,
given the returns of Citigroup (C) and S&P500 (SP). S&P500 index return is used
as a proxy for the market portfolio return.
Choosing market variables is crucial for the VaR/CoVaR estimation. For the
variables representing market states, we follow the most popular choices such as
VIX, short-term liquidity spread, etc. In particular, the variable we use for real
estate companies is the Dow Jones U.S. real estate index. The daily data date from
August 4, 2006 to August 4, 2011.
54 Quantile Regression in Risk Calibration 1471
0.0
0.0
−0.5
−0.5
Fig. 54.1 Goldman Sachs (GS) and Citigroup (C) weekly returns 0.05(left) and 0.1(right)
quantile functions. The y-axis is GS daily returns and the x-axis is the C daily returns.
The blue curve are the locally linear quantile regression curves (see Appendix 1). The locally
linear quantile regression bandwidth are 0.1026 and 0.0942. The red lines are the linear parametric
quantile regression line. The antique white dashed curves are the asymptotic confidence band
(see Appendix 2) with significance level 0.05. The sample size N ¼ 546
We call it “marginal contribution of risk” (MCR). Bae et al. (2003) and many others
have pointed out the phenomenon of financial contagion across national borders. This
motivates us to consider the stock indices of a few developed markets and explore
their risk contribution to the global stock market. MCR results show that when the
global market condition varies, the source of global market risk can be different. To
be more specific, when the global market return is bad, the risk contribution from the
USA is the largest. On the other hand, during financially stable periods, Hong Kong
and Japan are more significant risk contributors than the USA to the global market.
This study is organized as follows: Sect. 54.2 introduces the construction and
the estimation of the PLM model of CoVaR. The backtesting methods and our
risk contribution measure are also introduced in this section. Section 54.3 presents
the Goldman Sachs CoVaR time series and the backtesting procedure results.
Section 54.4 presents the conclusion and possible further studies. Appendices
describe the detailed estimation and statistical inference procedures used in
this study.
54.2 Methodology
d t ¼ ^a i þ ^g i Mt1 ,
VaRi,
AB
d
CoVaR a jji þ b^jji VaR
jji, t ¼ ^
d i, t þ ^gΤ Mt1 :
jji
where ð^a i ; ^g i Þ and ^a jji ; b^jji ; ^g jji are estimated from a linear model using standard
linear quantile regression.
We have motivated the need for more general functional forms for the quantile
curve. We therefore relax the model to a non- or semiparametric model. The market
54 Quantile Regression in Risk Calibration 1473
variable Mt is multidimensional, and the data frequency here is daily. The following
key variables are entering our analysis:
1. VIX: Measuring the model-free implied volatility of the market. This index is
known as the “fear gauge” of investors. The historical data can be found in the
Chicago Board Options Exchange’s website.
2. Short-term liquidity spread: Measuring short-term liquidity risk by the differ-
ence between the 3-month treasury repo rate and the 3-month treasury bill rate.
The repo data is from the Bloomberg database and the treasury bill rate data is
from the Federal Reserve Board H.15.
3. The daily change in the 3-month treasury bill rate: AB find that the changes have
better explanatory power than the levels for the negative tail behavior of asset returns.
4. The change in the slope of the yield curve: The slope is defined by the difference
of the 10-year treasury rate and the 3-month treasury bill rate.
5. The change in the credit spread between 10 years BAA-rated bonds and the
10 years treasury rate.
6. The daily Dow Jones U.S. Real Estate index returns: The index reflects the
information of lease rates, vacancies, property development, and transactions of
real estates in the USA.
7. The daily S&P500 index returns: The approximate of the theoretical market
portfolio returns.
The variables 3, 4, 5 are from the Federal Reserve Board H.15 and the data of
6 and 7 are from Yahoo Finance.
First we conduct a statistical check of the linearity between GS return and the
market variables using the confidence band as constructed in Appendix 2. As shown
in Fig. 54.2, except for some ignorable outsiders, the linear quantile regression line
lies in the LLQR asymptotic confidence band.
On the other hand, there is nonlinearity between two individual assets Xi and Xj.
To illustrate this, we regress Xj on Mt, and then take the residuals and regress them
on Xi. Again the Xj,t is GS daily return and Xi is C daily return. The result is shown
in Fig. 54.3. The linear QR line (red) lies well outside the LLQR confidence band
(magenta) when the C return is negative. The linear quantile regression line is fairly
flat. The risk of using a linear model is obvious in this figure: The linear regression
can “average out” the humped relation of the underlying structure (blue), and
therefore imply a model risk in estimation.
Based on the results of the linearity tests above, we construct a PLM model:
where Xi,t, Xj,t are asset returns of i, j firms. Mt is a vector of market variables at time
t as introduced before. If i ¼ S&P500, Mt is set to consist of the first 6 market
variables only. Notice the variable Xi,t enter the Eq. 54.8 nonlinearly.
1474 S.-K. Chao et al.
0.2
0.0
0.0
−0.4
−0.3
0.2 0.2
0.0
0.0
−0.3 −0.2
−0.5 0.0 0.5 0.00 0.01 0.02 0.03 0.04
Change in yields of 3 mon. TB Slope of yield curve
0.2 0.2
0.0 0.0
−0.3 −0.3
−0.001 0.001 0.003 −0.05 0.00 0.05 0.10
Credit Spread S&P500 Index Returns
0.2
0.0
−0.2
Fig. 54.2 The scatter plots of GS daily returns to the seven market variables with the LLQR
curves. The bandwidths are selected by the method described in Appendix 1. The LLQR band-
widths are 0.1101, 0.1668, 0.2449, 0.0053, 0.0088, 0.0295 and 0.0569. The data period is from
August 4, 2006, to August 4, 2011. N ¼ 1260. t ¼ 0.05
Applying the algorithm of Koenker and Bassett (1978) to Eq. n54.7 and the o
process described in Appendix 3 to Eq. 54.8, we get f^a i ; ^g i g and ^a jji , b^i , ^l ðÞ
with F1 1
ei, t ðtjMt1 Þ ¼ 0 for Eq. 54.7 and Fei, t tjM t1 , X i, t ¼ 0 for Eq. 54.8. Finally,
we estimate the PLM CoVaRjji,t by
54 Quantile Regression in Risk Calibration 1475
0.2
0.0
−0.2
−0.4
−0.6
−0.8
−0.15 −0.10 −0.05 0.00 0.05 0.10
Fig. 54.3 The nonparametric part l^GSjC ðÞ of the PLM estimation. The y-axis is the GS daily
returns. The x-axis is the C daily returns. The blue curve is the LLQR quantile curve. The red line
is the linear parametric quantile line. The magenta dashed curves are the asymptotic confidence
band with significance level 0.05. The data is from June 25, 2008, to December 23, 2009.
378 observations. Bandwidth ¼ 0.1255. t ¼ 0.05
d i, t ¼ ^a i þ ^g Τi Mt1 ,
VaR (54.9)
^Τ
a^jji þ bej Mt1 þ ^l jji VaR
PLM
d d i, t :
CoVaR jji, t ¼ e (54.10)
54.2.2 Backtesting
where VaRdt can be replaced by CoVaR d t in the case of CoVaR. It should form
t t
a sequence of martingale difference.
There is a large literature on martingale difference tests. We adopt
Ljung-Box test, Lobato test, and the CaViaR test. The Ljung-Box test and Lobato
test aim to check whether the time series is autocorrelated. If the time series is
autocorrelated, then we reject of course the hypothesis that the time series is a
martingale difference.
^ k be the estimated autocorrelation of lag k of the sequence
Particularly, let r
of violation {It} and n be the length of the time series. The Ljung-Box test
statistics is
1476 S.-K. Chao et al.
Xm
^ 2k
r L
LBðmÞ ¼ nðn þ 2Þ ! wðmÞ, (54.11)
k¼1
nk
as n ! 1.
This test is too strong though in the sense that the asymptotic distribution is
derived based on the i.i.d. assumption. A modified Box-Pierce test is proposed by
Lobato et al. (2001), who also consider the test of no autocorrelation, but their test
is more robust to the correlation of higher (greater than the first) moments.
(Autocorrelation in higher moments does not contradict with the martingale differ-
ence hypothesis.) The test statistics is given by
Xm
^ 2k L
r
Lð m Þ ¼ n ! wðmÞ,
^v
k¼1 kk
as n ! 1, where
Xnk 2
1
n ðyi yÞ2 yiþk y
^v kk ¼ i¼1
n Xn o2 :
2
1
N ð
i¼1 i
y y Þ
The CaViaR test, proposed by Berkowitz et al. (2011), is based on the idea that if
the sequence of violation is a martingale difference, there ought to be no correlation
between any function of the past variables and the current violation. One way to test
this uncorrelatedness is through a linear model. The model is
I t ¼ a þ b1 I t1 þ b2 VaRt þ ut ,
where VaRt can be replaced by CoVaRt in the case of conditional VaR. The residual
ut follows a Logistic distribution since It is binary. We get the estimates of the
Τ
coefficients b^1 ; b^2 . Therefore, the null hypothesis is b^1 ¼ b^2 ¼ 0 . This
hypothesis can be tested by Wald’s test.
We set m ¼ 1 or 5 for the Ljung-Box and Lobato tests. For the CaViaR test, two
data periods are considered separately. The first is the overall data from August
4, 2006, to August 4, 2011. The second is the data from August 4, 2008, to August
4, 2009, the period when the financial market reached its bottom. By separately
testing the two periods, we can gain more insights into the PLM model.
The risk contribution of one firm to the market is one of the top concerns among
central bankers. The regulator can restrict the risky behaviors of the financial
institution with high-risk contribution to the market, and reduce the
institution’s incentive to take more risk. AB propose the idea of DCoVaR, which
is defined by
54 Quantile Regression in Risk Calibration 1477
t
where CoVaRjji,t is defined as in the introduction. j, i represent the financial system
and an individual asset. t ¼ 0.5 corresponds to the normal state of the individual
asset i. This is essentially a sensitivity measure quantifying the effect to the
financial system from the occurrence of a tail event of asset Xi.
In this study, we adopt a mathematically intuitive way to measure the marginal
effect by searching the first order derivative of the quantile function. Because the
spillover effect from stock market to stock market has already got much attention, it
is important to investigate the risk contribution of a local market to the global stock
market. The estimation is conducted as follows:
First, one estimates the following model nonparametrically:
Xj, t ¼ f 0:05
j ðX t Þ þ ej , (54.13)
The quantile function fj 0.05() is estimated with local linear quantile regression
with t ¼ 0.05, described with more details in Appendix 1. Xj is the weekly return of
the stock index of an individual country and X is the weekly return of the global
stock market.
Second, with f^j ðÞ, we compute the “marginal contribution of risk” (MCR) of
0:05
institution j by
@ ^0:05 ðxÞ
f
MCRtj ¼
j 1 , (54.14)
@x x ¼ F^x ðtk Þ
1
where F^ ðtk Þ is a consistent estimator of the tk quantile of the global market return,
and it can be estimated by regressing Xt on the time trend. We put k ¼ 1, 2 with
t1 ¼ 0.5 and t2 ¼ 0.05. The quantity Eq. 54.14 is similar to the MES proposed by
Acharya et al. (2010) in the sense that the conditioned event belongs to the
information set of the market return, but we reformulate it in the VaR framework
instead of the expected shortfall framework.
There are some properties of the MCR to be described further. First, tk deter-
mines the condition of the global stock market. This allows us to explore the risk
contribution from the index j to the global market, given different global market
status. Second, the higher the value of MCR, the more risk factor j imposes on the
market in terms of risk. Third, since the function fj 0.05() is estimated by LLQR, the
quantile curve is locally linear, and therefore, the local first order derivative is
straightforward to compute.
We choose indices j ¼ S&P500, NIKKEI225, FTSE100, DAX30, CAC40, Hang
Seng as the approximate of the market returns of each developed country or market.
The global market is approximated by the MSCI World (developed countries)
market index. The data is weekly from April 11, 2004, to April 11, 2011, and
t ¼ 0.05.
1478 S.-K. Chao et al.
54.3 Results
The estimation results of VaR/CoVaR are shown in this section. We compute three
types of VaR/CoVaR of GS, with a moving window size of 126 business days and
t ¼ 0.05.
First, the VaR of GS is estimated using linear quantile regression:
d C, t ¼ ^a C þ^g ΤC Mt1 ;
VaR (54.16)
d AB
CoVaR a GSjC þ b^GSjC VaR
GSjC, t ¼ ^
d C, t þ ^gΤ Mt1 :
GSjC (54.17)
e
d SP, t ¼ ^a SP þ ^gΤ M
VaR SP t1 ; (54.18)
d AB
CoVaR a GSjSP þ b^GSjSP VaR
GSjSP, t ¼ ^
d SP, t þ ^gΤ M e
GSjSP t1 , (54.19)
d C, t ¼ ^a C þ ^gΤ Mt1 ;
VaR (54.20)
C
^Τ
d PLM ¼ e
CoVaR GSjC, t a^GSjC þ beGSjC Mt1 þ ^l GSjC VaR
d C, t : (54.21)
If SP replaces C:
e
d SP, t ¼ ^a SP þ ^gΤ M
VaR SP t1 ; (54.22)
^Τ
d PLM ¼ e
CoVaR GSjSP, t a^GSjSP þ beGSjSP M
e t1 þ ^ d SP, t :
l GSjSP VaR (54.23)
The coefficients in Eqs. 54.15–54.20, and 54.22 are estimated from the linear
quantile regression and those in Eqs. 54.21 and 54.23 are estimated from the
method described in Appendix 3.
d GS, t sequence. The VaR forecasts (red) seem to form
Figure 54.4 shows the VaR
a lower cover of the GS returns (blue). This suggests that the market variables Mt
54 Quantile Regression in Risk Calibration 1479
−0.2
0.2
0.0
−0.2
Fig. 54.5 The CoVaR of GS given the VaR of C. The gray dots mark the daily returns of GS. The
d PLM . The dark blue curve is the median LLQR smoother of
light green dashed curve is the CoVaR GSjC, t
d AB . The purple
the light green dashed curve with h ¼ 3.19. The cyan dashed curve is the CoVaR GSjC, t
curve is the median LLQR smoother of the cyan dashed curve with h ¼ 3.90. The red curve is the
d GS, t . t ¼ 0.05. The moving window size is 126 days
VaR
have some predictive power for the left tail quantile of the GS return distribution.
Figure 54.5 shows the sequences CoVaR d AB d PLM
GSjSP, t (cyan) and CoVaRGSjC, t (light
green). As the time series of the estimates is too volatile, we smooth it further by the
median LLQR. The two estimates are similar as the market state is stable, but
during the period of financial instability (from mid-2008 to mid-2009), the two
estimates have different behavior. The performances of these estimates are
evaluated by backtesting procedure in Sect. 54.3.2.
Table 54.1 shows the summary statistics of the VaR/CoVaR estimates. The first
three rows show the summary statistics of VaR d GS, t , VaR
d C, t , and VaR
d SP, t . The
d GS, t has lower mean and higher standard deviation than the other two. Partic-
VaR
ularly during 2008–2009, the standard deviation of the GS VaR is twice as much as
the other two. The mean and standard deviation of the VaR d C, t and VaR d SP, t
are rather similar. The last four rows show the summary statistics of CoVaRd PLM ,
GSjC, t
1480 S.-K. Chao et al.
Table 54.1 VaR/CoVaR summary statistics. The overall period is from August 4, 2006, to
August 4, 2011. The crisis period is from August 4, 2008, to August 4, 2009. The numbers in
the table are scaled up by 102
mean-overall sd-overall mean-crisis sd-crisis
d GS, t
VaR 3.66 3.08 7.43 4.76
d C, t
VaR 2.63 1.67 4.62 2.25
d SP, t
VaR 2.09 1.57 3.88 2.24
d PLM
CoVaR 4.26 3.84 8.79 5.97
GSjC, t
d AB
CoVaR 4.60 4.30 10.36 6.32
GSjC, t
d PLM
CoVaR 3.86 3.30 8.20 4.69
GSjSP, t
d AB
CoVaR 5.81 4.56 12.65 5.56
GSjSP, t
d AB , CoVaR
CoVaR d PLM , and CoVaR d AB
GSjC, t GSjSP, t GSjSP, t . This shows that the CoVaR
obtaining from the AB model has smaller mean and greater standard deviation
than the CoVaR obtaining from PLM model.
Figure 54.6 shows the bandwidth sequence of the nonparametric part of the
PLM estimation. The bandwidth varies with time. Before mid-2007, the bandwidth
sequence is stably jumping around 0.2. After that the sequence becomes very
volatile. This may have something to do with the rising systemic risk.
54.3.2 Backtesting
For the evaluation of the CoVaR models, we resort to the backtesting procedure
described in Sect. 54.2.2. In order to perform the backtesting procedure,
the sequences {It} (defined in Sect. 54.2.2) have to be computed for all
VaR/CoVaR estimates. Figure 54.7 shows the timings of the violations
{t : It ¼ 1} of CoVaRd PLM , CoVaR d AB d
GSjC, t GSjC, t and VaR GS, t . This figure shows the total
number of violations of PLM CoVaR and CoVaR is similar, while VaR d GS, t has more
d
violations than the both. The VaR GS, t has a few clusters of violations in both
financial stable and unstable periods. This may result from the failure VaR d GS, t to
d PLM
adapt for the negative shocks. The violations of CoVaRGSjC, t are more evenly
distributed. The violations of CoVaR d AB
GSjC, t have large clusters during financially
stable period, while the violation during financial crisis period is meager. This
d AB tend to overreact, as it is slack during the stable
contrast suggests that CoVaR GSjC, t
period but is too tight during the unstable period.
Figure 54.8 shows the timings of the violations {t : It ¼ 1} of CoVaR d PLM ,
GSjSP, t
d
CoVaR AB d d PLM
GSjSP, t , and VaR GS, t . The overall number of violations of CoVaRGSjSP, t is
more than that of VaR d GS, t , and it has many clusters. CoVaR d PLM
GSjSP, t behaves
d PLM
differently from CoVaRGSjC, t. The SP may not be more informative than C, though
54 Quantile Regression in Risk Calibration 1481
0.4
0.1
^
Fig. 54.6 LLQR bandwidth in the daily estimation of CoVaRPLM
GSjC, t . The average bandwidth is
0.24
Table 54.2 Goldman Sachs VaR/CoVaR backtesting p-values. The overall period is from August
4, 2006, to August 4, 2011. The crisis period is from August 4, 2008, to August 4, 2009. LB(1) and
LB(5) are the Ljung-Box tests of lags 1 and 5. L(1) and L(5) are the Lobato tests of lags 1 and
5. CaViaR-overall and CaViaR-crisis are two CaViaR tests described in Sect. 2.2 applied on the
two data periods
Measure LB(1) LB(5) L(1) L(5) CaViaR-overall CaViaR-crisis
Panel 1
d GS, t
VaR 0.3449 0.0253* 0.3931 0.1310 1.265 106*** 0.0024**
Panel 2
d AB
CoVaR 0.0869 0.2059 0.2684 0.6586 8.716 107*** 0.0424*
GSjSP, t
d PLM
CoVaR 0.0518 0.0006*** 0.0999 0.0117* 2.2 1016*** 0.0019**
GSjSP, t
Panel 3
d AB
CoVaR 0.0489* 0.2143 0.1201 0.4335 3.378 109*** 0.0001***
GSjC, t
d PLM
CoVaR 0.8109 0.0251* 0.8162 0.2306 2.946 109*** 0.0535
GSjC, t
* ** ***
, and denote significance at the 5 %, 1 % and 0.1 % levels
d AB
market and improve the performance of the estimates. The CoVaR GSjC, t is rejected
by the two CaViaR tests and the LB(1) test with 0.1 % and 5 % significant level.
d PLM is not rejected by the CaViaR-crisis test. This implies that
However, CoVaR GSjC, t
the nonparametric part in the PLM model captures the nonlinear effect of C returns
to GS returns, which can lead to better risk-measuring performance.
In this section, we present the MCR (defined in Sect. 54.2.3), which measures the
marginal risk contribution of risk factors. We choose t1 ¼ 0.5, associated to the
54 Quantile Regression in Risk Calibration 1483
0.2
0.0
−0.2
2005 2006 2007 2008 2009 2010 2011
1.5
1.0
0.5
54.4 Conclusion
In this study, we construct a PLM model for the CoVaR, and we compare it to the
AB model by backtesting. Results show that PLM CoVaR is preferable, especially
during a crisis period. The study of the MCR reveals the fact that the risk from each
country can vary with the state of global economy.
As an illustration, we only study the Goldman Sachs conditional VaR with
Citigroup and S&P500 as conditioned risk sources. In practice, we need to choose
variables. In Hautsch et al. (2011), the Least Absolute Shrinkage and Selection
Operator (LASSO) techniques are used to determine the most relevant systemic risk
sources from a pool of financial institutions. A VAR (Vector Autoregression) model
may be also suitable for capturing the asset dynamics, but the estimation may be
more involved. We may include other firm-specific variables such as corporate
bond yields as these variables can bear other information which is not included in
the stock returns or stock indices.
yi ¼ lðxi Þ þ ei ,
1
with Fe|x (t) ¼ 0. A locally linear kernel quantile estimator (LLQR) is estimated as
^
l ðx 0 Þ ¼ ^
a 0 from:
Xn x x
^a 0 ; ^b 0 ¼ argmin
i 0
K (54.24)
fa0 , b0 g i¼1 h
where h is the bandwidth, K(·) is a kernel, and rt(·) is the check function given by
rt ðuÞ ¼ t 1fu<0g u (54.26)
Figure 54.11 illustrates the check functions. Different loss functions give differ-
ent estimates. u2 corresponds to the conditional mean. rt(u) corresponds to the
conditional tth quantile.
It is shown by Fan et al. (1994) that the locally linear kernel estimator is asymptot-
ically efficient in a minimax sense. It also possesses good finite sampling
property which is adaptive to a variety of empirical density g(x) and has good boundary
property.
54 Quantile Regression in Risk Calibration 1485
1.0
0.5
0.0
−2 −1 0 1 2
Next, we describe the method to compute the bandwidths. The approach used
here follows Yu and Jones (1998). The bandwidth is chosen by
h 2 i1=5
ht ¼ hmean tð1 tÞ’ F1 ðtÞ , (54.27)
where hmean is the locally linear mean regression bandwidth, which can be com-
puted by the algorithm described in Ruppert and Wand (1995) or Ruppert
et al. (1995). ’(·) and F(·) are the pdf and cdf of the standard normal distribution.
Since we discuss the case for VaR, t is usually small. ht needs to be enlarged to
allow for more smoothing (usually taking 1.5ht or 2ht).
The approach is acceptable but not so flexible, because it is based on assuming the
quantile functions are parallel. A more flexible approach was developed by Spokoiny
et al. (2011). In order to stabilize the bandwidth choice, we first regress yi on the rank
of the corresponding xi and then rescale the resulted estimated values to the original
x space. Carroll and Hardle (1989) show that this local bandwidth estimator and the
global bandwidth estimator are asymptotically equivalent.
The uniform confidence band of the quantile estimator is based on the Theorem 2.2
and Corollary 2.1 presented in Hardle and Song (2010). The details are as follows.
1486 S.-K. Chao et al.
X
n
e n ðy; xÞ def
H ¼ n1 K h ðx Xi Þrt ðY i yÞ;
i¼1
ð
e ðy; xÞ def
H ¼ f ðx; yÞrt ðy yÞdy,
ℝ
8 1=2
logfc ðKÞg
> 1=2
< ð2d lognÞ þ ð2d logn
> þ 12flogd þ loglogng,
1
p1=2
dn ¼ if c1 ðK > 0;
>
: ð2d lognÞ1=2 þ ð2d logn1=2 logfc2 ðKÞgg, otherwise:
>
2p
Then
" ( ) #
jln ðxÞ lðxÞj
P ð2d lognÞ1=2 sup r ðxÞ dn <z
x2J lðK Þ1=2
! expf2expðzÞg,
as n ! 1, with
where fX(·) is the marginal pdf for X and f(·|x) is the conditional pdf of Y on X ¼ x.
The corollary followed by the theorem explicitly indicates how a uniform
confidence interval can be constructed.
Corollary 1 An approximate (1 a) 100 % confidence band is
n o1=2 1 n 1=2 o
ln ðnhÞ1=2 tð1 tÞlðK Þ=f^X ðtÞ f^ lðt t g dn þ cðaÞð2d logn ,
where c(a) ¼ log 2 log|log(1 a)| and f^X ðtÞ, f^flðtÞjtg are consistent estimates
for fX(t), f{l(t)|t}.
54 Quantile Regression in Risk Calibration 1487
0.5
0.5
0.0 0.0
Fig. 54.12 GS and C weekly returns 0.90(left) and 0.95(right) quantile functions. The y-axis is
GS daily returns and the x-axis is the C daily returns. The blue curves are the LLQR curves (see
Appendix 1). The LLQR bandwidths are 0.0942 and 0.1026. The red lines are the linear parametric
quantile regression line. The antique white curves are the asymptotic confidence band (see
Appendix 2) with significance level 0.05. N ¼ 546
For the PLM estimation, we adopt the algorithm described in Song et al. (2012).
Given data {(Xt, Yn)}Tn ¼ 1 bivariate and {Mt}Tn ¼ 1 multivariate random variables.
The PLM is:
Y t ¼ a þ bΤ Mt1 þ lðXt Þ þ et :
( )
X
n X
an
b^ ¼ argmin min rt Xj, t a b Mt1 lm 1 Xi, t 2 I nt
T
b l1 , ..., lan
t¼1 m¼1
References
Acharya, V. V., Pedersen, L. H., Philippon, T., & Richardson, M. (2010). Measuring systemic risk
(Working paper 10-02). Federal Reserve Bank of Cleveland.
Adams, Z., F€uss, R., & Gropp, R. (2012). Spillover effects among financial institutions: A state-
dependent sensitivity value-at-risk approach, SAFE Working Paper Series, No. 20, http://dx.
doi.org/10.2139/ssrn.2267853.
Adrian, T., & Brunnermeier, M. K. (2011). CoVaR (Staff reports 348). Federal Reserve Bank of
New York.
Bae, K.-H., Karolyi, G. A., & Stulz, R. M. (2003). A new approach to measuring financial
contagion. Review of Financial Studies, 16(3), 717–763.
Berkowitz, J., Christoffersen, P., & Pelletier, D. (2011). Evaluating value-at-risk models with
desk-level data management science, 57, 2213–2227.
Brownlees, C. T., & Engle, R. F. (2012). Volatility, correlation and tails for systemic risk
measurement. Available at SSRN: http://ssrn.com/abstract=1611229 or http://dx.doi.org/
10.2139/ssrn.1611229.
Brunnermeier, M., & Pedersen, L. H. (2008). Market liquidity and funding liquidity. Review of
Financial Studies, 22(2201), 2238.
Cai, Z., & Wang, X. (2008). Nonparametric estimation of conditional VaR and expected shortfall.
Journal of Econometrics, 147(120), 130.
Carroll, R., & Hardle, W. (1989). Symmetrized nearest neighbor regression estimates. Statistics &
Probability Letters, 7(315), 318.
Chernozhukov, V., & Umantsev, L. (2001). Conditional value-at-risk: Aspects of modeling and
estimation. Empirical Economics, 26(271), 292.
Engle, R., & Manganelli, S. (2004). CAViaR: Conditional autoregressive value at risk by regres-
sion quantiles. Journal of Business & Economic Statistics, 22(367), 381.
Fan, J., Hu, T.-C., & Truong, Y. K. (1994). Robust nonparametric function estimation. Scandina-
vian Journal of Statistics, 21, 433–446.
Härdle, W., & Song, S. (2010). Confidence bands in quantile regression. Econometric Theory, 26,
1180–1200.
Härdle, W., Liang, H., & Gao, J. (2000). Partially linear models. Heidelberg: Physica-Verlag.
Härdle, W., Muller, M., Sperlicli, S., & Werwatz, A. (2004). Nonparametric and semiparametric
models. Berlin: Springer.
Hautsch, N., Schaumburg, J., & Schienle, M. (2011). Financial network systemic risk contribu-
tions (Discussion paper 2011-072). CRC 649, Humboldt-Universitat zu Berlin.
Huang, X., Zhou, H., & Zhu, H. (2012). Systemic risk contributions Journal of Financial Services
Research, 42, 55–83. Springer US.
Koenker, R., & Bassett, G. S. (1978). Regression quantiles. Econometrica, 46, 33.
Kuan, C.-M., Yeh, J.-H., & Hsu, Y.-C. (2009). Assessing value at risk with CARE, the conditional
autoregressive expectile models. Journal of Econometrics, 150, 261–270.
Lobato, I., Nankervis, J. C., & Savin, N. (2001). Testing for autocorrelation using a modified
Box-Pierce Q test. International Economic Review, 42(1), 187–205.
54 Quantile Regression in Risk Calibration 1489
Ruppert, D., & Wand, M. P. (1995). Multivariate locally weighted least, squares regression.
Annals of Statistics, 23(1346), 1370.
Ruppert, D., Sheatlier, S. J., & Wand, M. P. (1995). An effective bandwidth selector for
local least squares regression. Journal of the American Statistical Association, 90(1257),
1270.
Schaumburg, J. (2011). Predicting extreme VaR: Nonparametric quantile regression with refine-
ments from extreme value theory (Discussion Paper 2010-009). CRC 649, Humboldt-
Universitat zu Berlin.
Song, S., Ritov, Y., & Härdle, W. (2012). Partial linear quantile regression and bootstrap
confidence bands. Journal of Multivariate Analysis, 107, 244–262.
Spokoiny, V., Wang, W., & Härdle, W. (2013). Local quantile regression. Journal of
Statistical Planning and Inference, 143(7), 1109–1129. doi:10.1016/j.jspi.2013.03.008,
with discussions.
Taylor, J. W. (2008). Using exponentially weighted quantile regression to estimate value at risk
and expected shortfall. Journal of Financial Econometrics, 6, 382–406.
Yu, K., & Jones, M. C. (1998). Local linear quantile regression. Journal of the American Statistical
Association, 98(228), 237.
Strike Prices of Options for Overconfident
Executives 55
Oded Palmon and Itzhak Venezia
Contents
55.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1492
55.2 Overconfidence and the Optimal Exercise Prices of Executive
Incentive Options . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1494
55.3 The Simulation Procedures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1497
55.4 Results and Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1498
55.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1503
Appendix 1 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1504
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1506
Abstract
We explore via simulations the impacts of managerial overconfidence on the
optimal strike prices of executive incentive options. Although it has been shown
that, optimally, managerial incentive options should be awarded in-the-money,
in practice most firms award them at-the-money. We show that the optimal strike
prices of options granted to overconfident executive are directly related to their
overconfidence level and that this bias brings the optimal strike prices closer to
the institutionally prevalent at-the-money prices. Our results thus support the
viability of the common practice of awarding managers with at-the-money
incentive options. We also show that overoptimistic CEOs receive lower com-
pensation than their realistic counterparts and that the stockholders benefit from
O. Palmon (*)
Department of Finance and Economics, Rutgers Business School – Newark and New Brunswick,
Piscataway, NJ, USA
e-mail: palmon@business.rutgers.edu; palmon@rbs.rutgers.edu
I. Venezia
School of Business, The Hebrew University, Jerusalem, Israel
Bocconi University, Milan, Italy
e-mail: msvenez@mscc.huji.ac.il
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1491
DOI 10.1007/978-1-4614-7750-1_55,
# Springer Science+Business Media New York 2015
1492 O. Palmon and I. Venezia
their managers bias. The combined welfare of the firm’s stakeholders is, how-
ever, positively related to managerial overconfidence.
The Monte Carlo simulation procedure described in Sect. 55.3 uses
a Mathematica program to find the optimal effort by managers and the optimal
(for stockholders) contract parameters. An expanded discussion of the simula-
tions, including the choice of the functional forms and the calibration of the
parameters, is provided in Appendix 1.
Keywords
Overconfidence • Managerial effort • Incentive options • Strike price • Simula-
tions • Behavioral finance • Executive compensation schemes • Mathematica
optimization • Risk aversion • Effort aversion
55.1 Introduction
1
According to Mahajan (2002), less than 1 % of firms used out-of-the-money strike prices.
Furthermore, in his study firms did not benefit from awarding such options to their managers.
2
Hall and Murphy (2000) did not show that at-the-money strike prices are optimal, just that they
possess the highest sensitivity to stock prices. They did not assume effort aversion by managers
either.
55 Strike Prices of Options for Overconfident Executives 1493
managers’ risk and effort aversion, that in-the-money options provide are optimal.
Such options provide managers a better risk-return trade-off and ultimately consti-
tute a better form of compensation than either out-of-the-money or at-the-money
options.3 Palmon et al. further argue that the asymmetric tax treatment of options
under the old (prior to 2006) tax system, which penalized the issuance of
in-the-money options, may have driven firms to use at-the-money options.
Whereas most studies of the issue of optimal incentive contracts assumed
that managers as well as stockholders are rational, there exists extensive literature
that documents that managers often are overconfident. Of the few studies that
explore the effect of cognitive biases on managerial compensation, none however
explores the effect of overconfidence on the optimal strike prices for the incentive
options. Gervais et al. (2011) investigate the optimal form of managerial
compensation under overconfidence but define overconfidence in the sense of
too-high-precision-of-estimates (calibration), and the managers in their model
exert effort to obtain better information on the investment parameters. There is an
abundant literature however that indicates the pervasiveness of overconfidence in
the optimism or “better than average” sense rather than in the calibration interpre-
tation (see, e.g., Malmendier and Tate 2005a, b, 2008; Roll 1986; Suntheim 2012).4
Oyer and Schaefer (2005) and Bergman and Jenter (2007) also consider the effect of
optimism, and other sentiments, on managerial compensation, but they do not
consider the effect of these sentiments on the optimal strike prices or on the
managers’ effort.
In this paper we investigate the hitherto unexplored question of the effect of
overconfidence on the optimal strike prices for risk-averse and effort-averse man-
agers. We show that overconfidence leads to higher optimal strike prices of
managerial incentive schemes, and that awarding overconfident CEOs at-the-
money options mitigates the stockholders’ vs. managers’ agency problem, leading
to higher managers’ productivity. Our results thus provide support for the viability
of the ubiquitous yet seemingly unoptimal practice of awarding CEOs with
at-the-money incentive options.
Whereas the main focus of the paper is the interaction between overconfidence
and the strike prices of managerial incentive options, it also sheds light on the effect
of overconfidence on the firm’s stakeholders (stockholders and managers). We
predict, as empirically shown by Otto (2011), that overoptimistic CEOs receive
lower compensation than their realistic counterparts. However, the stockholders
benefit from their managers bias since they pay less and enjoy the productivity of
the higher effort the overconfident manager exerts. We construct a measure of the
combined welfare of managers and stockholders and demonstrate that it is
3
Dittman et al. (2010) found that for a range of parameterizations, a principal-agent model with
loss-averse agents generates convex compensation contract but did not investigate the parameters
of the options to be used in the compensation package. Recently, however, Dittman and Yu (2011)
found that in-the-money options are optimal.
4
Glaser and Weber (2007) note that only overconfidence in the better than average sense affects
trading.
1494 O. Palmon and I. Venezia
where Y denotes the managerial extra effort (a managerial choice variable) and
m(Y) and s denote, respectively, the mean and the standard deviation of the
underlying normal distribution of the natural logarithm of X. We assume that
managerial effort increases cash flows and that overconfident managers
overestimate the impact of their effort on cash flows. Formally, we use the follow-
ing specification:
5
Palmon and Venezia (2012) explore the effect of managerial overconfidence on the firm’s
stockholders and show that overconfidence may improve welfare. However, that study does not
investigate the optimal strike price of managerial incentive options.
6
In our model we assume symmetry of information between the manager and the firm regarding
the distribution of cash flows of the firm except for the different view of the effect of the manager’s
effort on cash flows.
55 Strike Prices of Options for Overconfident Executives 1495
hlognormal2 distribution,
i
e½2mðYÞþs es 1 , respectively. Thus, it follows from Eq. 55.2 that a person
2
with a l overconfidence measure believes that the mean of the cash flows X is
emðYÞþ0:5s ¼ m0 þ 500lY and that their coefficient of variation is approximately s.7
2
Since managers and stockholders differ in their perception of the distributions of cash
flows, one must be careful in their use. In what follows we refer to the distribution of
cash flows as seen by stockholders as the realistic distribution, and will make a
special note whenever the manager’s overconfident beliefs are used.
Except for her overconfidence, the manager is assumed to be rational and to
choose her extra effort so as to maximize the expected value of the following utility
function which exhibits constant relative risk aversion (CRRA) with respect to
compensation:
1 1 1g
UðI; YÞ ¼ NYb þ I (55.3)
1g 1g
In Eq. 55.3, I denotes the manager’s monetary income, g denotes the constant
relative risk aversion measure, N is a scaling constant representing the importance
of effort relative to monetary income in the manager’s preferences, and the positive
parameter b is related to the convexity of the disutility of effort.
Since stockholders cannot observe the manager’s extra effort, they propose
compensation schemes that depend on the observed cash flows, but not on
Y. Stockholders, which we assume to be risk neutral, strive to make the compen-
sation performance sensitive in order to better align the manager’s incentives with
their own. Stockholders offer the manager a compensation package that includes
two components: a fixed wage (W) that she will receive regardless of her extra
effort and of the resulting cash flows and options with a strike price (K) for
a fraction (s) of the equity of the firm. We assume that stockholders offer the
contract that maximizes the value of their equity.
The following timeline of decisions is assumed. At the beginning of the period,
the firm chooses the parameters of the compensation contract (K, W, and s) and
offers this contract to the manager. Observing the contract parameters, and taking
into account the effects of her endeavors on firm cash flows and hence on her
compensation, the manager determines the extra-effort level Y that maximizes her
expected utility. At the end of the period, X is revealed, and the firm distributes the
cash flows to the manager and to the stockholders and then dissolves. The priority of
payments is as follows. The firm first pays the wages or only part of them if the cash
flows do not suffice. If the cash flows exceed the wage, W, but not (K + W), then the
h 2 i
7
More precisely the square of the coefficient of variation is es 1 which can be approximated
by s2 since for any small z, ez1 is close to z.
1496 O. Palmon and I. Venezia
managers just receive their fixed wage. The managers are paid the value of the
options s(X K W), in addition to W if X exceeds K + W. The manager therefore
receives the cash flows I(X) defined by
8 9 8 9
<X = <X W =
IðxÞ ¼ W when WXWþK (55.4)
: ; : ;
W þ sðX W KÞ WþKX
While the derivation of the optimal contract for any set of exogenous parameters
is conceptually straightforward, unfortunately, closed form solutions cannot be
obtained in our integrative model. Hence, following Hall and Murphy (2000), we
resort to simulations to evaluate the optimal contracts and analyze their properties.
In addition, we cannot use the Black-Scholes model to evaluate the executive stock
options since this model takes the values of the underlying asset as given, whereas
a crucial aspect of the managerial incentive scheme of our model is that managerial
extra effort and firm value are endogenously determined. We therefore introduce
a model that simultaneously simulates the manager’s optimal extra-effort level as
well as the expected values of the executive stock options and shareholders’ equity
8
Discounting the cash flows by an appropriate risk-adjusted discount rate would yield a linear
transformation of equity values. To simplify the presentation, and as is common in the literature,
we abstract from that.
55 Strike Prices of Options for Overconfident Executives 1497
for each compensation package. We check the robustness of our results by using
alternative parameters for the manager’s utility function and the distribution func-
tions of the cash flows.
We assume that managers have external employment opportunities and that stock-
holders offer managerial compensation packages that provide the managers with
a comparable expected utility.9 Without loss of generality (i.e., by an appropriate
definition of the wage units), we assume that these external employment opportu-
nities provide the manager an expected utility that equals the level of utility that is
obtained from a fixed compensation of 100 in the absence of any extra effort. Thus,
in all the simulations, we set the manager’s expected utility to correspond to the
level obtained from a fixed compensation of 100 (wage ¼ 100 and no option grants)
and no extra effort (which is the optimal extra-effort choice when no options are
granted).10 We then search over a grid of strike prices (using four-digit accuracy)
and find for each strike price the percentage of options that should be awarded so
that the manager’s expected utility equals the expected utility target when the
manager chooses the optimal extra-effort level. We identify the strike price that
is associated with the highest equity level and refer to this contract as the optimal
contract for the given set of parameters.
In calibrating the other parameters for the simulations, we try to approximately
conform to Hall and Murphy (2000) and Hall and Liebman (1998); to studies that
simulate decisions with effort aversion, such as Bitler et al. (2005); and to studies that
explore the effect of overconfidence on corporate decisions, such as Malmendier and
Tate (2005a, b, 2008).11 Accordingly, we set the parameters in our base case as
follows. The coefficient of variation, s, equals 0.3, and thus, the standard deviation is
0.3E(X). Since the expected cash flows serve as numeraire, the volatility is deter-
mined solely by the coefficient of variation. In our base case, we set the managerial
wage to equal 50.12 The expected cash flows as viewed by an overconfident manager
with an overconfidence measure of l are E(X) ¼ 45,000 + 500lY (i.e., m0 ¼ 45,000).
The risk aversion and effort aversion parameters are g ¼ 4 and b ¼ 3, respectively.
9
See Appendix 1 for more details.
10
When the manager is overconfident, this expected utility is calculated according to the man-
ager’s expectations.
11
See Appendix 1 for the explanation for the calibration of our model. To be on the safe side and in
stride with explanations for the risk premium puzzle, we use higher values for the risk aversion
parameter.
12
It should be noted that although the wage level in our base case equals half of the fixed
compensation that corresponds to the utility target, it equals only about 11 % of the expected
compensation under the optimal contract when managers are realistic. When managers are
overconfident, a wage of 50 consists of less than 11 % of total compensation according to the
manager’s expectations but more than 11 % according to the realistic expectations.
1498 O. Palmon and I. Venezia
In Table 55.1 we present the impact of overconfidence on the strike price that
stockholders choose to offer: the moneyness, the percentage of the firm given as
options, the effort choice of managers, the stockholders’ equity value, and the
expected managerial compensation. The expected compensation is calculated
both under the realistic distribution and under the subjective distribution of the
manager.
One observes from Table 55.1 that the strike price, the options’ moneyness, the
optimal managerial effort, the value of the stockholders’ equity, and the expected
compensation according to the managers’ expectations are directly related to
overconfidence. The optimal strike price (in thousands of dollars; strike prices
will be denoted in thousands of dollars in the rest of the study) for a rational
manager is 40.71, with a 0.60 moneyness (which can be described as deep-in-the-
money), but it rises to 63.74 with a 0.89 moneyness (closer to at-the-money) when
l ¼ 2.5.14 Managers also work harder the more overconfident they are
(Y increases from around 47 when they are realistic to around 52 when
l ¼ 2.5). Consequently, in order to hold the managers’ expected utility fixed,
their subjective expected monetary compensation must increase with
overconfidence to compensate for the extra risk resulting from the higher strike
price and for the additional effort they exert. The expected compensation the
stockholders perceive they pay according to the realistic expectation, however, is
inversely related to the overconfidence measure as they take advantage of
managers’ unrealistic expectations. The SEVs of the optimal contracts increase
as managerial overconfidence increases (see column 5, the SEV rises from 68,099
when l ¼ 1 to 70,958 when l ¼ 2.5, an increase of about 6 %). That is, the
stockholders benefit from the managers overestimating their powers.
This analysis suggests that stockholders are able to induce overconfident man-
agers to exert higher effort levels even though the objective contract parameters
they offer them are less favorable (the managers work harder but receive lower
13
Because of scaling there is no need to conduct robustness checks for the expected cash flows.
14
The moneyness measure depends on the strike price and the value of equity, which in turn
depends on effort. Thus, the moneyness measure varies with overconfidence because effort varies
with overconfidence, even when the strike price remains constant.
55
Table 55.1 Stockholders’ equity values (SEV), strike prices (K), moneyness (K/S), effort levels (Y), and other parameters of interest. Base case:
g ¼ 4 b ¼ 3, s ¼ 0.3, wage ¼ 50, N ¼ 4,000
Expected Sum of SEV and
Percentage of Stockholders compensation, Expected realistic
Strike price Moneyness firm given as Equity Value realistic compensation, expected
Overconfidence (K) in 1000s (K/S) options, (s) Effort (Y) (SEV) valuation managers valuation compensation
measure, l 1 2 3 4 5 6 7 8
1 40.71 0.5944 1.40% 47.083 68099.42 441.89 441.89 68541.31
1.5 48.16 0.6891 1.35% 49.867 69580.04 353.41 513.34 69933.45
Strike Prices of Options for Overconfident Executives
expected compensation). In particular the optimal strike prices of the options that
stockholders award overconfident managers are increasing with their
overconfidence. While realistic managers estimate that there is a substantial prob-
ability that options with an at-the-money strike price will be worthless regardless of
their effort, overconfident managers may believe that their efforts will enhance the
values of such options making them valuable.
In practice executive options usually are provided with at-the-money strike
prices. When overconfidence or some other behavioral biases are not present,
theory has shown (see, e.g., Dittmann et al. 2010; Dittmann and Yu 2011; Palmon
et al. 2008), contrary to Hall and Murphy, that at-the-money prices are not optimal.
Hall and Murphy argue that at-the-money prices are optimal because they provide
maximum sensitivity to stock prices, but their argument does not hold when the
managers are risk averse and effort averse. Managers must be adequately compen-
sated for their efforts and for risk taking, and a balance must be reached between
their efforts, risk taking, and their pay. As Palmon et al. have shown, the optimal
balance is reached by issuing in-the-money options which do not necessarily
provide maximum sensitivity to stock prices. If managers are overconfident, how-
ever, that makes them more amenable for stock price sensitivity, and hence, they
will prefer higher strike prices which are closer to the at-the-money options usually
awarded in practice.
We also note in Table 55.1 that managerial overconfidence increases stock-
holders’ equity value. Given that the compensation is determined so as to equate
the manager’s expected utility to the target expected utility, it follows from
Table 55.1 that consistent with the results of Palmon and Venezia (2012), the
total welfare of both the managers and the stockholders improves with increased
managerial overconfidence. The fixed-level expected utility of the manager is
determined according to their subjective, overoptimistic perception. However,
when evaluated according to the realistic view, expected managerial compensa-
tion falls with overconfidence. We note that, nonetheless, the difference between
the monetary expected compensations according to the overoptimistic and real-
istic expectations is smaller than the monetary gains to stockholders from
overconfidence, so that the sum of realistic compensation and SEV rises with
overconfidence (see column 8). Thus, also in terms of realistic monetary values,
the welfare of the stakeholders (stockholders and managers) increases with
overconfidence.
In Table 55.2 we provide sensitivity analysis examining the effect of each of the
parameters on the behavior of stockholders and managers. We present the results of
only one or two changes in each of the exogenous parameters, but we conduct many
other simulations, and all provide the same qualitative results.15 In all the panels,
higher overconfidence measure is associated with higher strike prices, moneyness
levels, optimal managerial effort, value of the stockholders’ equity, and expected
compensation according to the managers’ expectations. They also are associated
15
The results of these simulations can be obtained from the authors upon request.
55
Table 55.2 Stockholders’ equity values (SEV), strike prices (K), moneyness (K/S), effort levels (Y), and other parameters of interest. Departures from the
base case
Expected
compensation, Sum of SEV
Percentage of Stockholders’ realistic Expected and realistic
Strike price Moneyness firm given as Equity Value, stockholders’ compensation, expected
Overconfidence (K) in 1000s (K/S) options, (s) Effort (Y) SEV valuation managers’ valuation compensation
measure, l 1 2 3 4 5 6 7 8
Panel A: wage ¼ 25
1 31.56 0.4490 1.25 % 50.63 69,807.58 508.76 508.76 70,316.34
1.5 37.49 0.5252 1.14 % 52.83 71,000.72 413.11 563.28 71,413.83
2 43.57 0.6053 1.08 % 54.01 71,670.56 334.41 622.54 72,004.97
2.5 49.74 0.6875 1.04 % 54.75 72,106.28 269.37 688.74 72,375.65
Panel B: wage ¼ 75
1 51.70 0.7859 1.63 % 41.71 65,518.91 336.51 336.51 65,855.42
1.5 60.61 0.8965 1.70 % 45.36 67,411.64 269.97 420.45 67,681.61
2 69.91 1.0186 1.82 % 47.42 68,498.86 211.05 522.88 68,709.91
2.5 79.42 1.1462 2.01 % 48.72 69,195.89 166.20 653.44 69,362.09
Strike Prices of Options for Overconfident Executives
Panel C: s ¼ 0.2
1 48.97 0.6935 1.80 % 51.33 70,225.15 441.49 441.49 70,666.64
1.5 58.39 0.8138 1.83 % 53.60 71,489.22 312.64 542.64 71,801.86
2 68.10 0.9412 2.02 % 54.82 72,198.16 209.92 692.03 72,408.08
2.5 77.94 1.0716 2.35 % 55.56 72,642.57 137.08 912.80 72,779.65
Panel D: s ¼ 0.4
1 34.02 0.5090 1.19 % 43.77 66,438.75 444.93 444.93 66,883.68
1.5 40.00 0.5850 1.12 % 46.86 68,050.65 377.20 503.87 68,427.85
2 46.22 0.6675 1.09 % 48.59 68,975.32 317.47 566.82 69,292.79
2.5 52.58 0.7534 1.07 % 49.69 69,577.77 266.95 635.35 69,844.72
1501
(continued)
Table 55.2 (continued)
1502
Expected
compensation, Sum of SEV
Percentage of Stockholders’ realistic Expected and realistic
Strike price Moneyness firm given as Equity Value, stockholders’ compensation, expected
Overconfidence (K) in 1000s (K/S) options, (s) Effort (Y) SEV valuation managers’ valuation compensation
measure, l 1 2 3 4 5 6 7 8
Panel E: g ¼ 3
1 45.70 0.6801 2.09 % 44.49 66,732.83 513.41 513.41 67,246.24
1.5 53.94 0.7839 2.14 % 47.71 68,445.70 411.53 639.26 68,857.23
2 62.52 0.8969 2.28 % 49.52 69,438.59 319.32 795.50 69,757.91
2.5 71.28 1.0142 2.49 % 50.66 70,086.17 243.48 992.82 70,329.65
Panel F: g ¼ 5
1 36.91 0.5325 1.08 % 48.73 68,964.71 400.77 400.77 69,365.48
1.5 43.74 0.6198 1.01 % 51.24 70,295.63 323.30 449.43 70,618.93
2 50.78 0.7127 0.97 % 52.59 71,038.19 258.80 504.00 71,296.99
2.5 57.95 0.8085 0.96 % 53.44 71,515.08 206.66 567.94 71,721.74
Panel G: Ν ¼ 2,000
1 44.54 0.5886 1.36 % 61.44 75,242.70 477.96 477.96 75,720.66
1.5 54.10 0.7013 1.31 % 64.39 76,830.62 364.92 565.18 77,195.55
2 63.94 0.8204 1.32 % 65.98 77,714.30 274.85 672.50 77,989.15
2.5 73.93 0.9426 1.37 % 66.97 78,277.56 205.29 804.84 78,482.85
Panel H: Ν ¼ 6,000
1 38.87 0.5977 1.41 % 40.15 64,656.11 421.35 421.35 65,077.47
1.5 45.30 0.6826 1.36 % 42.83 66,067.86 346.87 486.08 66,414.74
2 51.97 0.7745 1.35 % 44.31 66,875.94 279.15 557.17 67,155.09
2.5 58.80 0.8701 1.38 % 45.25 67,401.85 223.48 643.22 67,625.33
O. Palmon and I. Venezia
55 Strike Prices of Options for Overconfident Executives 1503
55.5 Conclusion
Our study suggests an explanation for the puzzling questions of why most incentive
stock options are issued with at-the-money strike prices. This practice seems
arbitrary and beyond its institutional appeal and its expired tax advantages; its
main theoretical backing is that it provides the highest sensitivity to stock price.
Several studies however have shown that in many cases it is inferior to awarding
in-the-money options. Our analysis demonstrates that the optimal strike prices of
incentive stock options when managers are overconfident are higher than the
corresponding strike prices when managers are realistic, and are closer to the
at-the-money strike prices awarded in practice. This makes at-the-money options
more attractive to overconfident managers, and hence, given the ubiquity of
overconfident managers, it provides support for the popularity of awarding such
options. We also show that overoptimistic CEOs receive lower compensation than
their realistic counterparts and that the stockholders benefit from their managers’
bias. The combined welfare of the firm’s stakeholders however is positively related
to managerial overconfidence, hence providing support to the survival of manage-
rial overconfidence.
Assef and Santos (2005) interpret the strike price as an intermediate instrument
(between wages and stocks) in the incentive schemes for managers. Similarly one
1504 O. Palmon and I. Venezia
Appendix 1
16
We found additional estimates of effort disutility (leisure utility) in the following papers: Dowell
(1985), Kiker and Mendes de Oliveira (1990), and Prasch (2001). These estimates varied in the
functional form as well as in the level of effort aversion.
55 Strike Prices of Options for Overconfident Executives 1505
extent the effects of changes in the parameter values on the outcomes coincide with
economic intuition.
The parameter m0 serves as a numeraire for the other cash flows related param-
eters, and is chosen, without loss of generality, to equal 45,000. That, in the absence
of managerial extra effort, the expected value of the company’s cash flows is
45,000. Since the expected cash flows serve as numeraire, the ratio of the standard
deviation of the cash flows per share to their expected value is a surrogate for the
standard deviation of stock returns. Since Hall and Murphy (2000) used a standard
deviation of 0.3, we chose this value also for our base case coefficient of variation.
The appropriate measure of risk aversion is harder to agree upon. Early estimates
of risk aversion put this variable at around two (see, e.g., Mehra and Prescott 1985),
but they are based on aggregate data and not on CEO compensation data.17 In our
study, in line with more advanced econometric methods (see, e.g., Campbell
et al. 1996), we prefer using a base case risk aversion measure of four, slightly
higher than the measure of three suggested by Malmendier and Tate (2008) and
Hall and Liebman (1998). Our simulations (see, e.g., Glasserman 2003) and
sensitivity analysis, of course, cover these parameters as well.
The next step in the simulation process is to identify, for each overconfidence
level, the executive options’ strike price that is optimal for stockholders. All the
simulations were conducted using Mathematica. Because it is not possible to
express the equity value as an explicit function of the strike price, we search for
the optimal strike price by calculating the equity values that are associated with
a set of discrete strike prices. Our search was facilitated by assuming that the
stockholders know the manager’s reservation expected utility. We assume that
reservation utility to equal the utility obtained from a fixed salary of 100 with no
extra effort.
For any given wage, the strike price and the fraction of the company awarded to
the manager (which is a continuous variable representing the number of options
the manager receives; we will henceforth use the latter expression) determine the
value of the options to the managers and their cost to the stockholders. For each
strike price and number of options, we then find the effort that the manager chooses
to apply in order to maximize his/her expected utility Eq. 55.3. For each given
strike price, the stockholders, well aware of the managers’ reactions, will offer them
the number of options that yield their reservation utility. We calculate the value of
the stockholders’ equity for each strike price (in thousands of dollars, using two
digits beyond the decimal point) and identify the strike price that yields a maximum
for stockholders’ equity.
For each set of parameters for the cash flow distribution function and the
managerial utility function, as well as for the several values of fixed salary
(50 for the base case, 25 and 75 for the presented robustness simulations), we
obtain the optimal effort, stockholders’ equity value, and the expected managerial
17
Similar estimates are provided in other contexts by Carpenter (2000), Constantinides
et al. (2002), Epstein and Zin (1991), Friend and Blume (1975), and Levy (1994).
1506 O. Palmon and I. Venezia
Acknowledgment We thank Darius Palia, Orly Sade, and seminar participants at Rutgers
University and the Universitat Pompeu Fabra for helpful comments and suggestions. The financial
support of The Sanger Family Chair for Banking and Risk Management, The Galanter Fund, The
Mordecai Zagagi Fund, the Whitcomb Center for Research in Financial Services, and The School
of Accounting, the Hebrew University are gratefully acknowledged.
References
Assef, J. G., & Santos, M. (2005). Stock options and managerial optimal contracts. Economic
Theory, 26, 813–837.
Bergman, N. K., & Jenter, D. (2007). Employee sentiment and stock option compensation. Journal
of Financial Economics, 84, 667–712.
Bitler, M. P., Moskowitz, T. J., & Vissing-Jorgensen, A. (2005). Testing agency theory with
entrepreneur extra-effort and wealth. Journal of Finance, 60(2), 539–576.
Campbell, J., Lo, A., & MacKinlay, C. (1996). The econometrics of financial markets. New Jersey,
Princeton: Princeton University Press.
Carpenter, J. N. (2000). Does option compensation increase managerial risk appetite. Journal of
Finance, 55, 2311–2331.
Constantinides, G. M., Donaldson, J. M., & Mehra, R. (2002). Junior can’t borrow: A new
perspective on the equity premium puzzle. Quarterly Journal of Economics, 117, 269–296.
Dittmann, I., & Yu, K. C. (2011). How important are risk-taking incentives in executive compen-
sation? (Working Paper). Rotterdam: Erasmus University.
Dittmann, I., Maug, E., & Spalt, O. (2010). Sticks or carrots: Optimal CEO compensation when
managers are loss-averse. Journal of Finance, 65, 2015–2050.
Dowell, R. (1985). Risk preference and the work-leisure trade-off. Economic-Inquiry, 23,
691–701.
Epstein, L. G., & Zin, S. E. (1991). Risk aversion, and the temporal behavior of consumption and
asset returns: An empirical analysis. Journal of Political Economy, 99, 263–86.
Friend, I., & Blume, M. (1975). The demand for risky assets. American Economic Review, 65,
900–927.
Gervais, S., Heaton, J. B., & Odean, T. (2011). Overconfidence, compensation contracts, and
capital budgeting. Journal of Finance, 66, 1735–1777.
Glaser, M., & Weber, M. (2007). Overconfidence and trading volume. The Geneva Risk and
Insurance Review, 32, 1–36.
Glasserman, P. (2003). Monte Carlo methods in financial engineering. Berlin: Springer.
Hall, B., & Liebman, J. B. (1998). Are CEOs really paid like bureaucrats? Quarterly. Journal of
Economics, 113, 653–691.
Hall, B., & Murphy, K. J. (2000). Optimal exercise prices for executive stock options. American
Economic Review, 90, 209–14.
55 Strike Prices of Options for Overconfident Executives 1507
Hall, B. J., & Murphy, K. J. (2002). Stock options for undiversified executives. Journal of
Accounting and Economics, 33, 3–42.
Hek, P. A. (1999). Endogenous technological change with leisure-dependent utility. Economic
Theory, 14, 669–84.
Holmstrom, B. (1979). Moral hazard and observability. Bell Journal of Economics, 10, 74–91.
Kiker, B. F., & Mendes-de-Oliveira, M. (1990). Estimation and valuation of non-leisure time.
Oxford Bulletin of Economics and Statistics, 52, 115–141.
Levy, H. (1994). Absolute and relative risk aversion: An experimental study. Journal of Risk and
Uncertainty, 8, 289–307.
Mahajan, S. (2002). Role of out-of-money options in executive compensation (Working Paper).
New York: NYU, Stern School of Business.
Malmendier, U., & Tate, G. (2005a). Does overconfidence affect corporate investment? CEO
overconfidence measures revisited. European Financial Management, 11, 649–659.
Malmendier, U., & Tate, G. (2005b). CEO overconfidence and corporate investment. Journal of
Finance, 60, 2661–2700.
Malmendier, U., & Tate, G. (2008). Who makes acquisitions? CEO overconfidence and the
market’s reaction. Journal of Financial Economics, 89, 20–43.
Mehra, R., & Prescott, E. C. (1985). The equity premium: A puzzle. Journal of Monetary
Economics, 15, 145–161.
Otto, C. A. (2011). CEO optimism and incentive compensation (Working Paper). London:
Business School.
Oyer, P., & Schaefer, S. (2005). Why do some firms give stock options to all employees? An
empirical examinations of alternative theories. Journal of Financial Economics, 76, 99–133.
Palmon, O., & Venezia, I. (2012, forthcoming). A rationale for hiring irrationally overconfident
managers. In Encyclopedia of finance (2nd ed.). Berlin: Springer.
Palmon, O., Chen, R. R., Bar Yosef, S., & Venezia, I. (2008). Optimal strike prices of stock options
for effort averse executives. Journal of Banking and Finance, 32, 229–239.
Prasch, R. E. (2001). Work, leisure, and the labor supply curve: A reply to David Spencer. Journal
of Economic Issues, 35, 1001–1007.
Roll, R. (1986). The hubris hypothesis of corporate takeovers. Journal of Business, 59, 197–216.
Suntheim, F. (2012). CEO overconfidence in banking (Working Paper). Milano: Bocconi
University.
Density and Conditional
Distribution-Based Specification Analysis 56
Diep Duong and Norman R. Swanson
Contents
56.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1510
56.2 Setup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1513
56.2.1 Diffusion Models in Finance and Financial Econometrics . . . . . . . . . . . . . . . . . . 1513
56.2.2 Overview on Specification Tests and Model Selection . . . . . . . . . . . . . . . . . . . . . . 1519
56.3 Consistent Distribution-Based Specification Tests and Predictive
Density-Type Model Selection for Diffusion Processes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1521
56.3.1 One-Factor Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1521
56.3.2 Multifactor Models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1532
56.3.3 Model Simulation and Estimation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1537
56.3.4 Bootstrap Critical Value Procedures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1547
56.4 Summary of Empirical Applications of the Tests . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1553
56.5 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1558
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1558
Abstract
The technique of using densities and conditional distributions to carry out
consistent specification testing and model selection amongst multiple diffusion
processes has received considerable attention from both financial theoreticians
and empirical econometricians over the last two decades. In this chapter, we
discuss advances to this literature introduced by Corradi and Swanson
(J Econom 124:117–148, 2005), who compare the cumulative distribution
(marginal or joint) implied by a hypothesized null model with corresponding
empirical distributions of observed data. We also outline and expand upon
D. Duong (*)
Department of Business and Economics, Utica College, Utica, NY, USA
e-mail: dnduong@utica.edu
N.R. Swanson
Department of Economics, Rutgers, The State University of New Jersey, New Brunswick, NJ, USA
e-mail: nswanson@econ.rutgers.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1509
DOI 10.1007/978-1-4614-7750-1_56,
# Springer Science+Business Media New York 2015
1510 D. Duong and N.R. Swanson
further testing results from Bhardwaj et al. (J Bus Econ Stat 26:176–193, 2008)
and Corradi and Swanson (J Econom 161:304–324, 2011). In particular, para-
metric specification tests in the spirit of the conditional Kolmogorov test of
Andrews (Econometrica 65:1097–1128, 1997) that rely on block bootstrap
resampling methods in order to construct test critical values are first discussed.
Thereafter, extensions due to Bhardwaj et al. (J Bus Econ Stat 26:176–193,
2008) for cases where the functional form of the conditional density is unknown
are introduced, and related continuous time simulation methods are introduced.
Finally, we broaden our discussion from single process specification testing to
multiple process model selection by discussing how to construct predictive
densities and how to compare the accuracy of predictive densities derived
from alternative (possibly misspecified) diffusion models. In particular, we
generalize simulation steps outlined in Cai and Swanson (J Empir Financ
18:743–764, 2011) to multifactor models where the number of latent variables
is larger than three. We finish the chapter with an empirical illustration of model
selection amongst alternative short-term interest rate models.
Keywords
Multifactor diffusion process • Specification test • Out-of-sample forecasts •
Conditional distribution • Model selection • Block bootstrap • Jump process
56.1 Introduction
The last three decades have provided a unique opportunity to observe numerous
interesting developments in finance, financial econometrics, and statistics. For
example, although starting as a narrow subfield, financial econometrics has recently
transformed itself into an important discipline, equipping financial economic
researchers and industry practitioners with immensely helpful tools for estimation,
testing, and forecasting. One of these developments has involved the development
of “state-of-the-art” consistent specification tests for continuous time models,
including not only the geometric Brownian motion process used to describe the
dynamics of asset returns (Merton (1973)) but also a myriad of other diffusion
models used in finance, such as the Ornstein-Uhlenbeck process introduced by
Vasicek (1977); the constant elastic volatility process applied by Beckers (1980);
the square root process due to Cox et al. (1985); the so-called CKLS model by Chan
et al. (1992); various three-factor models proposed Chen (1996); stochastic vola-
tility processes such as generalized CIR of Andersen and Lund (1997); and the
generic class of affine jump diffusion processes discussed in Duffle et al. (2000).1
The plethora of available diffusion models allow decision makers to be flexible
when choosing a specification to be subsequently used in contexts ranging from
equity and option pricing, to term structure modeling and risk management.
1
For complete details, see Sect. 56.2.2.
56 Density and Conditional Distribution-Based Specification Analysis 1511
Moreover, the use of high-frequency data when estimating such model, in contin-
uous time contexts, allows investors to continuously update their dynamic trading
strategies in real time.2 However, for statisticians and econometricians, the vast
number of available models has important implications for formalizing model
selection and specification testing methods. This has led to several key papers
that have recently been published in the area of parametric and nonparametric
specification testing. Most of the papers focus on the ongoing “search” for correct
Markov and stationary models that “fit” historical data and associated dynamics. In
this literature, it is important to note that correct specification of a joint distribution
is not the same as that of a conditional distribution, and hence the recent focus on
conditional distributions, given that most models have an interpretation as condi-
tional models. In summary, the key issue in the construction of model selection and
specification tests of conditional distributions is the fact that knowledge of the
transition density (or conditional distribution) in general cannot be inferred from
knowledge of the drift and variance terms of a diffusion model. If the functional
form of the density is available parametrically, though, one can test the hypothesis
of correct specification of a diffusion via the probability integral transform
approach of Diebold et al. (1998); the cross-spectrum approach of Hong (2001),
Hong and Li (2005), and Hong et al. (2007); the martingalization-type Kolmogorov
test of Bai (2003); or the normality transformation approaches of Bontemps and
Meddahi (2005) and Duan (2003). Furthermore, if the transition density is
unknown, one can construct a nonparametric test by comparing a kernel density
estimator of the actual and simulated data, for example, as in Altissimo and Mele
(2009) and Thompson (2008), or by comparing the conditional distribution of the
simulated and the historical data, as in Bhardwaj et al. (2008). One can also use the
methods of Aı̈t-Sahalia (2002) and Aı̈t-Sahalia et al. (2009), in which they compare
closed form approximations of conditional densities under the null, using data-
driven kernel density estimates.
For clarity and ease of presentation, we categorize the above literature into two
areas. The first area, initiated by the seminal work of Aı̈t-Sahalia (1996) and later
followed by Pritsker (1998) and Jiang (1998), breaks new ground in the continuous
time specification testing literature by comparing marginal densities implied by
hypothesized null models with nonparametric estimates thereof. These sorts of
tests examine one-factor specifications. The second area of testing, as initiated in
Corradi and Swanson (2005), does not look at densities. Instead, they compare
cumulative distributions (marginal, joint, or conditional) implied by a hypothesized
null model with corresponding empirical distributions. A natural extension of these
sorts of tests involves model selection amongst alternative predictive densities
associated with competing models. While Corradi and Swanson (2005) focus on
cases where the functional form of the conditional density is known, Bhardwaj
et al. (2008) use simulation methods to examine testing in cases where the func-
tional form of the conditional density is unknown. Corradi and Swanson (2011) and
2
For further discussion, see Duong and Swanson (2010, 2011).
1512 D. Duong and N.R. Swanson
Cai and Swanson (2011) take the analysis of Bhardwaj et al. (2008) on Step further
and focus on the comparison of out-of-sample predictive accuracy of possibly
misspecified diffusion models, when the conditional distribution is not known in
closed form (i.e., they “choose” amongst competing models based on predictive
density model performance). The “best” model is selected by constructing tests that
compare both predictive densities and predictive conditional confidence intervals
associated with alternative models.
In this chapter, we primarily focus our attention on the second area of the model
selection and testing literature.3 One feature of all the tests that we shall discuss is
that, given that they are based on the comparison of CDFs, they obtain parametric
rates. Moreover, the tests can be used to evaluate single and multiple factor and
dimensional models, regardless of whether or not the functional form of the
conditional distribution is known.
In addition to discussing simple diffusion process specification tests of Corradi
and Swanson (2005), we discuss tests discussed in Bhardwaj et al. (2008) and
Corradi and Swanson (2011) and provide some generalizations and additional
results. In particular, parametric specification tests in the spirit of the conditional
Kolmogorov test of Andrews (1997) that rely on block bootstrap resampling
methods in order to construct test critical values are first discussed. Thereafter,
extensions due to Bhardwaj et al. (2008) for cases where the functional form of the
conditional density is unknown are introduced, and related continuous time simu-
lation methods are introduced. Finally, we broaden our discussion from single
dimensional specification testing to multiple dimensional selection by discussing
how to construct predictive densities and how to compare the accuracy of predictive
densities derived from alternative (possibly misspecified) diffusion models as in
Corradi and Swanson (2011). In addition, we generalize simulation and testing
procedures introduced in Cai and Swanson (2011) to more complicated multifactor
and multidimensional models where the number of latent variables is larger than
three. These final tests can be thought of as continuous time generalizations of the
discrete time “reality check” test statistics of White (2000), which are widely used
in empirical finance (see, e.g., Sullivan et al. (1999, 2001)). We finish the chapter
with an empirical illustration of model selection amongst alternative short-term
interest rate models, drawing on Bhardwaj et al. (2008), Corradi and Swanson
(2011) and Cai and Swanson (2011).
Of the final note is that the test statistics discussed here are implemented via use
of simple bootstrap methods for critical value simulation. We use the bootstrap
because the covariance kernels of the (Gaussian) asymptotic limiting distributions
of the test statistics are shown to contain terms deriving from both the contribution
of recursive parameter estimation error (PEE) and the time dependence of data.
Asymptotic critical value thus cannot be tabulated in a usual way. Several methods
can easily be implemented in this context. First one can use block bootstrapping
procedures, as discussed below. Second one can use the conditional p-value
3
For a recent survey on results in the first area of this literature, see Aı̈t-Sahalia (2007).
56 Density and Conditional Distribution-Based Specification Analysis 1513
approach of Corradi and Swanson (2002) which extends the work of Hansen (1996)
and Inoue (2001) to the case of nonvanishing parameter estimation error. Third is
the subsampling method of Politis et al. (1999), which has clear efficiency “costs,”
but is easy to implement. Use of the latter two methods yields simulated
(or subsample based) critical values that diverge at rate equivalent to the block
size length under the alternative. This is the main drawback to their use in our
context. We therefore focus on use of a block bootstrap that mimics the contribution
of parameter estimation error in a recursive setting and in the context of time series
data. In general, use of the block bootstrap approach is made feasible by
establishing consistency and asymptotic normality of both simulated generalized
method of moments (SGMM) and nonparametric simulated quasi-maximum like-
lihood (NPSQML) estimators of (possibly misspecified) diffusion models, in
a recursive setting, and by establishing the first-order validity of their bootstrap
analogs.
The rest of the paper is organized as follows. In Sect. 56.2, we present our
setup and discuss various diffusion models used in finance and financial economet-
rics. Section 56.3 outlines the specification testing hypotheses, presents the
cumulative distribution-based test statistics for one-factor and multiple-
factor models, discusses relevant procedures for simulation and estimation, and
outlines bootstrap techniques that can be used for critical value tabulation.
In Sect. 56.4, we present a small empirical illustration. Section 56.5 summarizes
and concludes.
56.2 Setup
For the past two decades, continuous time models have taken center stage in the
field of financial econometrics, particularly in the context of structural modeling,
option pricing, risk management, and volatility forecasting. One key advantage of
continuous time models is that they allow financial econometricians to use the full
information set that is available. With the availability of high-frequency data and
current computation capability, one can update information, model estimates, and
predictions in milliseconds. In this section, we will summarize some of the standard
models that have been used in asset pricing as well as term structure modeling.
Generally, assume that financial asset returns follow Ito-semimartingale processes
with jumps, which are the solution to the following stochastic differential equation
system:
ðt ð
Xðt Þ ¼ bðXðs Þ, y0 Þds l0 t yfðyÞdy
0 Y
ðt X
Jt (56:1)
þ sðXðs Þ, y0 ÞdW ðsÞ þ yj ,
0 j¼1
1514 D. Duong and N.R. Swanson
where X(t) is a cadlag process (right continuous with left limit) for t 2 ℜ+ and is an
N-dimensional vector of variables, W(t) is an N-dimensional Brownian motion, b(·)
is N-dimensional function of X(t), and s(·) is an N N matrix-valued function of
X(t), where y0 is an unknown true parameter. Jt is a Poisson process with intensity
parameter l0, l0 finite, and the N-dimensional jump size, yj, is i.i.d. with marginal
distribution given by f. Both Jt and yj are assumed
Ð to be independent of the driving
Brownian motion, W(t).4 Also, note that Yyf(y)dy denotes the mean jump size,
hereafter denoted by m0. Over a unit time interval, there are on average l0 jumps, so
that over the time span [0, t], there are on average l0t jumps. The dynamics of X(t)
is then given by
dXðtÞ ¼ bðXðt Þ, y0 Þ l0 my, 0 dt
ð
(56:2)
þ sðXðt Þ, y0 ÞdW ðtÞ þ ypðdy; dtÞ,
Y
where p(dy, dt) is a random Poisson measure giving point mass at y if a jump occurs
in the interval dt and b(·), s(·) are the “drift” and “volatility” functions defining the
parametric specification of the model. Hereafter, the same (or similar) notation is
used throughout when models are specified.
Through not an exhaustive list, we review some popular models. Models are
presented with the “true” parameters.
Hereafter, X(t) denotes the cadlag, while Xt denotes discrete skeleton for t ¼ 1, 2, . . . .
4
5
See Black and Scholes (1973) for details.
56 Density and Conditional Distribution-Based Specification Analysis 1515
where W(t) is a standard Brownian motion and a0, b0, and s0 are constants. b0 is
negative to ensure the mean reversion of X(t).
Cox et al. (1985) use the following square root process to model the term
structure of interest rates:
pffiffiffiffiffiffiffiffiffi
dXðtÞ ¼ kða0 XðtÞÞdt þ s0 XðtÞdW ðtÞ,
belongs to the linear exponential (or Pearson) family with a closed form cumulative
distribution. a0 and l0 are fixed parameters of the model.
The constant elasticity of variance or CEV model is specified as follows:
where W(t) is a standard Brownian motion and a0, s0, and b0 are fixed constants.
Of note is that the interpretation of this model depends on b0, i.e., in the case of
stock prices, if b0 ¼ 2, then the price process X(t) follows a lognormal diffusion; if
b0 < 2, then the model captures exactly the leverage effect as price and volatility
are inversely correlated.
Amongst other authors, Beckers (1980) used this CEV model for stocks, Marsh
and Rosenfeld (1983) apply a CEV parametrization to interest rates, and Emanuel
and Macbeth (1982) utilize this setup for option pricing.
The generalized constant elasticity of variance model is defined as follows:
dXðtÞ ¼ a0 XðtÞð1b0 Þ þ l0 XðtÞ dt þ s0 XðtÞb0 =2 dW ðtÞ,
where the notation follows the CEV case. l0 is another parameter of the model. This
process nests log diffusion when b0 ¼ 2 and nests square root diffusion when
b0 ¼ 1.
Brennan and Schwartz (1979) and Courtadon (1982) analyze the model:
where a0, b0, s0 are fixed constants and W(t) is a standard Brownian motion.
1516 D. Duong and N.R. Swanson
where W(t) is a standard Brownian motion and a0, b0, and g0 are fixed parameters.
Aїt-Sahalia (1996) looks at a general case with general drift and CEV diffusion:
dXðtÞ ¼ a0 þ b0 XðtÞ þ g0 XðtÞ2 þ 0 =XðtÞ dt þ s0 XðtÞb0 =2 dW ðtÞ:
In the above expression, a0, b0, g0, 0, s0, and b0 are fixed constants and W(t) is
again a standard Brownian motion.
pffiffiffiffiffiffiffiffiffi
dXðtÞ ¼ k0 ða0 XðtÞÞdt þ O0 DðtÞdW ðtÞ þ dJ ðtÞ,
In the above expressions, y0i and d0i 0 are constants. The jump intensity is
assumed to be a positive, affine function of X(t), and the jump size distribution
is assumed to be determined by its conditional characteristic function. The attrac-
tive feature of this class of affine jump diffusions is that, as shown in Duffie
et al. (2000), it has an exponential affine structure that can be derived in closed
form, i.e.,
FðXðtÞÞ ¼ exp aðtÞ þ bðtÞ0 XðtÞ ,
where the functions a(t) and b(t) can be derived from Riccati equations.6
Given a known characteristic function, one can use either GMM to estimate the
6
For details, see Singleton (2006), p. 102.
56 Density and Conditional Distribution-Based Specification Analysis 1517
! ! !
dXðtÞ b1 ðXðtÞ, y0 Þ s11 ðV ðtÞ, y0 Þ
¼ dt þ dW 1 ðtÞ
dV ðtÞ b2 ðV ðtÞ, y0 Þ 0
! (56:4)
s12 ðV ðtÞ, y0 Þ
þ dW 2 ðtÞ,
s22 ðV ðtÞ, y0 Þ
where W1(t)11 and W2(t)11 are independent standard Brownian motions and V(t)
is latent volatility process. b1(·) is a function of X(t) and b2(·), s11(·), s22(·), and
s22(·) are general functions of V(t), such that system of Eq. 56.4 is well defined.
Popular specifications are the square root model of Heston (1993), the GARCH
diffusion model of Nelson (1990), lognormal model of Hull and White (1987), and
the eigenfunction models of Meddahi (2001). Note that in this stochastic volatility
case, the dimension of volatility is d ¼ 1. More general setup can involve d driving
Brownian motions in V(t) equation.
As an example, Andersen and Lund (1997) study the generalized CIR model
with stochastic volatility, specifically
pffiffiffiffiffiffiffiffiffi
dXðtÞ ¼ kx0 ðx0 XðtÞÞdt þ V ðtÞdW 1 ðtÞ,
pffiffiffiffiffiffiffiffiffi
dXðtÞ ¼ kv0 ðv0 V ðtÞÞdt þ sv0 V ðtÞdW 2 ðtÞ,
where X(t) and V(t) are price and volatility processes, respectively, kx0, kv0 > 0 to
ensure stationarity, x0 is the long-run mean of (log) price process, and v0 and sv0 are
constants. W1(t) and W2(t) are scalar Brownian motions. However, W1(t) and W2(t)
are correlated such that dW1(t)dW2(t) ¼ rdt where the correlation r is some
constant r 2 [1, 1]. Finally, note that volatility is a square root diffusion process,
which requires that kv0 v0 > s2v0 .
1518 D. Duong and N.R. Swanson
where qu and qd are Poisson processes with jump intensity parameters lu and ld,
respectively, and are independent of the Brownian motions W1(t) and W2(t).
In particular, lu is the probability of a jump-up, Pr(dqu (t) ¼ 1) ¼ lu, and ld is
the probability of a jump-down, Pr(dqd(t) ¼ 1) ¼ ld. Ju and Jd are jump-up
and jump-down sizes and have exponential distributions: f ðJ u Þ ¼ B1u exp JBuu
and f ðJ d Þ ¼ B1d exp JBdd , where Bu, Bd > 0 are the jump magnitudes, which are
the means of the jumps, Ju and Jd.
Three-Factor Model (CHEN): The three-factor model combines various features
of the above models, by considering a version of the oft examined three-factor
model due to Chan et al. (1992), which is discussed in detail in Dai and Singleton
(2000). In particular,
pffiffiffiffiffiffiffiffiffi
dXðtÞ ¼ kx0 ðyðtÞ XðtÞÞdt þ pVffiffiffiffiffiffiffiffi
ðtÞdW ffi 1 ðtÞ,
dV ðtÞ ¼ kv0 ðv V ðtÞÞdt þ sv0 V ð t Þ
pffiffiffiffiffiffiffiffidW 2 ðtÞ, (56:5)
dyðtÞ ¼ ky0 yðtÞ yðtÞ dt þ sy0 yðtÞdW 3 ðtÞ,
where W1(t), W2(t) W3(t) are independent Brownian motions and V and y are the
stochastic volatility and stochastic mean of X(t), respectively. kx0, kv0, ky0, v0 , y0 ,
sv0, sy0 are constants. As discussed above, non-negativity for V(t) and y(t) requires
that 2kv0 v0 > s2v0 and 2ky0 y0 > s2y0 .
Three-Factor Jump Diffusion Model (CHENJ): Andersen et al. (2004) extend the
three-factor Chen (1996) model by incorporating jumps in the short rate process,
hence improving the ability of the model to capture the effect of outliers and to
address the finding by Piazzesi (2004, 2005) that violent discontinuous movements
in underlying measures may arise from monetary policy regime changes. The
model is defined as follows:
pffiffiffiffiffiffiffiffiffi
dXðtÞ ¼ kx0 ðyðtÞ XðtÞÞdt þ V ðtÞdW 1 ðtÞ þ J u dqu J d dqd ,
pffiffiffiffiffiffiffiffiffi
dV ðtÞ ¼ kv0 v0 V ðtÞÞdt þ sv0 V ðtÞdW 2 ðtÞ, (56:6)
pffiffiffiffiffiffiffiffi
dyðtÞ ¼ ky0 y0 yðtÞÞdt þ sy0 yðtÞdW 3 ðtÞ
where all parameters are similar as in Eq. 56.5; W1(t), W2(t), and W3(t) are
independent Brownian motions; and qu and qd are Poisson processes
with jump intensities lu0 and ld0, respectively, and are independent of the
Brownian motions Wr(t), Wv(t), and Wy(t). In particular, lu0 is the
56 Density and Conditional Distribution-Based Specification Analysis 1519
probability of a jump-up, Pr(dqu (t) ¼ 1) ¼ lu0, and ld0 is the probability of a jump-
down, Pr(dqd(t) ¼ 1) ¼ ld0. Ju and Jd are jump-up
and jump-down sizes and have
exponential distributions f ðJ u Þ ¼ B1u0 exp BJu0u and f ðJ d Þ ¼ B1d0 exp BJd0d , where
Bu0, Bd0 > 0 are the jump magnitudes, which are the means of the jumps Ju and Jd.
The focus in this chapter is specification testing and model selection. The “tools”
used in this literature have been long established. Several key classical contribu-
tions include the Kolmogorov-Smirnov test (see, e.g., Kolmogorov (1933) and
Smirnov (1939)), various results on empirical processes (see, e.g., Andrews
(1993) and the discussion in Chap. 19 of van der Vaart (1998) on the contributions
of Glivenko, Cantelli, Doob, Donsker, and others), the probability integral trans-
form (see, e.g., Rosenblatt (1952)), and the Kullback–Leibler information criterion
(see, e.g., White (1982) and Vuong (1989)). For illustration, the empirical distri-
bution mentioned above is crucial in our discussion of predictive densities because
it is useful in estimation, testing, and model evaluation. Let Yt is a variable of
interest with distribution F and parameter y0. The theory of empirical distributions
provides a result that
1 X T
pffiffiffi ð1fY t ug Fðujy0 ÞÞ
T t¼1
satisfies a central limit theorem (with a parametric rate) if T is large (i.e., asymp-
totically). In the above expression, 1{Yt u} is the indicator function which takes
value 1 if Yt u and 0 otherwise. In the case where there is parameter estimation
error, we can use more general results in Chap. 19 of van der Vaart (1998). Define
ð
1X T
PT ð f Þ ¼ f ðY i Þ and Pð f Þ ¼ fdP,
T i¼1
7
See Theorem 19.23 in van der Vaart (1998) for details.
1520 D. Duong and N.R. Swanson
of statistics that admit for dynamic misspecification under both hypotheses. This
differs from other classes of tests such as the framework used by Diebold
et al. (1998), Hong (2001), and Bai (2003) in which correction dynamic specifica-
tion under the null hypothesis is assumed. In particular, DGT use the probability
ð Yt
integral transform to show that Ft ðY t jℑt1 ; y0 Þ ¼ f t ðyjℑt1 ; y0 Þdy is identically
1
and independently distributed as a uniform random variable on [0; 1], where Ft(·)
and ft(·) are a parametric distribution and density with underlying parameter y0, Yt is
again our random variable of interest, and ℑt is the information set containing all
“relevant” past information. They
thus suggest using the difference between the
empirical distribution of Ft Y t ℑt1 ; y^t : and the 45 line as a measure of “goodness
of fit,” where y^t is some estimator of y0. This approach has been shown to be very
useful for financial risk management (see, e.g., Diebold et al. (1999)), as well as for
macroeconomic forecasting (see, e.g., Diebold et al. (1998) and Clements and
Smith (2000, 2002)). Similarly, Bai (2003) develops aKolmogorov-type
test of
^
Ft(Yt|ℑt1, y0) on the basis of the discrepancy between Ft Y t ℑt1 ; y t : and the CDF
of a uniform on [0; 1]. As the test involves estimator y^t , the limiting distribution
reflects the contribution of parameter estimation error and is not nuisance
parameter-free. To overcome this problem, Bai (2003) proposes a novel approach
based on a martingalization argument to construct a modified Kolmogorov test
which has a nuisance parameter-free limiting distribution. This test has power
against violations of uniformity but not against violations of independence. Hong
(2001) proposes another related interesting test, based on the generalized spectrum,
which has power against both uniformity and independence violations, for the case
in which the contribution of parameter estimation error vanishes asymptotically. If
the null is rejected, Hong (2001) also proposes a test for uniformity robust to
nonindependence, which is based on the comparison between a kernel density
estimator and the uniform density. Two features differentiate the tests surveyed in
this chapter from the tests outlined in the other papers mentioned above. First, the
tests discussed here assume strict stationarity. Second, they allow for dynamic
misspecification under the null hypothesis. The second feature allows us to obtain
asymptotically valid critical values even when the conditioning information set
does not contain all of the relevant past history. More precisely, assume that we are
interested in testing for correct specification, given a particular information set
which may or may not contain all of the relevant past information. This is important
when a Kolmogorov test is constructed, as one is generally faced with the problem
of defining ℑt1. If enough history is not included, then there may be dynamic
misspecification. Additionally, finding out how much information (e.g., how many
lags) to include may involve pre-testing, hence leading to a form of sequential test
bias. By allowing for dynamic misspecification, such pre-testing is not required.
Also note that critical values derived under correct specification given ℑt1 are not
in general valid in the case of correct specification given a subset of ℑt1. Consider
the following example. Assume that we are interested in testing whether
56 Density and Conditional Distribution-Based Specification Analysis 1521
In this section, we outline the setup for the general class of one-factor jump
diffusion specifications. All analyses carry through to the more complicated case
of multifactor stochastic volatility models which we will elaborate upon in the next
subsection. In the presentation of the tests, we follow a view that all candidate
models, either single or multiple dimensional ones, are approximations of reality
and can thus be misspecified. The issue of correct specification (or misspecification)
of a single model and the model selection test for choosing amongst multiple
competing models allow for this feature.
To begin, fix the time interval [0, T] and consider a given single one-factor
candidate model the same as Eq. 56.1, with the true parameters y0, l0, m0 to be
replaced by its pseudo true analogs y{, l, m, respectively, and 0 t T:
ðt ð ðt X
{
Jt
Xðt Þ ¼ b Xðs Þ, y ds lt yfðyÞdy þ s Xðs Þ, y{ dW ðsÞ þ yj ,
0 Y 0 j¼1
or
dXðtÞ ¼ b XðtÞ, y{ lm dt: ð
(56:7)
þ s XðtÞ, y{ dW ðtÞ þ ypðdy; dtÞ,
Y
where variables are defined the same as in Eqs. 56.1 and 56.2. Note that as the above
model is the one-factor version of Eqs. 56.1 and 56.2 where the dimension of X(t)
is 1 1, W(t) is a one-dimensional standard Brownian motion and jump size, and yj
1522 D. Duong and N.R. Swanson
is one-dimensional variable for all j. Also note both Jt and yj are assumed to be
independent of the driving Brownian motion.
If the single model is correctly specified, then b(X(t), y{) ¼ b0(X(t), y0),
s(X(t), y{) ¼ s0(X(t), y0), l ¼ l0, m ¼ m0, and f ¼ f0 where b0(X(t), y0),
s0(X(t), y0), l0, m0, f0 are unknown and belong to the true specification.
Now consider a different case (not a single model) where m candidate models are
involved. For model k with 1 k m, denote its corresponding specification to be
(bk(X(t), y{k ), sk(X(t), y{k ), lk, mk, fk). Two scenarios immediate arise. Firstly, if
the model k is correctly specified, then bk(X(t), y{k ) ¼ b0(X(t), y0), sk(X(t),
y{k ) ¼ s0(X(t), y0), lk ¼ l0, mk ¼ m0, and fk ¼ f0 which are similar to the case of
a single model. In the second scenario, all the models are likely to be misspecified
and modelers are faced with the choice of selecting the “best” one. This type of
problem is well fitted into the class of accuracy assessment tests initiated earlier by
Diebold and Mariano (1995) or White (2000).
The tests discussed hereafter are Kolmogorov-type tests based on the construc-
tion of cumulative distribution functions (CDFs). In a few cases, the CDF is known
in closed form. For instance, for the simplified version of the CIR model as in
Eq. 56.3, X(t) belongs to the linear exponential (or Pearson) family with the gamma
CDF of the form8
ð u 2ð1a=lÞ1
l
exp x= l2 dx
2
Fðu; a; lÞ ¼ 0 (56:8)
Gð2ð1 a=lÞÞ
Ð1
where G(x) ¼ 0 txexp(t)dt, and a, l are constants.
Furthermore, if we look at the pure diffusion process without jumps
dXðtÞ ¼ b XðtÞ, y{ dt þ s XðtÞ, y{ dW ðtÞ, (56:9)
where b(·) and s ¼ s(·) are drift and volatility functions, it is known that the
stationary density, say f(x, y{), associated with the invariant probability measure
can be expressed explicitly as9
ðx { !
{ c y{ 2b u; y
f x; y ¼ { exp du ,
s x; y
2 s2 u; y{
{ {
Ð u c(y{) is a constant ensuring that f integrates to one. The CDF, say F(u,y )
where
¼ f(x,y ) dx, can then be obtained using available numerical integration
procedures.
However, in most cases, it is impossible to derive the CDFs in closed form. To
obtain a CDF in such cases, a more general approach is to use simulation. Instead of
8
See Wong (1964) for details.
9
See Karlin and Taylor (1981) for details.
56 Density and Conditional Distribution-Based Specification Analysis 1523
estimating the CDF directly, simulation techniques estimate the CDF indirectly
utilizing its generated sample paths and the theory of empirical distributions. The
specification of a specific diffusion process will dictate the sample paths and
thereby corresponding test outcomes.
Note that in the historical context, many early papers in this literature are probability
density based. For example, in a seminal paper, Ait-Sahalia (1996) compares the
marginal densities implied by hypothesized null models with nonparametric estimates
thereof. Following the same framework of correct specification tests, Corradi and
Swanson (2005) and Bhardwaj et al. (2008), however, do not look at densities. Instead,
they compare the cumulative distribution (marginal or joint) implied by a hypothesized
null model with the corresponding empirical distribution. While Corradi and Swanson
(2005) focus on the known unconditional distribution, Bhardwaj et al. (2008) look at
the conditional simulated distributions. Corradi and Swanson (2011) make extensions
to multiple models in the context of out-of-sample accuracy assessment tests. This
approach is somewhat novel to this continuous time model testing literature.
Now suppose we observe a discrete sample path X1, X2, . . ., XT (also referred as
skeletons).10 The corresponding hypotheses can be set up as follows:
Hypothesis 1 Unconditional Distribution Specification Test of a Single Model
H0 : F(u,y{) ¼ F0(u,y0), for all u, a.s.
HA : Pr(F(u,y{) F0(u,y0) 6¼ 0) > 0, for some u 2 U, with nonzero Lebesgue
measure.
where F0(u, y0) is the true cumulativedistribution
implied by the above density,
{ y{
i.e., F0(u,y0) ¼ Pr(Xt u). F u; y ¼ Pr Xt u is the cumulative distribution of
{
the proposed model. Xyt is a skeleton implied by model (56.7).
Hypothesis 2 Conditional Distribution Specification Test of a Single Model
H0 : Ft(u|Xt, y{) ¼ F0,t(u|Xt, y0), for all u, a.s.
HA : Pr(Ft(u|Xt, y{) F0,t(u|Xt, y0) 6¼ 0) > 0, for some u 2 U, with nonzero
Lebesgue measure.
{ {
where Ft ujXt ; y{ ¼ Pr Xytþt u Xyt ¼ Xt is t-step ahead conditional
distributions and t ¼ 1, . . . , T t. F0,t(u|Xt, y0) is t-step ahead true conditional
distributions.
Hypothesis 3 Predictive Density Test for Choosing Amongst Multiple
Competing Models
The null hypothesis is that no model can outperform model 1 which is the
benchmark model.11
10
As mentioned earlier, we follow Corradi and Swanson (2005) by using notation X(·) when
defining continuous time processes and Xt for a skeleton.
11
See White (2000) for a discussion of a discrete time series analog to this case, whereby point
rather than density-based loss is considered; Corradi and Swanson (2007b) for an extension of
White (2000) that allows for parameter estimation error; and Corradi and Swanson (2006) for an
extension of Corradi and Swanson (2007b) that allows for the comparison of conditional distri-
butions and densities in a discrete time series context.
1524 D. Duong and N.R. Swanson
H0:
!
max EX F y
{ ð u2 Þ F y
{ ð u1 Þ
k¼2, ..., m X1,1 tþt ðXt Þ X 1 ðXt Þ
!2 1, tþt
ðF0 ðu2 jXt Þ F0 ðu1 jXt ÞÞ
!
EX F y
{ ðu2 Þ F y
{ ð u1 Þ
Xk,k tþt ðXt Þ X k ðXt Þ
!2k, tþt
ðF0 ðu2 jXt Þ F0 ðu1 jXt ÞÞ :
{
yk y{
F { ð uÞ ¼ Ftk { t
u Xt ; yk ¼ Py{ Xk, tþt u Xt ¼ Xt ,
k
y
Xk,k tþt ðXt Þ k
which is the conditional distribution of Xt+t, given Xt, and evaluated at u under the
y{
probability law generated by model k. Xk,k tþt ðXt Þ with 1 t T t is the
skeleton implied by model k, parameter y{k , and initial value Xt. Analogously, define
Ft0 ðujXt ; y0 Þ ¼ Pty0 ðXtþt ujXt Þ to be the “true” conditional distribution.
Note that the three hypotheses expressed above apply exactly the same to
the case of multifactor diffusions. Now, before moving to the statistics
description section, we briefly explain the intuitions in facilitating construction of
the tests.
In the first case (Hypothesis 1), Corradi and Swanson (2005) construct a
Kolmogorov-type test based on comparison of the empirical distribution and the
unconditional CDF implied by the specification of the drift, variance, and jumps.
Specifically, one can look at the scaled difference between
{ ðu
F u; y{ ¼ Pr Xyt u ¼ f x; y{ dx
and estimator of the true F0(u|Xt, y0), the empirical distribution of Xt defined as
1X T
1fXt ug,
T t¼1
Similarly for the second case of conditional distribution (Hypothesis 2), the test
statistic VT can be a measure of the distance between the t-step ahead conditional
{ {
distribution of Xytþt , given Xyt ¼ Xt , as
{ {
Ft u Xt ; y{ ¼ Pr Xytþt u Xyt ¼ Xt ,
1 X T t
1fXtþt ug,
T t t¼1
12
See Sect. 56.3.3.1 for model simulation details.
1526 D. Duong and N.R. Swanson
where
T
1 X
V T , N, h ¼ pffiffiffi 1fXt ug F u; y^T , N , h :
T t¼1
ð
In the above expression, U is a compact interval and pðuÞdu ¼ 1, 1fXt ug is
U
again the indicator function which returns value 1 if Xt u and 0 otherwise. Further,
as defined in Sect. 56.3.3, y^T , N, h hereafter is a simulated estimator where T is sample
size and h is the discretization interval used in simulation. In addition, with the abuse
of notation, N is a generic notation throughout this chapter, i.e., N ¼ L, the length of
each simulation path for (recursive) SGMM, and N ¼ M, the number of random
draws (simulated paths) for (recursive) NPQML estimator.13 Also note in our nota-
tion that as the above test is in sample specification test, the estimator and the
statistics are constructed using the entire sample, i.e., y^T , N, h .
It has been shown in Corradi and Swanson (2005) that under regular conditions
and if the estimator is estimated by SGMM, the above statistics converges
to a functional of Gaussian process.14 In particular, pick the choice T, N ! 1,
h ! 0, T/N ! 0, and Th2 ! 0.
13
M is often chosen to coincide with S, the number of simulated paths used when simulating
distributions.
14
For details and the proof, see Theorem 1 in Corradi and Swanson (2005).
56 Density and Conditional Distribution-Based Specification Analysis 1527
where Z is a Gaussian process with covariance kernel. Hence, the limiting distri-
2
bution of VT,N,h is a functional of a Gaussian process with a covariance kernel that
reflects both PEE and the time series nature of the data. As y^T , N, h is root-T
consistent, PEE does not disappear in the asymptotic covariance kernel.
Under HA, there exists an e > 0 such that
1 2
lim Pr V > e ¼ 1:
T!1 T T, N, h
For the asymptotic critical value tabulation, we use the bootstrap procedure. In
order to establish validity of the block bootstrap under SGMM with the presence of
PEE, the simulated sample size should be chosen to grow at a faster rate than the
historical sample, i.e., T/N ! 0.
Thus, we can follow the steps in appropriate bootstrap procedure in Sect. 56.3.4.
For instance, if the SGMM estimator is used, the bootstrap statistic is
ð
V T , N, h ¼ V 2
2
T , N , h ðuÞpðuÞdu,
U
where
T
1 X
V 2
T, N, h ¼ p ffiffiffi 1 Xt u 1fXt ug
T t¼1
^ ^
F u; y T , N, h F u; y T , N, h :
In the above expression, y^T , N, h is the bootstrap analog of y^T , N, h and is estimated
by the bootstrap sample X1, . . ., XT (see Sect. 56.3.4). With appropriate conditions,
* *
2*
Corradi and Swanson (2005) show that under the null, VT,N,h has a well-defined
2
limiting distribution which coincides with that of VT,N,h. We then can straightfor-
wardly derive the bootstrap critical value by following Steps 1–5 in Sect. 56.3.4. In
particular, in Step 5, the idea is to perform B bootstrap replications (B large) and
compute the percentiles of the empirical distribution of the B bootstrap statistics.
Reject H0 if V2T,N,h is greater than the (1a)th percentile of this empirical distribu-
tion. Otherwise, do not reject H0.
with U and V compact sets on the real line. y^T , N, h is the simulated estimator using
entire sample X1, . . . , XT, and S is the number of simulated replications used in the
estimation of conditional distributions as described in Sect. 56.3.3. If SGMM estimator
is used (similar to unconditional distribution case and the same as in Bhardwaj et al.
(2008)), then N ¼ L, where L is the simulation length used in parameter estimation.
The above statistic is a simulation-based version of the conditional Kolmogorov
test of Andrews (1997), which compare the joint empirical distribution:
1 X T t
pffiffiffiffiffiffiffiffiffiffiffi 1fXtþt ug1fXt vg,
T t t¼1
1 X T t
pffiffiffiffiffiffiffiffiffiffiffi F0, t ðujXt ; y0 Þ1fXt vg:
T t t¼1
Intuitively, if the null is not rejected, the metric distance between the two should
asymptotically disappear. In the simulation context with parameter estimation
error, the asymptotic limit of ZT however is a nontrivial one. Bhardwaj et al.
(2008) show that with the proper choice of T, N, S, h, i.e., T, N, S, T2/S ! 1
and h, T/N, T/S, Nh, h2T ! 0, then
d
ZT ! sup jZ ðu; vÞj,
uv2UV
15
In this chapter, we assume that X(·) satisfies the regularity conditions stated in Corradi and
Swanson (2011), i.e., assumptions A1–A8. Those conditions also reflect requirements A1–A2 in
Bhardwaj et al. (2008). Note that the SGMM estimator used in Bhardwaj et al. (2008) satisfies the
root-N consistency condition that Corradi and Swanson (2011) impose on their parameter estima-
tor (see Assumption 4).
16
See Sects. 56.3.3 and 56.3.4 for further details.
56 Density and Conditional Distribution-Based Specification Analysis 1529
of parameter estimation error, and (3) The correlation between the first two.
Furthermore, under HA, there exists some e > 0 such that
1
lim Pr pffiffiffi ZT > e ¼ 1:
P!1 T
As T/S ! 0, the contribution of simulation error is asymptotically negligible. The
limiting distribution is not nuisance parameter-free and hence critical values cannot
be tabulated directly from it. The appropriate bootstrap statistic in this context is
ZT ¼ sup ZT ðu; vÞ ,
uv2UV
where
1 X Tt
Z T ¼ ðu; vÞ ¼ pffiffiffiffiffiffiffiffiffiffiffi 1 Xt v
T t t¼1
!
1X S
y^T , N, h
n o
1 Xs, tþt u 1 Xtþt u
S s¼1
1 X T t
pffiffiffiffiffiffiffiffiffiffiffi 1fXt vg
T t t¼1
!
S y^
1X T, N, h
1 Xs, tþt u 1fXtþt ug :
S s¼1
In the above expression, y^ T , N, h is the bootstrap parameter estimated using the
y^
resampled data X*t , N, h , s ¼ 1, . . . , S and t ¼ 1, . . . , T t is
for t ¼ 1, . . . , T t. Xs,Ttþt
the simulated date under y^T, N, h and X*t ; t ¼ 1, . . ., T t is a resampled series
constructed using standard block bootstrap methods as described in Sect. 56.3.4.
Note that in the original paper, Bhardwaj et al. (2008) propose bootstrap SGMM
estimator for conditional distribution of diffusion processes. Corradi and Swanson
(2011) extend the test to the case of simulated recursive NPSQML estimator.
Regarding the generation of the empirical distribution of Z*T (asthmatically the
same as ZT), follow Steps 1–5 in the bootstrap procedure in Sect. 56.3.4. This
yields B bootstrap replications (B large) of Z*T. One can then compare ZT with the
percentiles of the empirical distribution of Z*T, and reject H0 if ZT is greater than the
(1 a)th percentile. Otherwise, do not reject H0. Tests carried out in this manner
are correctly asymptotically sized and have unit asymptotic power.
where U is a compact set so that the objective is evaluation of predictive densities for
a given range of values. In addition, in the recursive setting (not full sample is used to
estimate parameters), if we use the recursive NPSQML estimator, say y^1:t, N , h and
y^k:t, N, h , for models 1 and k, respectively, then the test statistic is defined as
where
"
S n o
1 X T t
1X y^ N, h
Dk, P, S ðu1 ; u2 Þ ¼ pffiffiffi 1 u1 X1,1,i,t,tþt ð X t Þ u2
P t¼R S i¼1
#2
1fu1 Xtþt u2 g
"
S n o
1X y^ N, h
1 u1 Xk,k,i,t,tþt ðXt Þ u2
S i¼1
#2 1
fu1 Xtþt u2 g A:
where, with an abuse of notation, mk(u1, u2) ¼ m1(u1, u2) mk(u1, u2), and
00 ! 12 1
B B F y{ ð u 2 Þ F y{ ð u1 Þ C C
mj ðu1 ; u2 Þ ¼ E@@ Xj,j tþt ðXt Þ Xj,j tþt ðXt Þ A A,
ðF0 ðu2 jXt Þ F0 ðu1 jXt ÞÞ
for j ¼ 1,. . ., m, and where (Z1(u1, u2), . . ., Zm(u1, u2)) is an m-dimensional Gaussian
random variable the covariance kernels that involves error in parameter estimation.
Bootstrap statistics are thereforerequired to reflect this parameterestimation errorissue.17
In the implementation, we can obtain the asymptotic critical value using a
recursive version of the block bootstrap. The idea is that when forming block
17
See Corradi and Swanson (2011) for further discussion.
56 Density and Conditional Distribution-Based Specification Analysis 1531
data to construct bootstrap estimator y^k, t, N, h. Recall that N is chosen in Corradi and
Swanson (2011) as the number of simulated series used to estimate the parameters
(N ¼ M ¼ S) and such as N/R, N/P ! 1. Under this condition, simulation error
vanishes and there is no need to resample the simulated series.
Corradi and Swanson (2011) show that
T
1 X
pffiffiffi y^k, t, N, h y^k, t, N, h Þ
P t¼R
has the same limiting distribution as
T
1 X
pffiffiffi y^k, t, N, h y{k Þ,
P t¼R
conditional on all samples except a set with probability measure approaching zero.
Given this, the appropriate bootstrap statistic is
( "
S n o
1 X Tt
1X y^ N, h
Dk, P, S ðu1 ; u2 Þ ¼ pffiffiffi 1 u1 X1,1,i,t,tþt X t u2
P t¼R S i¼1
#2
1 u1 Xtþt u2
"
S n o
1XT
1X y^ N , h
1 u1 X1,1,i,t,tþt Xj u2
T j¼1 S i¼1
#2
1 u1 Xjþt u2
"
S n o
1X y^ N , h
1 u1 Xk,k,i,t,tþt Xt u2
S i¼1
#2
1 u1 Xtþt u2
"
S n o
1X S
1X y^ N, h
1 u1 Xk,k,i,t,tþt X j u2
S i¼1 S i¼1
# 2 119
=
1 u1 Xjþt u2 AA :
;
1532 D. Duong and N.R. Swanson
Now, let us turn our attention to multifactor diffusion models of the form
(X(t), V(t))0 ¼ (X(t), V1(t), . . ., Vd(t))0 , where only the first element, the diffusion
process Xt, is observed while V(t) ¼ (V1(t), . . ., Vd(t))0 is latent. The most
popular class of the multifactor models is stochastic volatility model expressed as
below:
! ! !
dXðtÞ b1 XðtÞ,y{ s11 V ðtÞ,y{ s12 V ðtÞ,y{
¼ dt þ dW 1 ðtÞ þ dW 2 ðtÞ,
dV ðtÞ b2 V ðtÞ,y{ 0 s22 V ðtÞ,y{
(56:10)
where W1(t)11 and W2(t)11 are independent Brownian motions.18 For instance,
many term structure models require the multifactor specification of the above form
(see Dai and Singleton (2000)). In a more complicated case, the drift function can
also be specified to be a stochastic process which poses even more challenges to
testing. As mentioned earlier, the hypotheses (Hypothesis 1, 2, 3) and the test
18
Note that the dimension of X(·) can be higher and we can add jumps to the above specification
such that it satisfies the regularity conditions outlined in the one-factor case. In addition, Corradi
and Swanson (2005) provide a detailed discussion of approximation schemes in the context of
stochastic volatility models.
56 Density and Conditional Distribution-Based Specification Analysis 1533
construction strategy for multifactor models are the same as for one-factor model.
All theory essentially applies immediately to multifactor cases. In implementation,
the key difference is in the simulated approximation scheme facilitating parameter
and CDF estimation. X(t) cannot simply be expressed as a function
Ðt of d + 1 driving
Brownian motions but instead involves a function of (Wjt, 0 WjsdWis), i, j ¼ 1, .. . ,
d + 1 (see, e.g., Pardoux and Talay (1985, pp. 30–32) and Corradi and Swanson
(2005)).
For illustration, we hereafter focus on the analysis of a stochastic volatility
model (56.10) where drift and diffusion coefficients can be written as
!
b1 XðtÞ, y{
b¼
b2 V ðtÞ, y{
!
s11 V ðtÞ, y{ s12 V ðtÞ, y{
s¼
0 s22 V ðtÞ, y{
We also examine a three-factor model (i.e., the Chen model as in Eq. 56.5) and
a three-factor model with jumps (i.e., CHENJ as in Eq. 56.6). By presenting two-
and three-factor models as an extension of our above discussion, we make it clear
that specification tests of multiple factor diffusions with d 3 can be easily
constructed in similar manner.
In distribution estimation, the important challenge for multifactor models lies in
the missing variable issue. In particular, for simulation of Xt, one needs initial
values of the latent processes V1, . . . , Vd, which are unobserved. To overcome this
problem, it suffices to simulate the process using different random initial values for
the volatility process; then construct the simulated distribution using those initial
values and average them out. This allows one to integrate out the effect of
a particular choice of volatility initial value.
For clarity of exposition, we sketch out a simulation strategy for a general model
of d latent variables in Sect. 56.3.3. This generalizes the simulation scheme of
three-factor models in Cai and Swanson (2011). As a final remark before moving to
the statistic presentation, note that the class of multifactor diffusion processes
considered in this chapter is required to match the regular conditions as in previous
section (assumption from A1 to A8 in Corradi and Swanson (2011) with A4 being
replaced by A40 ).
T S
1 X 1X y^
SV T , S, h ¼ pffiffiffi 1fX t ug 1 Xt,Th, N, L, h u :
T t¼1 S t¼1
In the above expression, recall that S is the number of simulation paths used in
distribution simulation; y^T , N, L, h is a simulated estimator (see Sect. 56.3.3). N is
a generic notation throughout this chapter, i.e., N ¼ L, the length of each simulation
path for SGMM, and N ¼ M, the number of random draws (simulated paths) for
NPQML estimator. h is the discretization interval used in simulation. Note that
y^T , N, L, h is chosen in Corradi and Swanson (2005) to be SGMM estimator using
full sample and therefore N ¼ L ¼ S.19 To put it simply, one can write
y^T , S, h ¼ y^T , N, L, h .
Under the null, choose T, S to satisfy T, S ! 1, Sh ! 0, T/S ! 0, then
ð
SV 2T , S, h ! SV 2 ðuÞpðuÞ,
U
where Z is a Gaussian process with covariance kernel that reflects both PEE and the
time-dependent nature of the data. The relevant bootstrap statistic is
1 X T
SV 2
T , S, h ¼ p ffiffiffi 1 Xt u 1fXt ug
T t¼1
S ^
1 X T
1X y^ y
pffiffiffi 1 Xt,Th, N, L, h u 1 Xt,Th, N, L, h u ,
T t¼1 S t¼1
where y^T, S, h is the bootstrap analog of y^T , S, h . Repeat Steps 1–5 in the bootstrap
procedure in Sect. 56.3.4 to obtain critical values which are the percentiles of
the empirical distribution of Z*T. Compare SVT,S,h with the percentiles of the
empirical distribution of the bootstrap statistic and reject H0 if SVT,S,h is greater
than the (1a)th percentile thereof. Otherwise, do not reject H0.
19
As seen in assumption A40 in Corradi and Swanson (2011) and Sect. 56.3.3 of this chapter,
^
y T , N, L, h can be other estimators such as the NPSQML estimator. Importantly, y^T , N, L, h satisfies
condition A40 in Corradi and Swanson (2011).
56 Density and Conditional Distribution-Based Specification Analysis 1535
Note that for illustration, we again assume use of the SGMM estimator y^T , N, L, h ,
as in the original work of Bhardwaj et al. (2008) (namely, y^T , N, L, h is the
simulated estimator described in Sect. 56.3.3). Specifically, N is chosen as
the length of sample path L used in parameter estimation. The associated test
statistic is
y^
where Xj,Ti,,Ntþt
, L, h is t-step ahead simulated skeleton obtained by simulation procedure
for multifactor model in Sect. 56.3.3.1.
In a similar manner, the bootstrap statistic analogous to SZT is
SZ T ¼ sup SZ ðu; vÞ ,
T
uv2UV
1 X
T t
SZ T ðu; vÞ ¼ pffiffiffiffiffiffiffiffiffiffiffi 1 Xt v
T t t¼1
!
S ^
1 X N X
y T , N, L, h
1 Xj, i, tþt u 1 Xtþt u
NS j¼1 i¼1
1 X T t
pffiffiffiffiffiffiffiffiffiffiffi 1fX t v g
T t t¼1
!
S n o
1 X N X
y^T , N, L, h
1 Xj, i, tþt u 1fXtþt ug
NS j¼1 i¼1
where y^T , N, L, h is the bootstrap estimator described in Sect. 56.3.4. For the three-
factor model, the test statistic is defined as
1 X T t
MZ T ðu; vÞ ¼ pffiffiffiffiffiffiffiffiffiffiffi 1fX t v g
T t t¼1
!
S ^
1 X L X L X
y t, N , L , h
2 1 Xs, tþt u 1 Xtþt u
L S j¼1 k¼1 i¼1
1 X T t
pffiffiffiffiffiffiffiffiffiffiffi 1fX t v g
T t t¼1
!
S n o
1 X L X L X
y^t, N, L, h
2 1 Xs, tþt u 1fXtþt ug
L S j¼1 k¼1 i¼1
y^ N, L, h
y^ , N , L, h y^ , N, L, h 1, y^ 2, y^
where Xs,Ttþt ¼ Xs,Ttþt Xt ; V j T, N, L, h ; V k T, N, L, h and Xs,t,tþt ¼
y^ N , L, h 1, y^ 2, y^
Xs,t,tþt Xt ; V j t, N, L, h ; V k t, N, L, h .
The first-order asymptotic validity of inference carried out using bootstrap
statistics formed as outlined above follows immediately from Bhardwaj
et al. (2008). For testing decision, one compares the test statistics SZT,S,h and
MZT,S,h with the percentiles or the empirical distributions of SZ*T and MZT,S,h *
,
respectively. Then, reject H0 if the actual statistic is greater than the (1 a)th
percentile of the empirical distribution of the bootstrap statistic, as in Sect. 56.3.4.
Otherwise, do not reject H0.
DV P, L, S ¼ max DV k, P, L, S ðu1 ; u2 Þ,
k¼2, ..., m
where
1 X T t
DV k, P, L, S ðu1 ; u2 Þ ¼ pffiffiffi
P t¼R
!2
S n o
1 X L X
y^1, t, N , L, h y^1, t, N , L, h
1 u1 X1, tþt, i, j Xt ; V 1, j u2 1fu1 Xtþt u2 g
SL j¼1 i¼1
0 !2 1
XL X S n o
1 y^k, t, N , L, h y^k, t, N, L, h
@ 1 u1 Xk, tþt, i, j Xt ; V k, j u2 1fu1 Xtþt u2 g A:
SL j¼1 i¼1
56 Density and Conditional Distribution-Based Specification Analysis 1537
Critical values for these tests can be obtained using a recursive version of the
block bootstrap. The corresponding bootstrap test statistic is
where
1 X T t
DV k, P, L, S ðu1 ;u2 Þ ¼ pffiffiffi
P t¼R
80" #2
< X L X S
1 y^1, t, N , L, h y^1, t, N , L, h
@ 1 u1 X1, tþt, i, j Xt ;V 1, j
u2 1 u1 Xtþt u2
: SL j¼1 i¼1
" #2 1
S n o
1X T
1X L X
^
y , t , N , L, h ^
y 1, t, N , L, h
1 u1 < X1,1tþt , i, j Xl ;V 1, j u2 1fu1 Xlþt u2 g A
T l¼1 SL j¼1 i¼1
0" #2
L X
X S
^
1 y^k, t, N , L, h y
@ 1 u1 Xk, tþt, i, j Xt ;V k, j k , t , N , L , h
u2 1 u1 Xtþt u2
SL j¼1 i¼1
" # 2 19
1 XT
1 XL X S n
^ ^ o =
y , t, N , L, h y k, t, N , L, h A
1 u1 Xk,ktþt X ;V
, i, j l k , j u2 1fu 1 X lþt u2 g
T l¼1 SL j¼1 i¼1 ;
Of note is that we follow Cai and Swanson (2011) by adopting the recursive
NPSQML estimator y^1, t, N, L, h and y^k, t, N, L, h for model 1 and k, respectively, as
introduced in Sect. 56.3.3.4 with the choice N ¼ M ¼ S. y^1, t, N , L, h and y^k, t, N, L, h are
bootstrap analogs of y^1, t, N, L, h and y^k, t, N, L, h , respectively (see Sect. 56.3.4). In
addition, we do not need to resample the volatility process, although volatility is
simulated under both y^k, t, N, L, h and y^k, t, N, L, h , k ¼ 1, . . ., m.
Repeat Steps 1–5 in the bootstrap procedure in Sect. 56.3.4 to obtain critical
values. Compare DVP, L, S with the percentiles of the empirical distribution of
L, S, and reject H0 if DVP, L, S is greater than the (1 a)th percentile.
*
DVP,
Otherwise, do not reject H0. Again, in implementation, there is no need to
re-estimate y^k, t, N, L, h for each bootstrap replication if P/R ! 0, P, T, R ! 1, as
parameter estimation error vanishes asymptotically in this case.
where
iid
W qh W ðq1Þh ¼ ϵ qh N ð0; hÞ,
1 0
Xðyqþ1Þh Xyqh ¼ b Xyqh ; y h þ s Xyqh ; y ϵ ðqþ1Þh s Xyqh ; y s Xyqh ; y h
2
1 0
þ s Xyqh ; y s Xyqh ; y ϵ ð2qþ1Þh lmy h
2
XJ
þ yj 1 qh U j ðq þ 1Þh :
j¼1
(56:11)
The only difference between this approximation and that used for the pure diffusion
is the jump part. Note that the last term on the right-hand side (RHS) of Eq. 56.11
is nonzero whenever we have one (or more) jump realization(s) in the interval
[(q 1)h, qh]. Moreover, as neither the intensity nor the jump size is state
dependent, the jump component can be simulated without any discretization
error, as follows. Begin by making a draw from a Poisson distribution with intensity
parameter l^t , say J . This gives a realization for the number of jumps over the
simulation time span. Then draw J uniform random variables over [0, L], and sort
them in ascending order so that U 1 U 2 . . . U J . These provide realizations
for the J independent draws from f, say y1 , . . . , yJ .
SV models without jumps as in Eq. 56.4 (using a generalized Milstein scheme):
56 Density and Conditional Distribution-Based Specification Analysis 1539
Xðyqþ1Þh ¼ Xyqh þ eb 1 Xyqh ; y h þ s11 V yqh ; y ϵ 1, ðqþ1Þh þ s12 V yqh ; y ϵ 2, ðqþ1Þh
y @s11 V yqh ; y
1 y @s12, k V qh ; y 2
þ s22 V qh ; y ϵ 2, ðqþ1Þh þ s22 V yqh ; y
2ð @V @V
ðqþ1Þh ð s
dW 1, t dW 2:s ,
qh qh
(56:12)
V ðyqþ1Þh ¼ V yqh þ e
b 2 V yqh ; y h þ s22 V yqh ; y ϵ 2, ðqþ1Þh
y (56:13)
1 @s 22 V qh ; y
y
þ s22 V qh ; y ϵ 2, ðqþ1Þh ,
2
2 @V
where h 1/2ϵi,qh N(0,1), i ¼ 1, 2, E ϵ 1, qh ϵ 2, q0 h ¼ 0 for all q 6¼ q0 , and
0 1
1 @s12 ðV; yÞ
e
b 1 ðV; yÞ b
B 1 ð V; y Þ s22 ð V; y Þ C
e
b ðV; yÞ ¼ ¼@ 2 @V :
e
b 2 ðV; yÞ 1 @s22 ðV; yÞ A
b2 ðV; yÞ s22 ðV; yÞ
2 @V
The last terms on the RHS of Eq. 56.12 involve stochastic integrals and
cannot be explicitly computed. However, they can be approximated up to an
error of order o(h) by (see, e.g., Eq. 3.7, pp. 347 in Kloeden and Platen (1999))
ð ðqþ1Þh ð s
1 pffiffiffiffiffi
dW 1, t dW 2, s h x1 x2 þ rp m1, p x2 m2, p x1
qh qh 2
h X1 pffiffiffi pffiffiffi
p
þ B1, r 2x2 þ 2, r B2, r 2x1 þ 1, r ,
2p r
r¼1
where for j ¼ 1, 2, xj, mj, p, Bj, r, j, r are i.i.d. N(0, 1) random variables, rp ¼ 12
1
Xp 1
2p1 2 r¼1 r 2
, and p is such that as h ! 0, p ! 1.
One-factor Models
Consider the one-factor model as in Eq. 56.7. To estimate the simulated CDFs:
Step 1: Obtain y^T , N, h (using the entire sample) or y^t, N, h (recursive estimator) where
y^T , N, h and y^t, N, h are estimators as discussed in Sects. 56.3.3.3 and 56.3.3.4.
Step 2: Under y^T , N , h or y^t, N, h ,20 simulate S paths of length t, all having the same
starting value, Xt. In particular, for each path i ¼ 1, . . ., S of length t, generate
y^ , N, h
Xi,Ttþt ðXt Þ according to a Milstein schemes detailed in previous section, with
iid
y ¼ y^T , N, h or y^t, N, h. The errors used in simulation are eqh e N ð0; hÞ, and Qh ¼ t.
eqh is assumed to be independent across simulations, so that E(ei,qhej,qh) ¼ 0, for
all i 6¼ j and E(ei,qhei,qh) ¼ h, for any i, j. In addition, as the simulated diffusion is
ergodic, the effect of the starting value approaches zero at an exponential rate,
as t ! 1.
Step 3: If y^T , N, h y^t, N, h is used, an estimate for the distribution, at time t + t,
^ ^
conditional on Xt, with estimator y T , N, h y t, N, h , is defined as
n ^ o
^ ^ 1X S
y , N, h
F t u X t ; y T , N , h ¼ 1 Xi,Ttþt ðX t Þ u :
S i¼1
Bhardwaj et al. (2008) show that if the model is correctly specified, then
XS n y^T, N, h o
1
S i¼1
1 X iþtþt ðX t Þ u provides a consistent of the conditional distribution
Ft(u|Xt, y{) ¼ Pr(Xt+t
y{
u|Xy{
t ¼ Xt).
Specifically, assume that T, N, S ! 1. Then, for the case of SGMM estimator,
if h ! 0, T/N ! 0, and h2T ! 0, T2/S ! 1, the following result holds for any
Xt, t 1, uniformly in u
pr
F^t u Xt ; y^T , N, h Ft ujXt ; y{ ! 0,
20
Note that N ¼ L for the SGMM estimator while N ¼ M ¼ S for NSQML estimator.
56 Density and Conditional Distribution-Based Specification Analysis 1541
S n o pr
1X y^ N , h
1 u1 Xk,k,i,t,tþt ð X t Þ u 2 ! F y{ ð u 2 Þ F y{ ðu1 Þ, t ¼ R, . . . , T t,
S i¼1 Xk,k tþt ðXt Þ Xk,k tþt ðXt Þ
y{
where F y
{ ðuÞ is the marginal distribution of Xtþt
k
ðXt Þ implied by k model (i.e.,
Xk, tþt ðXt Þ
k
by the model used to simulate the series), conditional on the (simulation) starting
{
value Xt. Furthermore, the marginal distribution of Xytþt ðXt Þ is the distribution of Xt+t
conditional on the values observed at time t. Thus, FXy{ ðX Þ ðuÞ ¼ FTk u Xt ; y{k .
kþtþt t
y^ N , h
Of important note is that in the simulation of Xk,k,i,t,tþt ðXt Þ,
i ¼ 1, . . ., S, for each t,
t ¼ R, . . ., T t, we must use the same set of randomly drawn errors and similarly
the same draws for numbers of jumps, jump times, and jump sizes. Thus, we
only allow for the starting value to change. In particular, for each i ¼ 1, . . ., S,
y^ , N, h y^
we generate Xk,k,i,RRþt ðXR Þ, . . . , Xk,k,i,Tt
t
, N, h ðX Þ . This yields an S P matrix
Tt
of simulated values, where P ¼ T R t + 1 refers to the length of the out-of-
y^
sample period. Xk,k,i,Rþj , N, h
(at time R + j + t) can be seen as t periods
Rþjþt XRþj
ahead value “predicted” by model k using all available information up to
time R + jR+j, j ¼ 1, . . ., P (the initial value XR+j and y^k, Rþj, N, h estimated using
X1, . . ., XR+j). The key feature of this setup is that it enables us to compare
“predicted” t periods ahead values (i.e., Xk,k,i,Rþj
y^ , N, h
Rþjþt X Rþj ) with actual values
that are t periods ahead (i.e., XR+j+t), for j ¼ 1, . . ., P. In this manner,
simulation-based tests under ex-ante predictive density comparison framework
can be constructed.
1542 D. Duong and N.R. Swanson
Multifactor Model
Consider the multifactor model with a skeleton (Xt, V1t , . . .,Vdt )0 (e.g., stochastic
mean, stochastic volatility models, stochastic volatility of volatility) where only the
first element Xt is observed. For simulation of the CDF, the difficulty arises as we do
not know the initial values of latent variables (V1t , . . .,Vdt )0 at each point in time. We
generalize the simulation plan of Bhardwaj et al. (2008) and Cai and Swanson
(2011) to the case of d dimensions. Specifically, to overcome the initial value
difficulty, a natural strategy is to simulate a long path of length L for each latent
variable V1t , . . ., Vdt and use them to construct Xt+t and the corresponding simulated
CDF of Xt+t; and finally, we average out the volatility values. Note that there are Ld
combinations of the initial values V1t , . . ., Vdt . For illustration, consider the case of
stochastic volatility (d ¼ 1) and the Chen three-factor model as in Eq. 56.5 (d ¼ 2),
using recursive estimators.
For the case of stochastic volatility (d ¼ 1), i.e., (Xt, Vt)0 , the steps are as
follows:
Step 1: Estimate y^t, N, L, h using recursive SGMM or NSQML estimation methods.
y^
Step 2: Using the scheme in Eq. 56.13 with y ¼ y^t, N, L, h, generate the path V qht, N, L, h
y^
for q ¼ 1/h, . . ., Qh with Qh ¼ L and hence obtain V t, N, L, h j ¼ 1, . . . L.
j
Step 3: Using schemes in Eqs. 56.12, 56.13, simulate L S paths of length t, setting
the initial value for the observable state variable to be Xt. For the initial values of
y^
unobserved volatility, use V j,t,qhN, L, h j ¼ 1,. . ., L as retrieved in Step 2. Also,
Ð Ð
keep the simulated random innovations (i.e., e1,qh,. e1,qh, (q+1)h qh ( sqhdW1,t)dW2,s)
to be constant across ^each j and t. Hence, for each replication i, using
y , N, h y^ , L, h
initial values Xt and V j,Tqh , we obtain Xj,t,i,Ntþt ðXt Þ which is a t-step ahead
simulated value.
Step 4: Now the estimator of Ft(u|Xt,y{) is defined as
1 XL X S n ^ o
y ,h
F^t u Xt ; y^t, N, L, h ¼ 1 Xj,t,i,Ntþt ðX t Þ u :
LS j¼1 i¼1
Note that by averaging over the initial value of the volatility process, we have
1 X S n y t, N , h o
integrated out its effect. In other words, 1 X j, i, tþt ð X t Þ u is an
S i¼1
y^
estimate of Ft u Xt ; V j,t,hN, h , y{ .
Step 5: Repeat Steps 1–4 for t ¼ 1, . . ., T t. This yields T t conditional
distributions that are t-steps ahead which will be used in the construction of the
specification tests.
For three-factor model (d ¼ 2), i.e., (Xt, V1t , V2t ), consider model (56.5), where
Wt ¼ (W1t , W2t , W3t ) are mutually independent standard Brownian motions.
Step 1: Estimate y^t, N, L, h using SGMM or NSQML estimation methods.
56 Density and Conditional Distribution-Based Specification Analysis 1543
1, y ^
Step 2: Given the estimated parameter y^t, N, L, h , generate the paths V qh t, N, L, h and
2, y^ 1, y^
V ph t, N, L, h for q, p ¼ 1/h, . . ., Qh with Qh ¼ L, and hence, obtain V j T, N, L, h ,
2, y^
V k T, N, L, h , k ¼ 1, . . ., L.
j
1, y^
Step 3: Given the observable X and the L L simulated latent paths (V t, N, L, h and
t j
2, y^t, N, L, h y^t, N, L, h
Vk j, k ¼ 1, . . ., L) as the start values, we simulate t-step ahead Xtþt
1, y^ 2, y^
Xt ; V j t, N, L, h ; V k t, N, L, h . Since the start values for the two latent variables are
L L length, so for each Xt we have N2 path. Now to integrate out
the initial effect of latent variables, form the estimate of conditional
distribution as
1XL X L n ^ o
y N , L, h 1, y^ 2, y^
F^t, s u Xt y^ ¼ 2 1 Xs,t,tþt Xt ; V j t, N, L, h ; V k t, N, L, h u ,
L j¼1 k¼1
1X S
F^t u Xt y^ ¼ F^t, s u Xt ; y^T , N, h :
S i¼1
!
1 X t 1X L 0
y^t, L, h ¼ arg min g Xj g Xyj, h
y2Y t j¼1 L j¼1
!
Xt 1X L
1 1 y (56:14)
Wt g Xj g X j, h
t j¼1 L j¼1
¼ arg min Gt, L, h ðyÞ0 W t Gt, L, h ðyÞ,
y2Y
! !0
1X lt X
tlt
1X t 1X t
W 1
t ¼ wn g Xj g Xj g Xjn g Xj ,
t n¼l j¼nþ1þlt t j¼1 t j¼1
t
(56:15)
where wv ¼ 1 v/(lT + 1). Further, the pseudo true value, y{, is defined to be
where
0
∇y G1 y{ W { G1 y{ ¼ 0:
However, in the case for which the number of parameters and the number of
moment conditions are the same, ∇yG1(y{)0 W{ is invertible, and so the first-order
conditions also imply that G1(y{) ¼ 0.
Also note that other available estimation methods using moments include the
efficient method of moments (EMM) estimator as proposed by Gallant and Tauchen
(1996, 1997), which calculates moment functions by simulating the expected value
of the score implied by an auxiliary model. In their setup, parameters are then
computed by minimizing a chi-square criterion function.
One-Factor Models
Hereafter, let f(Xt|Xt1, y{) be the conditional density associated with the above
jump diffusion. If f is known in closed form, we can just estimate y{ recursively,
using standard QML as21
1X t
y^t ¼ arg max ln f Xj Xj1 ; y , t ¼ R, . . . R þ P 1: (56:16)
y2Y t j¼2
Note that, similarly to the case of SGMM, the pseudo true value y{ is optimal in
the sense:
For the case f is not known in closed form, we can follow Kristensen and Shin
(2008) and Cai and Swanson (2011) to construct the simulated analog f^of f and then
use it to estimate y{. f^ is estimated as function of the simulated sample paths
y
Xt,i(Xt 1), for t ¼ 2, . . ., T 1, i ¼ 1, . . ., M. First, generate T 1 paths of length
one for each simulation replication, using Xt1 with t ¼ 1, . . ., T as starting values.
y
Hence, at time t and simulation replication i, we obtain skeletons Xt,i (Xt 1), for
t ¼ 2, . . ., T 1; i ¼ 1,. . .M where M is the number of simulation paths (number of
y y
random draws or Xt,j (Xt1) and Xt,l (Xt1) are i.i.d.) for each simulation replication.
21
Note that as model k is, in general, misspecified, ∑ T1 t¼1 fk(Xt|Xt1,yk) is a quasi-likelihood and
fk(Xt|Xt 1,y{k ) is not necessarily a martingale difference sequence.
1546 D. Duong and N.R. Swanson
M is fixed across all initial values. Then the recursive NPSQML estimator is defined
as follows:
1X t
y^t, M, h ¼ arg max lnf^M, h ðXi jXi1 ; yÞtM f^M, h ðXi jXi1 ; yÞ , t R,
y2Y t i¼2
where
!
1 XM Xyt, i, h ðXt1 Þ Xt
f^M, h ðXt jXt1 ; yÞ ¼ K :
MxM i¼1 xM
Note that with abuse of notation, we define y^t, L, h for SGMM and y^t, M, h for
NPSQML estimators where L and M have different interpretations (L is the length
of each simulation path and M is number of random draws).
The function tM f^M, h ðXt jXt1 ; yÞ is a trimming function. It has some
characteristics such as positive and increasing, tM f^M, h ðXt ; Xt1 ; yÞ ¼ 0, if f^M, h
ðXt ; Xt1 ; yÞ < xdM and tM f^M, h ðXt ; Xt1 ; yÞ ¼ 1 if f^M, h ðXt ; Xt1 ; yÞ > 2xdM , for
some d > 0.22 Note that when the log density is close to zero, the derivative tends to
infinity and thus even very tiny simulation errors can have a large impact on the
likelihood. The introduction of the trimming parameter into the optimization
function ensures the impact of this case to be minimal asymptotically.
Multifactor Models
Since volatility is not observable, we cannot proceed as in the single-factor case
when estimating the SV model using NPSQML estimator. Instead, let Vyj be
generated according to Eq. 56.13, setting qh ¼ j, and j ¼ 1, . . ., L. The idea is to
simulate L different starting values for unobservable volatility, construct the sim-
ulated likelihood functions accordingly, and then average them out. For each
simulation replication at time t, we simulate L different values of Xt(Xt1, Vyj ) by
generating L paths of length one, using fixed observable Xt1 and unobservable Vyj ,
j ¼ 1, . . . , L as starting values. Repeat this procedure for any t ¼ 1, . . . , T 1 and
for any set j, j ¼ 1, . . . , L of random errors e1,t + (q + 1)h,j and e2,t+(q+1)h,j, q ¼ 1, . . . ,
1/h. Note that it is important to use the same set of random errors e1,t+(q+1)h,j and
e2, t+(q+1)h,j across different initial values for volatility. Denote the simulated value
22
Fermanian and Salanie (2004) suggest using the following trimming function:
4ðx aN Þ3 3ðx aN Þ4
tN ðxÞ ¼ ,
a3N a4N
for aN x 2aN.
56 Density and Conditional Distribution-Based Specification Analysis 1547
at time t and simulation replication i, under generic parameter y, using Xt1, Vyj as
starting values as Xyt,i,h(Xt1, Vyj ). Then
0 1
y y
1 XL
1 X M X t, i , h X t1 ; V j X t
f^M, L, h ðXt jXt1 ; yÞ ¼ K@ A,
L j¼1 MxM i¼1 xM
and note that by averaging over the initial values for the unobservable volatility, its
effect is integrated out. Finally, define23
1X t
y^t, M, L, h ¼ arg min ln f^M, L, h ðXs jXs1 ; yÞtM f^M, L, h ðXs jXs1 ; yÞ , t R:
y2Y t s¼2
Note that in this case, Xt is no longer Markov (i.e., Xt and Vt are jointly
Markovian, but Xt is not). Therefore, even in the case of true data generating
process, the joint likelihood cannot be expressed as the product of the conditional
and marginal distributions. Thus, y^t, M, L, h is necessarily a QML estimator. Further-
more, note that ∇yf(Xt|Xt1,y{) is no longer a martingale difference sequence;
therefore, we need to use HAC robust covariance matrix estimators, regardless of
whether the model is the “correct” model or not.
The test statistics presented in Sects. 56.3.1 and 56.3.2 are implemented
using critical values constructed via the bootstrap. As mentioned earlier, motivation
for using the bootstrap is clear. The covariance kernel of the statistic limiting
distributions contains both parameter estimation error and the data-related
time-dependence components. Asymptotic critical value cannot thus be tabulated
in a usual way. Several methods have been proposed to tackle this issue.
One is the block bootstrap procedures which we discuss. Others have been
mentioned above.
With regard to the validity of the bootstrap, note that, in the case of dependent
observations without PEE, we can tabulate valid critical value using a simple
empirical version of the K€unsch (1989) block bootstrap. Now, the difficulty in
our context lies in accounting for parameter estimation error. Goncalves and White
(2002) establish the first-order validity of the block bootstrap for QMLE
(or m-estimator) for dependent and heterogeneous data. This is an important result
for the class of SGMM and NSQML estimators surveyed in this chapter and allows
Corradi and Swanson in CS (2011) and elsewhere to develop asymptotically valid
version of the bootstrap that can be applied under generic model misspecification,
as assumed throughout this chapter.
For the SGMM estimator, as shown in Corradi and Swanson (2005), the first-
order validity of the block bootstrap is valid in the exact identification case, and
when T/S ! 0. In this case, SGMM is asymptotically equivalent to GMM, and
consequently, there is no need to bootstrap the simulated series. In addition, in the
exact identification case, GMM estimators can be treated the same way that QMLE
estimators are treated. For the NSQML estimator, Corradi and Swanson (2011)
point out that the NPSQML estimator is asymptotically equivalent to the QML
estimator. Thus, we do not need to resample the simulated observations as the
negligible contribution of simulation errors.
Also note that critical values for these tests can be obtained using a recursive
version of the block bootstrap. When forming block bootstrap samples in the
recursive case, observations at the beginning of the sample are used more
frequently than observations at the end of the sample. This introduces
a location bias to the usual block bootstrap, as under standard resampling with
replacement, all blocks from the original sample have the same probability of
being selected. Also, the bias term varies across samples and can be either
positive or negative, depending on the specific sample. A first-order valid
bootstrap procedure for nonsimulation-based m-estimators constructed using
a recursive estimation scheme is outlined in Corradi and Swanson (2007a).
Here we extend the results of Corradi and Swanson (2007a) by establishing
asymptotic results for cases in which simulation-based estimators are
bootstrapped in a recursive setting.
Now the details of bootstrap procedure for critical value tabulation can be
outlined in 5 steps as follows:
Step 1: Let T ¼ bl, where b denotes the number of blocks and l denotes the length of
each block. We first draw a discrete uniform random variable, I1, that can take
values 0, 1, . . ., T l with probability 1/(T l + 1). The first block is given by
XI1 þ1 , . . . , XI1 þl. We then draw another discrete uniform random variable, say I2,
and a second block of length l is formed, say XI2 þ1 , . . . , XI2 þl . Continue in the
same manner, until you draw the last discrete uniform, say Ib, and so the
last block is XIb þ1 , . . . , XIb þl . Let’s call the X*t the resampled series, and note
that X*1, X*2, . . ., X*T corresponds to XI1 þ1 , XI1 þ2 , . . . , XIb þl . Thus, conditional
on the sample, the only random element is the beginning of each block.
In particular,
!
1X T
1X T
E
X ¼ Xt þ OP ðl=T Þ (56:18)
T t¼1 t T t¼1
! ! !
1 X
T
1XTl X l
1X T
1X T
Var Xt ¼ Xt Xt Xtþi Xt þ OP l2 =T ,
T 1=2 t¼1
T t¼l i¼l T t¼1 T t¼1
(56:19)
where E* and Var* denote the expectation and the variance operators with
respect to P* (the probability law governing the resampled series or the proba-
bility law governing the i.i.d. uniform random variables, conditional on the
sample) and where OP*(l/T)(OP*(l2/T)) denotes a term converging in probability
P* to zero, as l/T ! 0(l2/T ! 0).
In the case of recursive estimators, we proceed the bootstrap similarly as
follows. Begin by resampling b blocks of length l from the full sample, with
lb ¼ T. For any given t, it is necessary to jointly resample Xt, Xt+1, . . ., Xt+t. More
precisely, let Zt,t ¼ (Xt, Xt+1, . . ., Xt+t), t ¼ 1,. . ., T t. Now, resample
b overlapping blocks of length l from Zt,t. This yields Zt,* ¼ (X*t , X*t+1, . . .,
*
Xt+t ), t ¼ 1, . . ., T t.
Step 2: Re-estimate y^t, N, h y^T , N, L, h using the bootstrap sample Zt,* ¼ (X*t , X*t+1,
. . .,Xt+t
*
), t ¼ 1, . . ., T t (or full sample X*1, X*2, . . ., X*T). Recall that if we use the
entire sample for the estimation, as the specification test in Corradi and Swanson
(2005) and Bhardwaj et al. (2008), then y^t, N, h is denoted as y^T , N, h. The bootstrap
estimators for SGMM and NPSQML are presented below:
Bootstrap (Recursive) SGMM Estimators
If the full sample is used in the specification test as in Corradi and Swanson
(2005) and Bhardwaj et al. (2008), the bootstrap estimator is constructed
straightforward as
!
1X T
1X L 0
y^T , N, h ¼ arg min
g Xj y
g X j, h
y2Y T j¼1 L i¼1
!
XT XL
1 1
W 1
T g Xj g Xyj, h ,
T j¼1 L i¼1
where W1T and g(.) are defined in Eq. 56.15 and L is the length of each
simulation path.
Note that it is convenient not to resample the simulated series as the
simulation error vanishes asymptotically. In implementation, we do not
mimic its contribution to the covariate kernel.
In the case of predictive density-type model selection where recursive
estimators are needed, define the bootstrap analog as
1550 D. Duong and N.R. Swanson
0 00 1
1 X t 1X T
y^t, L, h ¼ arg min @ @@g X
j g X j0 A
y2Y t j¼1 T j0 ¼1
!!!0
L 1X L
X ^t, L, h
y
y
L1 g X j, h g X j, h
i¼1
L i¼1
0 00 1
1 X t 1X T
O1 @ @@g X g Xj0 A
t j
t j¼1 T j0 ¼1
!!!
L 1X L
1X ^t, L, h
y
g Xyj, h g Xj, h
L i¼1 L i¼1
¼ arg min Gt, L, h ðyÞ0 O1
t Gt, L, h ðyÞ,
y2Y
where
2 32 3
1X lt X
tlt 1X T 1X T
O1 ¼ w n, t 4g X
g Xj0 54g Xjn
g X j0 5 :
t j
t n¼l j¼nþ1þl
T j0 ¼1 T j0 ¼1
t t
Note that each bootstrap term is recentered around the (full) sample mean.
The intuition behind the particular recentering in bootstrap recursive
SGMM estimator is that it ensures that the mean of the bootstrap moment
conditions, evaluated at y^t, L, h , is zero, up to a negligible term. Specifically,
we have
0 0 1
1 Xt 1X T
E@ @g X
j g Xj0 A
t j¼1 T j0 ¼1
!!
L ^ L
1X y t, L , h 1X y^t, L, h
g X j, h g Xj, h
L i¼1 L i¼1
1 X T
¼ E g Xj g X j0
T j0 ¼1
¼ Oðl=T Þ, with l ¼ o T 1=2 ,
where the O(l/T) term is due to the end block effect (see Corradi and Swanson
(2007b) for further discussion).
56 Density and Conditional Distribution-Based Specification Analysis 1551
!
1 X M Xyt, i, h Xt1 Xt
f^M, h Xt Xt1 , y ¼ K ,
MxM i¼1 xM
1X t
y^t, M, h ¼ arg max ln f^M, h Xl Xl1 ; y tM
y2Y t l¼2
f M, h Xl Xl1 , y
^
1X T
∇y f^M, h ðXl0 jXl0 1 ; yÞ
y0 y¼^yt, M, h
T l0 ¼2 f^M, h Xl0 jXtl0 y
tM f^M, h Xl0 Xl0 1 ; y^t, M, h
þt0M f^M, h Xl0 Xl0 1 ; y^t, M, h
∇y f^M, h ðXl0 jXl0 1 ; yÞ ^yt, M, h
!
lnf^M, h ðXl0 jXl0 1 y^t, M, h Þ ,
where tM 0 (
) denotes the derivative of tM(
) with respect to its argument. Note
that each term in the simulated likelihood is recentered around the (full)
sample mean of the score, evaluated at y^t, M, h. This ensures that the bootstrap
score has mean zero, conditional on the sample.
The recentering term requires
computation of ∇y f^M, h Xl0 Xl0 1 ; y^t, M, h , which is not known in closed
form. Nevertheless, it can be computed numerically, by simply taking the
numerical derivative of the simulated likelihood.
Bootstrap Estimators for Multifactor Model
The SGMM and the bootstrap SGMM estimators in the case of multifactor
model are similar as in one-factor model. The difference is that the simulation
schemes (56.12) and (56.13) are used instead of Eq. 56.11.
For recursive NPSQML estimators, to construct the bootstrap counterpart
y^t, M, L, h of y^t, M, L, h , since M/T ! 1 and L/T ! 1, the contribution of
simulation error is asymptotically negligible. Hence, there is no need to resample
the simulated observations or the simulated initial values for volatility. Define
1552 D. Duong and N.R. Swanson
0 1
y y
1 XL
1 XM X t, i, h X t1 ; V j Xt
f^M, L, h Xt Xt1 , y ¼
K@ A:
L j¼1 MxM i¼1 xM
1 X
t
y^t, M, L, h ¼ arg max
y2Y t l¼2
logf^t, M, L, h Xl Xl1 , y tM f^t, M, L, h Xl Xl1 , y
1X T
∇y f^t, M, L, h ðXl0 jXl0 1 ; yÞ
y0 yt, M, L, h
T l0 ¼2 f^t, M, L, h Xl0 Xtl0 ; y
tM ¼ f^t, M, L, h Xl0 Xl0 1 ; y^t, M, L, h
þt0M f^t, M, L, h Xl0 Xl0 1 ; y^t, M, L, h
∇y f^t, M, L, h ðXl0 jXl0 1 ; yÞ ^yt, M, L, h
!!
lnf^t, M, L, h ðXl0 jXl0 1 , y^t, M, L, h Þ ,
where tM 0 (
) denotes the derivative with respect to its argument.
Of note is that each bootstrap term is recentered around the (full) sample
mean. This is necessary because the bootstrap statistic is constructed using
the last P resampled observations, which in turn have been resampled from
the full sample. In particular, this is necessary regardless of the ratio, P/R. In
addition, in the case P/R ! 0, so that there is no need to mimic parameter
estimation error, the bootstrap statistics can be constructed using y^t, M, L, h
instead of y^t, M, L, h .
Step 3: Using the same set of random variables used in the construction of the
y^ N, h y^t, N , h
actual statistics, construct Xi,t,tþt , or X k, i, tþt, , i ¼ 1, . . ., S and t ¼ 1,. . ., T t.
Note that we do not need resample the simulated series (as L/T ! 1,
simulation error is asymptotically negligible). Instead, simulate the
series using bootstrap estimators and using bootstrapped values as starting
values.
2* * * 2 * *
Step 4: Corresponding bootstrap statistics VT,N,h (or ZT,N,h , Dk,P,S , SV
T,N,h, SZT,N,h,
*
SDk,P,S depending on the types of tests) which are built on y^t, N, h y^t, N, L, h Þ then
are followed correspondingly. For the numerical implementation, again, of
important note is that in the case where we pick the choice P/R ! 0, P, T,
56 Density and Conditional Distribution-Based Specification Analysis 1553
R ! 1, there is no need to re-estimate y^t, N, h y^t, N, L, h . y^t, N, h y^t, N, L, h can be
used in all the bootstrap replications.
Step 5: Repeat the bootstrap Steps 1–4 B times, and generate the empirical
distribution of the B bootstrap statistics.
ZT ¼ sup jZ T ðvÞj,
v2V
1554 D. Duong and N.R. Swanson
Table 56.1 Specification test hypotheses of continuous time spot rate processa
Reference and
Model Specification data Hypothesis
pffiffiffiffiffiffiffiffiffiffi
Wong drðtÞ ¼ ða l rðtÞÞdt þ ar ðtÞdW r ðtÞ Corradi and H1
Swanson (2005)
Simulated data
Bhardwaj
et al. (2008)
Eurodollar rate H2
(1971–2005)
pffiffiffiffiffiffiffiffiffi
CIR drðtÞ ¼ kr ðr rðtÞÞdt þ V ðtÞdW r ðtÞ, Cai and Swanson
(2011)
Eurodollar Rate H2, H3
(1971–2008)
Cai and Swanson
(2011)
CEV drðtÞ ¼ kr ðr rðtÞÞdt þ sr rðtÞr dW r ðtÞ Eurodollar rate H2, H3
(1971–2008)
pffiffiffiffiffiffiffiffiffi
SV drðtÞ ¼ kr ðr rðtÞÞdt þ V ðtÞdW r ðtÞ, Bhardwaj H2
et al. (2008)
pffiffiffiffiffiffiffiffiffi
dV ðtÞ ¼ kr ðv V ðtÞÞdt þ sv V ðtÞdW v ðtÞ, Cai and Swanson H2, H3
(2011)
pffiffiffiffiffiffiffiffiffi
SVJ drðtÞ ¼ kr ðr rðtÞÞdt þ V ðtÞdW r ðtÞ þ J u dqu J d dqd , Bhardwaj H2
et al. (2008)
pffiffiffiffiffiffiffiffiffi
dV ðtÞ ¼ kv ðv V ðtÞÞdt þ sv V ðtÞdW v ðtÞ, Cai and Swanson H2, H3
(2011)
pffiffiffiffiffiffiffiffiffi
CHEN drðtÞ ¼ kr ðyðtÞ rðtÞÞdt þ V ðtÞdW r , Cai and Swanson
(2011)
pffiffiffiffiffiffiffiffiffi
dV ðtÞ ¼ kv ðv V ðtÞÞdt þ sv V ðtÞdW v ðtÞ, Eurodollar rate H2, H3
pffiffiffiffiffiffiffiffi
dyðtÞ ¼ ky y yðtÞ dt þ sy yðtÞdW y ðtÞ, (1971–2008)
pffiffiffiffiffiffiffiffiffi
CHENJ drðtÞ ¼ kr ðyðtÞ rðtÞÞdt þ V ðtÞdW r ðtÞ þ J u dqu J d dqd , Cai and Swanson
(2011)
pffiffiffiffiffiffiffiffiffi
dV ðtÞ ¼ kv ðv V ðtÞÞdt þ sv V ðtÞdW v ðtÞ, Eurodollar rate H2, H3
pffiffiffiffiffiffiffiffi
dyðtÞ ¼ ky y yðtÞ dt þ sy yðtÞdW y ðtÞ, (1971–2008)
a
Note that the third column, “Reference and data,” provides the referenced papers and data used in
empirical applications. In the fourth column, H1, H2, and H3 denote Hypothesis 1, Hypothesis
2, and Hypothesis 3, respectively. The hypotheses are presented corresponding to the references in
the third column. For example, for CIR model, H2 corresponds to Bhardwaj et al. (2008) and H2
and H3 correspond to Cai and Swanson (2011)
where
1 X Tt
ZT ðvÞ ¼ pffiffiffiffiffiffiffiffiffiffiffi 1fXt vg
T t t¼1
!
S n o
1X y^T , N, h
1 u Xs, tþt u 1 u Xtþt u ,
S s¼1
56 Density and Conditional Distribution-Based Specification Analysis 1555
and
Z T ¼ sup Z T ðvÞ ,
v2V
Where
1 X T t
Z T ðvÞ ¼ pffiffiffiffiffiffiffiffiffiffiffi 1 Xt v
T t t¼1
!
S
1X
y^T , N, h
1 u Xs, tþt, u 1 u Xtþt u
S s¼1
1 X T t
pffiffiffiffiffiffiffiffiffiffiffi 1fXt vg:
T t t¼1
!
S n o
1X y^T , N , h
1 u Xs, tþt u 1 u Xtþt u :
S s¼1
where
1 X T t
SZT ðvÞ ¼ pffiffiffiffiffiffiffiffiffiffiffi 1fXt vg
T t t¼1
!
S n o
1 X L X
y^T , N, h
1 u Xj, s, tþt u 1 u Xtþt u ,
LS j¼1 s¼1
where
S
1 X T t
1 XL X
y^
Bhardwaj et al. (2008) carry out these tests using t-step ahead confidence inter-
vals. They set t ¼ {1, 2, 4, 12} which corresponds
to 1-week, 2-week,
1-month, and
one-quarter ahead intervals and set uu ¼ X 0:5sX , X sX , covering 46.3 %
and 72.4 % coverage, respectively. X and sX are the mean and variance of an initial
sample of data. In addition, S ¼ {10T, 20T} and l ¼ {5, 10, 20, 50}.
For illustrative purposes, we report one case from Bhardwaj et al. (2008). The test is
implemented by setting S ¼ 10T and l ¼ 25 for the calculation of both ZT and SZT. In
Table 56.2, single, double, and triple starred entries represent rejection using 20 %,
10 %, and 5 % size tests, respectively. Not surprisingly, the findings are consistent with
some other papers in the specification test literature such as Aı̈t-Sahalia (1996) and
Bandi (2002). Namely, the CIR model is rejected using 5 % size tests in almost all cases.
When considering SV and SVJ models, smaller confidence intervals appear to lead to
more model rejections. Moreover, results are somewhat mixed when evaluating the
SVJ model, with a slightly higher frequency of rejection than in the case of SV models.
Finally, turning to Hypothesis 3, Cai and Swanson (2011) use an extended
version of the above dataset, i.e., the 1-month Eurodollar deposit rate from January
1971 to April 2008 (1,996 weekly observations). Specifically, they examine
whether the Chen model is the “best” model amongst multiple alternative models
including those outlined in Table 56.1. The answer is “yes.” In this example, the test
was implemented using Dk,p,N(u1, u2), as described in Sects. 56.3.1 and 56.3.2,
where P ¼ T/2 and predictions are constructed using recursively estimated models
and the simulation sample length used to address latent variable initial values
is set
at L ¼ 10T. The choice of other inputs to the test such as t and interval u, u is the
same as in Bhardwaj et al. (2008). The number of replications S, the block length l,
and number of bootstrap replications are S ¼ 10T, l ¼ 20, and B ¼ 100.
Cai and Swanson (2011) also compare the Chen model with the so-called smooth
transition autoregression (STAR) model defined as follows:
whereut is a disturbance term; y1, b1,g, b2, and c are constants; G(·) is the logistic
CDF i:e:; G; ðg; zt ; cÞ; ¼; 1þexpð1gðzt cÞÞ and the number of lags; and p is selected via
the use of Schwarz information criterion. Test statistics and predictive density-type
“mean square forecast error” (MSFEs) values are again calculated as in Sects.
56.3.1 and 56.3.2.24 Their results indicate that at a 90 % level of confidence, one
cannot reject the null hypothesis that the Chen model generates predictive densities
at least as accurate as the STAR model, regardless of forecast horizon and confi-
dence interval width. Moreover, in almost all cases, the Chen model has lower
MSFE, and the magnitude of the MSFE differential between the Chen model and
STAR model rises as the forecast horizon increases. This confirms their in-sample
findings that the Chen model also wins when carrying out in-sample tests.
24
See Table 6 in Cai and Swanson (2011) for complete details.
56
Table 56.2 Empirical illustration of specification testing – CIR, SV, SVJ modelsa
u, u CIR SV SVJ
ZT 5 % CV 10 % CV SZT 5 % CV 10 % CV SZT 5 % CV 10 % CV
l ¼ 25
1 X 0:5sX 0.5274*** 0.2906 0.3545 0.9841*** 0.8729 0.9031 1.1319 1.8468 2.1957
X sX 0.4289*** 0.2658 0.3178 0.6870 0.6954 0.7254 1.2272* 1.1203 1.3031
***
2 X 0:5sX 0.6824 0.4291 0.4911 0.4113 1.3751 1.4900 0.9615* 0.8146 1.1334
*
X sX 0.4897 0.4264 0.5182 0.3682 1.1933 1.2243 1.2571 1.3316 1.4096
4 X 0:5sX 0.8662** 0.7111 0.8491 1.2840 2.3297 2.6109 1.5012* 1.1188 1.6856
X sX 0.8539* 0.7512 0.9389 1.0472 2.2549 2.2745 0.9901* 0.9793 1.0507
12 X 0:5sX 1.1631* 1.0087 1.3009 1.7687 4.9298 5.2832 2.4237* 2.0818 3.0640
X sX 1.0429 1.4767 2.0222 1.7017 5.2601 5.6522 1.4522 1.7400 2.1684
a
Tabulated entries are test statistics and 5 %, 10 %, and 20 % level critical values. Test intervals are given in the second column of the table, for t ¼ 1, 2, 4, 12.
All tests are carried out using historical 1-month Eurodollar deposit rate data for the period January 1971–September 2005, measured at a weekly frequency.
* **
Density and Conditional Distribution-Based Specification Analysis
, , and *** denote rejection at the 20 %, 10 %, and 5 % levels, respectively. Additionally, X and sX are the mean and standard deviation of the historical data.
See above for complete details
1557
1558 D. Duong and N.R. Swanson
56.5 Conclusion
This chapter reviews a class of specification and model selection-type tests devel-
oped by Corradi and Swanson (2005), Bhardwaj et al. (2008), and Corradi and
Swanson (2011) for continuous time models. We begin with outlining the setup
used to specify the types of diffusion models considered in this chapter. Thereafter,
diffusion models in finance are discussed, and testing procedures are outlined.
Related testing procedures are also discussed, both in contexts where models are
assumed to be either correctly specified under the null hypothesis or generically
misspecified under both the null and alternative test hypotheses. In addition to
discussing tests of correct specification and test for selecting amongst alternative
competing models, using both in-sample methods and via comparison of predictive
accuracy, methodology is outlined allowing for parameter estimation, model and
data simulation, and bootstrap critical value construction.
Several extensions that are left to future research are as follows. First, it remains
to construct specification tests that do not integrate out the effects of latent factors.
Additionally, it remains to examine the finite sample properties of the estimators
and bootstrap methods discussed in this chapter.
Acknowledgments The authors thank the editor, Cheng-Few Lee, for many useful suggestions
given during the writing of this chapter. Duong and Swanson would like to thank the Research
Council at Rutgers University for research support.
References
Ait-Sahalia, Y. (1999). Transition densities for interest rate and others nonlinear diffusions.
Journal of Finance, LIV, 1361–1395.
Aїt-Sahalia, Y. (1996). Testing continuous time models of the spot interest rate. Review of
Financial Studies, 9, 385–426.
Aїt-Sahalia, Y. (2002). Maximum likelihood estimation of discretely sampled diffusions: A closed
form approximation approach. Econometrica, 70, 223–262.
Aїt-Sahalia, Y. (2007). Estimating continuous-time models using discretely sampled data. In
R. Blundell, P. Torsten, & W. K. Newey (Eds.), Econometric Society World Congress invited
lecture, in Advances in economics and econometrics, theory and applications, Ninth World
Congress, Econometric Society Monographs. Cambridge University Press.
Aїt-Sahalia, Y., Fan, J., & Peng, H. (2009). Nonparametric transition-based tests for diffusions.
Journal of the American Statistical Association, 104, 1102–1116.
Altissimo, F., & Mele, A. (2009). Simulated nonparametric estimation of dynamic models with
application in finance. Review of Economic Studies, 76, 413–450.
Andersen, T. G., & Lund, J. (1997). Estimating continuous time stochastic volatility models of
short term interest rates. Journal of Econometrics, 72, 343–377.
Andersen, T. G., Benzoni, L., & Lund, J. (2004). Stochastic volatility, mean drift, and jumps in the
short-term interest rate (Working paper). Northwestern University.
Andrews, D. W. K. (1993). An introduction to econometric applications of empirical process
theory for dependent random variables. Econometric Reviews, 12, 183–216.
Andrews, D. W. K. (1997). A conditional Kolmogorov test. Econometrica, 65, 1097–1128.
Bai, J. (2003). Testing parametric conditional distributions of dynamic models. Review of
Economics and Statistics, 85, 531–549.
56 Density and Conditional Distribution-Based Specification Analysis 1559
Bandi, F. (2002). Short-term interest rate dynamics: A spatial approach. Journal of Financial
Economics, 65, 73–110.
Beckers, S. (1980). The constant elasticity of variance model and US implications for option
pricing. Journal of Finance, 35, 661–673.
Bhardwaj, G., Corradi, V., & Swanson, N. R. (2008). A simulation based specification test for
diffusion processes. Journal of Business and Economic Statistics, 26, 176–193.
Black, F., & Scholes, M. (1973). The pricing of options and corporate liabilities. Journal of
Political Economy, 81, 637–654.
Bontemps, C., & Meddahi, N. (2005). Testing normality: A GMM approach. Journal of Econo-
metrics, 124, 149–186.
Brennan, M. J., & Schwartz, E. S. (1979). A continuous-time approach to the pricing bonds.
Journal of Banking and Finance, 3, 133–155.
Cai, L., & Swanson, N. R. (2011). An empirical assessment of spot rate model stability. Journal of
Empirical Finance, 18, 743–764.
Chan, K. C., Karolyi, G. A., Longstaff, F. A., & Sanders, A. B. (1992). An empirical comparison
of alternative models of the short-term interest rate. Journal of Finance, 47, 1209–1227.
Chen, L. (1996). Stochastic mean and stochastic volatility – A three-factor model of term structure
of interest rates and its application to the pricing of interest rate derivatives. Oxford, UK:
Blackwell.
Clements, M. P., & Smith, J. (2000). Evaluating the forecast densities of linear and nonlinear
models: Applications to output growth and unemployment. Journal of Forecasting, 19,
255–276.
Clements, M. P., & Smith, J. (2002). Evaluating multivariate forecast densities: A comparison of
two approaches. International Journal of Forecasting, 18, 397–407.
Corradi, V., & Swanson, N. R. (2005). A bootstrap specification test for diffusion processes.
Journal of Econometrics, 124, 117–148.
Corradi, V., & Swanson, N. R. (2006). Predictive density and conditional confidence interval
accuracy tests. Journal of Econometrics, 135, 187–228.
Corradi, V., & Swanson, N. R. (2007a). Nonparametric bootstrap procedures for predictive
inference based on recursive estimation schemes. International Economic Review February,
2007(48), 67–109.
Corradi, V., & Swanson, N. R. (2007b). Nonparametric bootstrap procedures for
predictive inference based on recursive estimation schemes. International Economic Review,
48, 67–109.
Corradi, V., & Swanson, N. R. (2011). Predictive density construction and accuracy testing with
multiple possibly misspecified diffusion models. Journal of Econometrics, 161, 304–324.
Corradi, V. & Swanson, N. R. (2002). A consistent test for nonlinear out of sample predictive
accuracy. Journal of Econometrics, 110, 353–381.
Courtadon, G. (1982). The pricing of options on default-free bonds. Journal of Financial and
Quantitative Analysis, 17, 75–100.
Cox, J. C., Ingersoll, J. E., & Ross, S. A. (1985). A theory of the term structure of interest rates.
Economerica, 53, 385–407.
Dai, Q., & Singleton, K. J. (2000). Specification analysis of affine term structure models. Journal
of Finance, 55, 1943–1978.
Diebold, F. X., & Li, C. (2006). Forecasting the term structure of government bond yields. Journal
of Econometrics, 130, 337–364.
Diebold, F. X., & Mariano, R. S. (1995). Comparing predictive accuracy. Journal of Business and
Economic Statistics, 13, 253–263.
Diebold, F. X., Gunther, T., & Tay, A. S. (1998a). Evaluating density forecasts with applications to
finance and management. International Economic Review, 39, 863–883.
Diebold, F. X., Tay, A. S., & Wallis, K. D. (1998b). Evaluating density forecasts of inflation: The
survey of professional forecasters. In R. F. Engle & H. White (Eds.), Festschrift in honor of
C.W.J. Granger. Oxford: Oxford University Press.
1560 D. Duong and N.R. Swanson
Diebold, F. X., Hahn, J., & Tay, A. S. (1999). Multivariate density forecast evaluation and
calibration in financial risk management: High frequency returns on foreign exchange. Review
of Economics and Statistics, 81, 661–673.
Duan, J. C. (2003). A specification test for time series models by a normality transformation
(Working paper). University of Toronto.
Duffie, D., & Kan, R. (1996). A yield-factor model of interest rates. Mathematical Finance, 6,
379–406.
Duffie, D., & Singleton, K. (1993). Simulated moment estimation of Markov models of asset
prices. Econometrica, 61, 929–952.
Duffie, D., Pan, J., & Singleton, K. (2000). Transform analysis and asset pricing for affine
jump-diffusions. Econometrica, 68, 1343–1376.
Duong, D., & Swanson, N. R. (2010). Empirical evidence on jumps and large fluctuations in
individual stocks (Working paper). Rutgers University.
Duong, D., & Swanson, N. R. (2011). Volatility in discrete and continuous time models: A survey
with new evidence on large and small jumps. Missing Data Methods: Advances in
Econometrics, 27B, 179–233.
Emanuel, D., & Mcbeth, J. (1982). Further results on the constant elasticity of variance call option
pricing model. Journal of Financial Quantitative Analysis, 17, 533–554.
Fermanian, J. D., & Salanié, B. (2004). A non-parametric simulated maximum likelihood
estimation method. Econometric Theory, 20, 701–734.
Gallant, A. R., & Tauchen, G. (1996). Which moments to match. Econometric Theory, 12,
657–681.
Gallant, A. R., & Tauchen, G. (1997). Estimation of continuous time models for stock returns and
interest rates. Macroeconomic Dynamics, 1, 135–168.
Goncalves, S., & White, H. (2002). The bootstrap of the mean for dependent heterogeneous arrays.
Econometric Theory, 18, 1367–1384.
Hall, A. R., & Inoue, A. (2003). The large sample behavior of the generalized method of moments
estimator in misspecified models. Journal of Econometrics, 114, 361–394.
Hansen, B. E. (1996). Inference when a nuisance parameter is not identified under the null
hypothesis. Econometrica, 64, 413–430.
Heston, S. L. (1993). A closed form solution for option with stochastic volatility with applications
to bond and currency options. Review of Financial Studies, 6, 327–344.
Hong, Y. (2001). Evaluation of out-of-sample probability density forecasts with applications to
S&P 500 stock prices (Working paper). Cornell University.
Hong, Y. M., & Li, H. T. (2005). Nonparametric specification testing for continuous time models
with applications to term structure interest rates. Review of Financial Studies, 18, 37–84.
Hong, Y. M., Li, H. T., & Zhao, F. (2007). Can the random walk model be beaten in out-of-sample
density forecasts? Evidence from intraday foreign exchange rates. Journal of Econometrics,
141, 736–776.
Hull, J., & White, A. (1987). The pricing of options on assets with stochastic volatility. Journal of
Finance, 42, 281–300.
Inoue. (2001). Testing for distributional change in time series. Econometric Theory, 17, 156–187.
Jiang, G. J. (1998). Nonparametric modelling of US term structure of interest rates and implica-
tions on the prices of derivative securities. Journal of Financial and Quantitative Analysis, 33,
465–497.
Karlin, S., & Taylor, H. M. (1981). A second course in stochastic processes. San Diego: Academic.
Kloeden, P. E., & Platen, E. (1999). Numerical solution of stochastic differential equations.
New York: Springer.
Kolmogorov, A. N. (1933). Sulla Determinazione Empirica di una Legge di Distribuzione.
Giornale dell’Istitutodegli Attuari, 4, 83–91.
Kristensen, D., & Shin, Y. (2008). Estimation of dynamic models with nonparametric simulated
maximum likelihood (Creates Research paper 2008-58). University of Aarhus and Columbia
University.
56 Density and Conditional Distribution-Based Specification Analysis 1561
K€unsch, H. R. (1989). The Jackknife and the bootstrap for general stationary observations. Annals
of Statistics, 17, 1217–1241.
Marsh, T., & Rosenfeld, E. (1983). Stochastic processes for interest rates and equilibrium bond
prices. Journal of Finance, 38, 635–646.
Meddahi, N. (2001). An Eigenfunction approach for volatility modeling (Working paper).
University of Montreal.
Merton, C. R. (1973). Theory of rational option pricing. The Bell Journal of Economics and
Management Science, 4, 141–183.
Nelson, D. B. (1990). ARCH as diffusion approximations. Journal of Econometrics, 45, 7–38.
Pardoux, E., & Talay, D. (1985). Discretization and simulation of stochastic differential equations.
Acta Applicandae Mathematicae, 3, 23–47.
Piazzesi, M. (2001). Macroeconomic jump effects and the yield curve (Working paper). UCLA.
Piazzesi, M. (2004). Affine term structure models. Manuscript prepared for the Handbook of
Financial Econometrics. University of California at Los Angeles.
Piazzesi, M. (2005). Bond yields and the federal reserve. Journal of Political Economy, 113,
311–344.
Politis, D. N., Romano, J. P., & Wolf, M. (1999). Subsampling. New York: Springer.
Pritsker, M. (1998). Nonparametric density estimators and tests of continuous time interest rate
models. Review of Financial Studies, 11, 449–487.
Rosenblatt, M. (1952). Remarks on a multivariate transformation. Annals of Mathematical
Statistics, 23, 470–472.
Singleton, K. J. (2006). Empirical dynamic asset pricing – Model specification and econometric
assessment. Princeton: Princeton University Press.
Smirnov, N. (1939). On the estimation of the discrepancy between empirical curves of distribution
for two independent samples. Bulletin Mathematique de l’Universite’ de Moscou, 2, fasc. 2.
Sullivan, R., Timmermann, A., & White, H. (1999). Data-snooping, technical trading rule perfor-
mance, and the bootstrap. The Journal of Finance, 54(5), 1647–1691.
Sullivan, R., Timmermann, A., & White, H. (2001). Dangers of data-driven inference: The case of
calendar effects in stock returns. Journal of Econometrics, 105, 249–286.
Thompson, S. B. (2008). Identifying term structure volatility from the LIBOR-swap curve. Review
of Financial Studies, 21, 819–854.
van der Vaart, A. (1998). Asymptotic statistics. Cambridge, MA: Cambridge University Press.
Vasicek, O. A. (1977). An equilibrium characterization of the term structure. Journal of Financial
Economics, 5, 177–188.
Vuong, Q. (1989). Likelihood ratio tests for model selection and non-nested hypotheses.
Econometrica, 57, 307–333.
White, H. (1982). Maximum likelihood estimation of misspecified models. Econometrica, 50,
1–25.
White, H. (2000). A reality check for data snooping. Econometrica, 68, 1097–1126.
Wong, E. (1964). The construction of a class of stationary Markov processes. In R. Bellman (Ed.),
Sixtenn symposia in applied mathematics: Stochastic processes in mathematical physics and
engineering. Providence: American Mathematical Society.
Assessing the Performance of Estimators
Dealing with Measurement Errors 57
Heitor Almeida, Murillo Campello, and Antonio F. Galvao
Contents
57.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1564
57.2 Dealing with Mismeasurement: Alternative Estimators . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1570
57.2.1 The Erickson–Whited Estimator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1570
57.2.2 An OLS-IV Framework . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1574
57.2.3 GMM Estimator . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1578
57.3 Monte Carlo Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1581
57.3.1 Monte Carlo Design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1582
57.3.2 The EW Identification Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1585
57.3.3 Bias and Efficiency of the EW, OLS-IV, and AB-GMM Estimators . . . . . . . 1587
57.3.4 Heteroscedasticity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1593
57.3.5 Identification of the EW Estimator in Panel Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1594
57.3.6 Revisiting the OLS-IV Assumptions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1598
57.3.7 Distributional Properties of the EW and OLS-IV Estimators . . . . . . . . . . . . . . . 1600
57.4 Empirical Application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1602
57.4.1 Theoretical Expectations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1603
57.4.2 Data Description . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1604
57.4.3 Testing for the Presence of Fixed Effects and Heteroscedasticity . . . . . . . . . . . 1604
57.4.4 Implementing the EW Identification Test . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1605
This article is based on Almeida, H., M. Campello, and A. Galvao, 2010, Measurement Errors in
Investment Equations, Review of Financial Studies, 23, pages 3279–3382.
H. Almeida (*)
University of Illinois at Urbana-Champaign, Champaign, IL, USA
e-mail: halmeida@illinois.edu
M. Campello
Cornell University, Ithaca, NY, USA
e-mail: campello@cornell.edu
A.F. Galvao
University of Iowa, Iowa City, IA, USA
e-mail: antonio-galvao@uiowa.edu
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1563
DOI 10.1007/978-1-4614-7750-1_57,
# Springer Science+Business Media New York 2015
1564 H. Almeida et al.
Abstract
We describe different procedures to deal with measurement error in linear
models and assess their performance in finite samples using Monte Carlo
simulations and data on corporate investment. We consider the standard instru-
mental variable approach proposed by Griliches and Hausman (Journal of
Econometrics 31:93–118, 1986) as extended by Biorn (Econometric Reviews
19:391–424, 2000) [OLS-IV], the Arellano and Bond (Review of Economic
Studies 58:277–297, 1991) instrumental variable estimator, and the higher-order
moment estimator proposed by Erickson and Whited (Journal of Political Econ-
omy 108:1027–1057, 2000, Econometric Theory 18:776–799, 2002). Our anal-
ysis focuses on characterizing the conditions under which each of these
estimators produce unbiased and efficient estimates in a standard “errors-in-
variables” setting. In the presence of fixed effects, under heteroscedasticity, or in
the absence of a very high degree of skewness in the data, the EW estimator is
inefficient and returns biased estimates for mismeasured and perfectly measured
regressors. In contrast to the EW estimator, IV-type estimators (OLS-IV and
AB-GMM) easily handle individual effects, heteroscedastic errors, and different
degrees of data skewness. The IV approach, however, requires assumptions
about the autocorrelation structure of the mismeasured regressor and the mea-
surement error. We illustrate the application of the different estimators using
empirical investment models. Our results show that the EW estimator produces
inconsistent results when applied to real-world investment data, while the IV
estimators tend to return results that are consistent with theoretical priors.
Keywords
Investment equations • Measurement error • Monte Carlo simulations • Instru-
mental variables • GMM • Bias • Fixed effects • Heteroscedasticity • Skewness •
High-order moments
57.1 Introduction
OLS estimators are the workhorse of empirical research in many fields in applied
economics. Researchers see a number of advantages in these estimators. Most
notably, they are easy to implement and the results they generate are easy
to replicate. Another appealing feature of OLS estimators is that they easily
accommodate the inclusion of individual (e.g., firm and time) idiosyncratic effects.
Despite their popularity, however, OLS estimators are weak in dealing with the
problem of errors in variables. When the independent (right-hand side) variables of
an empirical model are mismeasured, coefficients estimated via standard OLS are
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1565
1
Naturally, if extraneous instruments are available, they can help solve the identification problem.
See Rauh (2006) for the use of discontinuities in pension contributions as a source of variation in
cash flows in an investment model. Bond and Cummins (2000) use information contained in
financial analysts’ forecasts to instrument for investment demand.
2
Lags of the well-measured variable may also be included in the instrument set if they are believed
to also contain information about the mismeasured one.
3
See, among others, Biorn (2000), Wansbeek (2001), and Xiao et al. (2008).
1566 H. Almeida et al.
use of shorter lags of observed variables in the instrument set, as the information
contained in short lags may be correlated with the current value of the measurement
error.
The EW estimator is based on high-order moments of residuals obtained by
“partialling out” perfectly measured regressors from the dependent, observed
mismeasured, and latent variables, as well as high-order moments of the innova-
tions of the model. The key idea is to create a set of auxiliary equations as a function
of these moments and cross-moments. Implementation then requires a high degree
of skewness in the distribution of the partialled out latent variable. Our analysis also
shows that the presence of individual fixed effects and heteroscedasticity also
impacts the performance of the EW estimator, particularly so if they are both
present in the data process.
We perform a series of Monte Carlo simulations to compare the performance of
the EW and IV estimators in finite samples. Emulating the types of environments
commonly found by empirical researchers, we set up a panel data model with
individual fixed effects and potential heteroscedasticity in the errors. Monte Carlo
experiments enable us to study those estimators in a “controlled environment,”
where we can investigate the role played by each element (or assumption) of an
estimator in evaluating its performance. Our simulations compare the EW and IV
(OLS-IV and AB-GMM) estimators in terms of bias and root mean squared error
(RMSE), a standard measure of efficiency.
We consider several distributional assumptions to generate observations and
errors. Experimenting with multiple distributions is important because researchers
often find a variety of distributions in real-world applications and because one
ultimately does not observe the distribution of the mismeasurement term. Since the
EW estimator is built around the notion of skewness of the relevant distributions,
we experiment with three skewed distributions (lognormal, chi-square, and
F-distribution), using the standard normal (non-skewed) as a benchmark. The
simulations also allow for significant correlation between mismeasured and well-
measured regressors (as in Erickson and Whited 2000, 2002), so that the attenuation
bias of the mismeasured regressor affects the coefficient of the well-measured
regressor.
Our simulation results can be summarized as follows. First, we examine the
identification test proposed by Erickson and Whited (2002). This is a test that
the data contain a sufficiently high degree of skewness to allow for the
identification of their model. We study the power of the EW identification
test by generating data that do not satisfy its null hypothesis of non-skewness.
In this case, even for the most skewed distribution (lognormal), the test rejects
the null hypothesis only 47 % of the time – this is far less than desirable, given
that the null is false. The power of the test becomes even weaker after we treat
the data for the presence of fixed effects in the true model (“within transfor-
mation”). In this case, the rejection rate under the lognormal distribution drops
to 43 %. The test’s power declines even further when we consider alternative
skewed distributions (chi-square and F-distributions). The upshot of this first
set of experiments is that the EW model too often rejects data that are
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1567
generated to fit its identifying assumptions. These findings may help explain
some of the difficulties previous researchers have reported when attempting to
implement the EW estimator.
We then study the bias and efficiency of the EW and IV estimators. Given that
the true model contains fixed effects, it is appropriate to apply the within transfor-
mation to the data. However, because most empirical implementations of the EW
estimator have used data in “level form” (i.e., not treated for the presence of fixed
effects),4 we also experiment with cases in which we do not apply the within
transformation. EW propose three estimators that differ according to the number
of moments used: GMM3, GMM4, and GMM5. We consider all of them in our
experiments.
In a first round of simulations, we impose error homoscedasticity. When we
implement the EW estimator with the data in level form, we find that the
coefficients returned are significantly biased even when the data have a high
degree of skewness (i.e., under the lognormal case, which is EW’s preferred
case). Indeed, for the mismeasured regressors the EW biases are in excess of
100 % of their true value. As should be expected, the performance of the EW
estimator improves once the within transformation is used. In the case of the
lognormal distribution, the EW estimator bias is relatively small. In addition,
deviations from the lognormal assumption tend to generate significant biases for
the EW estimator.
In a second round of simulations, we allow for heteroscedasticity in the data.
We focus our attention on simulations that use data that are generated using
a lognormal distribution after applying the within transformation (the best case
scenario for the EW estimator). Heteroscedasticity introduces heterogeneity to the
model and consequently to the distribution of the partialled out dependent
variable, compromising identification in the EW framework. The simulations
show that the EW estimator is biased and inefficient for both the mismeasured
and well-measured regressors. In fact, biases emerge even for very small amounts
of heteroscedasticity, where we find biases of approximately 40 % for the
mismeasured regressor. Paradoxically, biases “switch signs” depending on the
degree of heteroscedasticity that is allowed for in the model. For instance, for
small amounts of heteroscedasticity, the bias of the mismeasured regressor is
negative (i.e., the coefficient is biased downwards). However, the bias turns
positive for a higher degree of heteroscedasticity. Since heteroscedasticity is
a naturally occurring phenomenon in corporate data, our simulations imply that
empirical researchers might face serious drawbacks when using the EW
estimator.
Our simulations also show that, in contrast to the EW estimator, the bias in the
IV estimates is small and insensitive to the degree of skewness and heterosce-
dasticity in the data. Focusing on the OLS-IV estimator, we consider the case of
time-invariant correlation in the error structure and use the second lag of the
4
Examples are Whited (2001, 2006), Hennessy (2004), and Colak and Whited (2007).
1568 H. Almeida et al.
5
The results for the Arellano–Bond GMM estimator are similar to those of the OLS-IV estimator.
To save space and because the OLS-IV estimator is easier to implement, we focus on this
estimator.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1569
covariance), we should expect the q coefficient to go up and the cash flow coeffi-
cient to go down, when compared to standard (likely biased) OLS estimates.
Second, we would expect the q and cash flow coefficients to be nonnegative after
addressing the problem of mismeasurement. If the original q-theory of investment
holds and the estimator does a good job of addressing mismeasurement, then the
cash flow coefficient would be zero. Alternatively, the cash flow coefficient could
be positive because of financing frictions.6
Using data from Compustat from 1970 to 2005, we estimate investment equa-
tions in which investment is regressed on proxies for q and cash flow. Before doing
so, we perform standard tests to check for the presence of individual fixed effects
and heteroscedasticity in the data. In addition, we perform the EW identification
test to check whether the data contain a sufficiently high degree of skewness.
Our results are as follows. First, our tests reject the hypotheses that the data do
not contain firm-fixed effects and that errors are homoscedastic. Second, the EW
identification tests indicate that the data fail to display sufficiently high skewness.
These initial tests suggest that the EW estimator is not suitable for standard
investment equation applications. In fact, we find that, when applied to the data,
the EW estimator returns coefficients for q and cash flow that are highly unstable
across different years. Moreover, following the EW procedure for panel models
(which comprises combining yearly cross-sectional coefficients into single esti-
mates), we obtain estimates for q and cash flow that do not satisfy the conditions
discussed above. In particular, EW estimators do not reduce the cash flow coeffi-
cient relative to that obtained by standard OLS, while the q coefficient is never
statistically significant. In addition, those estimates are not robust with respect to
the number of moments used: EW’s GMM3, GMM4, and GMM5 models procedure
results that are inconsistent with one another. These results suggest that the pres-
ence of heteroscedasticity and fixed effects in real-world investment data hampers
identification when using the EW estimator.
In contrast to EW, the OLS-IV procedure yields estimates that are fairly sensi-
ble. The q coefficient goes up by a factor of 3–5, depending on the set of instruments
used. At the same time, the cash flow coefficient goes down by about two-thirds of
the standard OLS value. Similar conclusions apply to the AB-GMM estimator. We
also examine the robustness of the OLS-IV to variations in the set of instruments
used in the estimation, including sets that feature only longer lags of the variables in
the model. The OLS-IV coefficients remain fairly stable after such changes. These
results suggest that real-world investment data likely satisfies the assumptions that
are required for identification of IV estimators.
The remainder of the chapter is structured as follows. We start the next section
discussing in detail the EW estimator, clarifying the assumptions that are needed for
its implementation. Subsequently, we show how alternative IV models deal with
6
See Hubbard (1998) and Stein (2003) for comprehensive reviews. We note that the presence of
financing frictions does not necessarily imply that the cash flow coefficient should be positive. See
Chirinko (1993) and Gomes (2001) for arguments suggesting that financing frictions are not
sufficient to generate positive cash flow coefficients.
1570 H. Almeida et al.
where xi is the observed variable and ei is the measurement error. The observed
variables are yi, zi, and xi; and by substituting Eq. 57.2 in Eq. 57.1, we have
yi ¼ zi a þ xi b þ v i ,
yi zi my ¼ i b þ ui (57.3)
xi zi mx ¼ i þ ei , (57.4)
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1571
where (my, mx, mw) ¼ [E(z0izi)]1E[z0i(yi, xi, wi)] and i wi zimw. For the details of
this derivation, see Erickson and Whited (2002, p. 779). One can then consider
a two-step estimation approach, where the first step is to substitute least squares
Xn 0 1 Xn 0
estimates m^y ; m^x zz
i¼1 i i
z ðy ; x Þ into Eqs. 57.3 and 57.4 to
i¼1 i i i
obtain a lower dimensional errors-in-variables model. The second step consists
of estimating b using GMM using high-order sample moments of yi zi m^y and
xi zi m^x . Estimates of a are then recovered via my ¼ a + mxb. Thus, the estimators
are based on equations giving the moments of yi zimy and xi zimx as functions of
b and the moments of (ui, ei, i).
To give a concrete example of how the EW estimator works, we explore the case
of J ¼ 1. The more general case is discussed below. By substituting
!1
X
n X
n
m^y ; m^x z0i zi zi 0 ðyi ; xi Þ
i¼1 i¼1
into Eqs. 57.1 and 57.2, one can estimate b, E(u2i ), E(e2i ), and E(2i ) via
GMM. Estimates of the lth element of a are obtained by substituting the estimate
of b and the lth elements of m^y and m^x into
al ¼ myl mxl b, l 6¼ 0:
h 2 i
E yi zi m y ¼ b2 E 2i þ E u2i (57.5)
E yi zi my ðxi zi mx Þ ¼ bE 2i (57.6)
h i
E ðxi zi mx Þ2 ¼ E 2i þ E e21 : (57.7)
The left-hand side quantities are consistently estimable, but there are only three
equations with which to estimate four unknown parameters on the right-hand side.
The third-order product moment equations, however, consist of two equations in
two unknowns:
h 2 i
E yi zi my ðxi zi mx Þ ¼ b2 E 3i , (57.8)
h i
E yi zi my ðxi zi mx Þ2 ¼ bE 3i : (57.9)
It is possible to solve these two equations for b. Crucially, a solution exists if the
identifying assumptions b 6¼ 0 and E(3i ) 6¼ 0 are true, and one can test the contrary
1572 H. Almeida et al.
h 2 i
E yi zi my ðxi zi mx Þ2 ¼ b2 E 4i þ E 2i E u2i
þ E u2i E 2i þE e2i , (57.11)
h i
E yi zi my ðxi zi mx Þ3 ¼ b E 4i þ 3E 2i E e2i : (57.12)
The resulting eight-equation system Eqs. 57.5, 57.6, 57.7, 57.8, 57.9, 57.10,
57.11, and 57.12 contains the six unknowns (b, E(u2i ), E(e2i ), E(2i ), E(3i ), E(4i )).
It is possible to estimate this vector by numerically minimizing a quadratic form
that minimizes asymptotic variance.
The conditions imposed by EW imply restrictions on the residual moments of
the observable variables. Such restrictions can be tested using the corresponding
sample moments. EW also propose a test for residual moments that is based on
several assumptions.7 These assumptions imply testable restrictions on the residuals
from the population regression of the dependent and proxy variables on the
perfectly measured regressors. Accordingly, one can develop Wald-type partially
adjusted statistics and asymptotic null distributions for the test. Empirically, one
can use the Wald test statistic and critical values from a chi-square distribution to
test whether the last moments are equal to zero. This is an identification test, and if
in a particular application one cannot reject the null hypothesis, then the model is
unidentified and the EW estimator may not be used. We study the finite sample
performance of this test and its sensitivity to different data-generating processes in
the next section.
It is possible to derive more general forms of the EW estimator. In particular, the
EW estimators are based on the equations for the moments of yi zimy and xi zimx
as functions of b and the moments ui, ei, and i. To derive these equations, write
Eq. 57.3 as yi zimy ¼ ∑ jJ ¼ 1ijbj + ui, where J is the number of well-measured
regressors and the jth equation in Eq. 57.4 as xij zimxj ¼ ij + eij, where mxj is the jth
column of and mx and (ij, eij) is the jth row of (0ij, e0ij). Next write
7
First, the measurement errors, the equation error, and all regressors have finite moments of
sufficiently high order. Second, the regression error and the measurement error must be indepen-
dent of each other and of all regressors. Third, the residuals from the population regression of the
unobservable regressors on the perfectly measured regressors must have a nonnormal distribution.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1573
" !r0 #
h r0 YJ i X
J YJ
rj rj
E yi zi my j¼1
ðxi zi mx Þ ¼E i b þ ui j¼1
ði þ ei Þ ,
j¼1
(57.13)
where (r0, r1,. . ., rJ) are nonnegative integers. After expanding the right-hand side of
Eq. 57.13, using the multinomial theorem, it is possible to write the above moment
condition as
Xn
where g- i ðsÞ ^ is a positive definite matrix. Assuming a
g ðsÞ for all s, and W
t¼1 i
8
number of regularity conditions, the estimator is consistent and asymptotically normal.
It is important to notice that the estimator proposed by Erickson and Whited
(2002) was originally designed for cross-sectional data. To accommodate a panel-
like structure, Erickson and Whited (2000) propose transforming the data before the
estimation using the within transformation or differencing. To mimic a panel
structure, the authors propose the idea of combining different cross-sectional
GMM estimates using a minimum distance estimator (MDE).
The MDE estimator is derived by minimizing the distance between the auxiliary
parameter vectors under the following restrictions:
f b; y^ ¼ Hb y^ ¼ 0,
8
More specifically, these conditions are as follows: (zi, wi, ui, ei) is an independent and identically
distributed sequence; ui and the elements of zi, wi, and ei, have finite moments of every order; (ui, ei)
is independent of (zi, wi), and the individual elements in (ui, ei) are independent of each other;
E(ui) ¼ 0 and E(ei) ¼ 0; E[(zi, wi)0 (zi, wi)] is positive definite; every element of b is nonzero; and the
distribution of satisfies E[(ic)3] 6¼ 0 for every vector of constants c ¼ (c1, ,cJ) having at least
one nonzero element.
1574 H. Almeida et al.
h i 0 h i1 1
V^ b^ ¼ H V^ y^ H :
In this section, we revisit the work of Griliches and Hausman (1986) and Biorn
(2000) to discuss a class of OLS-IV estimators that can help address the errors-in-
variables problem.
Consider the following single-equation model:
where uit is independently and identically distributed, with mean zero and variance
s2u, and Cov(wit, uis) ¼ Cov(gi, uis) ¼ 0 for any t and s, but Cov(gi, wit) 6¼ 0, y is an
observable scalar, w is a 1 K vector, and b is K 1 vector. Suppose we do not
observe wit itself, but rather the error-ridden measure:
9
See Erickson and Whited (2002) Lemma 1 for the definition of their proposed influence function.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1575
yit yit1 ¼ ðxit xit1 Þb þ ½ðuit eit bÞ ðuit1 eit1 bÞ (57.17)
10
A more recent paper by Xiao et al. (2008) also shows how to relax the classical Griliches–
Hausman assumptions for measurement error models.
1576 H. Almeida et al.
This set of assumptions is weaker since (B1) and (B2) allow for a vector moving
average (MA) structure up to order t (1) for the innovations. Alternatively, one
can use the following assumptions:
(C1) E(e0itety) is invariant to t, y, t 6¼ y.
(C2) E(uituiy) is invariant to t, y, t 6¼ y.
Assumptions (C1) and (C2) allow for a different type of autocorrelation,
more specifically they allow for any amount of autocorrelation that is time
invariant. Assumptions (C1) and (C2) will be satisfied if the measurement
errors and the disturbances have individual components, say eit ¼ e1i + e2it,
uit ¼ u1i + u2it, where e1i, e2it, u1i, u2it i.i.d. Homoscedasticity of eit and/or uit across
i and t need not be assumed; the model accommodates various forms of
heteroscedasticity.
Biorn also considers assumptions related to the distribution of the latent regressor
vector wit:
(D1) E(wit) is invariant to t.
(D2) E(giwit) is invariant to t.
Assumptions (D1) and (D2) hold when wit is stationary. Note that wit and i need
not be uncorrelated.
To ensure identification of the slope coefficient vector when panel data are
available, it is necessary to impose restrictions on the second-order moments of
the variables (wit, uit, eit, gi). For simplicity, Biorn assumes that this distribution is
the same across individuals and that the moments are finite. More specifically,
C(wit, wit) ¼ ∑ ww wn 0 ee uu 2
ty , E(wit i) ¼ ∑ t , E(eiteit) ¼ ∑ty, E(uituit) ¼ sty , E( i ) ¼ s , where
C denotes the covariance matrix operator. Then, it is possible to derive the
second-order moments of the observable variables and show that they only depend
on these matrices and the coefficient b.11 In this framework, there is no need to use
assumptions based on higher-order moments.
Biorn proposes several strategies to estimate the slope parameter of interest.
Under the OLS-IV framework, he proposes estimation procedures of two kinds:
• OLS-IV A: The equation is transformed to differences to remove individual
heterogeneity and is estimated by OLS-IV. Admissible instruments for
this case are the level values of the regressors and/or regressands for other
periods.
• OLS-IV B: The equation is kept in level form and is estimated by OLS-IV.
Admissible instruments for this case are differenced values of the regressors
and/or regressands for other periods.
Using moment conditions from the OLS-IV framework, one can define
the estimators just described. In particular, using the mean counterpart and the
moment conditions, one can formally define the OLS-IV A and OLS-IV B
estimators.
If K > 1, the latter estimator is infeasible, but it is possible to modify the former
estimator by replacing one element in x0ip by yip.
The estimator for OLS-IV B (equation in level and instruments in difference) can
be defined as
" #1 " #
XN 0 XN 0
b^ ¼
xðpqÞt Dxipq xit Dxipq y : it
i¼1 i¼1
If K > 1, the latter estimator is infeasible, but it is possible to modify the former
estimator by replacing one element in Dxip by Dyip.
For some applications, it might be useful to impose weaker conditions on the
autocorrelation of measurement errors and disturbances. In this case, it is necessary
to restrict slightly further the conditions on the instrumental variables. More
formally, if one replaces assumptions (A1) and (A2), or (C1) and (C2), by the
weaker assumptions (B1) and (B2), then it is necessary to ensure that the IV set has
a lag of at least t 2 and/or lead of at least t + 1 periods of the regressor in order to
“clear” the t period memory of the MA process. Consistency of these estimators is
discussed in Biorn (2000).12
To sum up, there are two simple ways to relax the standard assumption of i.i.d.
measurement errors. Under the assumption of time-invariant autocorrelation, the set of
instruments can contain the same variables used under Griliches–Hausman.
12
In particular, if jt pj, jy pj > t, then (B1) and rank (E[w0ip(Dwity)]) ¼ K for some
p 6¼ t 6¼ 0 ensure consistency of OLS–IV B, b^xpðtyÞ , and (B2) and the same rank condition ensure
consistency of b^xyðtyÞ . In the same way, if jp tj, q t > t, (B1), (D1), (D2), and rank
(E[(Dwipq)0 wit)]) ¼ K for some p 6¼ q 6¼ t ensure consistency of OLS-IV B, b^xðpqÞt , and (B2),
(D1), (D2), and the same rank condition ensure consistency of b^yðpqÞt .
1578 H. Almeida et al.
For example, if one uses the OLS-IV A estimator (equation in differences and instru-
ments in levels), then twice-lagged levels of the observable variables can be used as
instruments. Under a moving average structure for the innovations in the measurement
error [assumptions (B1) and (B2)], identification requires the researcher to use longer
lags of the observable variables as instruments. For example, if the innovations follow
an MA(1) structure, then consistency of the OLS-IV A estimator requires the use of
instruments that are lagged three periods and longer. Finally, identification requires the
latent regressor to have some degree of autocorrelation (since lagged values are used as
instruments). Our Monte Carlo simulations will illustrate the importance of these
assumptions and will evaluate the performance of the OLS-IV estimator under
different sets of assumptions about the structure of the errors.
Within the broader instrumental variable approach, we also consider an entire class
of GMM estimators that deal with mismeasurement. These GMM estimators are
close to the OLS-IV estimator discussed above but may attain appreciable gains in
efficiency by combining numerous orthogonality conditions [see Biorn (2000) for
a detailed discussion]. GMM estimators that use all the available lags at each period
as instruments for equations in first differences were proposed by Holtz-Eakin
et al. (1988) and Arellano and Bond (1991). We provide a brief discussion in turn.
In the context of a standard investment model, Blundell et al. (1992) use GMM
allowing for correlated firm-specific effects, as well as endogeneity
(mismeasurement) of q. The authors use an instrumental variable approach on
a first-differenced model in which the instruments are weighted optimally so as to
form the GMM estimator. In particular, they use qit2 and twice-lagged investments
as instruments for the first-differenced equation for firm i in period t. The Blundell,
Bond, Devereux, and Schiantarelli estimators can be seen as an application of the
GMM instrumental approach proposed by Arellano and Bond (1991), which was
originally applied to a dynamic panel.
A GMM estimator for the errors-in-variables model of Eq. 57.17 based on IV
moment conditions takes the form
h 0 i1 0 0
b^ ¼
0
Dx Z V 1
N Z Dx Dx Z V 1
N Z Dy ,
13
In models with exogenous explanatory variables, Zi may consist of sub-matrices with the block
diagonal (exploiting all or part of the moment restrictions), concatenated to straightforward
one-column instruments.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1579
0 1
x1 0 0 0 0
B 0 x1 x2 0 0 C
Zi ¼ B
@⋮ ⋮
C:
⋮ ⋮ ⋮ ⋮ ⋮ A
0 0 0 x1 xT2
2 3 2 3 2 3
Dyi21 Dxi21 Dei21
6 Dyi32 7 6 Dxi32 7 6 Dei32 7
6 7 6 7 6 7
6 ⋮ 7 6 ⋮ 7 6 ⋮ 7
6 7 6 7 6 7
6 Dyi , T, T 1 7 6 Dxi T, T 1 7 6 Dei , T, T 1 7
6 7¼6 7b þ 6 7,
6 Dyi31 7 6 Dxi31 7 6 Dei31 7
6 7 6 7 6 7
6 Dyi42 7 6 Dxi42 7 6 Dei42 7
6 7 6 7 6 7
4 ⋮ 5 4 ⋮ 5 4 ⋮ 5
Dyi , T, T 2 Dxi , T, T 2 Dei , T, T 2
or compactly
The IV matrix is the ((2T 3) KT(T 2)) diagonal matrix with the
instruments in the diagonal defined by Z. Let
1580 H. Almeida et al.
h 0 0
i0 h 0 0
i0
Dy ¼ ðDy1 Þ ; . . . ; ðDyN Þ , D2 ¼ ðD21 Þ ; . . . ; ðD2N Þ
h 0 0
i0 0
DX ¼ ðDX1 Þ ; . . . ; ðDXN Þ , Z ¼ Z01 ; . . . ; Z0N :
V^ ZðD2Þ 1 X 0 c c 0
¼ Z D2 D2 Z,
N N i
where D2 c i ¼ Dy ðDXi Þb^ . This procedure assumes that (A1) and (A2) are
i Dx
satisfied. However, as Biorn (2000) argues, one can replace them by (B1) or
(B2) and then ensure that the variables in the IV matrix have a lead or lag of at
least t + 1 periods to the regressor, to “get clear of” the t period memory of the
MA(t) process. The procedure described below is also based on the same set of
assumptions and can be extended similarly.14
The procedure for estimation using equation in levels and IVs in difference is
similar. Consider the T stacked level equations for individual i:
2 3 2 3 2 3 2 3
yi1 c xi1 2i1
4 ⋮ 5 ¼ 4 ⋮ 5 þ 4 ⋮ 5b þ 4 ⋮ 5,
yiT c xiT 2iT
or more compactly,
yi ¼ eT c þ Xi b þ 2,
where eΤ denotes a (T 1) vector of ones. Let the (T T(T 2)K) diagonal matrix
of instrument be denoted by DZi. This matrix has the instruments in difference in
the main diagonal. In addition,
14
See Propositions 1* and 2* in Biorn (2000) for a formal treatment of the conditions.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1581
define:
0 0
y ¼ y01 ; . . . ; y0N , 2 ¼ 201 ; . . . ; 20N
0 h 0 0
i0
X ¼ X01 ; . . . ; X0N , DZ ¼ ðDZ 1 Þ ; . . . ; ðDZN Þ :
V^ ðDZÞ2 1 X 0 0
¼ ðDZÞ 2^ 2^ ðDZÞ,
N N i
where 2^ ¼ yi Xi b^Lx .
Finally, let us briefly contrast the OLS-IV and AB-GMM estimators. The advantages
of GMM over IV are clear: if heteroscedasticity is present, the GMM estimator is more
efficient than the IV estimator, while if heteroscedasticity is not present, the GMM
estimator is no worse asymptotically than the IV. Implementing the GMM estimator,
however, usually comes with a high price. The main problem, as Hayashi (2000, p. 215)
points out, concerns the estimation of the optimal weighting matrix that is at the core of the
GMM approach. This matrix is a function of fourth moments, and obtaining reasonable
estimates of fourth moments requires very large sample sizes. Problems also arise when
the number of moment conditions is high, that is, when there are “too many instruments.”
This latter problem affects squarely the implementation of the AB-GMM, since it relies
on large numbers of lags (especially in long panels). The upshot is that the efficient GMM
estimator can have poor small sample properties [see Baum et al. (2003) for a discussion].
These problems are well documented and remedies have been proposed by, among
others, Altonji and Segal (1996) and Doran and Schmidt (2006).
We use Monte Carlo simulations to assess the finite sample performance of the EW
and IV estimators discussed in Sect. 57.2. Monte Carlo simulations are an ideal
experimental tool because they enable us to study those two estimators in
a controlled setting, where we can assess and compare the importance of elements
1582 H. Almeida et al.
that are key to estimation performance. Our simulations use several distributions to
generate observations. This is important because researchers will often find
a variety of distributions in real-world applications and because one ultimately
does not see the distribution of the mismeasurement term. Our Monte Carlos
compare the EW, OLS-IV, and AB-GMM estimators presented in Sect. 57.2 in
terms of bias and RMSE.15 We also investigate the properties of the EW identifi-
cation test, focusing on the empirical size and power of this test.
A critical feature of panel data models is the observation of multiple data points
from the same individuals over time. It is natural to consider that repeat samples are
particularly useful in that individual idiosyncrasies are likely to contain information
that might influence the error structure of the data-generating process.
We consider a simple data-generating process to study the finite sample perfor-
mance of the EW and OLS-IV estimators. The response variable yit is generated
according to the following model:
15
The mean squared error (MSE) of an estimator y^ incorporates a component measuring the
variability of the estimator (precision) and another measuring its bias (accuracy). An estimator
with good MSE properties has small combined variance and bias. The MSE of y^ can be defined as
h i2
Var y^ þ Bias y^ . The root mean squared error (RMSE) is simply the square root of the
^ For an
MSE. This is an easily interpretable statistic, since it has the same unit as the estimator y.
approximately unbiased estimator, the RMSE is just the square root of the variance, that is, the
standard error.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1583
In all simulations, we set w*i, 50 ¼ 0 and generate wit for t ¼ 49, 48, . . .,T,
such that we drop the first 50 observations. This ensures that the results are not
unduly influenced by the initial values of the wit process.
Following Biorn (2000), we relax the assumption of i.i.d. measurement error.
Our benchmark simulations will use the assumption of time-invariant autocorrela-
tion [(C1) and (C2)]. In particular, we assume that uit ¼ u1i + u2it and vit ¼ v1i + v2it.
We draw all the innovations (u1i,u2it,v1i,v2it,) from a lognormal distribution; that is,
we exponentiate two normal distributions and standardize the resulting variables to
have unit variances and zero means (this follows the approach used by EW). In
Sect. 57.3.6, we analyze the alternative case in which the innovations follow an MA
structure.
The perfectly measured regressor is generated according to
mi ¼ e1i
1 X T
(57.22)
gi ¼ e2i þ pffiffiffi W it ,
T t¼1
where Wit is the sum of the explanatory variables. Our method of generating mi and
gi ensures that the usual random effects estimators are inconsistent because of the
correlation that exists between the individual effects and the error term or the explan-
atory variables. The variables (e1i, e2i) are fixed as standard normal distributions.16
We employ four different schemes to generate the disturbances (2it, eit). Under
Scheme 1, we generate them under a normal distribution, N(0,s2u). Under Scheme 2,
we generate them from a lognormal distribution, LN(0,s2u). Under Scheme 3, we use
a chi-square with 5 degrees of freedom, w25. Under Scheme 4, we generate the
innovations from a Fm,n-distribution with m ¼ 10 and n ¼ 40. The latter three
distributions are right-skewed so as to capture the key distributional assumptions
behind the EW estimator. We use the normal (non-skewed) distribution as
a benchmark.
Naturally, in practice, one cannot determine how skewed – if at all – is the
distribution of the partially out latent variable. One of our goals is to check how this
assumption affects the properties of the estimators we consider. Figure 57.1 pro-
vides a visual illustration of the distributions we employ. By inspection, at least, the
three skewed distributions we study appear to be plausible candidates for the
distribution governing mismeasurement, assuming EW’s prior that measurement
error must be markedly rightly skewed.
16
Robustness checks show that the choice of a standard normal does not influence our results.
1584 H. Almeida et al.
0.3
0.10
0.2
y
y
0.05
0.1
0.0 0.00
−3 −2 −1 0 1 2 3 0 5 10 15 20
x x
F Density Lognormal Density
0.8 0.6
0.6
0.4
y
0.4
0.2
0.2
0.0 0.0
0 5 10 15 20 0 5 10 15 20
x x
Fig. 57.1
ðb; a1 ; a2 ; a3 Þ ¼ ð1, 1, 1, 1Þ
f ¼ 0:6, s2u ¼ s2e1 ¼ s2e2 ¼ 1,
Notice that the parameter f controls the amount of autocorrelation of the latent
regressor. As explained above, this autocorrelation is an important requirement for
the identification of the IV estimator. While we set f ¼ 0.6 in the following
experiments, we also conduct simulations in which we check the robustness of
the results with respect to variations in f between 0 and 1 (see Sect. 57.3.6).
We study the EW identification test in a simple panel data setup. In the panel
context, it is important to consider individual fixed effects. If the data contain fixed
effects, according to Erickson and Whited (2000), a possible strategy is to transform
the data first and then apply their high-order GMM estimator. Accordingly,
throughout this section, our estimations consider data presented in two forms:
“level” and “within.” The first refers to data in their original format, without the
use of any transformation; estimations in level form ignore the presence of fixed
effects.17 The second applies the within transformation to the data – eliminating
fixed effects – before the model estimation.
We first compute the empirical size and power of the test. Note that the null
hypothesis is that the model is incorrectly specified, such that b ¼ 0 and/or
E[3i ] ¼ 0. The empirical size is defined as the number of rejections of the null
hypothesis when the null is true – ideally, this should hover around 5 %. In our case,
the empirical size is given when we draw the innovations (2it, eit) from a
non-skewed distribution, which is the normal distribution since it generates
E[3i ] ¼ 0. The empirical power is the number of rejections when the null hypoth-
esis is false – ideally, this should happen with very high probability. In the present
case, the empirical power is given when we use skewed distributions: lognormal,
chi-square, and F-distribution.
Our purpose is to investigate the validity of the skewness assumption once we
are setting b 6¼ 0. Erickson and Whited (2002) also restrict every element of b to be
nonzero. We conduct a Monte Carlo experiment to quantify the second part of this
assumption. It is important to note that we can compute the E[(i)3] since, in our
controlled experiment, we generate wi and therefore observe it.
Since the EW test is originally designed for cross-sectional data, the first
difficulty the researcher faces when implementing a panel test is aggregation.
Following EW, our test is computed for each year separately. We report the
average of empirical rejections over the years.18 To illustrate the size and power
of the test for the panel data case, we set the time series dimension of the
17
To our knowledge, all but one of the empirical applications of the EW model use the data in level
form. In other words, firm-fixed effects are ignored outright in panel setting estimations of
parameters influencing firm behavior.
18
The results using the median are similar.
1586 H. Almeida et al.
panel to T ¼ 10. Our tests are performed over 5,000 samples of cross-sectional size
equal to 1,000. We use a simple homoscedastic model with r ¼ 0, with the other
model parameters given as above.
Table 57.1 reports the empirical size and power of the statistic proposed by EW
for testing the null hypothesis H0 : E(y˙2i x˙i) ¼ E(y˙ix˙2i ) ¼ 0. This hypothesis is
equivalent to testing H0 : b ¼ 0 and/or E(3i ) ¼ 0. Table 57.1 reports the frequencies
at which the statistic of test is rejected at the 5 % level of significance for,
respectively, the normal, lognormal, chi-square, and F-distributions of the data-
generating process. Recall that when the null hypothesis is true, we have the size of
the test, and when the null is false, we have the power of the test.
The results reported in Table 57.1 imply an average size of approximately 5 %
for the test. In particular, the first two rows in the table show the results in the case
of a normal distribution for the residuals (implying that we are operating under the
null hypothesis). For both the level and within cases, the empirical sizes match the
target significance level of 5 %.
When we move to the case of skewed distributions (lognormal, chi-square, and F),
the null hypothesis is not satisfied by design, and the number of rejections delivers
the empirical power of the test. In the case when the data is presented in levels and
innovations are drawn from a lognormal distribution (see row 2), the test rejects
about 47 % of the time the null hypothesis of no skewness. Using within data, the
test rejects the null hypothesis 43 % of the time. Not only are these frequencies low,
but comparing these results, one can see that the within transformation slightly
reduces the power of the test.
The results associated with the identification test are more disappointing when
we consider other skewed distributions. For example, for the F-distribution, we
obtain only 17 % of rejections of the null hypothesis in the level case and only 28 %
for the within case. Similarly, poor statistical properties for the model identification
test are observed in the chi-square case.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1587
In this section we present simulation results that assess the finite sample performance of
the estimators discussed in Sect. 57.2. The simulations compare the estimators in terms
of bias and efficiency under several distributional assumptions. In the next subsection,
we consider the cross-sectional setting, focusing on the properties of the EW estimator.
Subsequently, we examine the panel case in detail, comparing the performance of the
EW, OLS-IV, and AB-GMM estimators in terms of bias and efficiency.
most basic cross-sectional setup. Crucially, for the other skewed distributions we
study, the EW estimator is significantly biased for both the mismeasured and the
well-measured variables. In addition, the RMSEs are quite high, indicating low
efficiency.
Table 57.4 reports results for the case in which we apply the within transforma-
tion to the data. Here, we introduce the OLS, OLS-IV, and AB-GMM estimators.
We first present the results associated with the set up that is most favorable for the
EW estimations, which is the lognormal case in Panel A. The EW estimates for the
lognormal case are relatively unbiased for the well-measured regressors (between 4 %
and 7 % deviation from true parameter values). The same applies for the mismeasured
regressors. Regarding the OLS-IV, Panel A shows that coefficient estimates are
unbiased in all models considered. AB-GMM estimates are also approximately
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1591
Table 57.4 OLS, OLS-IV, AB-GMM, and EW estimators: panel data after within transformation
b a1 a2 a3
Panel A. Lognormal distribution
OLS Bias 0.7126 0.1553 0.1558 0.1556
RMSE 0.7131 0.1565 0.1570 0.1568
OLS-IV Bias 0.0065 0.0019 0.0014 0.0015
RMSE 0.1179 0.0358 0.0357 0.0355
AB-GMM Bias 0.0248 0.0080 0.0085 0.0081
RMSE 0.0983 0.0344 0.0344 0.0340
EW-GMM3 Bias 0.0459 0.0185 0.0184 0.0183
RMSE 0.0901 0.0336 0.0335 0.0335
EW-GMM4 Bias 0.0553 0.0182 0.0182 0.0183
RMSE 0.1405 0.0320 0.0321 0.0319
EW-GMM5 Bias 0.0749 0.0161 0.0161 0.0161
RMSE 0.1823 0.0303 0.0297 0.0297
Panel B. Chi-square distribution
OLS Bias 0.7126 0.1555 0.1553 0.1556
RMSE 0.7132 0.1565 0.1563 0.1567
OLS-IV Bias 0.0064 0.0011 0.0017 0.001
RMSE 0.1149 0.0348 0.0348 0.0348
AB-GMM Bias 0.0231 0.0083 0.0077 0.0081
RMSE 0.0976 0.0339 0.0338 0.0342
EW-GMM3 Bias 0.3811 0.0982 0.0987 0.0982
RMSE 0.4421 0.1133 0.1136 0.1133
EW-GMM4 Bias 0.3887 0.0788 0.0786 0.0783
RMSE 0.4834 0.0927 0.0923 0.0919
EW-GMM5 Bias 0.4126 0.0799 0.0795 0.0798
RMSE 0.5093 0.0926 0.0921 0.0923
Panel C. F-distribution
OLS Bias 0.7123 0.1554 0.1549 0.1555
RMSE 0.7127 0.1565 0.1559 0.1566
OLS-IV Bias 0.0066 0.0013 0.0023 0.001
RMSE 0.1212 0.0359 0.0362 0.0361
AB-GMM Bias 0.0232 0.0079 0.0072 0.0085
RMSE 0.0984 0.0343 0.0342 0.0344
EW-GMM3 Bias 0.3537 0.0928 0.0916 0.0917
RMSE 0.4239 0.1094 0.1086 0.1095
EW-GMM3 Bias 0.3906 0.0802 0.0790 0.0791
RMSE 0.4891 0.0939 0.0930 0.0932
EW-GMM3 Bias 0.4188 0.0818 0.0808 0.0813
RMSE 0.5098 0.0939 0.0932 0.0935
(continued)
1592 H. Almeida et al.
unbiased, while standard OLS estimates are very biased. In terms of efficiency, the
RMSEs of the EW-GMM3 are somewhat smaller than those of the OLS-IV and
AB-GMM for the well-measured and mismeasured regressors. However, for the
mismeasured regressor, both OLS-IV and AB-GMM have smaller RMSEs than
EW-GMM4 and EW-GMM5.
Panel B of Table 57.4 presents the results for the chi-square distribution. One can
see that the EW yields markedly biased estimates in this case. The bias in the
mismeasured regressor is approximately 38 % (downwards), and the coefficients
for the well-measured variable are also biased (upwards). In contrast, the OLS-IV
and AB-GMM estimates for both well-measured and mismeasured regressors are
approximately unbiased. In terms of efficiency, as expected, the AB-GMM presents
slightly smaller RMSEs than the OLS-IV estimator. These IV estimators’ RMSEs
are much smaller than those associated with the EW estimators.
Panels C and D of Table 57.4 show the results for the F and standard normal
distributions, respectively. The results for the F-distribution in Panel C are essentially
similar to those in Panel B: the instrumental variable estimators are approximately
unbiased while the EW estimators are very biased. Finally, Panel D shows that deviations
from a strongly skewed distribution are very costly in terms of bias for the EW estimator,
since the bias for the mismeasured regressor is larger than 70 %, while for the well
measured, it is around 20 %. A comparison of RMSEs shows that the IV estimators are
more efficient in both the F and normal cases. In all, our simulations show that standard IV
methods almost universally dominate the EW estimator in terms of bias and efficiency.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1593
We reiterate that the bias and RMSE of the IV estimators in Table 57.4 are all
relatively invariant to the distributional assumptions, while the EW estimators are
all very sensitive to those assumptions. In short, this happens because the EW relies
on the high-order moment conditions as opposed to the OLS and IV estimators.
57.3.4 Heteroscedasticity
One way in which individual heterogeneity may manifest itself in the data is via error
heteroscedasticity. Up to this point, we have disregarded the case in which the data has
a heteroscedastic error structure. However, most empirical applications in corporate
finance entail the use of data for which heteroscedasticity might be relevant. It is important
that we examine how the EW and the IV estimators are affected by heteroscedasticity.19
The presence of heteroscedasticity introduces heterogeneity in the model and
consequently in the distribution of the partialled out dependent variable. This
compromises identification in the EW framework. Since the EW estimator is
based on equations giving the moments of (yi zimy) and (yi zimw) as functions
of b and moments of (ui, ei, i), the heteroscedasticity associated with the fixed
effects (ai) or with the perfectly measured regressor (zit) distorts the required
moment conditions associated with (yi zimy), yielding biased estimates. These
inaccurate estimates enter the minimum distance estimator equation and conse-
quently produce incorrect weights for each estimate along the time dimension. As
our simulations of this section demonstrate, this leads to biased MDE estimates,
where the bias is a function of the amount of heteroscedasticity.
We examine the biases imputed by heteroscedasticity by way of graphical analysis.
The graphs we present below are useful in that they synthesize the outputs of numerous
tables and provide a fuller visualization of the contrasts we draw between the EW and
OLS-IV estimators. The graphs depict the sensitivity of those two estimators with
respect to heteroscedasticity as we perturb the coefficient p in Eq. 57.18.
In our simulations, we alternatively set wit ¼ gi or wit ¼ zit. In the first case,
heteroscedasticity is associated with the individual effects. In the second, heterosce-
dasticity is associated with the well-measured regressor. Each of our figures
describes the biases associated with the mismeasured and the well-measured
regressors for each of the OLS-IV, EW-GMM3, EW-GMM4, and EW-GMM5
estimators.20 In order to narrow our discussion, we only present results for the
highly skewed distribution case (lognormal distribution) and for data that is treated
for fixed effects using the within transformation. As Sect. 57.3.3 shows, this is the
only case in which the EW estimator returns relatively unbiased estimators for the
parameters of interest. In all the other cases (data in levels and for data generated by
19
We focus on the OLS-IV estimator hereinafter for the purpose of comparison with the EW
estimator.
20
Since estimation biases have the same features across all well-measured regressors of a model,
we restrict attention to the first well-measured regressor of each of the estimated models.
1594 H. Almeida et al.
chi-square, F, and normal distributions), the estimates are strongly biased even
under the assumption of homoscedasticity.21
Figure 57.2 presents the simulation results under the assumption that wit ¼ gi as we
vary the amount of heteroscedasticity by changing the parameter r,22 the results for
the mismeasured coefficients show that biases in the EW estimators are generally
small for r equal to zero (this is the result reported in Sect. 57.3.3). However, as this
coefficient increases, the bias quickly becomes large. For example, for r ¼ 0.45, the
biases in the coefficient of the mismeasured variable are, respectively, 11 %,
20 %, and 43 %, for the EW-GMM3, EW-GMM4, and EW-GMM5 estimators.
Notably, those biases, which are initially negative, turn positive for moderate values
of r. As heteroscedasticity increases, some of the biases diverge to positive infinite.
The variance of the biases of the EW estimators is also large. The results regarding
the well-measured variables using EW estimators are analogous to those for the
mismeasured one. Biases are substantial even for small amounts of heterosce-
dasticity, they switch signs for some level of heteroscedasticity, and their variances
are large. In sharp contrast, the same simulation exercises show that the OLS-IV
estimates are approximately unbiased even under heteroscedasticity. While the EW
estimator may potentially allow for some forms of heteroscedasticity, it is clear that it
is not well equipped to deal with this problem in more general settings.
Our Monte Carlo experiments show that the EW estimator has a poor handle of
individual fixed effects and that biases arise for deviations from the assumption of
strict lognormality. Biases in the EW framework are further magnified if one allows
for heteroscedasticity in the data (even under lognormality). The biases arising from
the EW framework are hard to measure and sign, ultimately implying that it can be
very difficult to replicate the results one obtains under that framework.
To better understand these results, we now discuss in more mathematical details
the identification of the EW estimator for the panel data case for both the model in
level and after the within transformation. Extending the EW estimator to panel data
seems to be a nontrivial task. EW have proposed to break the problem for each time
series, estimate a cross-section model for each t, and after that combine the
estimates using a minimum distance estimator. In what follows we show that this
procedure might affect the identification condition.
Consider the following model:
21
Our simulation results (available upon request) suggest that introducing heteroscedasticity
makes the performance of the EW estimator even worse in these cases.
22
The results for wit ¼ zit are quite similar to those we get from setting wit ¼ gi. We report only one
set of graphs to save space.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1595
Bias 2
0 1 2 3 4 5
ρ
b Perfectly−Measured Variable
0.5
0.0
Bias
−0.5
−1.0
0 1 2 3 4 5
ρ
where uit is independent and identically distributed, with mean zero and variance
s2u. Assume that the Var(xit) ¼ s2w. The independent variable and unobserved effects
are exogenous, that is, Cov(wit, uis) ¼ Cov(ai, uit) ¼ 0 for any t and s. However,
Cov(ai, wit) 6¼ 0. Now, assume that we do not observe the true variable wit, but rather
a mismeasured variable, that is, you observe the following variable with an error:
where Cov(xit, eis) ¼ Cov(ai, eis) ¼ Cov(uit, eis) ¼ 0, and Var(eit) ¼ s2e ,
Cov(eit, eit 1) ¼ gs2e .
In addition, assume here that there is no variable zit (a ¼ 0 in Eq. 57.18) to
simplify the argument.
and
and
and
1 1 1
yi1 ðyi1 þ yi2 Þ ¼ b wi1 ðwi1 þ wi2 Þ þ ui1 ðui1 þ ui2 Þ ,
2 2 2
yil þ yi2 ¼ bðwi1 þ wi2 Þ þ ðui1 þ ui2 Þ,
and
1 1 1
xi1 ðxi1 þ xi2 Þ ¼ wi1 ðwi1 þ wi2 Þ þ ei1 ðei1 þ ei2 Þ ,
2 2 2
xi1 þ xi2 ¼ ðwi1 þ wi2 Þ þ ðei1 þ ei2 Þ:
and
Let’s now define Yi ¼ yi1 + yi2, Xi ¼ xi1 + xi2, Ui ¼ ui1 + ui2, vi ¼ wi1 + wi2, and
Ei ¼ ei1 + ei2. So, the model could be rewritten as
Y i ¼ bvi þ Ui
and
X i ¼ v i þ Ei :
Notice that the requirements for identification now are on the high-order
moments of (V, U, E). However, note that vi ¼ wi1 + wi2, which is a sum of two
random variables. As it is well known from the econometrics literature, convolution
of random variables is in general a nontrivial object.
One example of why the identification condition may worsen considerably
is the following. Consider a model where wi1 and wi2 are independent
chi-square distributions with 2 degrees ofpffiffiffiffiffiffiffi
freedom.
ffi The skewness of the
chi-square with k degrees of freedom is 8=k . Note that the sum of two
independent chi-squares with k degrees of freedom is a chi-square with 2k degrees
1598 H. Almeida et al.
of freedom. Therefore, the skewness of the vi ¼ wi1 + wi2 drops from two
for the model using one distribution to 1.41 for the model using the summation
of both wi1 and wi2.
From this simple analysis one could conclude that the identification conditions
required for EW estimator can deteriorate considerably when using the within
transformation to eliminate the fixed effects in panel data. Thus, the required
conditions to achieve unbiased estimates with EW are very strong.
Our Monte Carlo simulations show that the OLS-IV estimator is consistent even
when one allows for autocorrelation in the measurement-error structure. We have
assumed, however, some structure on the processes governing innovations. In this
section, we examine the sensitivity of the OLS-IV results with respect to our
assumptions about measurement-error correlation and the amount of autocorrela-
tion in the latent regressor. These assumptions can affect the quality of the instru-
ments and therefore should be examined in some detail.
We first examine conditions regarding the correlation of the measurement
errors and disturbances. The assumption of time-invariant autocorrelation for
measurement errors and disturbances implies that past shocks to measurement
errors do not affect the current level of the measurement error. One way to relax
this assumption is to allow for the measurement-error process to have a moving
average structure. This structure satisfies Biorn’s assumptions (B1) and (B2). In
this case, Proposition 1 in Biorn (2000) shows that for an MA(t), the instruments
should be of order of at most t t 2. Intuitively, the set of instruments must be
“older” than the memory of the measurement-error process. For example, if the
measurement error is MA(1), then one must use third- and longer-lagged
instruments to identify the model.
To analyze this case, we conduct Monte Carlo simulations in which we
replace the time-invariant assumption for innovation uit and vit with an MA(1)
structure for the measurement-error process. The degree of correlation in the
MA process is set to y ¼ 0.4. Thus, the innovation in Eqs. 57.18 and 57.19 has
the following structure:
with jyj 1, and u1it and v1it are i.i.d. lognormal distributions. The other parameters
in the simulation remain the same.
The results are presented in Table 57.5. Using MA(1) in the innovations and
the third lag of the latent regressor as an instrument (either on its own or in
combination with the fourth lag), the bias of the OLS estimator is very small
(approximately 2–3 %). The bias increases somewhat when we use only the fourth
lag. While the fourth is an admissible instrument in this case, using longer lags
decreases the implied autocorrelation in the latent regressor [which follows an
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1599
AR(1) process by Eq. 57.20]. This effect decreases somewhat the quality of the
instruments. Notice also that when we do not eliminate short lags from the instru-
ment set, the identification fails. For example, the bias is 60 % when we use the
second lag as an instrument. These results thus underscore the importance of using
long enough lags in this MA case. Table 57.5 also reports results based on an MA
(2) structure. The results are qualitatively identical to those shown in the MA
(1) case. Once again, the important condition for identification is to use long enough
lags (no less than four lags in this case).23
The second condition underlying the use of the OLS-IV is that the latent
regressor is not time invariant. Accordingly, the degree of autocorrelation in
the process for the latent regressor is an important element of the identification
strategy. We assess the sensitivity of the OLS-IV results to this condition by
varying the degree of autocorrelation through the autoregressive coefficient in the
AR(1) process for the latent regressor. In these simulations, we use a time-invariant
autocorrelation condition for the measurement error, but the results are very similar
for the MA case.
Figure 57.3 shows the results for the bias in the coefficients of interest for the well-
measured and mismeasured variables, using the second lag of the mismeasured
variable as an instrument. The results show that the OLS-IV estimator performs well
for a large range of the autoregressive coefficient. However, as expected, when the f
coefficient is very close to zero or one, we have evidence of a weak instrument
problem. For example, when f ¼ 1, then Dwit is uncorrelated with any variable
dated at time t 2 or earlier. These simulations show that, provided that one uses
adequately lagged instruments, the exact amount of autocorrelation in the latent
variable is not a critical aspect of the estimation.
The simulations of this section show how the performance of the OLS-IV
estimator is affected by changes in assumptions concerning measurement errors
and latent regressors. In practical applications, it is important to verify whether
the results obtained with OLS-IV estimators are robust to the elimination of
short lags from the instrumental set. This robustness check is particularly
important given that the researcher will be unable to pin down the process
followed by the measurement error. Our empirical application below incorporates
this suggestion. In addition, identification relies on some degree of autocorrelation
in the process for the latent regressor. While this condition cannot be directly
verified, we can perform standard tests of instrument adequacy that rely on
“first-stage” test statistics calculated from the processes for the observable
variables in the model.
Another important assumption in the OLS is non-autocorrelation in both uit
and vit. For example, these innovations cannot follow an autoregressive process.
23
We note that if the instrument set uses suitably long lags, then the OLS-IV results are robust to
variations in the degree of correlation in the MA process. In unreported simulations under MA(1),
we show that the OLS bias is nearly invariant to the parameter y.
1600 H. Almeida et al.
When this is the case, the IV strategy of using lags of mismeasured variable as valid
instruments is invalid (see Biorn 2000).
A natural question is whether our simulation results are rooted in the lack of
accuracy of the asymptotic approximation of the EW method. Inference in models
with mismeasured regressors is based on asymptotic approximations; hence infer-
ence based on estimators with poor approximations might lead to wrong inference
procedures. For instance, we might select wrong critical values for a test under poor
asymptotic approximations and make inaccurate statements under such circum-
stances. In this section, we use the panel data simulation procedure of Sect. 3.3.2 to
study and compare the accuracy of the asymptotic approximation of the EW and IV
methods. To save space, we restrict our attention to the mismeasured regressor
coefficient for the EW-GMM5 and OLS-IV cases. We present results where we
draw the data from the lognormal, chi-square, and F-distributions. The EW-GMM5
estimator is computed after the within transformation and the OLS-IV uses second
lags as instruments.
One should expect both the IV and EW estimators to have asymptotically normal
representations, such that when we normalize the estimator by subtracting the true
parameter and divide by the standard deviation, this quantity behaves asymptoti-
cally as a normal distribution. Accordingly, we compute the empirical density and
the distribution functions of the normalized sample estimators and their normal
approximations. These functions are plotted in Fig. 57.4. The true normal density
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1601
Mismeasured Variable
2
1
Bias
−1
−2
0.0 0.2 0.4 0.6 0.8 1.0
φ
Perfectly−Measured Variable
0.4
Bias
0.0
−0.4
Fig. 57.3
24
The empirical cumulative distribution function Fn is a step function with jumps i/n at observation
values, where i is the number of tied observations at that value.
1602 H. Almeida et al.
Fig. 57.4
25
However, financial constraints are not sufficient to generate a strictly positive cash flow
coefficient because the effect of financial constraints is capitalized in stock prices and may thus
be captured by variations in q (Chirinko 1993; Gomes 2001).
1604 H. Almeida et al.
due to fact that cash flow picks up variation in investment opportunities even after
we apply a correction for mismeasurement in q. Accordingly, one should observe:
Condition 2. bMEC 0 and aMEC 0. Notice that these conditions are fairly weak.
If a particular measurement-error consistent estimator does not deliver these basic
results, one should have reasons to question the usefulness of that estimator in
applied work.
Our data collection process follows that of Almeida and Campello (2007). We consider
a sample of manufacturing firms over the 1970–2005 period with data available from
Compustat. Following those authors, we eliminate firm years displaying asset or sales
growth exceeding 100 %, or for which the stock of fixed capital (the denominator of
the investment and cash flow variables) is less than $5 million (in 1976 dollars).
Our raw sample consists of 31,278 observations from 3,084 individual firms. Summary
statistics for investment, q, and cash flow are presented in Table 57.6. These statistics
are similar to those reported by Almeida and Campello, among other papers. To save
space we omit the discussion of these descriptive statistics.
Before estimating our investment models, we conduct a series of tests for the
presence of firm-fixed effects and heteroscedasticity in our data. As a general
rule, these phenomena might arise naturally in panel data applications and should
not be ignored. Importantly, whether they appear in the data can have concrete
implications for the results generated by different estimators.
We first perform a couple of tests for the presence of firm-fixed effects. We allow
for individual firm intercepts in Eq. 57.27 and test the null hypothesis that the
coefficients associated with those firm effects are jointly equal to zero (Baltagi
2005). Table 57.7 shows that the F-statistic for this test is 4.4 (the associated
p-value is 0.000). Next, we contrast the random effects OLS and the fixed effects
OLS estimators to test again for the presence of fixed effects. The Hausman test
statistic reported in Table 57.7 rejects the null hypothesis that the random effects
model is appropriate with a test statistic of 8.2 (p-value of 0.017). In sum, standard
tests strongly reject the hypothesis that fixed effects can be ignored.
We test for homoscedasticity using two different panel data-based methods.
First, we compute the residuals from the least squares dummy variables estimator
and regress the squared residuals on a function of the independent variables [see
Frees (2004) for additional details]. We use two different combinations of inde-
pendent regressors – (qit, CFit) and (qit, q2it, CFit, CFit, CF2it) – and both of them
robustly reject the null hypothesis of homoscedasticity. We report the results for the
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1605
first combination in Table 57.7, which yields a test statistic of 55.2 (p-value of
0.000). Our second approach for testing the null of homoscedasticity is the standard
random effects Breusch-Pagan test. Table 57.7 shows that the Breusch-Pagan test
yields a statistic of 7,396.2 (p-value of 0.000). Our tests hence show that the data
strongly reject the hypothesis of error homoscedasticity.
Our preliminary tests show that one should control for fixed effects when estimating
investment models using real data. In the context of the EW estimator, it is thus
appropriate to apply the within transformation before the estimation. However, in this
section, we also present results for the data in level form to illustrate the point made in
Sect. 57.3.2 that applying the within transformation compromises identification in the
EW context. Prior papers adopting the EW estimator have ignored (or simply
dismissed) the importance of fixed effects (e.g., Whited 2001, 2006).
We present the results for EW’s identification test in Table 57.8. Using the data in
level form, we reject the hypothesis of no identification in 12 out of 30 years (or 36 %
rejection). For data that is transformed to accommodate fixed effects (within trans-
formation), we find that in only 7 out of 33 (or 21 %) of the years between 1973 and
2005, one can reject the null hypothesis that the model is not identified at the usual
5 % level of significance. These results suggest that the power of the test is low and
decreases further after applying the within transformation to the data. These results
are consistent with Almeida and Campello’s (2007) use of the EW estimator.
Working with a 15-year Compustat panel, those authors report that they could only
find a maximum of 3 years of data passing the EW identification test.
1606 H. Almeida et al.
The results in Table 57.8 reinforce the notion that it is quite difficult to
operationalize the EW estimator in real-world applications, particularly in situa-
tions in which the within transformation is appropriate due to the presence of fixed
effects. We recognize that the EW identification test rejects the model for most of
the data at hand. However, recall from Sect. 57.3.2 that the test itself is likely to be
misleading (“over-rejecting” the data). In the next section, we take the EW estima-
tor to the data (a standard Compustat sample extract) to illustrate the issues applied
researchers face when using that estimator, contrasting it to an easy-to-implement
alternative.
We estimate Eq. 57.27 using the EW, OLS-IV, and AB-GMM estimators. For
comparison purposes, we also estimate the investment equation using standard OLS
and OLS with fixed effects (OLS-FE). The estimates for the standard OLS are likely to
be biased, providing a benchmark to evaluate the performance of the other estimators.
As discussed in Sect. 57.4.1, we expect estimators that improve upon the problem of
mismeasurement to deliver results that satisfy Conditions 1 and 2 above.
As is standard in the empirical literature, we use an unbalanced panel in our
estimations. Erickson and Whited (2000) propose a minimum distance estimator
(MDE) to aggregate the cross-sectional estimates obtained for each sample year,
but their proposed MDE is designed for balanced panel data. Following Riddick and
Whited (2009), we use a Fama–MacBeth procedure to aggregate the yearly EW
estimations.26
To implement our OLS-IV estimators, we first take differences of the model in
Eq. 57.27. We then employ the estimator denoted by OLS-IV A from Sect. 57.2.2,
using lagged levels of q and cash flow as instruments for (differenced) qit. Our
Monte Carlos suggest that identification in this context may require the use of
longer lags of the model variables. Accordingly, we experiment with specifications
that use progressively longer lags of q and cash flow to verify the robustness of our
results.
Table 57.9 reports our findings. The OLS and OLS-FE estimates, reported in
columns (1) and (2), respectively, disregard the presence of measurement error
in q. The EW-GMM3, EW-GMM4, and EW-GMM5 estimates are reported in
columns (3), (4), and (5). For the OLS-IV estimates reported in column (6), we
use qt—2 as an instrument.27 The AB-GMM estimator, reported in column (7), uses
lags of q as instruments. Given our data structure, this implies using a total of
465 instruments. We account for firm-fixed effects by transforming the data.
26
Fama–MacBeth estimates are computed as a simple standard errors for yearly estimates. An
alternative approach could use the Hall–Horowitz bootstrap. For completeness, we present in the
appendix the actual yearly EW estimates.
27
In the next section, we examine the robustness of the results with respect to variation in the
instrument set.
57
performed on a yearly basis. In the last columns, we collect the number of years in which the null hypothesis is rejected (sum) and compute the percentage of
years in which the null is rejected. The data are taken from the annual Compustat industrial files over the 1970–2005 period. See text for details
1609
1610
When using OLS and OLS-FE, we obtain the standard result in the literature that
both q and cash flow attract positive coefficients [see columns (1) and (2)]. In the
OLS-FE specification, for example, we obtain a q coefficient of 0.025 and a cash
flow coefficient of 0.121. Columns (3), (4), and (5) show that the EW estimator does
not deliver robust inferences about the correlations between investment, cash flow,
and q. The q coefficient estimate varies significantly with the set of moment
conditions used, even flipping signs. In addition, none of the q coefficients is
statistically significant. The cash flow coefficient is highly inflated under EW, and
in the case of the EW-GMM4 estimator, it is more than three times larger than the
(supposedly biased) OLS coefficient. These results are inconsistent with Conditions
1 and 2 above. These findings agree with the Monte Carlo simulations of Sect.
57.3.3, which also point to a very poor performance of the EW estimator in cases in
which fixed effects and heteroscedasticity are present.
By comparison, the OLS-IV delivers results that are consistent with Conditions
1 and 2. In particular, the q coefficient increases from 0.025 to 0.063, while the cash
flow coefficient drops from 0.131 to 0.043. These results suggest that the proposed
OLS-IV estimator does a fairly reasonable job at addressing the measurement-error
problem. This conclusion is consistent with the Monte Carlo simulations reported
above, which show that the OLS-IV procedure is robust to the presence of fixed
effects and heteroscedasticity in simulated data. The AB-GMM results also gener-
ally satisfy Conditions 1 and 2. Notice, however, that the observed changes in the
q and cash flow coefficients (“corrections” relative to the simple, biased OLS
estimator) are less significant than those obtained under the OLS-IV estimation.
This suggests that the researcher should look for longer lagging schemes, lags that
“erase” the MA memory of the error structure.
Our next set of estimations use longer lagging structures for our proposed
instruments and even an instrumental set with only lags of cash flow, the exogenous
regressors in the model. We use combinations of longer lags of q (such as the fourth
and fifth lags) and longer lags of cash flow (fourth and fifth lags). This set of tests
yields estimates that more clearly meet standard tests for instrument validity.28
Specifically, the J-statistics now indicate we do not reject the hypothesis that the
exclusion restrictions are met. The results reported in columns (2) through (7) of
Table 57.10 also remain consistent with Conditions 1 and 2. In particular, the
q coefficient varies from approximately 0.040 to 0.091, while the cash flow
coefficient varies roughly from 0.044 to 0.046. These results are consistent with
our simulations, which suggest that these longer lag structures should deliver
relatively consistent, stable estimates of the coefficients for q and cash flow in
standard investment regressions.
28
All of the F-statistics associated with the first-stage regressions have p-values that are close to
zero. These statistics (reported in Table 57.10) suggest that we do not incur a weak instrument
problem when we use longer lags in our instrumental set.
1614 H. Almeida et al.
References
Agca, S., & Mozumdar, A. (2007). Investment-cash flow sensitivity: Myth or reality? Working
paper, George Washington University.
Almeida, H., & Campello, M. (2007). Financial constraints, asset tangibility and corporate
investment. Review of Financial Studies, 20, 1429–1460.
Altonji, J., & Segal, L. (1996). Small-sample bias in GMM estimation of covariance structures.
Journal of Business & Economic Statistics, 14, 353–366.
Arellano, M., & Bond, S. (1991). Some tests of specification for panel data: Monte Carlo evidence
and an application to employment equations. Review of Economic Studies, 58, 277–297.
1616 H. Almeida et al.
Bakke, T., & Whited, T. (2009). What gives? A study of firms’ reactions to cash shortfalls.
Working paper, University of Rochester.
Baltagi, B. (2005). Econometric analysis of panel data (3rd ed.). Chichester: Wiley.
Baum, C., Schaffer, M., & Stillman, S. (2003). Instrumental variables and GMM: Estimation and
testing. Stata Journal, 3, 1–31.
Bertrand, M., & Mullainathan, S. (2005). Bidding for oil and gas leases in the Gulf of Mexico:
A test of the free cash flow model?. mimeo, University of Chicago and MIT.
Bertrand, M., & Schoar, A. (2003). Managing with style: The effect of managers on firm policies.
Quarterly Journal of Economics, 118, 1169–1208.
Bertrand, M., Duflo, E., & Mullainathan, S. (2004). How much should we trust differences-in-
differences estimates? Quarterly Journal of Economics, 119, 249–275.
Biorn, E. (2000). Panel data with measurement errors: Instrumental variables and GMM pro-
cedures combining levels and differences. Econometric Reviews, 19, 391–424.
Blanchard, O., Lopez-de-Silanes, F., & Shleifer, A. (1994). What do firms do with cash windfalls?
Journal of Financial Economics, 36, 337–360.
Blundell, R., Bond, S., Devereux, M., & Schiantarelli, F. (1992). Investment and Tobin’s Q:
Evidence from company panel data. Journal of Econometrics, 51, 233–257.
Bond, S., & Cummins, J. (2000). The stock market and investment in the new economy. Brookings
Papers on Economic Activity, 13, 61–124.
Chirinko, R. (1993). Business fixed investment spending: Modeling strategies, empirical results,
and policy implications. Journal of Economic Literature, 31, 1875–1911.
Colak, G., & Whited, T. (2007). Spin-offs, divestitures, and conglomerate investment. Review of
Financial Studies, 20, 557–595.
Cragg, J. (1997). Using higher moments to estimate the simple errors-in-variables model. RAND
Journal of Economics, 28, 71–91.
Doran, H., & Schmidt, P. (2006). GMM estimators with improved finite sample properties using
principal components of the weighting matrix, with an application to the dynamic panel data
model. Journal of Econometrics, 133, 387–409.
Erickson, T., & Whited, T. (2000). Measurement error and the relationship between investment
and Q. Journal of Political Economy, 108, 1027–1057.
Erickson, T., & Whited, T. (2002). Two-step GMM estimation of the errors-in-variables model
using high-order moments. Econometric Theory, 18, 776–799.
Fazzari, S., Hubbard, R. G., & Petersen, B. (1988). Financing constraints and corporate invest-
ment. Brooking Papers on Economic Activity, 1, 141–195.
Frees, E. (2004). Longitudinal and panel data analysis and applications in the social sciences.
New York: Cambridge University Press.
Gan, J. (2007). Financial constraints and corporate investment: Evidence from an exogenous shock
to collateral. Journal of Financial Economics, 85, 709–734.
Gomes, J. (2001). Financing investment. American Economic Review, 91, 1263–1285.
Griliches, Z., & Hausman, J. A. (1986). Errors in variables in panel data. Journal of Econometrics,
31, 93–118.
Hadlock, C. (1998). Ownership, liquidity, and investment. RAND Journal of Economics, 29, 487–508.
Hayashi, F. (1982). Tobin’s marginal q and average q: A neoclassical interpretation.
Econometrica, 50, 213–224.
Hayashi, F. (2000). Econometrics. Princeton: Princeton University Press.
Hennessy, C. (2004). Tobin’s Q, debt overhang, and investment. Journal of Finance, 59,
1717–1742.
Holtz-Eakin, D., Newey, W., & Rosen, H. S. (1988). Estimating vector autoregressions with panel
data. Econometrica, 56, 1371–1395.
Hoshi, T., Kashyap, A., & Scharfstein, D. (1991). Corporate structure, liquidity, and investment:
Evidence from Japanese industrial groups. Quarterly Journal of Economics, 106, 33–60.
Hubbard, R. G. (1998). Capital market imperfections and investment. Journal of Economic
Literature, 36, 193–227.
57 Assessing the Performance of Estimators Dealing with Measurement Errors 1617
Kaplan, S., & Zingales, L. (1997). Do financing constraints explain why investment is correlated
with cash flow? Quarterly Journal of Economics, 112, 169–215.
Lamont, O. (1997). Cash flow and investment: Evidence from internal capital markets. Journal of
Finance, 52, 83–110.
Lyandres, E. (2007). External financing costs, investment timing, and investment-cash flow
sensitivity. Journal of Corporate Finance, 13, 959–980.
Malmendier, U., & Tate, G. (2005). CEO overconfidence and corporate investment. Journal of
Finance, 60, 2661–2700.
Petersen, M. (2009). Estimating standard errors in finance panel data sets: Comparing approaches.
Review of Financial Studies, 22, 435–480.
Poterba, J. (1988). Financing constraints and corporate investment: Comment. Brookings Papers
on Economic Activity, 1, 200–204.
Rauh, J. (2006). Investment and financing constraints: Evidence from the funding of corporate
pension plans. Journal of Finance, 61, 33–71.
Riddick, L., & Whited, T. (2009). The corporate propensity to save. Journal of Finance, 64,
1729–1766.
Shin, H., & Stulz, R. (1998). Are internal capital markets efficient? Quarterly Journal of Eco-
nomics, 113, 531–552.
Stein, J. (2003). Agency information and corporate investment. In G. Constantinides, M. Harris, &
R. Stulz (Eds.), Handbook of the economics of finance. Amsterdam: Elsevier/North-Holland.
Wansbeek, T. J. (2001). GMM estimation in panel data models with measurement error. Journal of
Econometrics, 104, 259–268.
Whited, T. (2001). Is it inefficient investment that causes the diversification discount? Journal of
Finance, 56, 1667–1691.
Whited, T. (2006). External finance constraints and the intertemporal pattern of intermittent
investment. Journal of Financial Economics, 81, 467–502.
Xiao, Z., Shao, J., & Palta, M. (2008). A unified theory for GMM estimation in panel data models
with measurement error. Working paper, University of Wisconsin Madison.
Realized Distributions of Dynamic
Conditional Correlation and Volatility 58
Thresholds in the Crude Oil, Gold, and
Dollar/Pound Currency Markets
Contents
58.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1620
58.2 Literature Review . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1622
58.3 Data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1623
58.4 Dynamic Conditional Correlation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1625
58.4.1 Dynamic Conditional Correlation Between Gold, Oil, and
the Dollar/Pound . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1625
58.4.2 Empirical Results of Dynamic Conditional Correlation . . . . . . . . . . . . . . . . . . . . . 1628
58.4.3 Volatility Threshold Dynamic Conditional Correlation . . . . . . . . . . . . . . . . . . . . . 1636
58.4.4 Does Investors’ Behavior Change over Subperiods? . . . . . . . . . . . . . . . . . . . . . . . . 1639
58.5 Conclusions and Implications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1643
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1644
Abstract
This chapter proposes a modeling framework for the study of co-movements in
price changes among crude oil, gold, and dollar/pound currencies that are
conditional on volatility regimes. Methodologically, we extend the dynamic
T.-L. Shih
Department of Hospitality Management, Ming Dao University, Changhua Peetow, Taiwan
e-mail: tungli@mdu.edu.tw
H.-C. Yu (*)
Department of International Business, Chung Yuan University, Chungli, Taiwan
e-mail: haichin@cycu.edu.tw; haichinyu@hotmail.com
D.-T. Hsieh
Department of Economics, National Taiwan University, Taipei, Taiwan
e-mail: dthsieh@ccms.ntu.edu.tw
C.-J. Lee
College of Business, Chung Yuan University, Chungli, Taiwan
e-mail: g9604601@cycu.edu.tw
C.-F. Lee, J. Lee (eds.), Handbook of Financial Econometrics and Statistics, 1619
DOI 10.1007/978-1-4614-7750-1_58,
# Springer Science+Business Media New York 2015
1620 T.-L. Shih et al.
Keywords
Dynamic conditional correlation • Volatility threshold • Realized distribution •
Currency market • Gold • Oil
58.1 Introduction
This widely expanded and complex volatility in commodity prices increases the
importance of modeling real volatility and correlation, because a good estimate
helps facilitate portfolio optimization, risk management, and hedging activities.
Although some of the literature assumes volatility and correlation to be constant
in the past years, it is widely recognized that they indeed vary over time. This
recognition has spurred a vibrant body of work regarding the dynamic properties of
market volatility. To date, very little is known about the volatility dynamics
between the commodity and currency markets, for instance, in the case of gold
and its possible correlations with oil and major currencies. This chapter intends to
address this gap.
The main purpose of this study is to examine the dynamic relationships among
gold, oil, and the dollar/pound to further understand the hedging ability of gold
relative to another commodity or currency. That is to say, if gold acts as a financial
safe haven against the dollar (or oil), it allows for systematic feedback between
changes in the price of gold, oil, and the dollar/pound exchange rate. Specifically,
this chapter asks, does gold act as a safe haven against the dollar/pound, as a hedge,
or as neither? Are gold and oil highly correlated with each other? Movements in the
price of gold, oil, and the dollar/pound are analyzed using a model of dynamic
conditional correlation covering 20 years of daily data.
Studies related to this issue are few. Capie et al. (2005) point out that gold acts as
an effective hedge against the US dollar by estimating elasticity relative to changes
in the exchange rate. However, their approach involves the use of a single-equation
model in which the exchange rate is assumed to be unaffected by the time of
the dependent variable, the price of gold. Our chapter improves their work
by employing a dynamic model of conditional correlations in which all variables
are treated symmetrically. Besides, although Baur and Lucey (2010) find evidence
in support of gold providing a haven from losses incurred in the bond and
stock markets, they neglect the interactions with the currency market and,
like Capie et al. (2005), do not consider feedback in their model of returns. Nikos
(2006) uses correlation analysis to estimate the correlation of returns between gold
and the dollar and shows that the correlation between the dollar and gold is 0.19
and 0.51 for two different periods. These findings imply that gold is
a contemporaneous safe haven in extreme currency market conditions. Steinitz
(2006) utilizes the same method to estimate the correlations of weekly
returns between gold and Brent oil for two periods of 1 year and 5 years,
respectively, and shows that the correlations between gold and Brent oil are 0.310
and 0.117, respectively. Boyer and Fillion (2007) report on the financial
determination of Canadian oil and gas stock returns and conclude that
a weakening of the Canadian dollar against the US dollar has a negative impact
on stock returns.
If correlations and volatilities vary over time, the hedge ratio should be adjusted
to account for the new information. Other work, such as Baur and McDermott
(2010), similarly neglects feedback in its regression model. Further studies
1622 T.-L. Shih et al.
investigate the concept of a safe-haven asset without reference to gold. For example,
Ranaldo and Soderlind (2009) and Kaul and Sapp (2006) examine safe-haven
currencies, while Upper (2000) examines German government bonds as safe-haven
instruments. Andersen et al. (2007) show that exchange rate volatility outstrips bond
volatility in the US, British, and German markets. Thus, currency risk is worth
exploring and being hedged.
As a general rule, commodities are priced in US dollars. Since the US currency
has weakened that a bull run of commodity prices appeared, the question arises as to
which the increases in commodity prices have been a product of the depreciation in
the US dollar. Furthermore, it would be interesting to examine how to provide
a hedge against the dollar that varies across different commodities. It also needs to
be asked which investment instruments are more suitable for diversification pur-
poses to protect against changes in the US currency.
This chapter investigates the following issues. First, how do the time-varying
correlations and associated distributions appear in the crude oil, gold, and dollar/
pound markets? Second, what is the shape of each separate distribution of various
volatility levels among the crude oil, gold, and dollar/pound markets? Third, by
employing the volatility threshold DCC model put forward by Kasch and Caporin
(2012), is the high volatility (exceeding a specified threshold) of the assets associ-
ated with an increasing degree of correlation?
We find that the volatility thresholds of oil and gold correspond to two major
events – the First Gulf War in 1990 and the 911 event in 2001. We also find that
the increase in commodity (crude oil and gold) prices was a reflection of the falling
US dollar, especially after the 911 event. The evidence shows that the DCC
between crude oil and gold was 0.1168, while those for the gold/dollar/pound and
oil/dollar/pound markets were 0.2826 and 0.0369, respectively, with the latter
being significantly higher than in the other subperiods.
This remainder of this chapter is organized as follows. Section 58.2 provides
a review of the literature. Section 58.3 describes the data and summary statistics for
crude oil, gold, and the dollar/pound exchange rate. Section 58.4 presents the
dynamic conditional correlation model and reports the results of its volatility
threshold. And also provides the results for subperiods separated by the thresholds
found. Finally, Sect. 58.5 discusses the results and concludes.
Engle et al. (1994) investigate how the returns and volatilities of stock indices
between Tokyo and New York are correlated and find that, except for a lagged
return spillover from New York to Tokyo after the crash, there was no significant
lagged spillover in returns or in volatilities. Ng (2000) examines the size and the
impact of volatility spillover from Japan and the USA to six Pacific Basin equity
markets. Using four different specifications of correlation by constructing volatility
spillover models, he distinguishes the volatility between local idiosyncratic shock,
regional shock from Japan, and global shock from the USA and finds significant
58 Realized Distributions of Dynamic Conditional Correlation 1623
spillover effects from regional to Pacific Basin economies. Andersen et al. (2001a)
find strong evidence that volatilities and correlations move together in a manner
broadly consistent with the latent factor structure. Andersen et al. (2001b) found
that volatility movements are highly correlated across the deutsche mark and yen
against the US dollar. Furthermore, the correlation between the two exchange rates
increases with volatility. Engle (2002) finds that the breakdown of the correlations
between the deutsche mark and the pound and lira in August 1992 is very apparent.
In addition, after the euro is launched, the estimated currency correlation essentially
moves to 1.
Recently, Doong et al. (2005) examined the dynamic relationship and pricing
between stocks and exchange rates for six Asian emerging markets. They found that
the currency depreciation is accompanied by a fall in stock prices. The conditional
variance-covariance process of changes in stock prices and exchange rates is time
varying. Lanza et al. (2006) estimate the dynamic conditional correlations in the
daily returns for West Texas Intermediate (WTI) oil forward and future prices from
January 3, 1985 to January 16, 2004, and find that the dynamic conditional
correlations vary dramatically. Chiang et al. (2009) investigate the probability
distribution properties, autocorrelations, dynamic conditional correlations, and
scaling analysis of Dow-Jones and NASDAQ Intraday returns from August 1,
1997 to December 31, 2003. They find the correlations to be positive and to mostly
fluctuate in the range of 0.6–0.8. Furthermore, the variance of the correlation
coefficients has been declining and appears to be stable during the post-2001
period. Pérez-Rodrı́guez (2006) applies a multivariate DCC-GARCH technique to
examine the structure of the short-run dynamics of volatility returns on the euro,
yen, and British pound against the US dollar over the period from 1999 to 2004 and
finds strong dynamic relationships between currencies. Tastan (2006) applies
multivariate GARCH to capture the time-varying variance-covariance matrix for
stock market returns (Dow-Jones Industrial Average Index and S&P500 Index) and
changes in exchange rates (euro/dollar exchange rates). He also plots news impact
surfaces for variances, covariances, and correlation coefficients to sort out the
effects of shocks. Chiang et al. (2007a) apply a dynamic conditional correlation
model to nine Asian daily stock-return series from 1990 to 2003 and find evidence
of a contagion effect and herding behavior. Chiang et al. (2007b) examine A-share
and B-share market segmentation conditions by employing a dynamic multivariate
GARCH model and show that stock returns in both A- and B-shares are positively
correlated with the daily change in trading volume or abnormal volume.
58.3 Data
Our data consist of the daily prices of crude oil and gold, and the US dollar/British
pound exchange rate, and are obtained from the AREMOS database over the period
from January 1, 1986 to December 31, 2007 for a total of 5,165 observations.
The West Texas Intermediate crude oil price is chosen to represent the oil
spot market, and the price of 99.5 % fine gold, the London afternoon fixing,
1624 T.-L. Shih et al.
40 500
400
20
300
0 200
1986 1988 1990 1992 1994 1996 1998 2000 2002 2004 2006
The Price Movement of Foreign Exchange Market
0.75
0.70
0.65
USD/Pound
0.60
0.55
0.50
0.45
1986 1988 1990 1992 1994 1996 1998 2000 2002 2004 2006
Time
Fig. 58.1 The price movement for the sampled markets from January 1, 1986 to December
31, 2007, for a total of 5,405 observations
is chosen to represent the gold spot market. The daily dollar/pound exchange rate,
which represents the major currencies, is selected to estimate the volatility of
the FX market.
In our sample period, the crude oil price was testing the $100 per barrel threshold
by November 2007. Meanwhile, the price of gold was relatively stable varying
between $415 and $440 per ounce from January to September 2005. However, in
the fourth quarter of 2005, the gold price jumped dramatically and hit $500 per
ounce. In April 2006, the gold price broke through the $640 level. 2007 was a strong
year, with the price steadily rising from $640 on January 2 with a closing London
fixed price of over $836 on December 31, 2007. Since then, prices have continued
to increase to reach new record highs of over $1,700 in 2011.
Figure 58.1 displays the price movements for oil, gold, and the dollar/pound over
the sample period. As shown in Fig. 58.1, gold traded between a low of $252
(August 1999) and a high of $836 (December 31, 2007) per ounce at the fixing,
while oil traded between a low of $10 (in late 1998, in the wake of the Asian
58 Realized Distributions of Dynamic Conditional Correlation 1625
Table 58.1 Summary statistics of the daily returns among crude oil, gold, and dollar/pounda
(January 1, 1986 to December 31, 2007)
Crude oil Gold Dollar/pound
Mean 0.024 0.017 0.006
(0.536) (0.159) (0.4769)
Max 0.437 0.070 0.0379
Min 0.404 0.063 0.0329
Standard dev. 0.029 0.009 0.006
Skewnessb 0.012 0.031 0.164**
(0.729) (0.354) (0.0000)
Kurtosisb 37.915** 5.968** 2.439**
(0.0000) (0.0000) (0.0000)
Jarque-Berac 323,692.99** 8,019.20** 1,363.59**
(0.0000) (0.0000) (0.0000)
a
The table summarizes the daily returns of estimates for the West Texas Intermediate crude oil,
gold, and dollar/pound markets. The sample covers the period from January 1, 1986 through
December 31, 2007 for a total of 5,405 observations
b
The three markets are far away from the skewness and kurtosis of 0 and 3, respectively, implying
that the three markets are not normally distributed
c
Jarque-Bera is the Jarque-Bera test statistic, distributed w22
**Denotes significance at the 0.05 level
Financial Crisis and the United Nations’ oil-for-food program) and a high of $99.3
(November 2007) per barrel. These large variations in the price of both gold and oil
indicate that the DCC and realized distribution are better approaches for detecting
the trading pattern of investors.
Table 58.1 reports the statistics of daily returns for crude oil, gold, and the dollar/
pound exchange rate. The daily returns are calculated as the first differences of the
natural log of the prices times 100. The results show that the crude oil has the
highest return, followed by gold and the dollar/pound.
Table 58.2 The correlation among crude oil, gold, and FX of dollar/pound (January 1, 1986 to
December 31, 2007)
Oil Gold FX
Oil 1
Gold 0.7488 1
FX 0.5592 0.6260 1
Et1 r i,t r j,t
rij,t ¼ rffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
ffi (58.1)
Et1 r 2i,t Et1 r 2j,t
Since the correlation coefficients among crude oil, gold, and dollar/pound FX
markets provide useful measures of the long-term relationship between each pair of
markets, Table 58.2 presents a simple correlation matrix in which the calculation is
based on the constant coefficient given by Eq. 58.1. Some preliminary information
is obtained below. First, the crude oil and gold are highly correlated with
a coefficient of 0.7488, a result that is in line with Steinitz (2006). Secondly, both
gold and crude oil are highly negatively related to the dollar/pound with coefficients
of 0.6260 and 0.5592, respectively, which is consistent with the report of
Nikos (2006).
As the autoregressive conditional heteroskedasticity (ARCH) model has become
the most useful model in investigating the conditional volatility since Engle (1982),
we then follow this model in our analysis. The ARCH model adopts the effect of
past residuals that helps explain the phenomenon of volatility clustering. Bollerslev
(1986) proposed the generalized autoregressive conditional heteroskedasticity
(GARCH) model, which has created a new field in the research on volatility and
is widely used in financial and economic time series. Some of his research attempts
to discuss the effects of more than one variable simultaneously. For instance,
Bollerslev (1990) proposed the constant conditional correlation (CCC) model
which makes a strong assumption, namely, that the correlation among the variables
remains constant in order to simplify the estimation. Engle (2002) later proposed
58 Realized Distributions of Dynamic Conditional Correlation 1627
where t is a time interval, which can be a day, an hour, or 1 min. Here t is a daily
interval. Vt,t is a symmetric conditional correlation matrix of et and Dt,t is a (2 2)
matrix with the conditional variances ht,ii,t for two stock returns (where i ¼ gold,
oil, or the dollar/pound exchange rate) on the diagonal. That is,
hqffiffiffiffiffiffiffiffiffiffi i
Dt,t ¼ diag s2t,ii,t . Equation 58.3 suggests that the dynamic properties of
ð2;2Þ
the covariance matrix Ht,t are determined by Dt,t and Vt,t for a given t, a time
interval that can be 1 min, 1 day, or 1 week and so on. The DCC model proposed by
Engle (2002) involves a two-stage estimation of the conditional covariance matrix
Ht in Eq. 58.3. In the first stage, univariate volatility models are fitted for each of the
qffiffiffiffiffiffiffiffiffiffiffi
returns, and estimates of s2t, ii, t ði ¼ 1, 2, and 3Þ are obtained by using Eq. 58.4.
In the second stage, return residuals are transformed by their estimated standard
qffiffiffiffiffiffiffiffiffiffi
deviations from the first stage. That is t,i,t ¼ et,i,t = s2t,ii,t , where t,i,t is used to
estimate the parameters of the conditional correlation. The evolution of the corre-
lation in the DCC model is given by Eq. 58.5:
where
h Qt,t ¼i (qt,ij,t) is the 2 2 time-varying covariance matrix of t, i, t , Qt ¼
E t, i, t 0t, i, t is the 2 2 unconditional variance matrix of t,i,t, and at,i and bt,i are
non-negative scalar parameters satisfying (at,i + bt,i) < 1. Since Qt does
not generally have ones on its diagonal, we scale it to obtain a proper correlation
matrix Vt,t Thus,
1=2 1=2
V t, t ¼ diag Qt, t Qt, t diag Qt, t , (58.6)
1=2
pffiffiffiffiffiffiffiffiffiffiffi pffiffiffiffiffiffiffiffiffiffiffi
where diag Qt, t ¼ diag 1= qt,11,t , 1= qt,22,t :
Here Vt,t in Eq. 58.6 is a correlation matrix with ones on the diagonal and
off-diagonal elements of less than one in absolute value terms, as long as Qt,t is
positive definite. A typical element of Vt,t takes the form:
pffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
rt, 12, t ¼ qt, 12, t = qt, 11, t qt, 22, t (58.7)
The dynamic correlation coefficient, rt,12,t, can be obtained by using the element
of Qt,t in Eq. 58.5, which is given by Eq. 58.8 below:
0
qt, ij, t ¼ 1 at, i bt, i rt, ij þ at, i t, i, t1 t, j, t1 þ bt, i qt, ij, t1 , (58.8)
The mean reversion requires that (at,i + bt,i) < 1. In general terms, the essence of
this concept is the assumption that both an asset’s high and low prices are temporary
and that the asset’s price will tend to move toward the average price over time.
Besides, the estimates of the dynamic correlation coefficients, rij,t, between each
pair of the three markets have been specified as in Eq. 58.1.
In this section, we present the estimation results of the models outlined above. The
estimation results are presented in Table 58.1, which provides the dynamic corre-
lations of returns across crude oil, gold, and the dollar/pound foreign exchange rate
with each other. The estimated a and b for three markets are listed in Tables 58.3
and 58.4. The likelihood ratio does not support the rejection of the null hypothesis
of the scalar dynamic conditional correlation. It can be seen that the sum of the
estimated coefficients in the variance equations (a + b) is close to 1 for all of the
cases, implying that the volatility appears to be highly persistent. As for the Ljung-
Box Q-statistic of the serial correlation of the residuals, the results show that that
the serial correlations in the error series are regarded as adequate.
Calvert et al. (2006) observed that through the dynamic conditional correlation
distribution, we can more fully understand the real impacts in international markets.
58 Realized Distributions of Dynamic Conditional Correlation 1629
Table 58.3 DCC estimates: three marketsa (January 1, 1986 to December 31, 2007)
DCC
a 0.0202** b 0.9651**
(52.3020) (1,050.395)
a
The t-statistic is given in parentheses
**Denotes significance at the 0.05 level
It can also help with portfolio selection and risk management. In Fig. 58.2, which
reports the results of the dynamic conditional correlation, the estimated correlation
coefficients are time varying, reflecting some sort of portfolio shift between each
two items.
The correlation between crude oil and gold was estimated using the DCC
integrated method, and the results, shown in Fig. 58.2a, are quite interesting.
The correlations are found to be generally positive around 0.2 except for
mid-1990 which turns out to be highly correlated with a correlation of around
0.6605. The possible interpretation for the high correlation is due to the Iraqi
invasion of Kuwait and the Gulf War. The crude oil price jumped from about $15
to over $33 per barrel during that time, so that investors channeled their money into
the gold market because of their fear of inflation. This fact accords with the “flight
to quality” concept, which represents the action of investors moving their capital
away from riskier or more volatile assets to the ones considered to be safer and less
volatile. The correlation between gold and the dollar/pound exchange rate is shown
in Fig. 58.2b for the integrated DCC in the last 20 years. Whereas for most of the
period the correlations were between 0.1 and 0.3, there were two notable drops,
where the stock market crashed in October 1987 and in late 2002, and we also find
two peaks, one in the middle of 1990 and the other in late 1998 where the gold price
dropped to $252 per ounce. Fig. 58.2c shows the correlation between crude oil and
1630 T.-L. Shih et al.
a 0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
−0.1
−0.2
1986 1988 1990 1992 1994 1996 1998 2000 2002 2004 2006
b 0.3
0.2
0.1
−0.0
−0.1
−0.2
−0.3
−0.4
−0.5
−0.6
1986 1988 1990 1992 1994 1996 1998 2000 2002 2004 2006
c 0.40
0.32
0.24
0.16
0.08
0.00
− 0.08
− 0.16
− 0.24
− 0.32
1986 1988 1990 1992 1994 1996 1998 2000 2002 2004 2006
Fig. 58.2 The time series of dynamic conditional correlations (DCC) among each pair of three
markets: (a) daily DCC between crude oil and gold, (b) daily DCC between gold and dollar/pound,
and (c) daily DCC between crude oil and dollar/pound
58 Realized Distributions of Dynamic Conditional Correlation 1631
the dollar/pound that was estimated using the DCC integrated method. Except in the
mid-1990s when they are highly correlated with a coefficient of 0.32, the correla-
tion between crude oil and the dollar/pound is generally negative (with a coefficient
of 0.08 at the beginning of 1986 and a coefficient of 0.16 at the beginning of
2003, respectively).
Key issues relevant in financial economic applications include, for example,
whether and how volatility and correlation move together. It is widely recognized
among both finance academics and practitioners that they vary importantly over
time (Andersen et al. 2001a, b; Engle 2002; Kasch and Caporin 2012). Such
questions are difficult to answer using conventional volatility models, and so we
wish to use the dynamic conditional correlation model to explain the
phenomenon. From Fig. 58.3, the bivariate scatter plots of volatilities and
correlations, it is hard to tell if there is a strong positive association between
each of the two markets sampled. The correlation between two financial assets
will affect the diversification of the portfolio. If two financial assets are highly
negatively correlated, the effect of diversification will be significant, meaning that
the portfolio can balance the returns. This is the so-called idea of not putting all
one’s eggs in the same basket.
According to the empirical data, the dynamic conditional correlation for the
overall gold and dollar/pound (at 0.1986) or the overall oil and dollar/pound
(at 0.0116) can moderately diversify investment risks and can thereby increase
the rate of return. Investors can add commodities and their related derivatives to
portfolios, in an effort to diversify away from traditional investments and assets.
These results are in line with Capie et al. (2005) who found a negative relationship
between the gold price and the sterling/dollar and yen/dollar foreign exchange rates
and Nikos (2006) who found that the correlation between the dollar and gold is
significantly negative. The conclusion we can draw from the results is that gold is
by far the most relevant commodity in hedging against the US dollar. Capie
et al. (2005) observed that gold has served as a hedge against fluctuations in the
foreign exchange value of the dollar. Secondly, gold has become particularly
relevant during times of US dollar weakness. In addition to that, the dynamic
conditional correlation for the overall crude oil and gold markets is 0.0889, and
a similar correlation was documented for the Brent crude oil and gold markets by
Steinitz (2006).
To characterize the distributions of dynamic conditional correlation among the
sampled markets, the summary statistics of the probability distributions for DCC
are shown in Table 58.5, and the associated distributions of DCC for the sampled
markets are shown in Fig. 58.4. We can find that the average DCC between the
crude oil and gold markets is 0.0889 with a standard deviation of 0.0916. The
distribution of the daily DCC between the crude oil and gold markets reflects
a slightly right-skewed (at 1.2021) and leptokurtic distribution (at 4.6799), imply-
ing that a positive DCC occurs more often than a negative DCC between the
crude oil and gold markets. Furthermore, the average DCC between the gold and
1632
a DCC between oil and gold markets b DCC between oil and gold markets
0.7 0.7
0.6 0.6
0.5 0.5
0.4 0.4
0.3 0.3
DCC
DCC
0.2 0.2
0.1 0.1
0.0 0.0
−0.1 −0.1
−0.2 −0.2
0.00 0.01 0.02 0.03 0.04 0.05 0.06 0.0000 0.0001 0.0002 0.0003 0.0004 0.0005 0.0006 0.0007
oil volatility gold volatility
DCC between gold and FX markets DCC between FX and gold markets
c 0.6 d 0.6
0.5 0.5
0.4 0.4
0.3 0.3
0.2 0.2
DCC
DCC
0.1 0.1
−0.0 −0.0
−0.1 −0.1
−0.2 −0.2
−0.3 −0.3
0.0000 0.0001 0.0002 0.0003 0.0004 0.0005 0.0006 0.0007 0.000000 0.000050 0.000100 0.000150
gold volatility FX volatility
DCC between oil and FX markets DCC between oil and FX markets
e 0.36 f 0.36
0.24 0.24
0.12 0.12
0.00 0.00
DCC
DCC
−0.12 −0.12
−0.24 −0.24
−0.36 −0.36
0.000000 0.000050 0.000100 0.000150 0.00 0.01 0.02 0.03 0.04 0.05 0.06
FX volatility oil volatility
Fig. 58.3 Bivariate scatter plots of volatilities and correlations. The scatter plots of the daily DCC between (a) crude oil and gold and (b) crude oil and dollar/
pound against the crude oil volatility; the daily DCC between (c) gold and crude oil and (d) gold and dollar/pound against the gold volatility; and the daily
DCC between (e) dollar/pound and crude oil and (f) dollar/pound and gold against the dollar/pound volatility have been shown
T.-L. Shih et al.
58 Realized Distributions of Dynamic Conditional Correlation 1633
Table 58.5 Summary statistics of probability distributions of DCC for each pair of oil, gold,
and FX
Panel A The DCC distributions for the sampled markets from January 1, 1986 to December 31, 2007
Mean Standard dev. Max Min Skewness Kurtosis
DCC between oil and gold 0.0889 0.0916 0.6605 0.1744 1.2021 4.6799
DCC between gold and FX 0.1986 0.1197 0.2019 0.5596 0.1529 0.3304
DCC between FX and oil 0.0116 0.0799 0.3349 0.3076 0.0201 0.8678
Panel B The DCC distributions for the sampled markets from January 1, 1986 through July 31, 1990
Mean Standard dev. Max Min Skewness Kurtosis
DCC between oil and gold 0.0910 0.0669 0.3003 0.0783 0.3752 0.1478
DCC between gold and FX 0.2567 0.1052 0.0246 0.5506 0.2751 0.0666
DCC between FX and oil 0.0030 0.0734 0.2308 0.3076 0.5582 1.9547
Panel C The DCC distributions for the sampled markets from August 1, 1990 through
August 31, 2001
Mean Standard dev. Max Min Skewness Kurtosis
DCC between oil and gold 0.0720 0.1040 0.6605 0.1744 1.8367 6.5149
DCC between gold and FX 0.1258 0.0857 0.2019 0.3222 0.7307 0.8036
DCC between FX and oil 0.0004 0.0772 0.3349 0.2469 0.1648 1.3984
Panel D The DCC distributions for the sampled markets from September 1, 2001 through
December 31, 2007
Mean Standard dev. Max Min Skewness Kurtosis
DCC between oil and gold 0.1168 0.0758 0.3270 0.0772 0.0521 0.4706
DCC between gold and FX 0.2826 0.1003 0.0688 0.5596 0.2534 0.4195
DCC between FX and oil 0.0369 0.0832 0.2413 0.2459 0.1549 0.0225
7 7
6 6
Probability Density
Probability Density
5 5
4 4
3 3
2 2
1 1
0 0
−0.6 −0.4 −0.2 0 0.2 0.4 0.6 −0.6 −0.4 −0.2 0 0.2 0.4 0.6
The DCC between oil and gold market The DCC between gold and FX market
7 7
WTI and GOLD
6 6
GOLD and FX
WTI and FX
Probability Density
Probability Density
5 5
4 4
3 3
2 2
1 1
0 0
−0.6 −0.4 −0.2 0 0.2 0.4 0.6 −0.6 −0.4 −0.2 0 0.2 0.4 0.6 0.8
The DCC between oil and FX market DCC
Fig. 58.4 The distributions of dynamic conditional correlations among gold, oil, and FX from
1986 to 2007
1
Following Andersen et al. (2001b), the authors classify the days into two groups: low volatility
days and high volatility days. The empirical results show that the distribution of correlations shifts
rightward when volatility increases.
58 Realized Distributions of Dynamic Conditional Correlation 1635
2
Kasch and Caporin (2012) extended the multivariate GARCH dynamic conditional correlation of
Engle to analyze the relationship between the volatilities and correlations. The empirical results
indicated that high volatility levels significantly affect the correlations of the developed markets,
while high volatility does not seem to have a direct impact on the correlations of the transition blue
chip indices with the rest of the markets. It is easy to see that the volatility and correlation move
together.
58 Realized Distributions of Dynamic Conditional Correlation 1637
a 12 12
High Volatility Days High Volatility Days
Low Volatility Days Low Volatility Days
10 10
Probability Density
Probability Density
8 8
6 6
4 4
2 2
0 0
−0.1 −0.05 0 0.05 0.1 0.15 0.2 0.25 0.3 −0.1 −0.05 0 0.05 0.1 0.15 0.2 0.25 0.3
The DCC between oil and gold markets The DCC between oil and gold markets
b 8 10
High Volatility Days High Volatility Days
7 9
Low Volatility Days Low Volatility Days
8
Probability Density
6
Probability Density
7
5 6
4 5
3 4
3
2
2
1 1
0 0
−0.5 −0.4 −0.3 −0.2 −0.1 0 −0.6 −0.5 −0.4 −0.3 −0.2 −0.1 0
The DCC between gold and FX markets The DCC between gold and FX markets
c 14 10
Low Volatility Days Low Volatility Days
9
12 High Volatility Days High Volatility Days
8
Probability Density
Probability Density
10 7
6
8
5
6 4
4 3
2
2
1
0 0
−0.25 −0.2 −0.15 −0.1 −0.05 0 0.05 0.1 0.15 −0.2 −0.1 0 0.1 0.2 0.3
The DCC between oil and FX markets The DCC between oil and FX markets
Fig. 58.5 (a) The distributions of DCC between crude oil and gold on low and high volatility
days. (b) The distributions of DCC between gold and dollar/pound on low and high volatility days.
(c) The distributions of DCC between oil and dollar/pound on low and high volatility days
The three subperiods based on our estimated thresholds are before the first Gulf War
(January 1, 1986 to July 31, 1990), after the first Gulf War up to the 911 attack
(August 1, 1990 to August 31, 2001), and after the 911 attack (September 1, 2001 to
December 1, 2007). We then examine the dynamic co-movement in each pair of
markets in various subperiods (Rigobon and Sack 2005; Guidi et al. 20073).
Our sampled period covers economic hardship and soaring energy prices.
Soaring energy prices make gold an attractive hedging asset against inflation in
that a positive correlation with oil is expected over time, especially after the
911 event. The evidence in Table 58.10 shows that oil and gold are highly
3
Guidi et al. (2007) examined the impact of relevant US decisions on oil spot price movements
from January 1986 to December 2005. They identified the following conflict periods: the Iran-Iraq
conflict, January 1985 until July 1988; Iraq’s invasion of Kuwait, August 1990 until February
1991; and the US-led forces’ invasion of Iraq, March 2003 until December 2005.
1640 T.-L. Shih et al.
Table 58.10 Simple correlation matrix of oil, gold, and dollar/pound markets among the three
subperiods
Panel A The correlation coefficients between crude oil, gold, and dollar/pound markets from January 1,
1986 through July 31, 1990
Oil Gold FX
Oil 1
Gold 0.1648 1
FX 0.1517 0.5228 1
Panel B The correlation coefficients between crude oil, gold, and dollar/pound markets from August 1,
1990 through August 31, 2001
Oil Gold FX
Oil 1
Gold 0.2465 1
FX 0.0628 0.2096 1
Panel C The correlation coefficients between crude oil, gold, and dollar/pound markets from September 1,
2001 through December 31, 2007
Oil Gold FX
Oil 1
Gold 0.9264 1
FX 0.8124 0.8142 1
Probability Density
1986 to July 31, 1990, (b) 20
August 1, 1990 to August 31,
2001, and (c) September 1, 15
2001 to December 31, 2007
10
0
−0.5 −0.4 −0.3 −0.2 −0.1 0 0.1 0.2 0.3
DCC during 19860101-19900731
b 12
WTI and GOLD
GOLD and FX
WTI and FX
10
Probability Density
0
−0.4 −0.2 0 0.2 0.4 0.6 0.8 1
DCC during 19900801-20010831
c 8
WTI and GOLD
GOLD and FX
7 WTI and FX
6
Probability Density
0
−0.6 −0.5 −0.4 −0.3 −0.2 −0.1 0 0.1 0.2 0.3 0.4
DCC during 20010901-20071231
1642 T.-L. Shih et al.
Probability Density
dollar/pound over three 20
subperiods
15
10
0
−0.1 0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8
DCC between crude oil and gold markets
b 25
19860101–19900731
19900801–20010831
20010901–20071231
20
Probability Density
15
10
0
−0.6 −0.5 −0.4 −0.3 −0.2 −0.1 0 0.1 0.2 0.3
DCC between gold and FX markets
c 30
19860101–19900731
19900801–20010831
20010901–20071231
25
Probability Density
20
15
10
0
−0.3 −0.2 −0.1 0 0.1 0.2 0.3 0.4
DCC between crude oil and FX markets
58 Realized Distributions of Dynamic Conditional Correlation 1643
Using the dynamic conditional correlation model, we have estimated the cross
correlation and volatility among crude oil, gold, and dollar/pound currencies from
1986 to 2007. After exploring the time-varying correlations and realized distribu-
tions, several regularities have been found that help illustrate the characteristics of
the crude oil, gold, and currency markets.
First, the correlation coefficients between each pair of the three assets are found
to be time varying instead of constant. As such, besides considering the mean and
standard deviation of the underlying assets, investors need to follow the
co-movement in the relevant assets in order to make better portfolio hedging and
risk management decisions across these assets. The results of the dynamic correla-
tion coefficients between the gold and dollar/pound show that gold is by far the
most relevant commodity in terms of serving as a hedge against the US dollar.
These results are in line with the reports suggested by Nikos (2006) and Capie
et al. (2005). Our findings are helpful in terms of arriving at a more optimal
allocation of assets based on their multivariate returns and associated risks.
Besides, the distributions of low volatility days are found to approximate
leptokurtic distributions in the gold, oil, and dollar/pound markets, whereas the
high volatility days approximate platykurtic distributions. Furthermore, the DCC
between oil and gold is increasing with volatility, indicating that the distribution of
DCC shifts rightward when volatility increases. By contrast, the DCCs between
gold and the dollar/pound and crude oil and the dollar/pound are decreasing with
the volatility. Our findings in terms of oil and gold are consistent with the reports of
Sonlinik et al. (1996) and Andersen et al. (2001b) who use different approaches.
Moreover, by estimating the volatility threshold dynamic conditional correlation
model addressed by Kasch and Caporin (2012), we find that high volatility values
(exceeding some specified thresholds) are associated with an increase in correlation
values in various subperiods. Remarkably, investors’ behaviors are seen to have
changed in different subperiods. During periods of market turmoil, such as the First
Gulf War in 1990 and the 911 terror attack in 2001, an increase in correlation
between the prices of oil and gold, as well as a decrease in correlation between the
oil (gold) and dollar/pound currencies, is observed. These behaviors make gold
an attractive asset against major currencies for value-preserving purposes.
For market participants from long-term hedging perspective, our results provide
useful information on asset allocation across commodity and currency markets
during market turmoil.
Acknowledgments This chapter was presented at the Seventh International Business Research
Conference in Sydney, Australia, and the Taiwan Finance Association Annual Meeting in
Taichung, Taiwan. The authors are grateful for the helpful and suggestive comments from Ken
Johnson, C. L. Chiu, Ming-Chi Lee, and other participants. The authors also appreciate the
financial grants for attending the conference from the National Science Council.
1644 T.-L. Shih et al.
References
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Ebens, H. (2001a). The distributions of realized
stock return volatility. Journal of Financial Economics, 61, 43–76.
Andersen, T. G., Bollerslev, T., Diebold, F. X., & Labys, P. (2001b). The distribution of realized
exchange rate volatility.