CH 4 Valuation and Risk Models
CH 4 Valuation and Risk Models
By AnalystPrep
1
© 2014-2020 AnalystPrep.
Table of Contents
2
© 2014-2020 AnalystPrep.
Reading 45: Measures of Financial Risk
Compare the normal distribution with the typical distribution of returns of risky
Define the VaR measure of risk, describe assumptions about return distributions and
Explain and calculate Expected Shortfall (ES), and compare and contrast VaR and ES.
Define the properties of a coherent risk measure and explain the meaning of each
property.
Describe spectral risk measures and explain how VaR and ES are special cases of
The mean-variance framework uses the expected mean and standard deviation to measure the
financial risk of portfolios. Under this framework, it is necessary to assume that returns follow a
The normal distribution is particularly common because it concentrates most of the data around
the mean return. 66.7% of returns occur within plus or minus one of the standard deviations of
the mean. A whopping 95% of the returns occur within plus or minus two standard deviations of
the mean.
3
© 2014-2020 AnalystPrep.
Investors are normally concerned with downside risk and are therefore interested in
Note the expected does not imply the anticipated return but rather the average returns. On the
other hand, the risk is measured using the standard deviation of returns.
The expected returns for an asset with corresponding probabilities are given below:
Calculate the expected return and standard deviation of the asset return.
4
© 2014-2020 AnalystPrep.
Solution
To calculate the expected return, we weight the expected return by their corresponding
n
R̄ = ∑ p iR
i=1
R̄ = (0.10 × 0.25) + (−0.20 × 0.09) + (0.15 × 0.40) + (0.07 × 0.06) + (0.30 × 0.20)
= 0.1312 = 13.12%
σR = √E (R2 ) − [E (R)]2
Therefore,
Combinations of Investments
Consider two investments with respective means μ1 and μ 2 . Suppose that an investor wishes to
5
© 2014-2020 AnalystPrep.
invest in both investments with a proportion of w 1 in the first investment and w2 in the second
The portfolio expected return is equivalent to weighted returns from individual investments. That
is:
μp = w 1 μ1 + w 2 μ2
Where
Cov (R 1 ,R 2 )
Cov (R 1 , R2 ) = ρ =
σ1 σ2
⇒ Cov (R 1 , R2 ) = ρσ1 σ2
6
© 2014-2020 AnalystPrep.
An investor invests in two assets X and Y, with an expected return of 10% and 15%. The investor
invests 45% of his funds in asset X and the rest in asset Y. The correlation coefficient is 0.45.
Given that the standard deviation of asset X is 15% and Y is 30%, what are the expected return
Solution
μp = wX μX + wY μ Y
= 0.10 × 0.45 + 0.15 × 0.55
= 0.1275 = 12.75%
Calculating the portfolio expected return and standard deviation can be extended to a portfolio
with n investments. The portfolio expected return for n returns is given by:
n
μ p = ∑ w i μi
i=1
Where μ i and wi are the mean return and weight of ith investment
n n
σP = ∑ ∑ ρij w i wj σiσj
i=1 j=1
where ρij is the correlation coefficient between investments i and j. Other variables are intuitively
definitive.
Efficient Frontier
7
© 2014-2020 AnalystPrep.
The efficient frontier represents the set of optimal portfolios that offers the highest expected
return for a defined level of risk or the lowest risk for a given level of expected return. This
concept can be represented on a graph pitting the expected return (Y-axis) against the standard
deviation (X-axis).
For every point on the efficient frontier, at least one portfolio can be constructed from all
available investments that have the expected risk and return corresponding to that point.
Portfolios that do not lie on the efficient frontier are suboptimal: those that lie below the line do
not provide enough return for the level of risk, and those that lie on the right of the line have a
8
© 2014-2020 AnalystPrep.
The choice between optimal portfolios A, B, and D above will depend on an individual investor's
appetite for risk. A very risk-averse investor will choose portfolio A because it offers an optimal
return at the lowest risk, whereas an investor with room for more risk might pick D. After all, it
The efficient frontier above considers only the risky assets. Now, consider when we introduce a
risk-free investment with a return of RF . It can be shown that the efficient frontier is a straight
line. That is, there is a linear relationship between expected return and the standard deviation of
return.
9
© 2014-2020 AnalystPrep.
F represents the risk-free return on the diagram above. Note that risk-free asset sits on the
efficient frontier because you cannot get a higher return with no risk, and you cannot have less
risk than zero. Consider the tangent line FM. By proportioning our investment between the risk-
free asset F and the risky asset M (market portfolio), we can obtain a risk-return that lies on the
Denote the risk-free return by R F (with a standard deviation of 0). Also, let the market portfolio
return be RM , and its standard deviation is σM . Let the proportion of funds in a risky portfolio be
μp = w 1 μ1 + w 2 μ2
10
© 2014-2020 AnalystPrep.
We have w1 = 1 − β, w2 = β, μ1 = R F, μ2 = R M so that return from the portfolio is given by
μ p = R F (1 − β) + βR M
μp − RF
⇒β=
RM − RF
σ = √w21 σF2 + w 22 σM
2 + 2ρw w σ σ
1 2 F M
But σF = 0
⇒ σ = √ 0 + w 22σM
2 + 0 = w 2 σM = βσ M
Therefore,
μP − R F
σ = σM ( )
RM − RF
σM σMR F
⇒ σ = μp ( )−
RM − RF RM − RF
The efficient frontier involving risk-free asset also shows that the investor should invest in risky
assets (in this case, M) by borrowing and lending at a risk-free rate rF . For instance, we assume
that an investor borrows at the rate rF so that now we are considering the efficient frontier
beyond M. If this is the case, then β > 1 and the proportion of amount borrowed will be β − 1,
and the total amount available is β multiplied by available funds. Assume now that we invest in
βrM − (β − 1) RF = (1 − β) rF + βrM
The standard deviation can be shown to be βσM , which is similar to arguments for the points
below point M.
Therefore, it is safe to say risk-averse investors will invest in points on line FM and close to F,
11
© 2014-2020 AnalystPrep.
and those investors that are risk-seeking will invest on points close to M or even points beyond M
on line FM.
The mean-variance framework is unreliable when the assumption of normality is not met. If the
return distribution is not symmetric, the standard deviation is not a reliable or relevant measure
The normality assumption is only strictly appropriate in the presence of a zero-skew (symmetric)
distribution. If the distribution leans to the left or the right – something that often happens with
financial returns – the mean-variance framework churns out misleading estimates of risk.
12
© 2014-2020 AnalystPrep.
The Normal Distribution
The normal distribution, also called Gaussian distribution, is a widely used continuous
distribution with two parameters: mean denoted by μ and the standard deviation denoted by σ .
(x − μ)2
1 −
f (x) = e 2σ2
√2πσ 2
The height of the normal distribution is equivalent to the probability that a value X has occurred.
This is commonly stated as X ∼ N(μ, σ 2). Values close to the center of the distribution are most
likely to occur while those values at the tails of the distribution are less likely to occur.
Note that, similar to other probability distributions, the probability that value lies between the
value a and b is equivalent to the area under the curve between the values a and b. This can be
13
© 2014-2020 AnalystPrep.
thought of as the cumulative distribution up to a point B less the cumulative distribution up to
point A.
A standard normal distribution has a mean of 0 and a standard deviation of 1. In other words, μ
=0 and σ=1. As such, the normal distribution density function reduces to:
x2
1 −
f (x) = f (x) = e 2σ 2
√2π
The normal tables give a cumulative distribution of the standard normal distribution. For normal
distribution with the mean μ and the standard deviation σ , it can be transformed into z-scores,
which gives cumulative probability up to a value x for standard normal. The z-score is defined by:
x −μ
z=
σ
Where z ∼N(0,1).
For example, consider a normal distribution with a mean of 4 and a standard deviation of 5. What
is the probability that a value X is less than 7? Using standard normal transformation,
X −μ 7− 4
Pr(X < 7) = P r( < ) = 0.6
σ 5
= P r(z < 0.6) = Φ(0.6) = 1 − 0.2743 = 0.7257
Note that this is taken from the reference table you will be given on your exam:
14
© 2014-2020 AnalystPrep.
A normal distribution is usually assumed to apply to financial data because financial analysts are
mostly concerned with the mean and standard deviation. However, financial variables have fatter
15
© 2014-2020 AnalystPrep.
tails than the normal distribution. For instance, the means created can have fatter tails. Consider
The diagram shows that the normal distribution and the actual distribution have equal standard
deviations, but the actual distribution has more peaked and has fatter tails than the normal
distribution. In other words, the actual distribution suggests that small and large changes
Assuming a normal distribution for financial variables (by use of mean and standard deviation)
may likely underestimate the probability of the adverse events. The standard deviation can be a
perfect measure of risk, but it does not capture the tails of the probability distribution.
16
© 2014-2020 AnalystPrep.
VaR is a risk measure that is concerned with the occurrence of adverse events and their
corresponding probability. VaR is built from two parameters: the time horizon and the
confidence level. Therefore, we can say that VaR is the loss that we do not anticipate to be
For example, consider a time horizon of 30 days and a confidence interval of 98%. Therefore 98%
VaR of USD 5 million implies that we are 98% certain that over the next 30 days, the loss will be
less than USD 5 million. Similarly, we can say that we are 2% certain that over the next 30 days,
Therefore, calculating X% VaR involves finding the loss that has an X% chance of being
17
© 2014-2020 AnalystPrep.
The investment return over a period of time has a normal loss distribution with a mean of -200
Solution
P (X < t) = 0.99
Standardizing the normal distribution with a given mean and standard deviation, we have:
P (X < t) = 0.99
Standardizing the normal distribution with a given mean and standard deviation, we have:
t − −200
P (z < ) = 0.99
300
t + 200
⇒ Φ( ) = (0.99)
300
t + 200 −1
⇒ = Φ (0.99)
300
Now, Φ −1(0.99) is the inverse of standard normal cumulative probability. To do this using a
standard table, look for 0.99 (or closest value) in the table and read the corresponding vertical
and horizontal values and add them. In other words, we are reversing the reading of the
18
© 2014-2020 AnalystPrep.
And thus:
19
© 2014-2020 AnalystPrep.
Φ−1 (0.99) = 2.33 + 0.02 = 2.32
t + 200
∴ = 2.33 ⇒ t = 499
300
Solution
To find the 99%, we need to find the cumulative probability distribution and locate 99%:
Therefore, with the confidence level of 99%, the VaR value is USD 17 million because 99% falls
Note that if we reduce our confidence level to 95%, VaR will change to USD 13 million because
However, if the confidence level is 97%, then we could have two VaR values: USD 13 million and
USD 17 million. This will be ambiguous, and so the best estimate is the average of the values,
Limitations of VaR
20
© 2014-2020 AnalystPrep.
I. It does not describe the worst possible loss. Indeed, as seen from the example above,
we would expect the $100 million loss mark to be breached 5 times out of a hundred for
II. VaR does not describe the losses in the left tail. It indicates the probability of a value
occurring but stops short of describing the distribution of losses in the left tail.
III. Two arbitrary parameters are used in its calculation – the confidence level and the
holding period. The confidence level indicates the probability of obtaining a value greater
than or equal to VaR. The holding period is the time span during which we expect the
loss to be incurred, say, a week, month, day, or year. VaR increases at an increasing rate
as the confidence level increases. VaR also increases with increases in the holding
period.
IV. VaR estimates are subject to both model risk and implementation risk. Model risks arise
from incorrect assumptions, while implementation risk is the risk of errors from the
implementation process.
Recall the VaR does not describe the worst possible loss. For instance, if 99% VaR is USD 10
million, we know that we are 1% certain that the loss will exceed USD 10 million. From the VaR
level, we cannot say that the loss is greater than 20 million or USD 50 million. Therefore, VaR
sets a risk measure equal to a certain percentile of the loss distribution and does not consider
Expected shortfall (ES) is a risk measure that considers the expected losses beyond the VaR
level. In other words, ES is the expected loss conditional that the loss is greater than the VaR
level.
Exam tip: Expected shortfall is also called conditional value at risk (CVaR), average value at risk
(AVaR), or expected tail loss (ETL). Think about this as the average loss beyond the VaR.
When the losses are normally distributed with the mean μ and standard deviation σ, then the ES
is given by:
U2
21
© 2014-2020 AnalystPrep.
U2
⎛ − ⎞
⎜⎜ e 2 ⎟⎟
ES = μ + σ ⎜
⎜⎜ (1 − X) √2π ⎟⎟⎟
⎝ ⎠
Where
U = the point in the standard normal distribution that has a probability of X% of being exceeded.
The investment return over a period of time has a normal loss distribution with a mean of -200
Solution
We know that:
U2
⎛ e− 2 ⎞
ES = μ + σ ⎜ ⎟
⎝ (1 − X) √2π ⎠
(2.32) 2
⎛ e− 2 ⎞
⎜
ES = −200 + 300 ⎜ ⎟⎟ = 611.44
⎝ (1 − 0.99) √ 2π⎠
ES should always be greater than the VaR level because the ES gives us the average of the
22
© 2014-2020 AnalystPrep.
Amount of Loss Probability
USD 20 Million 2%
USD 17 Million 8%
US 13 million 12%
US 10 million 78%
Solution
At a 95% confidence level, we need to answer the question, "given that we are at 5% of the loss
distribution, what is the value of the expected loss?".Now looking at the probability column, it is
clear to see that 5% tail distribution consists of a 2% probability that the loss is USD 20 million
and 3% that the loss is USD 17 million. Conditioned that we are dealing with tail distribution,
then there is a 2/5 chance that the loss is USD 20 million and 3/5 chance that the loss is USD 17
2 3
× 20 + × 17 = 18.20
5 5
Again, note that the 97% VaR is USD 10 million, which is less than ES.
A risk measure summarizes the entire distribution of dollar returns X by one number, ρ (X). There
are four desirable properties every risk measure should possess. These are:
I. Monotonicity: If X1 ≤ X2 , ρ (X1 ) ≥ ρ (X 2 )
Interpretation: If a portfolio has systematically lower values than another, it must have a
greater risk in each state of the world. In other words, if a portfolio gives undesirables
23
© 2014-2020 AnalystPrep.
Interpretation: When two portfolios are combined, their total risk should be less than (or
equal to) the sum of their risks. Merging of portfolios ought to reduce risk. This property
then the overall risk is the sum of their risk when considered separately. However, if the
two portfolios are not perfecdtly correlated, their overall risk should decrease due to
diversification benefits.
proportionate scale in its risk measure. For instance, if we increase the portfolio size by a
measured in dollars. This property reflects that more cash acts as a "loss absorber" and
If a risk measure satisfies all four properties, then it is a coherent risk measure. Expected
Value at risk is not a coherent risk measure because it fails the subadditivity test. Here's an
illustration:
Suppose we want to calculate the VaR of a portfolio at 95% confidence over the next year of two
24
© 2014-2020 AnalystPrep.
The recovery rate upon default is 30%.
Given these conditions, the 95% VaR for holding either of the bonds is 0 because the probability
of default is less than 5%. Now, what is the probability 'P' that at least one bond defaults?
So if we held a portfolio that consisted of 50% A and 50% B, then the 95%
This violates the subadditivity principle, and VaR is, therefore, not a coherent risk measure.
A spectral risk measure is a risk measure given as a weighted average of return quantiles from
the loss distribution. It's given as a weighted average of outcomes where bad outcomes typically
have more weight. A spectral risk measure is always a coherent risk measure.
A risk measure is a coherent risk measure if and only if the weights are a non-decreasing
function of the percentile of the loss distribution. To demonstrate this, let w (p) be a weighted
w (p1 ) ≥ w (p 2 ) if p 1 > p2
VaR is also a particular case of risk spectrum measurement, but it places no weight on tail losses.
For instance, given a VaR confidence level of 95%, all the weight is concentrated on the 95th
percentile. If we denote p2 =95 % and p 1=98%, w (p2 ) = 1 and w (p 2 )=0 implying that rule above
The expected shortfall is a particular case of risk spectrum measurement where the weighting
function is set to [1/(1−confidence level)] for tail losses, with all the other quantiles weighting
zero. In other words, no weight is assigned to losses that are less than the VaR level and that all
losses greater than VaR loss are given the same weight. Conclusively, ES is a coherent risk
25
© 2014-2020 AnalystPrep.
measure where the weights assigned to percentiles increase in a way that risk aversion is
1− p
−
e λ
Where:
λ is a constant that reflects the degree of aversion for a user (decreases with an increase in risk
aversion).
26
© 2014-2020 AnalystPrep.
Practice Question
Ann Conway, FRM, has spent the last several months trying to develop a new risk
Prior to its use, her supervisor has asked her to demonstrate that it is a coherent risk
Given:
P (x) and P (y) are risk measures for portfolio x and portfolio y.
Which of the following equations shows that Conway's risk measure is not coherent?
A. P (kx) = kP (x)
C. P (x) ≤ P (y) if x ≤ y
D. P (x+l) = P (x) − l
Option C, as represented above, shows that the risk measure does not satisfy the
has systematically lower values than another, in each state of the world, it must have
a greater risk.)
property.
27
© 2014-2020 AnalystPrep.
Reading 46: Calculating and Applying VaR
Describe and explain the historical simulation approach for computing VaR and ES.
Describe the delta-normal approach for calculating VaR for nonlinear derivatives.
Explain structured Monte Carlo and stress testing methods for computing VaR, and
Describe the worst-case scenario (WCS) analysis and compare WCS to VaR.
A linear portfolio linearly depends on the changes in the values of its corresponding variables
(risk factors). For instance, consider a portfolio consisting of 100 shares, each valued at USD
100. Therefore, the change in portfolio value (ΔP) is attributed to change in stock (share price)
which can be denoted by ΔS, and thus the change in portfolio value is given by :
ΔP = 100ΔS
The value of the portfolio is USD 10,000 (=100×100). Now, if we introduce the effect of the
interest rate, the change in the value of the portfolio will be given by:
28
© 2014-2020 AnalystPrep.
ΔP = 100Δr
Generally, consider a portfolio consisting of long and short positions in stocks. The change in a
ΔP = ∑ n iΔSi
i
Where:
Si
Now, if we multiply the above equation by , we have:
Si
Si ΔSi
ΔP = ∑ n iΔSi × = ∑ ni ΔSi
i Si i Si
ΔSi
Let qi = ni Si and Δri = , then
Si
ΔP = ∑ q iΔri
i
ΔS i
Note that q i is the amount invested in stock i, and Δri = is the return on stock i.
Si
Therefore, we can say the portfolio change is a linear function of change in stock price or change
in stock returns.
Nonlinear portfolios contain complex securities that are not linear. For instance, consider a
portfolio made of call options. The payoff from the call option is nonlinear because the payoff is
zero if the stock price at maturity is less than the strike price and S-K if the stock price is higher
29
© 2014-2020 AnalystPrep.
However, a forward contract is an example of a derivative whose value is a linear function of the
asset because even if the contracts do give a payoff. The holder is obligated to buy the asset at a
future time T at agreed price K. As such, the forward contract's value is given by:
S − PV (K)
Where S is the current asset price, and PV (K) denoted the present value of the future price K.
30
© 2014-2020 AnalystPrep.
VaR for Linear Derivatives
Where Δ represents the sensitivity of the derivative's price to the price of the underlying asset. It
Suppose the permitted lot size of S&P 500 futures contracts is 300 and multiples thereof. What is
31
© 2014-2020 AnalystPrep.
the VaR of an S&P 500 futures contract?
Historical simulation is used to calculate one-day VaR and ES. However, for longer periods T it is
assumed that,
VaR (T, X) = value at risk for a time horizon of T days and confidence level X.
ES (T, X) = expected shortfall for a time horizon of T days and confidence level X.
The above estimates assume that the portfolios' changes are normally distributed with a mean of
In a historical simulation, market variables (risk factors) on which the portfolio value depends.
Examples of such variables include commodity prices, equity prices, and volatilities. After this,
the data on the movement of these risk factors in the past is collected. After data collection,
scenarios are built by assuming that each risk factor's change over the next day corresponds to a
i. Those whose past percentage change is to define the future percentage, for example,
ii. Those whose past actual change is used to define an actual change in the future, for
32
© 2014-2020 AnalystPrep.
A portfolio is assumed to depend on many risk factors. For simplicity, let us assume three risk
factors (exchange rates, interest rates, and stock price) over the past 300 days (longer periods
are usually considered, such as 500 days). The most recent 301 days of historical data is as
follows:
Assuming that today is the 300th day, we need to know what will happen between today and
tomorrow (301st day). To achieve this, we use the above data to create 300 scenarios (that is why
In the first scenario, we will assume that the risk factors behave between the days 300 and 301
in a similar manner as they did between days 0 and 1. For instance, in the first scenario, the
34
stock price increased by 13% (= − 1), and thus the stock price on day 101 is USD 68
30
[= ( 34 1.400
− 1) × 60] . For the interest rate, it increased by 0.7% (= − 1) , and thus, we expect the
30 1.3901
exchange rate for the 101st day to be 1.4121 (= 1.4021 [ 1.400 − 1]). For the interest rate, it
1.3901
decreased by 0.87% (2.64%-3.51%) and thus 101st interest rate is 1.63% (2.50%-0.87%).
The values for the second and subsequent scenarios' risk factors are calculated similarly as the
first scenario. For the second scenario, assume that the risk factors behave in a similar manner
as they did between days 1 and 2. This will create the following table.
33
© 2014-2020 AnalystPrep.
Scenario Stock Price Exchange Interest Portfolio Loss
(USD) rates rate (%) Value
(USD/CAD) (USD
millions)
1 68 1.4121 3.37 70.25 1.0
2 71 1.3922 2.38 72.15 0.9
.. .. .. .. ..
299 57.14 1.3910 2.55 71.25 0
300 90 1.4021 2.55 73.25 2.0
The risk factor values for the scenario table are directly calculated from the historical data table.
For the scenario portfolio values, we must generate based on the risk factors. We assume that
the current portfolio value is USD 71.25 million (300th day's value). After generating the
portfolio values, we then calculate the losses while attaching a negative to create a loss
distribution.
Assume that the first scenario's portfolio value is 70.25, 72.15 for the second scenario, and so on.
Therefore, the loss for the first portfolio is 1.00 (=71.25-70.25), and the second scenario is 0.9
In order to calculate the VaR and the expected shortfall, we ought to arrange the scenario losses
from the largest to the smallest. Assume that in our example, we wish to calculate one day VaR
Scenario Loss
200 3.9
10 3.0
25 2.5
100 2.0
.. . .. .
.. . .. .
In this case, VaR is equivalent to third-worst loss since the third-worst loss is the first percentile
3
point of the distribution, i.e., = 0.01. Therefore, VaR=2.5 million.
100
By definition, the expected shortfall is calculated as the average of the losses that is worse than
1
34
© 2014-2020 AnalystPrep.
1
ES = (3.9 + 3.0) = 3.45 million
2
The following are hypothetical ten worst returns for an asset B from 120 days of data for 6
-3.45%, -14.12%, -15.72%, -10.92%, -5.50%, -3.56%, -6.90%, -2.50%, -5.30%, -4.31%.
Solution
First, we rearrange starting with the worst day, to the least bad day, as shown below:
-15.72%, -14.12%, -10.92%, -6.90%, -5.50%, -5.30%, -4.31%, -3.56%, -3.45%, -2.50%.
The VaR corresponds to the (5% × 120)=6th worst day = -5.30%. However, recall that VaR need
This implies that there is a 95% probability of getting at most 5.3% loss.
The expected shortfall (ES) is calculated as the average of the losses that is worse than the VaR.
In this case,
Under the full revaluation approach, the VaR of a portfolio is established by fully repricing the
to use Greek letters. The Greek letters are the hedging parameters used by the analysts to
35
© 2014-2020 AnalystPrep.
One of the crucial Greek letter deltas (δ) which is defined as:
ΔP
δ=
ΔS
Where ΔS is a small change in risk factors such as stock price and ΔP the corresponding change
in the portfolio value. Therefore, delta can be defined as a change of the portfolio value with
For instance, consider stock price as a risk factor. If the delta of a portfolio with respect to the
stock price is USD 100, it implies that the portfolio value changes by USD 100 if the stock price
changes by 1 USD.
From the delta formula, we express the change in the portfolio value as :
ΔP = δΔS
Generally, if we have multiple risk factors, we find each risk factor's effect and sum it up. That is,
ΔP = ∑ δi ΔSi
i
However, the delta concept gives relatively accurate estimates in linear portfolios as compared
to nonlinear portfolios.
The accuracy of nonlinear portfolios can be enhanced by including another Greek letter gamma
1
ΔP = δΔS + γ(ΔS)2
2
1 2
ΔP = ∑ δi ΔSi + ∑ γi (ΔSi)
i 2 i
36
© 2014-2020 AnalystPrep.
Risk Factors Expressed as Term Structures
We have considered risk factors such as stock and prices, which are expressed with a single
number and hence easy to manage. However, some of the risk factors, such as interest rates,
credit spreads, and at-the-money implied volatility are described using term structures.
Interest rates such US treasury rate are usually given in points such as a one-month rate, three-
month rate, and so on. Linear interpolation is usually used to approximate the rates between
these points.
Credit spreads are excess of the borrowing rate of a company over the set limit (benchmark
level). Credit spreads are stated in terms of basis points. As such, the credit spread for a
company might be stated as 100-basis points for a 2-year maturity and 200-basis points for a 6-
year maturity. Respectively, this implies that a firm will pay an extra 1% more than the risk-free
rate in case it borrows for two years and 2% more for six years.
At-the-money implied volatility is also expressed by the term structure. At the money implied
We usually use recent data to develop scenarios to calculate VaR and ES. Nevertheless, bank
regulators have suggested the stressed VaR and stressed ES to evaluate the capital requirements
of a bank.
Stressed VaR and stressed ES are calculated from one-year data that is particularly stressful to
the bank's current portfolio. For instance, a stressful period that might be considered in the
The scenarios are created from the percentage change or actual changes from a stressful period.
Therefore, 250 scenarios are generated from the one-year data. All the calculations in computing
37
© 2014-2020 AnalystPrep.
The delta-normal model is based on the equation (as seen earlier):
ΔP = ∑ δi ΔSi
i
Recall this equation gives an exact value in a linear portfolio and an approximate value in
nonlinear portfolios.
1. Those whose past percentage change is to define the future percentage; and
2. Those whose past actual change is used to define an actual change in the future.
In order to accommodate both types of risk factors, the equation above is written as:
38
© 2014-2020 AnalystPrep.
ΔP = ∑ aix i
i
ΔS i
For risk factors where percentage changes are used, ai = and x i = δiSi .
Si
And for the risk factors where actual changes are considered ai = ΔSi and x i = δi.
From the resulting equations, the mean and the standard deviation of the change in portfolio
n
μ P = ∑ aiμ i
i=1
n n
σP2 = ∑ ∑ ai aj ρij σi σj
i=1 j=1
Where:
Assuming that the portfolio changes are normally distributed, then we can comfortably compute
VaR = μ P + σP U
U2
⎛ e− 2 ⎞
ES = μ P + σP ⎜ ⎟
⎝ (1 − X) √2π ⎠
Where:
X = confidence level
39
© 2014-2020 AnalystPrep.
For instance, if X=95% then U=Φ−1 (0.05) = −1.645.
At this point, you might guess where the name "delta-normal" name comes from: the model uses
deltas of the risk factors and assumes that the portfolio changes are normally distributed.
A typical assumption is that the mean change in the risk factor is zero. This assumption is
sometimes not reasonable but is useful when dealing with short time periods because the mean
is less than the standard deviation for short periods when dealing with portfolio value changes.
VaR = σP U
U2
−
⎛ e 2 ⎞
ES = σP ⎜ ⎟
⎝ (1 − X) √2π ⎠
The method has several disadvantages, chief among them being that:
methods. Put more precisely, it may underestimate the occurrence of extreme losses
This method is accurate for small moves of the underlying, but quite inaccurate for
large moves. As we can see from the following graph, the slope of the green line is
40
© 2014-2020 AnalystPrep.
For large changes in a nonlinear derivative, we must use the delta + gamma approximation, or
full revaluation.
Monte Carlo approach is similar to that of the historical simulation, but Monte Carlo simulation
produces scenarios by randomly selecting samples from the distribution assumed for the risk
factors instead of historical data. Monte Carlo simulations work for both linear and nonlinear
portfolios.
Now, if for instance, we assume that risk factor changes have a multivariate normal distribution
41
© 2014-2020 AnalystPrep.
Step 1: Calculate the value of the portfolio today using the current values of the risk factors.
Step 2: Sample once from the multivariate normal probability distribution of Δxi . Sampling
should be consistent with the assumed standard deviations and correlations, which are usually
Step 3: Using the sample values of Δx i , determine the values of the risk factors at the end of the
Step 4: Revalue the portfolio using these new risk factor values.
Step 5: Subtract the revaluated portfolio from the current portfolio value to determine the loss.
Step 6: Repeat step 2 to step 5 multiple times to come up with a probability distribution for the
loss.
For instance, a total of 500 trials are conducted in a Monte Carlo simulation, then 99% VaR for
the period under consideration will be the fifth-worst loss, and thus the expected shortfall will be
Like other approaches, Monte Carlo simulation computes one-day VaR, and thus the following
equations apply when we want to compute T-day time horizon VaR and ES:
Monte Carlo simulation is slow because it is computationally intensive. This can be explained by
the fact that portfolios considered are usually huge, and evaluating each one of them in each
trial is quite time-consuming. To address this challenge, the delta-gamma approach can be used
(as discussed earlier) to determine the change in the portfolio value. This is called partial
simulation.
Delta-normal model assumes normal for the risk factors. However, Monte Carlo simulation uses
any distribution for the risk factors only if the correlation between the risk factors can be
defined.
42
© 2014-2020 AnalystPrep.
To execute the Monte Carlo simulation or delta-normal model, an approximation of the standard
deviations and correlations of either percentage or actual changes of the risk factors is
necessary. The approximation of these parameters is made using recent historical data. More
weights can be applied to more recent data using models such as GARCH (1,1), which we will
In the case of stressed VaR and ES, standard deviations and correlations should be estimated
from the past period, which would be considered stressful to the current portfolio.
Correlation Breakdown
During the stressed market conditions, standard deviations increase as well as correlations. This
phenomenon was witnessed during the 2007-2008 financial crisis, where default rates of
Therefore, correlations in a high volatility period are quite different from those of normal market
conditions. Thus, when calculating VaR or ES, risk managers might need to determine what will
Stress Testing
In finance, contagion is the spread of an economic shock in an economy or region, so that price
movements in one market are gradually but increasingly noticed in other markets. During a
contagion, both volatility and correlations increase, rendering diversification less effective as a
risk mitigation strategy. Stress testing is an attempt to model the contagion effect that would
Some of the historical events that have been used to stress test by various firms include the
Mexican crisis of 1994, the Gulf war of 1990, and the near-collapse of LTCM in 1998. The
underlying question among analysts while stress testing is quite simple: If a similar event
occurred right now, how would it affect the firm's current position?.
43
© 2014-2020 AnalystPrep.
The other approach to stress testing may not necessarily turn to historical events. It can also be
done using predetermined stress scenarios. For example, the firm could seek to find out the
effect on its current positions if interest rates short up by, say, 300 bps.
Strength:
Weakness:
Stress testing based on history is limited to a few past events; therefore, there is a cap
on the usefulness of such a strategy. In fact, a constant historical event problem is that
obtaining accurate data on the exact happenings can be a tough ask. Furthermore, as
is the case with all past events, there can be no guarantees that what happened back
Worst-case scenario analysis focuses on extreme losses at the tail end of the distribution. First,
firms assume that an unfavorable event is certain to occur. They then attempt to establish the
WCS analysis dissects the tail further to establish the range of worst-case losses that could be
incurred. For example, within the lowest 5% of returns, we can construct a "secondary"
WCS analysis complements the VaR, and here is how. Recall that the VaR specifies the minimum
loss for a given percentage, but it stops short of establishing the severity of losses in the tail.
WCS analysis goes a step further to describe the distribution of extreme losses more precisely.
44
© 2014-2020 AnalystPrep.
Practice Questions
Question 1
II. Stress testing can complement VaR estimation by helping managers establish
III. The larger the number of test scenarios, the better the understanding of the
IV. Stress testing allows users to include scenarios that did not occur in the
B. I, II, and IV
D. Only II and IV
In stress testing, the fewer the scenarios under consideration, the easier it is to
understand and interpret results. Too many scenarios make it difficult to interpret the
risk exposure.
Question 2
A risk manager wishes to calculate the VaR for a Nikkei futures contract using the
historical simulation approach. The current price of the contract is 955, and the
multiplier is 250. For the last 300 days, the following return data have been recorded:
-7.8%, -7.0%, -6.2%, -5.2%, -4.6%, -3.2%, -2.0%, …, 3.8%, 4.2%, 4.8%, 5.1%, 6.3%,
6.8%, 7.0%
45
© 2014-2020 AnalystPrep.
What is the VaR of the position at 99% using the historical simulation methodology?
A. $12,415
B. $16,713
C. $18,623
D. $14,803
The 99% return among 300 observations would be the third-worst observation among
Among the returns given above, the third-worst return is −6.2%. As such,
A is incorrect. This answer incorrectly uses the fourth-worst observation as the 99%
B is incorrect. This answer incorrectly uses the second-worst observation as the 99%
C is incorrect. This answer incorrectly uses the worst observation as the 99% return
Question 3
Bank X and Bank Y are two competing investment banks that are calculating the 1-
day 99% VaR for an at-the-money call on a non-dividend-paying stock with the
following information:
46
© 2014-2020 AnalystPrep.
Option delta: 0.7
To compute VaR, Bank X uses the linear approximation method, while Bank Y uses a
Monte Carlo simulation method for full revaluation. Which bank will estimate a
A. Bank X
B. Bank Y
The option’s return function is convex with respect to the value of the underlying;
therefore the linear approximation method will always underestimate the true value
of the option for any potential change in price. Therefore the VaR will always be
higher under the linear approximation method than a full revaluation conducted by
Monte Carlo simulation analysis. The difference is the bias resulting from the linear
approximation, and this bias increases in size with the change in the option price and
47
© 2014-2020 AnalystPrep.
48
© 2014-2020 AnalystPrep.
Reading 47: Measuring and Monitoring Volatility
Explain how asset return distributions tend to deviate from the normal distribution.
Explain reasons for fat tails in a return distribution and describe their implications.
Apply the exponentially weighted moving average (EWMA) approach and the GARCH
Explain and apply approaches to estimate long-horizon volatility/VaR and describe the
Constant volatility is easily approximated from historical data. However, volatility varies through
time. Therefore, an alternative to constant normality of asset returns is to assume that asset
returns are normally distributed conditioned on a known volatility. More specifically, high
volatility indicates that the daily asset return is normally distributed with high standard
deviation, and when the volatility is low, the daily returns are normally distributed with low
49
© 2014-2020 AnalystPrep.
standard deviation.
Monitoring volatility is made possible through two methodologies (as will be covered later in this
chapter): the exponentially weighted moving average (EMWA) model and he GARCH (1,1) model.
We will also apply the estimation of the volatility and application of volatility monitoring to
correlation.
There are several reasons as to why the normality framework has been adopted for most risk
estimation attempts. For starters, the normal distribution is relatively easy to implement. We
need two parameters – the mean, μ, and the standard deviation, σ, of returns. With these two, it
is possible to characterize the distribution fully and even come up with Value at Risk measures at
specified levels of confidence. In the real world, however, asset returns are not normally
distributed, and this can be observed empirically. Asset return distributions exhibit several
that most data points are concentrated around the center (mean), with very few points at
the tails. The distribution is symmetrical (has equal and opposite halves) with outliers
50
© 2014-2020 AnalystPrep.
2. Instability of Parameter Values: Unlike the normal distribution, asset return
distributions have unstable parameters (mean and standard deviation). This tendency
arises from market conditions that are continually changing. This instability is evident in
asset volatility where asset returns turn out to be way more volatile than predictions
3. The Asset Returns can be Non-symmetrical: As seen in the figure below, the equity
returns lean further to the right than the normal distribution. This phenomenon is known
as negative skewness. A direct consequence of negative skewness is that the left slope of
equity returns is longer than the left slope of the normal distribution, indicating a greater
51
© 2014-2020 AnalystPrep.
Fat Tails and their Implications
For this reading, it’s essential to keep in mind that the term ‘fat tails’ is relative: it refers to the
tails of one distribution relative to the normal distribution. If a distribution has fatter tails
relative to the normal distribution, it has a similar mean and variance, but probabilities at the
When modeling asset returns, analysts focus on extreme events – those that have a low
non-extreme events that would reasonably not be expected to cause severe losses.
If we were to assume that the normal distribution holds as far as asset returns are concerned, we
would not expect not even a single daily move of 4 standard deviations or more per year. The
normal distribution predicts that about 99.7% of outcomes lie within 3 standard deviations from
the mean.
52
© 2014-2020 AnalystPrep.
In reality, every financial market experience one or more daily price moves of 4 standard
deviations or more every year. There’s at least one market that experiences a daily price move of
10 or more standard deviations per year. What then does this suggest?
The overriding argument is that the true distribution of asset returns has fatter tails.
There is a higher probability of extreme events than the normal distribution would suggest. This
means that reliance on normality results in inaccurately low estimates of VaR. As a result, firms
tend to be unprepared for such tumultuous events and are heavily affected, sometimes leading to
foreclosure and/or large-scale ripple effects that reverberate and destabilize the entire financial
system.
Unconditional normal distribution manifests when the mean and standard deviation of asset
returns in a model is the same for any given day, regardless of market and economic conditions.
In other words, even if there is information about the distribution of asset returns that suggests
the existence of different parameters, we ignore that information and assume that the same
Conditionally normal distribution occurs when a model contains asset returns that are
normal every day, but the standard deviation of the returns varies over time. That is the standard
The underpinning idea here is that if we collect daily return data, we will observe the
unconditional distribution and that by monitoring volatility, we can estimate the conditional
distribution for daily return. For instance, consider that from the (historical) data we have
collected, we estimate volatility to be 0.5% per day, but from volatility monitoring, we estimate
the volatility to be 1.5% per day. Therefore, it will be more accurate to assume that asset returns
are normally distributed with a 1.5% standard deviation and thus use this 1.5% measure to
calculate VaR and the Expected Shortfall (ES). In other words, results from monitoring volatility
are better than using the results from the fat-tailed distribution or by assuming a normal
53
© 2014-2020 AnalystPrep.
distribution with volatility from the observed data.
Approaches for estimating the value at risk can be broadly categorized as either historical-based
or implied-volatility-based.
Historical-based Approaches
Historical-based approaches can be further subdivided into three categories: parametric, non-
returns are normally or lognormally distributed with volatility that varies with time.
historical simulation.
3. Hybrid approach: A hybrid approach borrows from both parametric and non-parametric
models. A result is a wholesome approach that still makes use of historical data.
Whereas historical approaches use historical data to gauge the volatility of an asset, the implied
volatility approach is based on current market data, i.e., it is forward-looking, which is its most
significant advantage. It is informed by the idea that historical events are not always very
That said, it is essential to note just like historical approaches; this method does not guarantee
certain results. Implied volatility is a product of probabilities and other factors such as current
investor sentiments, and therefore there can be no assurances that market prices will follow the
54
© 2014-2020 AnalystPrep.
predicted pattern. Another disadvantage of the model is that it is model-dependent.
They do not require any assumptions regarding the distribution of returns to estimate
VaR.
Multivariate density estimation allows for weights to vary to reflect data relevance of
the data in light of current market conditions, regardless of the timing of the data.
Problems posed by Fat tails, skewness, and other deviations from the assumed
economic variables.
Subdividing the full sample data into different market regimes reduces the amount of
Multivariate density estimation requires a large amount of data that has a direct
Multivariate density estimation may lead to data snooping or overfitting when working
In an attempt to model asset returns better, analysts may subdivide a specified time period into
55
© 2014-2020 AnalystPrep.
regimes, where each regime has a clearly noticeable set of parameters that are markedly
different from those of other regimes. For example, volatility could rise sharply during a market
crash only to stabilize when conditions improve. This is the idea behind the regime-switching
volatility model.
For illustration, let us use real interest rates of a developed nation between 1990 and 2015:
Given the graph above, it would be difficult to identify different states of the economy. Now,
suppose we make use of an econometric model to try and identify the different economic states.
From the econometric model above, we can identify three distinct states of the economy. These
are precisely what we would call regimes. As we switch from one regime to another, at least one
56
© 2014-2020 AnalystPrep.
parameter has to change. So, how exactly does the regime-switching model help to better
measure volatility?
The conditional distribution of returns is always normal with either low or high volatility but a
constant mean. The regime-switching model captures the conditional normality and, in so doing,
Measuring Volatility
Conventionally volatility is defined as a change of a variable value over a period of time. In the
context risk management, volatility is defined as the standard deviation of an asset return in one
day.
Si − Si −1
ri =
Si−1
Where
For a day n, the variance (σn2 ) from the previous m days is given by
1 m
σn2 = ∑ (rn−i − r̄)2
m − 1 i=1
1 m
2
σn = ⎷( ∑ (rn−i − r̄) )
m − 1 i=1
57
© 2014-2020 AnalystPrep.
Where:
1 m
r̄ = ∑ rn−i
m i=1
estimate of volatility, but when we use m, we get the maximum likelihood estimator,
Assuming that r̄ = 0: For a short period data, the standard deviation is far more
significant than the mean return, and we need expected return and not the historical
mean.
Therefore, defining the variables as before, the formula for the standard deviation changes to:
1 m
σn2 = ∑ r2n−i
m i =1
1 m
σn = ⎷( ∑ r2n−i )
m i =1
The square of the volatility is defined as the variance rate, which is typically the mean of
squared returns.
The asset returns over five days are 10%, -5%. 6%, -3%, and 12%. What is the volatility of the
asset returns?
58
© 2014-2020 AnalystPrep.
Solution
1 m
σn = ⎷( ∑ r2n−i )
m i =1
Now,
5
∑ r2n−i = 0.12 + (−0.05)2 + 0.062 + (−0.03)2 + 0.12 2 = 0.0314
i=1
1
⇒ σn = ⎷ (0.0314) = 0.07925 = 7.925%
5
An alternative method of calculating volatility is to use absolute returns rather than squared
returns when calculating volatility. This method is suitable when dealing with non-normal data
because it provides an appropriate prediction for fat-tailed distribution. However, the commonly
used method uses squared returns, which we have considered and continue to do.
For the effective use of the conditional normal model, it is essential to estimate current volatility.
1 m
σn = ⎷( ∑ r2n−i )
m i =1
For large values of ?, it would not capture current volatility due to variations of the volatility over
the period when data was collected. We can argue that the small data sample might be relatively
reliable, but the resulting estimate may not be accurate due to small data.
Alternatively, we can incorporate the effect of the standard error of the estimate. Recall that the
standard error of an estimate is the difference between the volatility estimate and the actual
59
© 2014-2020 AnalystPrep.
value. The standard error is estimated volatility divided by 2(m-1) where m is the number of
1 1 m
S.E.Eσn = ( ∑ r2 )
2 (m − 1)⎷ m i=1 n−i
1
S.E.Eσn = (0.07925) ≈ 0.01
2 (4)
We can then improve the accuracy of our volatility estimate by computing the confidence
intervals. Typically, confidence intervals are usually two standard errors from the estimated
value. In our case, the confidence interval is 5.925% to 9.925% (=0.07925± 2×0.01) per day. We
can reduce the confidence interval’s width by increasing the size of the observation, but we will
still face the disadvantage of the variability of the asset return volatility.
To solve this problem, we use a technique called exponential smoothing, also called an
exponentially weighted moving average (EWMA) used by RiskMetrics to estimate volatilities for
a wide range of market variables. Also, we use GARCH (1,1) as an exponential smoothing
technique.
1 1
Recall that from the formula σn2 = m ∑ m 2
i=1 r n−i , equal weight ( m ) is applied to the squared returns.
However, in EWMA, the weights given to the squared returns are not equal and must sum up to
1.
Weight in EWMA is defined as weight applied to squared return from the k days ago denoted by
λ multiplied by weight applied to squared return from k-1 days ago. λ is a constant positive and
For example, denote the weight from the recent squared return, that is, squared return from day
60
© 2014-2020 AnalystPrep.
n-1 by w0 . Therefore, the weight on day n-2 is λw0 , and for the squared return for day n-3 is λ2
The suitable value of λ is the one that leads to an estimate that gives the lowest error. Now
assume that we have K days of data. Then, the sum of the weight applied is given by:
w 0 + w 0 λ + w 0λ 2 + ⋯ + w 0 λk−1
2 k−1
=w 0 (1 + λ + λ + ⋯ + λ )
Practically, EWMA can be applied to data spanning one to two years, but data from a long time
ago carries less weight; thus, we can assume that data that spans to infinite past is the result of
w0 + w 0 λ + w0 λ 2 + w 0 λ3 + ⋯
∞
w 0 ∑ λk
i=0
∞ 1
∑ λk =
i =0 1 −λ
∞ 1
⇒ w 0 ∑ λk = w0 ( )
i=0 1 −λ
However, as stated earlier, the sum of weights in EWMA must be one. Therefore,
61
© 2014-2020 AnalystPrep.
1
w0 ( )=1
1−λ
∴ w0 = 1 − λ
By definition of the EWMA model, the estimated volatility on day n is computed by applying the
2
σn−1 = w0 r2n−2 + w 0 r2n−3 + w 0 r2n−4 + ⋯ (Eq2)
The last formula results from the EWMA model, which gives the estimate of the variance rate on
day n is the weighted average of the estimated variance rate for the previous day (n-1) and the
is termed as adaptive volatility because it incorporates prior information about the volatility into
the new information. One merit of the EWMA model is that less data is required once the EWMA
model has been built for a particular market variable. Only recent volatility estimates need to be
Assume that you estimate recent volatility to be 3%, with a corresponding return of 2%. Given
62
© 2014-2020 AnalystPrep.
that λ=0.84, what is the value of new volatility?
Solution
RiskMetrics determined the value of λ to be 0.94. This estimate, however, may not be
appropriate for today’s use, nor values higher than 0.94. Choosing a higher value of λ makes
EWMA less responsive to new data. For instance, let λ=0.99. From the equation
it implies that the new volatility is assigned a weight of 99%, and the new squared return is
assigned a weight of 1%. Clearly, for a number of days of high volatility will not change the value
of volatility significantly. Moreover, using a small value of λ, such 2% will make the volatility
There are two methods of determining the value of λ. One of them is to compute realized
volatility for a given day using 20 or 30 days of subsequent returns and then determine the value
of λ that minimizes the difference between the estimated volatility and the realized volatility.
The second method is the maximum likelihood method, where the λ that maximizes the
probability of observed data occurring is determined. For instance, if a trial leads to a value of λ
that results in predicting low volatility for a particular day, but the observed data is large, then
Historical Simulation
63
© 2014-2020 AnalystPrep.
Similar to EWMA, exponentially declining weights are applied when using historical simulation.
Exponentially declining weights are appropriate for determining VaR or expected shortfalls from
recent preceding observations because scenarios from the immediate data are more relevant
While weighting in historical simulation, the scenarios are arranged from the largest to the
smallest loss, after which the weights are accumulated to determine the VaR.
Recall that in the EWMA model, the weights were exponentially declining. An alternative
determine which periods in the past bear the same feature as the current period, after which
weights are assigned to the day’s historical data depending on the level of similarity of that day
For instance, the volatility of an interest rate varies depending on the level of the interest rates:
decreases when the interest rates increases and vice versa. We can calculate the volatility of an
interest rate by assigning the weights to interest data that is similar to the present interest rates
and decreasing the weight as the difference between past and today’s interest rate decreases.
Determining the level of similarity between one period and another is done using conditioning
variables. For example, while determining the interest rate weights, we could use GDP growth
rates as conditioning variables for the interest rate volatility. For conditional variables
X1 , X2 , … , Xn the similarity between today and the previous period is determined by calculating
the measure:
n 2
^i − X* )
∑ ai ( X i
i =1
Where
64
© 2014-2020 AnalystPrep.
ai : a constant variable reflecting the importance of the ith variable.
The measure given by the formula becomes smaller as the similarity between current and
the EWMA model, where we apply a weight to the recent variance rate estimate and the latest
squared return. According to the GARCH(1,1) model, the updated model for the variance rate is
given by:
Where:
α+β≤1
And,
γ = 1− α −β
It is easy to see that EWMA is a particular case of GARCH (1,1) where γ=0, α =1-γ , and β = γ .
Both GARCH (1,1) and EWMA are called first-order autoregressive (AR(1)) models since the
65
© 2014-2020 AnalystPrep.
forecast for the variance rate depends on the immediately preceding variable.
Similar to the EWMA model, the weights in GARCH (1,1) decline exponentially. Now, by
And,
2
σn-1 = αr2n−2 + βr2n−2 + γ VL
2
σn-2 = αr2n−3 + βr2n−3 + γ VL
2
σn-3 = αr2n−4 + βr2n−4 + γ VL
2
Starting from the equation defining GARCH (1,1), substitute σn-1 2 2
, σn-2 , σn-3, … , we define:
w = γ VL
ω
⇒ VL =
γ
So,
ω
Now since α + β + γ = 1 and VL = , then
γ
ω
VL =
1 − α− β
Consider a GARCH (1,1) model where ω=0.00005, α=0.15, and β=0.75. Given that the current
volatility is 3% and the new return is -5%, (a) what is the long-run variance rate?
Solution
We know that:
66
© 2014-2020 AnalystPrep.
α+ β + γ = 1
⇒ γ = 1− α − β
= 1 − 0.15 − 0.75 = 0.10
ω 0.00005
VL = = = 0.0005
1 − α− β 0.10
(b) Assuming the GARCH (1,1) model, what is the new volatility?
Solution
This corresponds to the volatility of 2.24% (=√ 0.0005). Now according to GARCH (1,1) model,
Mean reversion is simply the assumption that a stock's price (or any other variable) will tend to
67
© 2014-2020 AnalystPrep.
The simplest form of mean reversion can be illustrated by a first-order autoregressive process
[AR(1)], where the current value is based on the immediately preceding value.
Xt+1 = a + bX t + et+1
E [Xt+1 ] = a + bXt
a
E [Xt+1 ] = (1 − b) × + bXt
1− b
a
The long-run mean of this model is evaluated as [ (1−b) ]. The parameter of utmost interest in this
long-run mean equation is b, often termed “the speed of reversion” parameter. There are two
scenarios:
If b = 1, the long-run mean is infinite (i.e., the process is a random walk, nonstationary
process with an undefined long-run mean. That implies that the next period’s expected
68
© 2014-2020 AnalystPrep.
If b is less than 1, then the process is mean-reverting (i.e., the time series will trend
toward its long-run mean). That implies that when Xt is above the LRM, it will be
value.
Xt+1 = a + bX t + et+1
In GARCH (1,1), the long-run variance rate (VL ) provides a “pull” mechanism towards the long-
run average mean. Not that this is absent in the EWMA model due to a lack of the VL factor.
69
© 2014-2020 AnalystPrep.
Consider the figure above. When the volatility is above the long-run average mean, VL “pulls”
down toward it. On the other hand, when the volatility is below the long-run mean, VL pulls it
towards the long-run mean. This tendency of the “pull” mechanism is termed as the mean
reversion.
Therefore, GARCH (1,1) exhibits mean reversion, while EWMA does not. Mean reversion applies
to market variables. More importantly, market variables that are traded should not exhibit
Notably, volatility cannot be traded, and thus it exhibits mean reversion. In this case, mean
reversion implies that when the volatility is high, we do not expect to remain in that position
Our discussion has been based on approximating one-day volatility. We might want to know what
will happen over a longer period, such as one year. To achieve this, we assume that the variance
rate over T days is equivalent to T days times variance over one day. That is,
70
© 2014-2020 AnalystPrep.
T − days variance rate = one-day variance rate × T
This means that volatility over days is equivalent to the volatility over one day multiplied by the
Mean reversion provides an improvement to the square root rule of calculating long-horizon
volatility. Note that according to mean reversion, if the current daily volatility is high, we expect
it to decline. Therefore, using square root rule, in this case, overstates VaR, and if the volatility is
low, we expect it to rise a thus, the square root rule understates VaR. This applies to an expected
shortfall (ES).
According to GARCH (1,1), the expected variance rate on day t can be shown to be:
σn+t
2 = VL + (α + β)t (σn2 − VL )
We can use this formula to compute the average variance rate over the next T-days and
Assume that the daily variance rate is 0.000025. What is the 30-day volatility?
Implied Volatility
Implied volatility is an alternative measure of volatility that is constructed using the option
71
© 2014-2020 AnalystPrep.
valuation. The options (both put and call) have payouts that are non-linear functions of the price
of the underlying asset. For instance, the payout from the put option is given by:
max(K − PT )
Where:
Therefore, the price payout from an option is sensitive to the variance in the asset’s return.
The Black-Scholes-Merton model is commonly used for option pricing valuation. The model
relates the price of an option to the risk-free rate of interest, the current price of the underlying
asset, the strike price, time to maturity, and the variance of return. For instance, the price of the
C t = f (rf , T , Pt , σ 2 )
Where:
T=Time to maturity
The implied volatility σ relates the price of an option with the other three parameters. The
implied volatility is an annualized value but can be converted by dividing by 252, which is an
estimated number of trading days in a year. For instance, if annual volatility is 30%, then the
30%
daily implied volatility is 1.89% (= ).
√ 252
72
© 2014-2020 AnalystPrep.
The difference between the implied volatility and historical volatility (such as the one estimated
by GARCH(1,1) and EWMA models) is that implied volatility is forward-looking while historical
Options are not actively traded, and thus finding reliable volatility is an uphill task. However, risk
The volatility index (VIX) measures the volatility on the S&P 500 over the coming 30 calendar
days. VIX is constructed from a variety of options with different strike prices. VIX applies to a
large variety of assets such as gold, but it is only applicable to highly liquid derivative markets
There are two main advantages of implied volatility over historical volatility:
It is model dependent;
Options on the same underlying asset may trade at different implied volatilities. For
example, deep out of the money and deep in the money options trade at higher
It has limited availability because it can only be deduced when there are current
It assumes volatility will remain constant over a period of time, but volatility will
Monitoring Correlation
73
© 2014-2020 AnalystPrep.
Recall that in a delta-normal model, correlations between the daily asset returns are needed to
compute the VaR or expected shortfall for a linear portfolio. Therefore, it is crucial to monitor
correlations.
Updating the correlations is analogous to that of volatilities. Recall that while updating the
volatilities, we use the variances. In the case of relationships, we use covariances. Now,
assuming that mean of daily returns is zero, then the covariance is defined as the expectation of
According to the EWMA model, updating the correlation between returns X and Y is given by:
Where:
Cov (X, Y)
Corr (X, Y) =
σX σY
Where:
σY : standard deviation of Y.
Therefore, if the EWMA model has been used to estimate the standard deviations of the returns,
we can comfortably estimate the correlation coefficient. However, the same value of λ is used
74
© 2014-2020 AnalystPrep.
Example: Updating the correlation coefficient
Assume that recent volatilities for returns on X and Y are 2% and 5%, respectively. Their
corresponding correlation coefficient is 0.4. Moreover, their recent returns are 3% and 4%,
respectively.
Solution
i. Updated volatilities
Cov (X, Y)
Corr (X, Y) =
σX σY
75
© 2014-2020 AnalystPrep.
covn−1 = σXn−1 × σY n−1 × Corr (Xn−1, Y n−1 )
= 0.02 × 0.05 × 0.4
= 0.0004
Now using:
we have:
cov n
corr (x n ,y n ) =
σXn σY n
0.000464
= = 0.4489
0.02098 × 0.04927
76
© 2014-2020 AnalystPrep.
Practice Question
The current estimate of daily volatility is 2.3 percent. The closing price of an asset
yesterday was CAD 46. Today, the asset ended the day at CAD 47.20. Using log-
returns and the exponentially weighted moving average model with λ = 0.94,
A. 2.319%
B. 0.0537%
C. 2.317%
D. 2.315%
77
© 2014-2020 AnalystPrep.
Reading 48: External and Internal Credit Ratings
Describe external rating scales, the rating process, and the link between ratings and
default.
Describe the impact of time horizon, economic cycle, industry, and geography on
external ratings.
Define and use the hazard rate to calculate unconditional default probability of a credit
asset.
Define the recovery rate and calculate the expected loss from a loan.
approaches.
Explain the potential impact of rating changes on bond and stock prices.
Explain historical failures and potential challenges to the use of credit ratings in
An external rating scale is a scale used as an ordinal measure of risk. The highest grade on the
scale represents the least risky investments, but as we move down the scale, the amount of risk
78
© 2014-2020 AnalystPrep.
zero-coupon bond issued by a corporate entity. An issuer-specific credit rating, on the other
hand, conveys information about the entity behind an issue. The latter usually incorporates a lot
Here are S&P’s and Moody’s credit rating scores for long-term obligations:
79
© 2014-2020 AnalystPrep.
successive move down the scale represents an increase in risk. In the case of Moody’s
80
© 2014-2020 AnalystPrep.
ratings, Baa and above are said to be investment-grade while those below this level are said
to be non-investment-grade.
In the case of S&P’s, ratings BBB and above are investment-grade. All the others are non-
investment-grade.
The process leading up to the issuance of a credit rating follows certain steps. These are:
IV. A meeting of the rating agency committee assigned to rating the firm
VII. The rated firm has a window to appeal the assigned rating or offer new information
Apart from the ratings themselves, the rating agencies also provide outlooks which shows the
A developing outlook is an evolving one in which we can’t tell the direction of the
change.
When a rating is placed on a watchlist, it shows that a very small short-term change is expected.
81
© 2014-2020 AnalystPrep.
Rating stability
Rating stability is necessary since ratings are majorly used by bond traders. If the ratings were
to change, then the bond traders are required to trade more frequently and, in this case, they are
Rating stability is important because ratings are also used in financial contracts, and if the
ratings vary for different bonds, it would be difficult to administer the underlying contracts.
Time Horizon
The probability of default given any rating at the beginning of a cycle increases with the time
horizon. Non-investment bonds are the worst hit. Their default probabilities can dramatically
Economic Cycle
Since ratings are generally produced with an eye on a long-term period, they must take into
account any economic/industrial cycle on the horizon. Rating agencies make efforts to
incorporate the effects associated with an economic cycle in their ratings. Although this practice
economic cycle doesn’t play out exactly as expected. Put precisely, the probability of default can
occurs. In addition, the default rate of lower-grade bonds is correlated with the economic cycle,
Industry
Two firms in different industries – say, banking and manufacturing – could have the same rating,
but the probability of default may be higher for one of the firms than for the other. What does
that mean? The implication here is that for a given rating category, default rates can vary from
82
© 2014-2020 AnalystPrep.
industry to industry. However, there’s little evidence to support the notion that geographic
Hazard rates
The task is to answer the question, “What is the conditional probability of a firm defaulting
between time t and time t+δ t given that there is no default before time t?” We can denote this by
Suppose that h is the average hazard rate between time 0 and time t.
1 − exp (−ht)
exp (−ht)
and the unconditional probability between time t1 and t2 is given by the expression;
Calculate:
Solution
83
© 2014-2020 AnalystPrep.
a. The probability of default at the end of the 2nd year is given by:
1 − exp (−ht)
= 1 − exp (−0.05 × 2) = 0.09516
Recovery Rates
In the event that a firm runs bankruptcy or defaults, it may pay part of the amount of the total
loan to the lender. This amount that is repaid, expressed as a percentage, is known as the
recovery rate.
Since the loan is not fully repaid, then we can calculate the expected loss from the loan over a
Suppose the debt instrument has a notional value of $100 million, then the expected loss when
The interest rate on a given risky bond and the yield on an equivalent risk-free security will
84
© 2014-2020 AnalystPrep.
always. This extra interest charged over the risk-free rate is known as a credit spread. This extra
From the following image, we can see the spread between the high yield index (risky bonds) and
Since the default rates keep fluctuating, giving rise to a non-diversifiable risk, bond investors, in
addition to the credit spread, will require a risk premium. Difficulties are also likely to be
experienced in selling off the bonds. Since risky bonds sometimes have low liquidity, the
Bonds
There’s overwhelming evidence that a rating downgrade triggers a decrease in bond prices. In
85
© 2014-2020 AnalystPrep.
fact, bond prices sometimes decrease just because there’s a strong possibility of a downgrade.
A rating upgrade triggers an increase in bond prices, although there’s relatively less market
Therefore, the underperformance of bonds whose credit quality has been downgraded is more
Stocks
There’s moderate evidence to support the view that a rating downgrade will lead to a stock price
decrease. A ratings upgrade, on the other hand, is somewhat likely to trigger an increase in bond
prices.
In practice, the relationship between changes in rating and stock prices can be quite complex
and will usually be heavily impacted by the reason behind the changes. Furthermore,
downgrades tend to have more impact on the stock price compared to upgrades.
External ratings are produced by independent rating agencies and aim at revealing the financial
stability of both lenders and borrowers. For example, Moody’s periodically releases ratings for
big banks around the globe. Such ratings are important because banks usually rely on customer
deposits and money raised through the issuance of various assets such as bonds to sustain
lending. The funds raised this way to create a pool of money that is then loaned to borrowers in
smaller chunks. Thus, depositors and bond owners use such ratings to assess the riskiness of
Sometimes, however, banks also need their own ratings so as to undertake an independent
In modern times, internal credit ratings are usually developed based on the techniques used to
86
© 2014-2020 AnalystPrep.
develop external credit ratings. The same indicators are used, albeit with a few adjustments
One way of carrying out an internal rating is by use of a statistical technique known as the
Altman’s Z-score. The following ratios need to be provided when using this technique:
A Z-score above 3 means that the firm is not likely to default and when the Z-score is below 3,
Nowadays, machine learning algorithms use more than five input variables as compared to
Altman’s Z-score. Also, the functions used in machine learning algorithms can be non-linear.
Some of the factors that have contributed to the increased sophistication of modern internal
I. The ever-growing use of external credit rating agency language in financial markets
Alternative to Rating
Apart from the commonly known rating agencies, that is, Moody’s, S&P, and Fitch, we have some
organizations such as KMV and Kamakura which use some models to come up with default
probabilities and hence can then use probabilities to provide important information to clients.
87
© 2014-2020 AnalystPrep.
The amount of debt the firm has in its capital structure.
In the underlying model, a company defaults if the value of its debt exceeds the value of its
assets.
Suppose v is the value of the asset and d is the value of the debt, the firm defaults when v < d.
This implies that equity in a company is a call option on the assets of the firm with a strike price
equal to the face value of the debt. The firm defaults if the option is not exercised.
At-the-point internal ratings, also called point-in-time ratings, evaluate the current situation
of a customer by taking into account both cyclical and permanent effects. As such, they are
At-the-point ratings try to assess the customer’s quantitative financial data (e.g. balance sheet
information), qualitative factors (e.g. quality of management), and information about the state of
the economic cycle. Using statistical procedures such as scoring models, all that information is
At-the-point internal ratings are only valid for the short-term or medium term, and that’s
largely because they take into account cyclic information. They are usually valid for a period not
88
© 2014-2020 AnalystPrep.
Through-the-cycle Internal Ratings
Through-the-cycle (ttc) internal ratings try to evaluate the permanent component of default risk.
Unlike at-the-point ratings, they are said to be nearly independent of cyclical changes in the
creditworthiness of the borrower. They are not affected by credit cycles, i.e. they are through-
the-cycle. As a result, they are less volatile than at-the-point ratings and are valid for a much
I. They are much more stable over time compared to at-the-point ratings
II. Because of their low volatility, ttc ratings help financial institutions to better manage
customers. Too many rating changes necessitate changes in the way a bank handles a
One of the disadvantages of ttc ratings over at-the-point ratings is that they can at times be too
conservative if the stress scenarios used to develop the rating are frequently materially different
from the firm’s current condition. If the firm’s current condition is worse than the stress
scenarios simulated, then the ratings may be too optimistic. In fact, ttc ratings have very low
A rating transition matrix gives the probability of a firm ending up in a certain rating category at
some point in the future, given a specific starting point. The matrix, which is basically a table,
uses historical data to show exactly how bonds that begin, say, a 5-year period with an Aa rating,
change their rating status from one year to the next. Most matrices show one-year transition
probabilities.
Transition matrices demonstrate that the higher the credit rating, the lower the probability of
default.
The table below presents an example of a rating transition matrix according to S&P’s rating
categories:
89
© 2014-2020 AnalystPrep.
One−year transition matrix
Exam tips:
Each column corresponds to a rating at the end of 1 year. For example, a bond initially
rated BB has an 8.84% chance of moving to a B rating by the end of the year.
The sum of the probabilities of all possible destinations, given an initial rating, is equal
to 1 (100%)
You will need to recall the rules of probability from mathematics to come up with n-year
Credit ratings are their most stable over a one-year horizon. Stability decreases with
longer horizons.
Building
To build an internal rating system, banks try to replicate the methodology used by rating agency
analysts. Such a methodology consists of identifying the most meaningful financial ratios and
risk factors. After that, these ratios and factors are assigned weights such that the final rating
90
© 2014-2020 AnalystPrep.
estimate is close to what a rating agency analyst would come up with. Weights attached to
financial ratios or risk factors are either defined qualitatively following consultations with an
Calibrating
II. To determine the value of inputs used in the modeling of capital required as per the
For these reasons, internal ratings have to be calibrated. This involves establishing a link
between the internal rating scale and tables displaying the cumulative probabilities of default.
The timeline of such tables must capture all maturities, from, say, 1 year to 30 years. Sometimes,
it may be necessary to build different transition matrices that are specific to the asset classes
Backtesting
Before linking default probabilities to internal ratings, backtesting of the current internal rating
system is vital. The question is: Just how many years are needed to pull this off?
A historical sample of between 11 and 18 years is considered sufficient to test the validity of
ratings.
91
© 2014-2020 AnalystPrep.
Bias Description
Time horizon bias Using a combination of at-the-point and through-the-cycle
approaches to score a company.
Information bias Assigning rating based on insufficient information
Homogeneity bias Inability to maintain consistent rating methods.
Principal-agent bias Rating developers fail to act in the best interest of the
management.
Backtesting bias Incorrectly linking rating systems to default rates.
Distribution bias Modeling the probability of a default using an inappropriate
distribution.
Scale bias Producing ratings that are not stable with the passage of
time
92
© 2014-2020 AnalystPrep.
Questions
Question 1
Determine the probability that a B –rated firm will default over a two-year period.
A. 5%
B. 4.25%
C. 1%
D. 10.25%
Required probability = Sum of probabilities of all possible paths that could lead to a
In other words, in how many ways can a B-rated firm default over a two-year period?
Path Probability
B→ default 0.05
B→ B → default 0.85 x 0.05= 0.0425
B→ CCC → default 0.05 x 0.20= 0.01
Total 0.1025
Question 2
ABC Co., currently rated BBB, has an outstanding bond trading in the market.
93
© 2014-2020 AnalystPrep.
Suppose the company is upgraded to A. What will be the most likely effect on the
bond’s price?
A. Positive and stronger than the negative effect triggered by a bond downgrade
downgrade
C. Positive and weaker than the negative effect triggered by a bond downgrade
Rating downgrades tend to have more impact on the stock price compared to
upgrades. This can be explained by the fact that firms tend to release good news a lot
more often than bad news, and thus the expectations among investors are generally
94
© 2014-2020 AnalystPrep.
Reading 49: Country Risk
Explain how a country’s position in the economic growth life cycle, political risk, legal
Evaluate composite measures of risk that incorporate all types of country risk and
Compare instances of sovereign default in both foreign currency debt and local
Describe factors that influence the level of sovereign default risk; explain and assess
Describe characteristics of sovereign credit spreads and sovereign credit default swap
Country risk is the risk associated with investing in a given country. In most cases, an investor is
exposed to more risk by investing in some countries than others. When Apple, for example,
pushes for a bigger market presence in Latin America, they are exposed to the political and
95
© 2014-2020 AnalystPrep.
Disproportionate dependence on certain products or services.
The value of all goods and services produced within a country’s borders is known as GDP. When
we look at the economic growth rate with regard to an increase in the price of goods and
services, then this is called real GDP. The GDP, therefore, gives us a clear snapshot of the
Countries in early growth are more exposed to risk than larger, more mature countries. For
example, a global recession hits small, emerging markets harder than it does mature markets.
For instance, while countries like Japan and the U.K. experienced a 1-2% dip in GDP following a
recession, early growth economies like Kenya and Panama can record a dip as high as 5%.
Emerging markets are also hit hard in case of an economic shock. Even in the face of a robust
legal framework and good governance, there’s an upper cap on the powers that countries have
over their risk exposure. Some risks may simply be unavoidable. This is why it’s important to
thoroughly analyze a country’s risk profile before making critical investment decisions.
Factors of production;
The availability of other related industries that produce high-quality products; and
The political environment in a country can have a major bearing on its risk exposure. This can
I. Continuous vs. Discontinuous risk: Countries deeply rooted in democracy and free
96
© 2014-2020 AnalystPrep.
speech have continuous risk in the sense that rules and regulations governing are
continuously being challenged and amended. As such, regulations enacted for the long-
term may remain in place only for a while before new laws and bills are introduced and
signed into law. This means that an investor may have to contend with new capital
requirements, for example, at a time when they are faced with a crippling cash crunch.
risk. That means it may be difficult to amend rules and regulations however good or bad
they might be from the perspective of the investor. In practice, the strength and
II. Corruption and Side costs: The rules and regulations governing business and
operations in a country are only as good as the systems put in place to enforce them.
High levels of corruption make it easy to circumvent regulations or ignore them outright.
have to part with huge sums of money to get the job done.
III. Physical violence: Internal conflicts or civil war expose investors to both physical harm
and heavy operational costs, including high insurance costs and depreciation of physical
assets.
IV. Expropriation risk: Expropriation risk is the risk that a government may seize
ownership of a firm’s assets or impose certain rights that collectively reduce the firm’s
value. This could also happen when firms are subjected to specific taxes either by virtue
of their presence in a country or the nature of their core business. Compensation for
expropriation may be well below the value of rights or assets relinquished. Mining firms
Legal Risk
Legal risk has much to do with the enforcement of property/contractual rights and fidelity to the
rule of law. Laxity toward enforcement of rules and regulations not only disadvantages current
investors; it also serves to discourage potential investors from coming in. At best, those who
97
© 2014-2020 AnalystPrep.
decide to invest in the face of a failure of the legal system have to incorporate similar outcomes
Legal risk is also a function of how efficiently the system operates. For example, if enforcing a
contractual right takes years in a given country, investors will most likely shun that country.
Businesses and individual investors would not tolerate legal limbo for too long.
Some countries are known to have (and enforce) very robust property rights, especially in North
America. Others have very weak enforcement, especially African and South American countries.
A country that depends too much on one product or service exposes investors to additional risk.
A decline in the price or demand of the product or service can create severe economic shocks
that may reverberate well beyond the companies immediately affected. For example, a country
whose oil proceeds amount to, say, 50% of the GDP, exposes all investors within the country to
economic pain if the price of oil tumbles. In most cases, prices of other commodities shoot up.
A good measure of country risk should incorporate all the dimensions of such risk, be they
political, financial, or economic. Collectively these risks make up what we call total risk. There
are several professional organizations around the world that offer country risk measurement
extensive risk analysis of over 100 countries based on three core dimensions: political,
financial, and economic. These three dimensions are made up of 22 distinct variables.
PRS gives both dimensional and composite scores. The maximum score is 100, while the
minimum score is zero. For example, in its report dated July 2015, Switzerland was rated
the least risky country with a score of 88.5. Syria took the bottom spot with a score of
35.3
98
© 2014-2020 AnalystPrep.
economists to come up with a composite risk score of a country or region. The score
III. The Economist- The popular media house develops in-house country risk scores built
IV. The World Bank- The World Bank develops country risk scores based on six key
quality, rule of law, and accountability. The WB’s scores are scaled around zero, with
negative numbers indicating more risk and positive numbers less risk.
Risk services greatly help to understand and keep track of country risk, but they are not devoid
I. Some of the models used to churn out the final risk score incorporate risks that have
very little impact on business. The final score could be appropriate for policy-making or
II. There’s no standardization of scores, and each service uses its own protocol and
calibration techniques.
III. The scores can be quite misleading when used to measure relative risk. For example, if a
country has a PRS score of 80, that doesn’t imply it’s twice as safe as another country
Sovereign Default
A sovereign default is the failure or refusal of the government of a sovereign state to pay back its
debt in full. Sovereign debt can be denominated in either local or foreign currency.
Foreign currency defaults are defaults that occur on the sovereign currency that’s denominated
in foreign currency. In most cases, governments find themselves unable to raise the amount of
foreign currency required to meet contractual obligations. And the worst thing about foreign
99
© 2014-2020 AnalystPrep.
currency debt is that governments cannot print foreign currency to make up for the shortfall.
In the last few decades, a majority of sovereign defaults have been based on foreign currency. In
addition, countries have been more likely to default on bank debt owed than on sovereign bonds
issued. And in dollar value terms, Latin American countries have accounted for much of
Some countries have previously defaulted on debt denominated in local currency. Examples
include Argentina (2002-2004) and Russia (1998-1999). Local currency defaults could be traced
I. Mandatory gold backups: In the years prior to 1971, the printed currency had to be
backed up with gold reserves. Without enough reserves, countries could not print enough
II. Presence of shared currencies: When a country shares a currency with other
countries, it lacks the freedom to print cash to prevent a sovereign default. This scenario
played out in 2015 whereby the Greek government could not print more Euros even in
the face of a crippling economic meltdown punctuated by a huge sovereign debt. The
country defaulted on a USD1.7 billion IMF payment – becoming the first country to do so
III. A reluctance to print cash purely for debt repayment: Printing cash to meet debt
obligations is fraught with dangers, including reputation risk, political instability, and the
very real possibility of an economic recession. The local currency can dramatically lose
value, forcing investors to shun financial investments in favour of real assets such as real
diplomatic and economic ties between states. The defaulting government suffers a loss of
100
© 2014-2020 AnalystPrep.
II. Political instability: Following a sovereign default event, the populace may lose its
confidence in their leadership, leading to wave after wave of demos, unrest, as well as
III. Real output declines:? As investors increasingly shun financial assets in favour of real
assets, domestic consumption may also decrease. This, in turn, may lead to a drop-in
production
IV. Capital markets are thrust into a state of chaos and turmoil: Following a default
investments. The demand for long-term securities such as bonds declines, making it
difficult for firms to raise funds for expansion and other operations.
I. Degree of indebtedness: The larger the debt a sovereign state has, the more likely it is
II. The size of revenue: A high amount of revenue reduces the chances of a sovereign
default event occurring. If a government has regular cash inflows in the form of taxes
and other revenue streams such as income earned from the sale of state-owned natural
III. Stability of revenue: Countries with more stable revenue streams have less default
activities become more diversified. Countries that depend too much on a specific source
IV. Political risk: If the leadership of a country is somewhat immune from public pressure –
something common in autocracies like Iran and other Middle East nations, default events
can easily occur. In such situations, the leadership enjoys the stability of tenure, giving
them a free hand to dictate the nation’s financial priorities. Democracies, on the other
hand, are less likely to default because their leadership is constantly under pressure to
deliver.
Countries that form part of a regional economic block usually offer each other some
101
© 2014-2020 AnalystPrep.
implicit backing on matters debt. For example, an EU member state is unlikely to default
on a sovereign debt because other EU countries are likely to chip in. In fact, when
countries like Spain and Portugal joined the EU, renowned credit rating agencies
reduced their assessment of default risk in these countries. However, such support is not
guaranteed. That’s why investors don’t read too much from this kind of backing.
Rating agencies offer opinions on a firm or country’s ability to incur and/or pay back the debt.
Political and social risks – including issues like public participation in political decision
making, respect for the rule of law, transparency in government, and long-term stability
of political institutions.
Regime legitimacy
I. Ratings are upward biased: In the aftermath of the 2007/2008 financial crisis, rating
agencies were widely criticized for having awarded unduly high credit ratings to banks
and other financial institutions. In general, rating agencies have been accused of being
far too optimistic in their assessment of sovereign ratings. An upward bias on corporate
ratings could be explained by the fact that the same corporates double up as the credit
agencies’ remunerators. This argument, however, does not hold when it comes to
sovereign ratings because individual governments are not required to pay the rating
agencies.
II. They are at times reactive rather than proactive: Rather than updating their ratings
102
© 2014-2020 AnalystPrep.
before an actual credit event occurs, rating agencies sometimes downgrade countries
after a problem has become evident. This does little to protect investors.
III. Too much interdependence: Although rating agencies claim to work independently
albeit using similar risk indicators, they have been accused of exhibiting herd behaviour,
IV. Vicious cycle: Sometimes rating agencies have been accused of worsening a crisis by
unduly downgrading ratings – painting a situation as being worse than it actually is.
Sovereign default spread describes the difference between the rate of interest on a sovereign
bond denominated in a foreign currency and the rate of interest on a riskless investment in that
currency. For example, suppose that country A has a 10-year dollar-denominated bond with a
market interest rate of 8%. At the same time, the 10-year U.S. Treasury bond is trading at 2%.
This implies that the sovereign default spread for country A is 6%. The sovereign default spread
Here is a table representing the sovereign default spread in some specific countries as compared
Compared to rating agencies, the market differentiation for risk is more granular. In
other words, the market has an even more refined understanding of country risk
compared to credit ratings. For example, countries A and could have the same Moody’s
rating, but the sovereign spread for A may be greater than the spread for B. That could
103
© 2014-2020 AnalystPrep.
mean the market sees more default risk in A than in B.
Market-based spreads reflect changes in real-time, unlike credit ratings which can be
reactive rather than proactive. As such, market-based spreads can be more effective at
Market-based spreads tend to be far more volatile than credit ratings and can be affected by
variables with little or no correlation to default. For example, liquidity and market forces of
supply and demand can trigger shifts in spreads that have nothing to do with default.
Conclusion
Both credit ratings and market spreads are useful measures of default. Sovereign bond markets
usually price bonds guided by credit ratings. In addition, credit rating agencies also leverage
104
© 2014-2020 AnalystPrep.
Questions
Question 1
In the context of sovereign default spread, which of the following statements is least
accurate?
spreads
Sovereign default spread describes the difference between the rate of interest on a
riskless investment in that currency. For example, suppose that country A has a 20-
year dollar-denominated bond with a market interest rate of 8%. At the same time,
the 20-year U.S. Treasury bond is trading at 2%. This implies that the sovereign
Question 2
B. Young, growth companies are more exposed to risk partly because they have
C. In markets, a shock to global markets will travel across the world, but mature
105
© 2014-2020 AnalystPrep.
market equities will often show much greater reactions, both positive and
D. A country that is still in the early stages of economic growth will generally
Young, growth companies are more exposed to risk partly because they have limited
resources to overcome setbacks and also because they rely too much on a stable
Choice A is inaccurate. Mature growth countries are not insured from economic
shocks. They are still exposed to risk, but on a lower scale compared to young,
emerging markets.
Choice C is inaccurate. In markets, a shock to global markets will travel across the
world, but early growth market equities will often show much greater reactions,
Choice D is also inaccurate. A country that is still in the early stages of economic
growth will generally have more risk exposure than a mature country, even if it has
106
© 2014-2020 AnalystPrep.
Reading 50: Measuring Credit Risk
Explain the distinctions between economic capital and regulatory capital, and describe
how economic capital is derived. Identify and describe important factors used to
calculate economic capital for credit risk: the probability of default, exposure, and loss
rate.
Estimate the mean and standard deviation of credit losses assuming a binomial
distribution.
Describe and apply the Vasicek model to estimate default rate and credit risk capital
for a bank.
Describe the CreditMetrics model and explain how it is applied in estimating economic
capital.
Describe and use the Euler’s theorem to determine the contribution of a loan to the
Explain why it is more difficult to calculate credit risk capital for derivatives than for
loans.
Economic capital of a bank is the approximate amount of capital that the bank requires to
107
© 2014-2020 AnalystPrep.
absorb the losses from the loan portfolios. In other words, it is the cushion that a bank estimates
Regulatory capital is the amount of capital the regulators require the banks to maintain. For
instance, the Global bank requirements are determined by the Basel Committee on Banking
Supervision (BCBS) based in Switzerland. Supervisors then implement the BCBS requirements in
each country.
Banks also have what is called debt capital, funded by bondholders. In case the incurred losses
deplete the equity capital, the debt holders should incur losses before the depositors.
The equity capital is described as “going concern capital” because the bank is solvent if its
capital is positive. On the other hand, the debt capital is described as the “gone concern capital”
because it acts as a cushion to the depositors when the bank becomes insolvent (no longer a
going concern).
Banks face many risks through their transactions, which need to be quantified. Credit risk is a
primary risk that the banks have concentrated since the inception of the banking industry. It is
majorly explained by the fact that the banks’ activities mainly involve taking deposits and making
loans. The loans made are predisposed to some level of default of risk and hence some level of
Credit risk is quantified using different models. This chapter discusses three models:
i. The mean and standard deviation of the loss from a loan portfolio determined from the
ii. The Vasicek model – used by the bank regulators to approximate the extreme percentile
The term credit risk describes the risk that arises from nonpayment or rescheduling of any
108
© 2014-2020 AnalystPrep.
promised payment. It can also arise from credit migration – events related to changes in the
credit quality of the borrower. These events have the potential to cause economic loss to the
bank.
The expected loss is the amount a bank can expect to lose, on average, over a predetermined
period when extending credits to its customers. Unexpected loss is the volatility of credit losses
Once a bank determines its expected loss, it sets aside credit reserves in preparation. However,
for unexpected loss, the bank must estimate the excess capital reserves needed subject to a
predetermined confidence level. This excess capital needed to match the bank’s estimate of
To safeguard its long-term financial health, a lender must match its capital reserves with the
amount of credit risk borne. Economic capital is primarily determined by (i) level of confidence,
and (ii) level of risk. An increase in any of these two parameters causes the economic capital also
to increase.
Probability of Default
The probability of default (PD), describes the probability that a borrower will default on
contractual payments before the end of a predetermined period. This probability is in and of
itself not the most significant concern to a lender because a borrower may default but then
bounce back and settle the missed payments soon afterward, including any imposed penalties. It
is expressed as a percentage.
Exposure Amount
Exposure amount (EA), also known as exposure at default (EAD), is the loss exposure of a bank
at the time of a loan’s default, expressed as a dollar amount. It is the predicted amount of loss in
109
© 2014-2020 AnalystPrep.
EAD is a dynamic amount that keeps on changing as the borrower continues to make payments.
Loss Rate
The loss rate, also known as the loss given default (LGD), is the percentage loss incurred if the
borrower defaults. It can also be described as the expected loss expressed as a percentage. The
loss rate is the amount that’s impossible to recover after selling (salvaging) the underlying asset
The expected loss, EL, is the average credit loss that we would expect from an exposure or a
portfolio over a given period. It is the anticipated deterioration in the value of a risky asset. In
mathematical terms,
EL = EA × PD × LGD
Credit loss levels are not constant but rather fluctuate from year to year. The expected loss
represents the anticipated average loss that can be statistically determined. Businesses will
typically have a budget for the EL and try to bear the losses as part of the standard operating
cash flows.
Exam tip: The expected loss of a portfolio is equal to the summation of expected losses of
personal losses.
ELP = ∑ EA i × PD i × LGD i
A Canadian bank recently disbursed a CAD 2 million loan, of which CAD 1.6 million is currently
outstanding. According to the bank’s internal rating model, the beneficiary has a 1% chance of
110
© 2014-2020 AnalystPrep.
defaulting over the next year. In case that happens, the estimated loss rate is 30%. The
probability of default and the loss rate have standard deviations of 6% and 20%, respectively.
EL = EA × P D × LR
Where:
EA = CAD 1,600,000
P D = 1%
LR = 30%
Thus,
Unexpected Loss
Unexpected loss is the average total loss over and above the expected loss.
111
© 2014-2020 AnalystPrep.
It is the variation in the expected loss. It is calculated as the standard deviation from the mean at
Let UHL denote the unexpected loss at the horizon for asset value VH . Then,
You will usually apply the following formula to determine the value of the unexpected loss:
2 + LR2 × σ 2
UL = EA × √ PD × σLR PD
Where
2
σPD = PD × (1 − PD)
112
© 2014-2020 AnalystPrep.
The Mean and Standard Deviation of Credit Losses
L i – the amount borrowed in the ith assumed to be constant throughout the year.
R i – the recovery rate in the event of the default by the ith loan in the event of the default by the
ith loan
pij – the correlation between losses on the ith and jth loan
a – the standard deviation of portfolio loss as a fraction of the size of the portfolio
L i (1 − Ri )
Intuitively, the probability distribution for the loss from the ith loan is made of a probability p i
that there will be a loss of this amount and the probability 1 − pi that there is no loss, which is
typically a binomial distribution. Therefore, we can present the mean and the standard deviation
of the loss.
p i × L i (1 − R i) + (1 − p i) × 0 = p iL i (1 − Ri )
Now, recall that for a random variable X, the variance is defined as:
σx = E (x2 ) − [E (x)]2
113
© 2014-2020 AnalystPrep.
Where E denotes the expectation. Intuitively,
σi = √p i − p 2i [L i (1 − R i )]
Note that we can also calculate the standard deviation of the loan portfolio from the losses on
n n
σP2 = ∑ ∑ p ij σi σj … … … … … Eq1
i=1 j=1
Now, the standard deviation expressed as the percentage of the size of the portfolio is
n n
√∑i=1 ∑j=1 pij σi σj
α= … … … … … . . Eq2
∑ni=1 L i
Assume that all loans have the same principal L, all recovery rate R are equal, and all default
probabilities are equal and denoted by p, and the correlation coefficient is defined as:
1 when i = j
ρij = {
ρ when i ≠ j
Where ρ is constant.
Therefore, the standard deviation of the loss from loan i is the same for all i so that the common
σ = √p − p 2 [L (1 − R)]
114
© 2014-2020 AnalystPrep.
Therefore, Eq1 reduces to:
σP2 = nσ 2 + n (n − 1) ρσ 2
Also, the standard deviation of the loss from the loan portfolio as a percentage of its size (Eq2)
reduces to:
σP σ√1 + (n − 1) ρ
α= =
nL L√n
The bank of Baroda has a portfolio consisting of 10,000 loans each loan amounting to $2 million
and has a 1% probability of default over the following year. The recovery rate is 40%, and the
correlation coefficient is 0.1. Calculate α, the standard deviation of the loss from the loan
Solution
σ√1 + (n − 1) ρ
α=
L √n
L = $2 million
ρ = 0.1
n = 10,000
R = 0.4
σ = √p − p2 [L (1 − R)]
115
© 2014-2020 AnalystPrep.
Therefore,
A Gaussian copula maps the marginal distribution of each variable to the standard normal
distribution, which, by definition, has a mean of zero and a standard deviation of one. Copula
correlation models create a joint probability distribution for two or more variables while still
preserving their marginal distributions. The joint probability of the variables of interest is
implicitly defined by mapping them to other variables whose distribution properties are known.
Let us define two variables V1 and V 2 that have unknown distributions and unique marginal
distributions. V1 and V2 are mapped into new variables U1 and U 1 that have standard normal
copula.
116
© 2014-2020 AnalystPrep.
For example, the one-percentile point of the V 1 distribution is mapped to the one-percentile point
of the U 1 distribution. Similarly, the 50-percentile point of the V1 distribution is mapped to the
Before mapping variables V 1 and V2 to the normal distribution, it is very difficult to define a
relationship between them since their marginal distributions are unknown and are pretty much
incomprehensible structures. Once they have been mapped to the standard normal distribution
as new variables U1 and U 2, respectively, we can now define a relationship between them since
the standard normal distribution has a known structure. The Gaussian copula, therefore, helps us
to define a correlation between variables when it is not possible to define a correlation directly.
Assume now that we have many variables Vi for all i=1,2,…,n for which each of the Vi can be
117
© 2014-2020 AnalystPrep.
mapped to a standard normal distribution Ui . The main challenge that remains is to define the
correlation between U i distributions because the presence of many unique distributions means
that we have to specify numerous correlation parameters. To address this issue, we use the one-
factor model.
U i = ai F + √1 − a2i Zi
Where F is a common factor for all Ui and Zi is a component of Ui that is unrelated to the factor
F and uncorrelated to each other. The ai are the parameter values that lie between -1 and +1,
The variables F and Zi have the standard normal distributions, that is, F ∼ N (0 , 1) and
Zi ∼ N(0, 1). Therefore, U i is a sum of two independent normal distributions, and it is, therefore,
a normal variable with a mean of 0 and a standard deviation of 1. The variance of Ui is 1 since F
So in a nutshell, the one-factor model takes one standard normally distributed variable Ui and
defines it in relation to two other variables, which are both are standard normally distributed.
The coefficient of correlation between Ui and Uj comes in as a result of the shared factor F and
this ai aj. The correlation coefficient between Ui and Uj is defined from the basic statistics as
118
© 2014-2020 AnalystPrep.
But E (Ui )=E (U j)=0 and σUi = σUj = 1 and thus the above equation reduces to
E (U i Uj ) − 0
ρUi ,Uj = = E (U iU j)
1
Now,
Note that the immediate result stems from the fact that E(F2 ) = 1 because F is a standard normal
A notable example of a one-factor model is the capital asset pricing model (CAPM). In CAPM, the
correlation coefficient between two assets is assumed to arise from the dependence on the
common factor – the return from the market index. However, CAPM is significantly palatable
since it specifies the correlation between the returns of the different assets.
The Vasicek model uses the Gaussian copula model to define the correlation between the
119
© 2014-2020 AnalystPrep.
defaults when determining the capital for loan portfolios. The Vasicek model has an advantage
over the standard deviation of the loss from the loan portfolio as a percentage of its size, α in
Now, assume that the probability of default (PD) is the same for all firms in a huge portfolio so
that the PD for a company i for one year is mapped to a standard normal U i as described earlier.
The values at the further left (shaded region) tail of this standard normal distribution is the
Ui ≤ N−1 (PD)
Where N −1 is the inverse of the inverse cumulative normal distribution. That is if the probability
of default for the loan portfolio of a bank is 1%, then the bank defaults if:
120
© 2014-2020 AnalystPrep.
U i ≤ N −1 (0.01) = −2.326
Ui ∈ (−∞, −2.326)
Ui ∈ (−2.326 , ∞)
121
© 2014-2020 AnalystPrep.
U i = ai F + √1 − a2i Zi
To make the model more palatable, ai ’s are assumed to be equal for all i so that ai = a and thus,
Ui = aF + √ 1 − a2 Zi
ρUi,Uj = ai aj = a. a = a2
Now, the factor F can be taken as an index of the recent economic health. That is, if the F is high
the doing economy is healthy implying that U i is high, making the default relatively low.
Otherwise, if F is relatively low, then Ui is also relatively low, and thus making default most likely.
For a large portfolio, the default rate defined as the probability that:
Ui ≤ N−1 (PD)
which was discussed earlier. Now using the properties of a normal distribution:
N −1 (PD) − aF
Default Rate as a Function of F = N ( )
√1 − a2
We anticipate that the default rate is not exceeded with a 99.9% likelihood and thus given by the
low value of F. Moreover, we require that the probability of the true value of F will be worse than
other F' and it will be 0.1%. Now, recall that F is normally distributed so that:
′
F = N−1 (0.001)
and thus,
122
© 2014-2020 AnalystPrep.
Now, ρ is defined as the correlation between each pair of Ui given by:
ρ = a2 ⇒ a = √ρ
Therefore,
N (0.001) = −N (0.999)
Now consider a loan portfolio with the same default probability. The last equation allows us to
convert the average default rate (PD) into a portfolio default rate, which is interestingly only
Moreover, when ρ = 0,
This makes much sense because if the firms default independently, the “law of large numbers”
makes sure that the default rate in a large portfolio is always the same.
Example: Calculating the 99.9 percentile for default rate under the
Vasicek model
The Bank of Baroda has a loan portfolio that has a default rate of 2% and a correlation coefficient
123
© 2014-2020 AnalystPrep.
of 0.3. Under the Vasicek Model, what is the 99.9 percentile for the portfolio default rate?
Solution
124
© 2014-2020 AnalystPrep.
If a loan has the same PD, the same ρ and the same loss given default (LGD) and the same
principle, the Basel II capital requirement for banks under the IRB approach is given by:
125
© 2014-2020 AnalystPrep.
(WCDR − PD) × LGD × EAD
Where WCDR is defined as the worst-case default rate, and it is 99.9 percentile of the default
rate distribution defined as in Vasicek model. LGD is the loss given the default, which is defined
as one minus the recovery rate. EAD is the total exposure at default, which is the sum of the
The Basel II equation provides a way of calculating the unexpected loss with a 99.9% confidence
level. This can be seen by the fact that WCDR×LGD is the percentile point of loss rate
distribution, while WCDR×LGD×EAD is the loss at the 99.9 percentile. Therefore, the expected
PD × LGD × EAD
For a non-homogeneous loan portfolio, the one-factor model equation can be rewritten so that for
loan i:
The last equation gives a way of calculating the capital for each loan separately, and then the
results are added. Moreover, the equation can be adjusted to include the maturity adjustment
factor.
Basel II defines correlation ρ in that the banks must assume different conditions. Based on the
IRB approach, the banks estimate PD while EAD and LGD estimates are approximated in
accordance with Basel II rules or using the bank’s internal model but subject to circumstances
CreditMetrics Model
The CreditMetrics model is used by the banks to calculate economic capital where each
126
© 2014-2020 AnalystPrep.
borrower is given an external or internal credit rating. A one-year transition table is then utilized
The loan portfolio of the bank is determined at the beginning of the one year, and the Monte
Carlo simulation is used to define how the rating changes over the year. In each of the simulation
experiments, the ratings of all borrowers are determined at the end of the year, which allows the
Reevaluation of the portfolio involves calculating the credit loss of the portfolio, which is defined
as the value of the portfolio at the beginning of the year, less the value of the portfolio at the end
of the year. Through numerous simulation trials, complete credit loss distribution is produced.
To demonstrate the CreditMetrics model, consider a bank X whose credit ratings are A, B, C, and
default. The probabilities of the rating transition are as follows: a B-rated can transition to rating
A with a probability of 5%, 85% chance of staying in B, 14% of transitioning to C, and 3% chance
In each Monte Carlo simulation experiment, a number is sampled from a standard normal
N −1 (0.06) = −1.555, N −1 (0.90) = 1.282, N −1 (0.97) = 1.881. From these values, the corresponding
127
© 2014-2020 AnalystPrep.
However, since the bank borrowers do not default independently, sampling should be done in a
way that reflects the correlation between the samples. As such, a factor model (such as the one-
factor model) is used to capture the relationship between the normal distributions. In other
words, the Monte Carlo simulation should be based on the Gaussian model such that
probabilities of the rating transitions for each borrower are converted to a normal distribution,
and the correlations are defined on those distributions and not rating transitions themselves.
Notably, the correlations between the traded equities are usually used in CreditMetrics, which
can be justified by the Monte Carlo simulation, where the company defaults if its market value
Lastly, as can be seen in the CreditMetrics model, as opposed to the Vasicek model, it includes
Leonhard Euler developed a model that can be used to divide risk measures. The Euler Theorem
as
F (λx 1 , λx2 , . . . , λx n ) = λF (x 1 , x2 , . . . , x n )
Where λ is a constant.
Now, define:
ΔFi
ΔFi Δx i
Q i = xi =
Δxi xi
Where
Δx i = small change in x i
Δx
128
© 2014-2020 AnalystPrep.
Δx i
Qi = ratio of ΔFi to a proportional change in x i
xi
n
F = ∑ Qi
i=1
So, many risk measures are homogeneous functions, which is a property of a coherent risk
measure.
Now, if a portfolio is adjusted such that each position is multiplied by some constant λ, a risk
measure is typically multiplied by λ. So the Euler’s theorem provides a way to allocate a risk
As an application of the Euler’s theorem on the credit risk, we can determine the contribution of
each loan in a portfolio to the overall risk measure. Consider the following example.
The Basley bank has three loans, A, B, and C. Losses from the loans are 1.0, 1.2, and 1.3,
respectively. The correlations between the losses are as shown by the table below:
Now assume that the size of the loan A is increased by 1% so that its new standard deviation is
now 1 × 1.01 = 1.01. The increase in the standard deviation of the loan portfolio is:
129
© 2014-2020 AnalystPrep.
= (√1.012 + 1.22 + 1.32 + (2 × 0.8 × 1.2 × 1.3) − (√1.02 + 1.22 + 1.32 + (2 × 0.8 × 1.2 × 1.3)) = 0.0039
ΔFi
ΔFi Δx i
Q i = xi =
Δxi xi
It implies that,
0.0039
QA = = 0.39
0.01
Now if the size of loan B is also increased by 1%, so that its standard deviation of loss is now
1.2 × 1.01 = 1.2120 and thus the increase in the loan portfolio is given by:
= (√1.02 + 1.21202 + 1.32 + 2 × 0.8 × 1.2120 × 1.3) − (√ 1.02 + 1.22 + 1.32 + 2 × 0.8 × 1.2 × 1.3) = 0.01045
And thus,
0.01115
QB = = 1.045
0.01
QC = 1.142
So,
As per the context of Euler’s theorem, we have divided the total loss of 2.57 into the loan
contributions from loans A, B, and C. It is easy to see that that the contribution of loan A is low
because it is uncorrelated with loans B and C, hence contributes less risk to the entire portfolio.
130
© 2014-2020 AnalystPrep.
Credit Risk Capital for Derivatives
Similar to loans, derivatives (such as options and swaps) generate credit risk. For instance, if
company X purchases an option from company Y, the credit risk to company X arises in the sense
that the company Y might default and thus fail to honor its obligation. In the case of a swap, if
company X enters into an interest rate swap with company Y, the credit risk to company X occurs
The credit risk capital for derivatives can be similarly calculated using the equation:
However, it is challenging to compute EAD for the derivatives transactions. Note that in the case
of loans, EAD is the amount that has been advanced or expected to be advanced to the borrower.
Obviously, in derivatives, exposures varies with the value of the derivative. The solution this
challenge is addressed by the Basel Committee’s rules of computing EAD – setting the exposure
at default for the derivatives equal the current exposure plus add-on amount. The current
exposure is the maximum amount of capital that might be lost in case of default today. The add-
on amount is an additional amount for the possibility of the exposure worsening by the time
default occurs.
An additional challenge for the derivatives involves netting agreements so that all the
outstanding derivatives with a given counterparty may be considered a single derivative in case
The analysis of the credit risk needs many estimates, such as PD. Just like through-the-cycle and
point-in-time for credit ratings, we can differentiate between the through-the-cycle PD (mean of
the PD over the economic cycle) and point-in-time PD (indicates the current economic times).
The banks are required to estimate the through-the-cycle PD for regulatory reasons, but point-in-
131
© 2014-2020 AnalystPrep.
time PD is estimated for internal uses.
The accounting standards, such as IFRS 9, requires that the loans should be valued for
accounting reasons. As such, the expected losses over one year or loan’s life must be computed
and subtracted from the loan amount, at which point-in-time estimate is required. Therefore, the
bank is faced with the difficulty of estimating both through-the-cycle estimates for regulators and
The recovery rate or loss given default is usually negatively correlated with default rate so that
in adverse economic conditions contribute to credit risk is twofold: the default rate increases
The exposure to default (the amount the borrower owes at the time of default) is such that in
borrowing limit given to the customer. Usually, for a term loan, the exposure to default is the
expected principal during a given year. However, for a portfolio of derivatives, the exposure to
default is complex to be calculated during a year, which might lead to wrong-way risk. Wrong-
way risk occurs where a counterparty to an institution is most likely to default if the value of the
The correlations are challenging to approximate. Despite the fact the Gaussian model is easy to
use, there is no surety that it will reflect how bad the amount lost by a loan in one-in-thousand
Apart from the credit risk, a bank must also get concerned about other risks it faces. Such risks
include market, operational, liquidity, and strategic risks. Typically, these risks are assigned to
different arms in a bank but not necessarily independent of each other. These risks interact and
consequently impacting both economic and regulatory capital requirements of the bank.
132
© 2014-2020 AnalystPrep.
Questions
Question 1
Big Data Inc., a U.S. based cloud technology and computing firm, has been offered a
USD 10 million term loan fully repayable in exactly two years. The bank behind the
offer estimates that it will be able to recover 65% of its exposure if the borrower
defaults, and the probability of that happening is 0.8%. The bank’s expected loss one
A. USD 52,000
B. USD 26,000
C. USD 14,000
D. USD 28,000
EL = EA × P D × LR
P D = 0.8%
LR = 35%
Option A is incorrect. The loss given default is taken to be 65%. Note that
Option B is incorrect. The loss given default is taken to be 65% and the final result
dividend by two.
133
© 2014-2020 AnalystPrep.
134
© 2014-2020 AnalystPrep.
Reading 51: Operational Risk
Describe the different categories of operational risk and explain how each type of risk
can arise.
Compare the basic indicator approach, the standardized approach, and the advanced
Describe the standardized measurement approach and explain the reasons for its
Describe the common data issues that can introduce inaccuracies and biases in the
Describe how to identify causal relationships and how to use Risk and Control Self-
Assessment (RCSA) and Key Risk Indicators (KRIs) to measure and manage operational
risks.
Explain the risks of moral hazard and adverse selection when using insurance to
According to the Basel Committee, operational risk is “the risk of direct and indirect loss
resulting from inadequate or failed internal processes, people, and systems or from external
events.”
The International Association of Insurance Supervisors describes the operational risk as to the
135
© 2014-2020 AnalystPrep.
risk of adverse change in the value of the capital resource as a result of the operational
Operational risk emanates from internal functions or processes, systems, infrastructural flaws,
human factors, and outside events. It includes legal risk but leaves out reputational and strategic
This chapter primarily discusses the methods of computing the regulatory and economic capital
for operational risk and how the firms can reduce the likelihood of adverse occurrence and
severity.
1. Internal fraud: Internal fraud encompasses acts committed internally that diverge from
2. External fraud: External fraud encompasses acts committed by third parties. Commonly
encountered practices include theft, cheque fraud, hacking, and unauthorized access to
information.
3. Clients, products, and business practices: This category has much to do with
clients. That includes issues such as fiduciary breaches, improper trading, misuse of
4. Employment practices and work safety: These are acts that go against laws put in
place to safeguard the health, safety, and general well-being of both employees and
customers. Issues covered include unethical termination, discrimination, and the coerced
5. Damage to physical assets: These are losses incurred to either natural phenomena like
6. Business disruption and system failures: This included supply-chain disruptions and
system failures like power outages, software crashes, and hardware malfunctions.
136
© 2014-2020 AnalystPrep.
7. Execution, delivery, and process management: This describes the failure to execute
transactions and manage processes correctly. Issues such as data entry errors and
The three large operational risks faced by financial institutions include cyber risk, compliance
Cyber Risk
The banking industry has developed technologically. This development is evident through online
banking, mobile banking, credit and debit cards, and many other advanced banking technologies.
Technological advancement is beneficial to both the banks and their clients, but it can also be an
The cyber-attack can lead to the destruction of data, theft of money, intellectual property and
personal and financial data, embezzlement, and many other effects. Therefore, the financial
institutions have developed defenses mechanisms such as account controls and cryptography.
However, financial institutions should be aware that they are vulnerable to attacks in the future;
thus, they should have a plan that can be executed on short notice upon the attack.
Compliance Risks
Compliance risks occur when an institution incurs fines due to knowingly or unknowingly
ignoring the industry’s set of rules and regulations, internal policies, or best practices. Some
examples of compliance risks include money laundering, financing terrorism activities, and
Compliance risks not only lead to hefty fines but also reputational damage. Therefore, financial
institutions should put in place structures to ensure that the applicable laws and regulations are
adhered to. For example, some banks have developed a system where suspicious activities are
137
© 2014-2020 AnalystPrep.
detected as early as possible.
Rogue trader risk occurs when an employee engages in unauthorized activities that consequently
lead to large losses for the institutions. For instance, an employee can trade in highly risky assets
To protect itself from rogue trader risk, a bank should make the front office and back office
independent of each other. The front office is the one that is responsible for trading, and the
back office is the one responsible for record-keeping and the verifications of transactions.
Moreover, the treatment of the rogue trader upon discovery matters. Typically, if the
unauthorized trading occurs and leads to losses, the trader will most likely be disciplined (such
as lawful prosecution). On the other hand, if the trader makes a profit from the unauthorized
trading, this violation should not be ignored because it breeds the culture of risk ignorance,
The Basel Committee on Banking Supervision develops the global regulations which are
instituted by the supervisors of each of the banks in each country. Basel II, which was drafted in
1999, revised the methods of computing the credit risk capital. Basel II regulation includes the
The Basel Committee recommends three approaches that could be adopted by firms to build a
capital buffer that can protect against operational risk losses. These are:
138
© 2014-2020 AnalystPrep.
Basic Indicator Approach
Under the basic indicator approach, the amount of capital required to protect against
operational risk losses is set equal to 15% of annual gross income over the previous three years.
Standardized Approach
To determine the total capital required under the standardized approach is similar to the
primary indicator method, but a bank’s activities are classified into eight distinct business lines,
with each of the lines having a beta factor. The average gross income for each business line is
then multiplied by the line’s beta factor. After that, the capital results from all eight business
lines are summed up. In other words, the percentage applied to gross income varies business
lines.
Below are the eight business lines and their beta factors:
To use the standardized approach, a bank has to satisfy several requirements. The bank must:
III. Regularly report operational risk losses incurred in all business lines.
139
© 2014-2020 AnalystPrep.
both internal and external auditors.
The AMA approach is much more complicated compared to other approaches. Under this
method, the banks are required to treat operational risk as credit risk and set the capital equal
to the 99.9 percentile of the loss distribution less than the expected operational loss, as shown by
Moreover, under the AMA approach, the banks are required to take into consideration every
combination of the eight business lines mentioned in the standardized approach. Combining the
seven categories of operational risk with the eight business lines gives a total of (7 × 8 =) 56
potential sources of operational risk. The bank must then estimate the 99.9 percentile of one-
year loss for each combination and then aggregate each combination together to determine the
To use the AMA method, a bank has to satisfy all the requirements under the standardized
140
© 2014-2020 AnalystPrep.
approach, but the bank must also:
I. Be able to estimate unexpected losses, guided by the use of both external and internal
data.
II. Have a system capable of allocating economic capital for operational risk across all
business lines in a way that creates incentives for these business lines to manage
Currently, the Basel Committee has replaced the three approaches with a new standardized
measurement (SMA) approach. Despite being abandoned by the Basel Committee, some aspects
of the AMA approach are still being used by some of the banks to determine economic capital.
The AMA approach opened the eyes of risk managers to operational risk. However, bank
regulators found flaws in the AMA approach in that there is a considerable level of variation in
the calculation done by different banks. In other words, if different banks are provided with the
same data, there is a high chance that each will come up with different capital requirements
The Basel Committee announced in March 2016 to substitute all three approaches for
determining operational risk capital with a new approach called standardized measurement
approach (SMA).
The SMA approach first defines a quantity termed as Business Indicator (BI). BI is similar to
gross income, but it is structured to reflect the size of a bank. For instance, trading losses and
operating expenses are treated separately so that they lead to an increase in BI.
The BI Component for a bank is computed from the BI employing a piecewise linear relationship.
7X + 7Y + 5Z
Where X, Y, and Z are the approximations of the average losses from the operational risk over
141
© 2014-2020 AnalystPrep.
the past ten years defined as:
X – all losses
The computations are structured so that the losses component and the BI component are equal
for a given bank. The Basel provides formula used the required capital for the loss and BI
components.
Computations of the economic capital require the distributions for various categories of
operational risk losses and the aggregated results. The operational risk distribution is
The term “average loss frequency” is defined as the average number of times that large losses
occur in a given year. The loss frequency distribution shows just how the losses are distributed
If the average losses in a given year are λ, then the probability of n losses in a given year is given
e−λ λn
Pr (n) =
n!
The average number of losses of a given bank is 6. What is the probability that there are
142
© 2014-2020 AnalystPrep.
Solution
e−6 615
Pr (15) = ≈ 0.001
15!
Loss Severity
Loss severity is defined as a probability distribution of the size of each loss. The mean and
variance of the loss severity are modeled using the lognormal distribution. That is, suppose the
standard deviation of the loss size is σ, and the mean is μ then the mean of the logarithm of the
μσ
ln ( )
√1 + w
ln (1 + w)
Where:
2
σ
w =( )
μ
The estimated mean and standard deviation of the loss size is 50 and 20, respectively. What is
the mean and standard deviation of the logarithm of the loss size?
Solution
We start by calculating w,
143
© 2014-2020 AnalystPrep.
2
20
w= ( ) = 0.16
50
μσ 20 × 50
ln ( ) = ln ( ) = 6.8335
√1 + w √1.16
ln (1 + w) = ln (1.16) = 0.1484
After estimating λ , μ, and σ, Monte Carlo simulation can be utilized to determine the probability
144
© 2014-2020 AnalystPrep.
The necessary steps of the Monte Carlo Simulations are as follows:
I. Sampling is done from the Poisson distribution to determine the number of loss events
(=n) in a year. For instance, we can sample Percentile of Poisson distribution as a random
II. Sample n times from the lognormal distribution of the loss size for each of the n loss
145
© 2014-2020 AnalystPrep.
occurrences.
Assume the average loss frequency is six, and a number sampled in step I is 0.29. Therefore,
0.29 corresponds to three loss events in a year because using the formula.
e−λ λn
Pr (n) =
n!
Pr (0) + Pr (1) + Pr (2) = 0.284 and Pr (0) + Pr (1) + Pr (2) + Pr (3) = 0.4660
Also, assume that loss size it has a mean of 60 and a standard deviation of 5 using the formulas
μσ
ln ( ) and ln (1 + w).
√1 + w
In step II we will sample three times from the lognormal distribution using the mean
⎛ ⎞
2
⎜⎜ 60 × 5 ⎟⎟ ⎛ 5 ⎞
ln ⎜⎜ ⎟⎟ = 5.70 and standard deviation ln 1 + ( ) = 0.0069
⎜⎜ ⎟⎟ ⎝ 60 ⎠
5 2
√ 1 + ( )
⎝ 60 ⎠
Now, assume the sampled numbers are 4.12, 4.70, and 5.5. Note that the lognormal distribution
gives the logarithm of the loss size. Therefore we need to exponentiate the sampled numbers to
get the actual losses. As such, the three losses are e4.12=61.56, e4.70 =109.95 and e5.5 =244.69.
146
© 2014-2020 AnalystPrep.
This gives the total loss of 416.20 (61.56+109.95+244.69) in the trial herein.
Step 4 requires that the same process be repeated many times to generate the probability
distribution for the total loss, from which the desired percentile can be computed.
The estimation of the loss frequency and loss severity involves the use of data and subjective
judgment. Loss frequency is estimated from the banks’ data or subjectively estimated by
In the case that the loss severity cannot be approximated from the bank’s data, the loss incurred
by other financial institutions may be used as a guide. The methods in which the banks share
information have been laid out. Moreover, there exist data vendor services (such as Factiva),
which is useful at supplying data on publicly reported losses incurred by other banks.
I. Inadequate historical records: The data available for operational risk losses –
including loss frequency and loss amounts – is grossly inadequate, especially when
compared to credit risk data. This inadequacy creates problems when trying to model the
II. Inflation: When modeling the loss distribution using both external and internal data, an
adjustment must be made for inflation. The purchasing power of money keeps on
changing so that a $10,000 loss recorded today would not have the same effect as a
III. Firm-specific adjustments: No two firms are the same in terms of size, financial
structure, and operational risk management. As such, when using external data, it is
the source and your bank. A simple proportional adjustment can either underestimate or
147
© 2014-2020 AnalystPrep.
overestimate the potential loss.
Suppose that Bank A has revenues of USD 20 billion and incurs a loss of USD 500 million.
Another bank B has revenues of USD 30 billion and experiences a loss of USD 300 million. What
Solution
0.23
Bank A Revenue
Estimated Losss for Bank A = Observed Loss for Bank B × ( )
Bank B Revenue
0.23
20
= 300 × ( ) = USD 273.29 million.
30
Scenario analysis aims at estimating how a firm would get along in a range of scenarios, some of
which have not occurred in the past. It’s particularly essential when modeling low-frequency
high-severity losses, which are essential to determine the extreme tails of the loss distribution.
The objective of the scenario analysis is to list events and create a scenario for each one.
148
© 2014-2020 AnalystPrep.
The risk management unit in liaison with senior management.
For each scenario, loss frequency and loss severity are approximated. Monte Carlo simulations
are used to determine a probability distribution for total loss across diverse types of losses. The
loss frequency estimates should capture the existing controls at the financial institution and the
type of business.
Estimation of the probability of rare events is challenging. One method is to state several
categories and ask operational risk experts to put each loss to a category. For instance, some of
the categories might be a scenario that happens once every 1,000 years on average, which is
equivalent to λ = 0.001. The bank could also use a scenario happening once every 100 years on
Operational risk experts estimate the loss severity, but rather than in the form of the mean and
standard deviation, it is more suitable to estimate the 1 percentile to 99 percentile range of the
loss distribution. These estimated percentiles can be fitted with the lognormal distribution. That
is, if 1 percentile and 99 percentiles of the loss are 50 and 100 respectively, then 3.91 (ln(50))
and 4.61 (ln(100)) are 1 and 99 percentiles for the logarithm of the loss distribution respectively.
The concluding point in scenario analysis is that it takes into consideration the losses that have
never been incurred by a financial institution but can occur in the future. Managerial judgment
is used to analyze the loss frequency and loss severity which can give hints on how such loss
event may appear, which in turn assist the firms in setting up plans to respond to loss event or
Typically, economic capital is allocated to business units, after which the return on capital is
computed. Similar to credit risk, the same principles are used in the allocation of operational risk
capital. The provision of operational risk capital to business units acts as an incentive to the
business unit manager to reduce the operational risk because if the manager reduces the loss
frequency and severity, less operational capital will be allocated. Consequently, the profit from
149
© 2014-2020 AnalystPrep.
In a nutshell, the allocation of operational risk capital should sensitize the manager on the
benefits of operational risk. Operational risk reduction does not necessarily reach an optimal
point because there exists operational risk in a firm that cannot be avoided. Therefore, cost-
benefit analysis is carried out when operational risk is reduced by increasing the operational
cost.
The power-law states that, if v is a random variable and another x is a higher value of v (v>x),
The power holds for some probability distributions, and it describes the fatness of the right tail of
the probability distribution of v. K is a scale factor, and α depends on the fatness of the right tail
of the distribution. That is, the fatness of the right tail increases with a decrease in α.
According to the mathematician G.V Gnedenko, the power for many distributions increases as x
tends to infinity. Practically, the power-law is usually taken to be true for the values of x at the
top 5% of the distribution. Some of the distributions in which the power-law holds to be true are
the magnitude of earthquakes, trading volume of the stocks, income of individuals, and the sizes
of the corporations.
Generally, the power-law holds for the probability distributions of random variables resulting
variables, we usually get a normal distribution, and fat tails arise when the distribution is a
According to Fontnouvelle (2003), the power-law holds for the operational risk losses, which
turns to be crucial.
150
© 2014-2020 AnalystPrep.
Example: Measuring Operational Risk Using the Power-law
A risk manager has established that there is a 95% probability that losses over the next year will
not exceed $50 million. Given that the power-law parameter is 0.7, calculate the probability of
the loss exceeding (a) 20 million, (b) 70 million, and (c) 100 million.
Solution
Thus,
Now,
when x=20
when x=70
when x=100
151
© 2014-2020 AnalystPrep.
It is crucial to measure operational risk and to compute the required amount of operational risk
capital. However, it is also imperative to reduce the likelihood of significant losses and severity in
case an event occurs. More often, financial institutions learn from each other. That is, if
significant losses are incurred in one of the financial institutions, risk managers other financial
institutions will try and study what happened so that they can make necessary plans to avoid a
similar event.
Some of the methods of reducing operational risk include: reducing the cause of losses, risk
control, and self-assessment, identifying key risk indicators (KRI’s), and employee education.
Causal Relationship
Causal relationships describe the search for a correlation between firm actions and operational
risk losses. It is an attempt to identify firm-specific practices that can be linked to both past and
future operational risk losses. For example, if the use of new computer software coincides with
losses, it is only wise to investigate the matter in a bid to establish whether the two events are
Once a causal relationship has been identified, the firm should then decide whether or not to act
Risk and control self-assessment (RCSA) involves asking departmental heads and managers to
single out the operational risks in their jurisdiction. The underlying argument is that unit
managers are the focal point of the flow of information and correspondence within a unit. As
such, they are the persons best placed to understand the risks pertinent to their operations.
III. Carrying out interviews with line managers and their staff
152
© 2014-2020 AnalystPrep.
V. Executing brainstorming in a workshop environment
VI. Analyzing the reports from third parties such as auditors and regulators
RCSA should be done periodically, such as yearly. The problem with this approach is that
managers may not divulge information freely if they feel they are culpable or the risk is out of
control. Also, a manager’s perception of risk and its potential rewards may not conform to the
firm-wide assessment. For these reasons, there is a need for independent review.
Key risk indicators seek to identify firm-specific conditions that could expose the firm to
operational risk. KRIs are meant to provide firms with a system capable of predicting losses,
giving the firm ample time to make the necessary adjustments. Examples of KRIs include:
Staff turnover
The hope is that key risk indicators can identify potential problems and allow remedial action to
Education
It is essential to educate the employees on the prohibited business practices and breeding risk
culture where such unacceptable practices might be executed. Moreover, the legal branch of a
financial institution educates the employees to be cautious when writing emails and answering
phone calls. Essentially, employees should be mindful that their emails and recorded calls could
153
© 2014-2020 AnalystPrep.
Earlier in the reading, we saw that a bank using the AMA approach could reduce its capital
charge, subject to extensive investment in operational risk management. One of the ways
through which a bank can achieve this is by taking an insurance cover. That way, the firm is
For all its advantages, taking an insurance policy comes with two problems:
1. Moral Hazard: Moral hazard describes the observation that an insured firm is likely to
increasingly take high-risk positions in the knowledge that they are well protected from
heavy losses. Without such an insurance policy, the traders would be a bit more
In a bid to tame the moral hazard problem, insurers use a range of tactics. These may
include deductibles, coinsurance, and policy limits. Stiff penalties may also be imposed in
A firm can intentionally keep insurance cover private. This ensures that its traders do not
2. Adverse Selection: Adverse selection describes a situation where the risk seller has
more information than the buyer about a product, putting the buyer at a disadvantage.
For example, a company providing life assurance may unknowingly attract heavy
smokers, or even individuals suffering from terminal illnesses. If this happens, the
company effectively takes on many high-risk persons but very few low-risk individuals.
This may result in a claim experience that’s worse than initially anticipated.
On matters trading, firms with poor internal controls are more likely to take up insurance
policies compared to firms with robust risk management frameworks. To combat adverse
risk controls. The premium payable can then be adjusted to reflect the risk of the policy.
154
© 2014-2020 AnalystPrep.
Question 1
Melissa Roberts, FRM, has observed 12 losses in her portfolio over the last four
years. She believes the frequency of losses follows a Poisson distribution with a
parameter λ. The probability that she will observe a total of 4 losses over the next
A. 17%
B. 16%
C. 20%
D. 0.53%
12 losses
λ= = 3 losses per year
4 years
e−λ λn
Pr (n) =
n!
e−3 34
Pr (n = 4) = = 0.168
4!
Question 2
According to the Basel Committee, a bank has to satisfy certain qualitative standards
economic capital required. Which of the following options is NOT one of the
standards?
A. The bank must have a system capable of allocating economic capital for
155
© 2014-2020 AnalystPrep.
operational risk across all business lines in a way that creates incentives for
B. Internal and external auditors must regularly and independently review all
operational risk management processes. The review must include the policy
function.
C. The bank’s operational risk measurement system should only make use of
internally generated data to avoid the bias associated with external data.
D. The bank must have an operational risk management function tasked with
The Basel committee does not rule out the use of external data by banks. In fact, the
committee recommends the use of a combination of both external and internal data to
estimate the unexpected loss. External data may not conform to a particular firm, but
firms are allowed to scale the data to fit their profiles. In some cases, internal data
may be either insufficient or entirely unavailable, forcing the firm to look elsewhere.
Question 3
Melissa Roberts, FRM, has observed 12 losses in her portfolio over the last four
years. She believes the frequency of losses follows a Poisson distribution with a
parameter λ . The probability that she will observe a total of 4 losses over the next
A. 17%
B. 16%
C. 20%
D. 0.53%
12 losses
156
© 2014-2020 AnalystPrep.
12 losses
λ (1) = =3 losses per year
4 years
(λT )n
P r (n) = e−λT
n!
34
Pr (n = 4) = e−3
4!
= 0.168
157
© 2014-2020 AnalystPrep.
Reading 52: Stress Testing
Describe the rationale for the use of stress testing as a risk management tool.
Describe the relationship between stress testing and other risk measures, particularly
Explain the importance of stressed inputs and their importance in stressed VaR and
stressed ES.
Describe the responsibilities of the Board of directors and senior management in stress
testing activities.
Identify areas of validation and independent review for stress tests that require
Describe the important role of the internal audit in stress testing governance and
control.
Describe the Basel stress testing principles for banks regarding the implementation of
stress testing.
Stress testing is a risk management tool that involves analyzing the impacts of the extreme
scenarios that are unlikely but feasible. The main question for financial institutions is whether
158
© 2014-2020 AnalystPrep.
they have adequate capital and liquid assets to survive stressful times. Stress testing is done for
regulatory purposes or for internal risk management by financial institutions. Stress testing can
be combined with measurement of the risk such as the Vale-at-Risk (VaR) and the Expected
Shortfall (ES) to give a detailed picture of the risks facing a financial institution.
This chapter deals with the internally generated stress testing scenarios, regulatory
requirements of stress testing, governance issues of stress testing, and the Basel stress testing
principles.
Stress testing serves to warn a firm’s management of potential adverse events arising
from the firm’s risk exposure and goes further to give estimates of the amount of
capital needed to absorb losses that may result from such events.
Stress tests help to avoid any form of complacency that may creep in after an extended
period of stability and profitability. It serves to remind management that losses could
still occur, and adequate plans have to be put in place in readiness for every
eventuality. This way, a firm is able to avoid issues like underpricing of products,
Stress testing is a key risk management tool during periods of expansion when a firm
introduces new products into the market. For such products, there may be very limited
loss data or none at all, and hypothetical stress testing helps to come up with reliable
loss estimates.
Under pillar 1 of Basel II, stress testing is a requirement for all banks using the
Internal Models Approach (IMA) to model market risk and the internal ratings-based
approach to model credit risk. These banks have to employ stress testing to determine
Stress testing supplements other risk management tools, helping banks to mitigate
159
© 2014-2020 AnalystPrep.
risks through measures such as hedging and insurance. By itself, stress testing cannot
address all risk management weaknesses, nor can it provide a one-stop solution.
Recall that the VaR and ES are estimated from a loss distribution. VaR enables a financial
institution to conclude with X% likelihood that the losses will not exceed the VaR level during
time T. On the other hand, ES enables the financial institutions to conclude whether the losses
exceed the VaR level during a given time T and hence the expected loss will be the ES amount.
VaR and ES are backward-looking. That is, they assume that the future and the past are the
same. On the other hand, stress testing is forward-looking. It asks the question, “what if?”.
While stress testing largely does not involve probabilities, VaR, and ES models are founded on
probability theory. For example, a 99.9% VaR can be viewed as a 1-in-1,000 event.
The backward-looking ES and VaR consider a wide range of scenarios that are potentially good
or bad to the organization. However, stress testing considers a relatively small number of
Specifically, for the market risk, VaR/ES analysis often takes a short period of time, such as a day,
The primary objective of stress testing is to capture the enterprise view of the risks impacting a
financial institution. The scenarios used in the stress testing are often defined based on the
macroeconomic variables such as the unemployment rates and GDP growth rates. The effect of
these variables should be considered in all parts of an institution while considering interactions
Conventional VaR and ES are calculated from data spanning from one to five years, where a daily
variation of the risk factors during this period is used to compute the potential future
movements.
160
© 2014-2020 AnalystPrep.
However, in the case of the stressed VaR and stressed ES, the data is obtained from specifically
stressed periods (12-month stressed period on current portfolios according to Basel rules). In
other words, stressed VaR and stressed ES generates conditional distributions and conditional
risk measures. As such, they are conditioned to a recurrence of a given stressed period and thus
Though stressed VaR and stressed ES might be objectively similar, they are different. Typically
the time horizon for the stressed VaR/ES is short (one to ten days), while for the stress testing, it
For instance, assume that a stressed period is the year 2007. The stressed VaR would conclude
that if there was a repeat of 2007, then there is an X% likelihood that losses over a period of T
days will not surpass the stressed VaR level. On the other hand, stressed ES would conclude that
if the losses over T days do not exceed the stressed VaR level, then the expected loss is the
stressed ES.
However, stress testing would ask the questions “if the following year (2008) is the same as in
2007, will the financial institution survive?” Alternatively, what if the conditions of the next year
are twice as adverse as that of 2007, will the financial institution survive? Therefore, stress
testing does not consider the occurrence of the worst days of 2008 but rather the impact of the
whole year.
There is also a difference between conventional VaR and the stressed VaR. Conventional VaR can
be back-tested while stressed VaR cannot. That is, if we can compute one-day VaR with 95%
confidence, we can go back and determine how effective it would have worked in the past. We
are not able to back-test the stressed VaR output and its results because it only considers the
The basis of choosing a stress testing scenario is the selection of a time horizon. The time
horizon should be long enough to accommodate the full analysis of the impacts of scenarios.
Long time horizons are required in some situations. One-day to one-week scenarios can be
161
© 2014-2020 AnalystPrep.
considered, but three months to two-year scenarios are typically preferred.
The regulators recommend some scenarios, but in this section, we will discuss internally chosen
scenarios. They include using historical scenarios, stressing key variables, and developing ad hoc
Historical Scenarios
Historical scenarios are generated by the use of historical data whose all relevant variables will
behave in the same manner as in the past. For instance, variables such as interest rates and the
credit rate spreads are known to repeat past changes. As such, actual changes in the stressed
period will be assumed to repeat themselves while proportional variations will be assumed for
others. A good example of a historical scenario is the 2007-2008 US housing recession, which
In some cases, a moderately adverse scenario is made worse by multiplying variations of all risk
factors by a certain amount. For instance, we could multiply what happened in the loss-making
one-month period and increase the frequency of movement of all relevant risk movements by ten.
As a result, the scenario becomes more severe to financial institutions. However, this approach
assumes linear relationships between the movements in risk factors, which is not always the case
Other historical scenarios are based on one-day or one-week occurrences of all market risk
factors. Such events include terrorist attacks (such as 9/11 terrorist attacks) and one-day
massive movement of interest rates (such as on April 10, 1992, when ten-year bond yields
A scenario could be built by assuming that a significant change occurs in one or more key
162
© 2014-2020 AnalystPrep.
A 25% decrease in equity prices
Some other significant variations could occur in factors such as money exchange rates, prices of
In the case of the market risk, small changes in measured using the Greek letters (such as delta
and gamma). The Greek letters cannot be used in stress testing because the changes are usually
large. Moreover, Greeks are used to measure risk from a unit market variable over a short period
of time, while stress testing incorporates the interaction of the different market variables over a
The stress testing scenarios we have been discussing above are performed regularly, after which
the results are used to test the stability of the financial structure of a financial institution in case
of extreme conditions. However, the financial institutions need to develop ad hoc scenarios that
capture the current economic conditions, specific exposures facing the firm, and update analysis
of potential future extreme events. The firms either generate new scenarios or modify the
An example of an event that will prompt the firms to develop an ad hoc scenario is the change in
the government policy on an important aspect that impacts the financial institutions or change in
Basel regulation that requires increment of the capital within short periods of time.
The boards, senior management, and economic experts use their knowledge in markets, global
politics, and current global instabilities to come with adverse scenarios. The senior management
carries out a brain-storming event, after which they recommend necessary actions to avoid
unabsorbable risks.
163
© 2014-2020 AnalystPrep.
Using the Stress Testing Results
While stress testing, it is vital to involve the senior management for it to be taken seriously and
thus used for decision making. The stress-testing results are not only to satisfy the “what if”
question, but also the Board and management should analyze the results and decide whether a
certain class of risk mitigation is necessary. Stress testing makes sure that the senior
management and the Board do not base their decision making on what is most likely to happen,
but also consider other alternatives less likely to happen that could have a dramatic result on the
firm.
Model Building
It is possible to see how the majority of the relevant risk factors behave in a stressed period
while building a scenario, after which the impact of the scenario on the firm is analyzed in an
almost direct manner. However, scenarios generated by stressing key variables and ad hoc
scenarios capture the variations of a few key risk factors or economic variables. Therefore, in
order to exhaust the scenarios, it is necessary to build a model to determine how the “left out”
variables are expected to behave in a stressed market condition. The variables stated in the
context of the stress testing are termed as core variables, while the remaining variables are
One method is performing analysis, such as regression analysis, to relate the peripheral
variables to the core variables. Note that the variables are based on the stressed economic
conditions. Using the data of the past stressed periods is most efficient in determining
appropriate relationships.
For example, in the case of the credit risk losses, data from the rating agencies, such as default
rates, can be linked to an economic variable such as GDP growth rate. Afterward, general default
rates expected in various stressed periods are determined. The results can be modified (scaled
up or down) to determine the default rate for different loans or financial institutions. Note that
the same analysis can be done to the recovery rates to determine loss rates.
164
© 2014-2020 AnalystPrep.
Apart from the immediate impacts of a scenario, there are also knock on-effects that reflect how
financial institutions respond to extreme scenarios. In its response, a financial institution can
For instance, during the 2005-2006 US housing price bubble, banks were concerned with the
credit quality of other banks and were not ready to engage in interbank lending, which made
Recall that stress testing involves generating scenarios and then analyzing their effects. Reverse
stress testing, as the name suggests, takes the opposite direction by trying to identify
By using historical scenarios, a financial institution identifies past extreme conditions. Then, the
bank determines the level at which the scenario has to be worse than the historical observation
to cause the financial institution to fail. For instance, a financial institution might conclude that
twice the 2005-2006 US housing bubble will make the financial institution to fail. However, this
kind of reverse stress testing is an approximation. Typically, a financial institution will use
complicated models that take into consideration correlations between different variables to make
Finding an appropriate combination of risk factors that lead the financial institution to fail is a
challenging feat. However, an effective method is to identify some of the critical factors such as
GDP growth rate, unemployment rates, and interest rate variations, then build a model that
relates all other appropriate variables to these key variables. After that, possible factor
US, UK, and EU regulators require banks and insurance companies to perform specified stress
tests. In the United States, the Federal Reserve performs stress tests of all the banks whose
consolidated assets are over USD 50 billion. This type of stress test is termed as Comprehensive
Capital Analysis and Review (CCAR). Under CCAR, the banks are required to consider four
165
© 2014-2020 AnalystPrep.
scenarios:
I. Baseline Scenario
III. Scenario
The baseline scenario is based on the average projections from the surveys of the economic
predictors but does not represent the projection of the Federal Reserve.
The adverse and the severely adverse scenarios describe hypothetical sets of events which are
structured to test the strength of banking organization and their resilience. Each of the above
scenarios consists of the 28 variables (such as the unemployment rate, stock market prices, and
interest rates) which captures domestic and international economic activity accompanied by the
Board explanation on the overall economic conditions and variations in the scenarios from the
past year.
Banks are required to submit a capital plan, justification of the models used, and the outcomes of
their stress testing. If a bank fails to stress test due to insufficient capital, the bank is required to
raise more capital while restricting the dividend payment until the capital has been raised.
Banks with consolidated assets between USD 10 million and USD 50 million are under the Dodd-
Fank Act Stress Test (DFAST). The scenarios in the DFAST is similar to that in the CCAR.
However, in the DFAST, banks are not required to produce a capital plan.
Therefore, through stress tests, regulators can evaluate the banks to determine their ability to
extreme economic conditions consistently. However, they recommend that banks develop their
scenarios.
Stress testing governance is important for determining the extent to which financial institutions
do the exercise. Moreover, governance ensures that the assumption is a stress testing has been
well evaluated, and the senior management analyzes the stress testing outcomes and the
166
© 2014-2020 AnalystPrep.
necessary actions taken out of relevant results.
Stress testing is a broad term used to describe a firm’s examination of its risks and the potential
impact of stressful events on its financial condition. Stress testing seeks to lay bare all the
financial circumstances a firm may find itself in. It helps to understand the range of potential risk
exposures.
The importance of stress testing was highlighted in the post-analysis of the 2007/2008 financial
crises. At many firms, stress testing was insufficiently integrated into their risk management
practices. The few that had shown more commitment to stress testing did not examine
sufficiently severe scenarios, and test results had little or no consideration in decision making.
Most importantly, the crisis brought to the fore the key role of governance in stress testing. In
The buck stops with senior management and the Board of directors. The management
should have an in-depth understanding of the institution’s risk profile and the potential
The stress testing program should be guided by clear, well-detailed policies that outline
the procedure to follow from start to finish, as well as describing the role played by
various employees.
There should be an independent review of not just the testing procedure but also the
The ensuing learning outcomes seek to break down these key elements even further.
167
© 2014-2020 AnalystPrep.
For effective operation of stress testing, the Board of directors and senior management should
have distinct responsibilities. What’s more, there should be some shared responsibilities,
although a few roles can be set aside exclusively for one of the two groups.
1. The buck stops with the Board: The Board of directors is “ultimately” responsible for
a firm’s stress tests. Even if board members do not immerse themselves in the technical
details of stress tests, they should ensure that they stay sufficiently knowledgeable about
information on stress tests, including results from every scenario. Members should then
evaluate these results to ensure they take into account the firm’s risk appetite and
overall strategy.
3. Continuous review: Board members should regularly review stress testing reports with
a view to not just critic key assumptions but also supplement the information with their
4. Integrating stress testing results in decision making: The Board should make key
decisions on investment, capital, and liquidity based on stress test results along with
other information. While doing this, the Board should proceed with a certain level of
caution in cognizance of the fact that stress tests are subject to assumptions and a host
of limitations.
with guidelines on stress testing, such as the risk tolerance level (risk appetite).
1. Implementation oversight: Senior management has the mandate to ensure that stress
testing guidelines authorized by the Board are implemented to the letter. This involves
establishing policies and procedures that help to implement the Board’s guidelines.
2. Regularly reporting to the Board: Senior management should keep the Board up-to-
date on all matters to do with stress testing, including test designs, emerging issues, and
168
© 2014-2020 AnalystPrep.
compliance with stress-testing policies.
3. Coordinating and Integrating stress testing across the firm: Members of senior
across the firm, making sure that all departments understand its importance.
contradictions, and possible gaps in stress tests to make improvements to the whole
process.
5. Ensuring stress tests have a sufficient range: In consultations with the Board of
directors, senior management has to ensure that stress testing activities are sufficiently
severe to gauge the firm’s preparation for all possible scenarios, including low-frequency
high-impact events.
6. Using stress tests to assess the effectiveness of risk mitigation strategies: Stress
tests should help the management to assess just how effective risk mitigation strategies
are. If such strategies are effective, significantly severe events will not cause significant
financial strain. If the tests predict significant financial turmoil, it could be that the
7. Updating stress tests to reflect emerging risks: As time goes, an institution will
gradually gain exposure to new risks, either as a result of market-wide trends or its
A financial institution should set out clearly stated and understandable policies and procedures
governing stress testing, which must be adhered to. The policies and procedures ensure that the
169
© 2014-2020 AnalystPrep.
Describe the procedures of stress testing;
Describe the roles and responsibilities of the parties involved in stress testing;
Describe how the independent reviews of the stress testing will be done;
Give clear documentation on stress testing to third parties (e.g., regulators, external
Explain how the results of the stress testing will be used and by whom;
They were amended as the stress testing practices changes as the market conditions
change;
Accommodate tracking of the stress test results as they change through time; and
Document the activities of models and the software acquired from the vendors or other
third parties.
Documentation
A firm should ensure that its stress-testing activities are well documented, including clear and
precise details about key assumptions, methodologies, limitations, uncertainties, and test results.
It provides for continuity of stress tests even in the face of significant changes in
It helps stress test developers to further improve the process by forcing them to think
170
© 2014-2020 AnalystPrep.
clearly about their methods, assumptions, and choices.
The stress testing governance covers the independent review procedures, which are expected to
unbiased and provide assurance to the Board that stress testing is carried out while following the
firm’s policies and procedures. Financial institutions use diverse models that are subject to
independent review to make sure that they serve the intended purpose.
Areas of Validation and Independent Review for Stress Tests that Require Attention from a
Ensuring that validation and independent review are conducted on an ongoing basis;
Ensuring that subjective or qualitative aspects of a stress test are also validated and
Ensuring that there is sufficient independence in both validation and review of stress
tests;
Ensuring that third-party models used in stress-testing activities are validated and
Ensuring that stress tests results are implemented rigorously, and verifying that any
171
© 2014-2020 AnalystPrep.
activities;
Ensure that stress tests across the organization are conducted in a sound manner and
Assess the skills and expertise of the staff involved in stress-testing activities;
To accomplish all the above, internal audit staff must be well qualified. They should be well-
When stress-testing for capital and liquidity adequacy, the firm’s overall strategy and
annual planning cycle should be taken into account. Results should be refreshed in
light of emerging strategic decisions that have the potential to affect the financial
Stress tests for capital and liquidity accuracy should seek to reveal how cash flows –
For entities with subsidiaries, tests should assess the likelihood of individual
Stress tests should simulate the effects of capital problems and illiquidity pressures
happening at the same time. Also, there should be efforts to show if one of these
Stress tests should help the firm to come up with contingency plans to deal with
172
© 2014-2020 AnalystPrep.
crippling capital/liquidity problems. The tests should make clear the specific steps
taken, as well as those that cannot be taken in the face of the aforementioned
problems.
The Basel Committee emphasizes that stress testing is a crucial aspect by requiring that the
market risk calculations are based on the internal VaR and the Expected Shortfall (ES) models,
which should be accompanied by “ rigorous and comprehensive” stress testing. Moreover, Banks
that use the internal rating approach of the Basel II to calculate the credit risk capital should
Influenced by the 2007-2008 financial crisis, the Basel Committee published the principles of the
stress-testing for the banks and corresponding supervisors. The overarching emphasis of the
Basel committee was the importance of stress testing in determining the amount of the capital
Therefore, the Basel committee recognized the importance of stress testing in:
Facilitating the development of risk mitigation, or any other plans to reduce risks in
When the Basel committee considered the stress tests done before 2007-2008, they concluded
that:
It is crucial to involve the Board and the senior management in stress testing. The
173
© 2014-2020 AnalystPrep.
Board and the senior management should be involved in stress testing aspects such as
choosing scenarios, setting stress testing objectives, analysis of the stress testing
results, determining the potential actions, and strategic decision making. During the
crisis, banks that have had senior management that had an interest in developing a
stress test, which eventually affected their decision making, performed fairly well.
The approaches of the stress-testing did not give room for the aggregation of different
exposures in different parts of a bank. That is, experts from different parts of the bank
The scenarios chosen in the stress tests were too moderate and were based on a short
period of time. The possible correlations between different risk types, products, and
markets were ignored. As such, the stress test relied on the historical scenarios and left
out risks from new products and positions taken by the banks.
Some of the risks were not considered comprehensively in the chosen scenarios. For
example, counterparty credit risk, risks related to structured products, and product
awaiting securitizations were partially considered. Moreover, the effect of the stressed
The stress testing frameworks should involve a governance structure that is clear,
oversight bodies, and those concerned with stress testing operation should be clearly
stated.
174
© 2014-2020 AnalystPrep.
stakeholders and the appropriate communication to stakeholders of the stress testing
The stress testing frameworks should satisfy the objectives that are documented and
approved by the Board of an organization or any other senior governance. The objective
should be able to meet the requirements and expectations of the framework of the bank
and its general governance structure. The staff mandated to carry out stress testing
Stress testing should reflect the material and relevant risk determined by a robust risk
identification process and key variables within each scenario that is internally consistent.
A narrative should be developed explaining a scenario that captures risks, and those
risks that are excluded by the scenario should be described clearly and well documented.
Stress testing is typically a forward-looking risk management tool that potentially helps a
bank in identifying and monitoring risk. Therefore, stress testing plays a role in the
formulation and implementation of strategic and policy objectives. When using the
results of stress testing, banks and authorities should comprehend crucial assumptions
and limitations such as the relevance of the scenario, model risks, and risk coverage.
Lastly, stress testing as a risk management tool should be done regularly in accordance
with a well-developed schedule (except ad hoc stress tests). The frequency of a stress
175
© 2014-2020 AnalystPrep.
The size and complexity of the financial institution; and
Stress testing frameworks should have adequate organizational structures that meet the
objectives of the stress test. The governance processes should ensure that the resources
for stress testing are adequate, such that these resources have relevant skill sets to
Stress tests identify risks and produce reliable results if the data used is accurate and
complete, and available at an adequately granular level and on time. Banks and
processing, and reporting of information used in stress tests. The data infrastructure
should be able to provide adequate quality information to satisfy and objectives of the
stress testing framework. Moreover, structures should put place to cover any material
information deficiencies.
The models and methodologies utilized in stress testing should serve the intended
purpose. Therefore,
granularity of the data and the types of risks based on the objectives of the
The complexity of the models should be relevant to both the objectives of the
176
© 2014-2020 AnalystPrep.
stress testing and target portfolios being assessed using the models; and
The model building should be a collaborative task between the different experts. As such,
the model builders engage with stakeholders to gain knowledge on the type of risks
being modeled and understand the business goals, business catalysts, risk factors, and
other business information relevant to the objectives of the stress testing framework.
Periodic review and challenge of stress testing for the financial institutions and the
understanding of results’ limitations, identifying the areas that need improvement and
ensuring that the results are utilized in accordance with the objectives of the stress
testing framework.
banks or authorities improves the market discipline and motivate the resilience of the
Banks and authorities who choose to disclose stress testing results should ensure that
the method of delivery should make the results understandable while including the
limitations and assumptions to which the stress test is based. Clear conveyance of stress
test results prevents inappropriate conclusions on the resilience of the banks with
different results.
177
© 2014-2020 AnalystPrep.
Question 1
Hardik and Simriti compare and contrast stress testing with economic capital and
A. Stress tests tend to calculate losses from the perspective of the market, while
conditional scenarios
C. While stress tests examine a long period, typically spanning several years, EC
models focus on losses at a given point in time, say, the loss in value at the
D. Stress tests tend to use cardinal probabilities while EC/VaR methods use
ordinal arrangements
accounting, while EC/VaR methods compute losses based on a market point of view.
D is also inaccurate. Stress tests do not focus on probabilities. Instead, they focus on
ordinal arrangements like “severe,” “more severe,” and “extremely severe.” EC/VaR
methods, on the other hand, focus on cardinal probabilities. For instance, a 95% VaR
Question 2
One of the approaches used to incorporate stress testing in VaR involves the use of
178
© 2014-2020 AnalystPrep.
genuine disadvantage of relying on risk metrics that incorporate stressed inputs?
insufficient
D. The risk metrics primarily depend on portfolio composition and are not
The most common disadvantage of using stressed risk metrics is that they do not
respond to current issues in the market. As such, significant shocks in the market can
Question 3
Sarah Wayne, FRM, works at Capital Bank, based in U.S. The bank owns a portfolio of
corporate bonds and also has significant equity stakes in several medium-size
companies across the United States. She was recently requested to head a risk
establish how well prepared the bank is for destabilizing events. Which of the
following scenario analysis options would be best for the purpose at hand?
analysis
Scenario analyses should be dynamic and forward-looking. This implies that historical
179
© 2014-2020 AnalystPrep.
combined. Pure historical scenarios can give valuable insights into impact but can
underestimate the confluence of events that are yet to occur. What’s more, historical
scenario analyses are backward-looking and hence neglect recent developments (risk
should take into account both specific and systematic changes in the present and
near future.
Question 4
I. Ensuring that stress testing policies and procedures are followed to the letter
II. Assessing the skills and expertise of the staff involved in stress-testing
activities
IV. Making key decisions on investment, capital, and liquidity based on stress
A. I, II, and IV
B. I and V
C. III and IV
D. V only
Roles II and III belong to internal audit. Role IV belongs to the board of directors.
180
© 2014-2020 AnalystPrep.
Reading 53: Pricing Conventions, Discounting, and Arbitrage
Define discount factor and use a discount function to compute present and future
values.
Define the “law of one price,” explain it using an arbitrage argument, and describe how
Identify arbitrage opportunities for fixed income securities with certain cash flows.
Identify the components of a US Treasury coupon bond, and compare and contrast the
STRIPS.
Construct a replicating portfolio using multiple fixed income securities to match the
Differentiate between “clean” and “dirty” bond pricing and explain the implications of
A discount factor for a particular term gives the value today of one unit of currency due at the
end of that term. It's essentially a discount rate. The discount factor for t years is denoted as
d(t). For example, if d(1) =0.85, then the present value of, say, $1 to be received a year from
Treasury bond prices. The discount factor d(t) is the factor which, when multiplied by the total
amount of money to be received (principal + interest), gives the price (present value) of the
bond. However, when performing these calculations, it’s important to note that cash flows with
different timings have different discount factors, in line with the time value of money. For
181
© 2014-2020 AnalystPrep.
example, the discount factor that applies to interest due in six months will be different from the
discount factor for interest due in a year, i.e., d(0.5) ≠ d(1), and d(1) < d(0.5).
A bond quote is the last price at which a bond is traded, expressed as a percentage of par value
(100). Those bonds sold at a discount are priced at less than 100, and another group, although
fewer, are sold at a premium and are priced at more than 100.
US T-bonds are quoted in dollars and fractions of a dollar – paving the way for the so-called
"32nds" convention. And as the wording suggests, 32 portions of a dollar are considered. For
example, if we have a T-bond quoted at 98–16, this means 98 "full" dollars plus 16/32 of a dollar,
Corporate or municipal bonds, on the other hand, use dollars and eight fractions of a dollar.
A “+” sign at the end of a quote represents half a tick. For example,
16.5
98 − 16 + implies 98 +
32
Treasury Bills
Treasury bills are short term debt obligation issued by the government, which usually last for
one year or less and do not pay coupons. They are usually quoted at a discount to the face value
of 100. The cash price is the face value minus the quoted discount rate.
Let Q be the quoted price of a T-bill and C as the cash price. If the cash price is 97 and there are
360
Q= (100 − C)
n
360
Q= (100 − 97)
n
Q = 12
182
© 2014-2020 AnalystPrep.
This means that we would pay 97 today to get the face value of 100 in 90 days. Our discount is
12
12 for every 100 of face value, which means our annual discount rate is around = 12%.
100
Treasury Bonds
Treasury bonds last for more than one year and usually pay coupons. The accrued interest is the
amount of coupon payment accrued between two coupon dates. When we are talking of treasury
bonds,
The Law of one price states that the price of a security, commodity, or asset should be the same
in two different markets, say, A and B. In other words, if two securities have the same cash flows,
they must have the same price. Otherwise, a trader can generate a risk-free profit by buying on
market A and selling on market B in one risk-free move. Such a possibility is called an arbitrage
opportunity.
The Law of one price describes security price quite well because, in case of an arbitrage
opportunity, traders rush en masse to take advantage of it. Within no time, market forces of
Consider a 1-year maturity bond with a face value of $100, a coupon rate of 10%, paying coupon
semi-annually. Assume that the borrowing (bank) interest rate is 5% per annum.
The present value of the cash flows from this bond is:
5 105
PV = + = $104.82
1.025 1.025 2
If the bond has a price of $100, an investor can borrow $100 from a bank and buy the bond. After
183
© 2014-2020 AnalystPrep.
six months, they will be able to repay $5 after receiving the first coupon. At this point, the debt
At the end of the year, the investor will pocket the principal ($100) as well as the second coupon
Thus, after fully repaying the debt, they will be left with $105 − $99.94 = $5.06, which would
To exploit this situation, eagle-eyed investors in an efficient market would attempt to buy this
bond by borrowing funds from banks. Increased demand would drive the price up so that at the
Liquidity
Liquidity refers to how easily an asset converts to cash. It affects the price of a bond since it
Liquidity issues have, at times, causing a violation of the Law of one price, and therefore we can
A US Treasury coupon can have strips in two distinct securities: The principal security, also
known as the P-STRIP, and the detached coupons, also called C-STRIPS. The two types of
For instance, suppose we have a 10-year bond with a $100,000 face value and a 10% annual
interest rate. Assuming it initially pays coupons semi-annually, 21 zero-coupon bonds can be
184
© 2014-2020 AnalystPrep.
created. That's the 20 C-STRIPS plus the principal strip (P-STRIP). Each of the C-STRIPS has a
$%5,000 face value, which is the amount of each coupon. The P-STRIP to be received at maturity
They have a very high credit quality because US Treasury securities back them.
Because they are sold at a discount, investors do not need a large stash of money to
purchase them
The payout is known in advance as long as the investor holds them to maturity
They offer a range of maturity dates and can, therefore, be used to match liabilities due
STRIPS are eligible for inclusion in tax-deferred retirement plans and non-taxable
accounts such as pension funds, in which their value would grow tax-free until your
retirement.
Disadvantages of STRIPS
Shorter-term STRIPS tend to trade rich while longer-term STRIPS tend to trade cheap
They typically trade very close to the fair value, thus potential profits are small
Here’s an example of how a replicating portfolio can be created from multiple fixed income
securities:
Assume we have a 2-year fixed-income security with $100 face value and a 20% coupon rate,
185
© 2014-2020 AnalystPrep.
paid on a semi-annual basis. Assume further that the security has a yield to maturity of 5%.
10 10 10 10
PBB1 = + + + = $128.21
1 2 3
1.025 1.025 1.025 1.0254
If this bond is determined to be trading cheap, a trader can carry out an arbitrage trade by
purchasing the undervalued bond and shorting a portfolio that mimics (replicates) the bond’s
cash flows. Assume that in addition to our bond above, which we shall call bond 1, we have four
Using the above multiple fixed-income securities, we can create a replicating portfolio. However,
we must first determine the percentage face amounts of each bond to purchase,Fi , where
i=1,2,3,4, which match bond one cash flows in every semi-annual period.
In these types of calculations, the most straightforward approach to obtaining the values of Fi
involves starting from the end and then working backward. The logic here is simple. At 24
months, only bond 5 makes a payment. Hence at this point, all other values are equal to zero.
12
$110 = F2 × 0 + F3 × 0 + F4 × 0 + F5 × (100 + )%
2
110
186
© 2014-2020 AnalystPrep.
110
F5 = = 103.77%
106%
At 18 months, only bonds 4 and 5 make a payment. We can, therefore, obtain the value of F4 as
follows:
10 12%
$10 = F2 × 0 + F3 × 0 + F4 × (100 + ) % + 103.77 ×
2 2
10 − 103.77 × 0.06
F4 =
1.05
24 10% 12%
$10 = F2 × 0 + F3 × (100 + ) % + 3.59 × + 103.77% ×
2 2 2
10 − 0.18 − 6.23
F3 = = 3.21%
1.12
Similarly,
We can create Cash flows from the replicating portfolio as the product of each bond's initial cash
flows and the face amount percentage. For example, the cash flow from bond five at 12 months is
equal to:
(12%)
× $100 × 103.77% = $6.22
2
187
© 2014-2020 AnalystPrep.
Similarly, the cash flow from bond two at six months
(14%)
× $100 × 2.99% = $0.21
2
As can be seen above, the cash flows from the four bonds replicate bond one cash flows.
The dirty price of a bond is a bond pricing quote that's equal to the present value of all future
cash flows, including interest accruing on the next coupon payment date. Bonds do trade in the
secondary-market before paying any coupon, or after clearing several coupons. In other words,
the day a trader buys or sells the bond could be in between coupon payment dates.
In line with the principle of the time value of money, it's only fair to compensate the seller of a
bond for the number of days they have held the bond between coupon payment dates. We call
this compensation the accrued interest – the interest earned in between any two coupon dates.
(number of days that have elapsed since the last coupon was paid
Accrued interest = c ( )
number of days in the coupon period)
For example, suppose a $1,000 par value bond pays semi-annual coupons at a rate of 20%, and
we've had 120 days since the last coupon was paid. Assuming that there are 30 days in a month,
120
Accrued interest = × $100 = $66.70
180
188
© 2014-2020 AnalystPrep.
The seller would be compensated to the tune of $67.70, while the buyer would see out the
The clean price of a bond is the price that doesn’t include any coupon payments.
The dirty and clean prices are also known as the full and quoted prices, respectively.
Day-Count Conventions
When computing the accrued interest, we use one of several day-count conventions. These
include:
Actual/actual
Actual/360
Actual/365
30/360
convention considers the actual number of days between two coupon dates. The 30/360
convention assumes there are 30 days in any given month and 360 days in a year.
Exam tips:
If coupons are paid semi-annually, the denominator should be 180 in both actual/360
and 30/360 conventions. Similarly, the denominator would be 90 for quarterly coupons.
Almost all US Treasury trades settle T + 1, which means that the exchange of bonds for
189
© 2014-2020 AnalystPrep.
cash happens one business day after the trade date.
C C C C+F
Price = + + +⋯
k k+1 k+2
(1 + y) (1 + y) (1 + y) (1 + y)(k+n−1)
Where:
P = price
C = semi-annual coupon
the number of days in the coupon period, determined as per the relevant day-count convention.
190
© 2014-2020 AnalystPrep.
Question 1
A $1,000 par value US corporate bond pays coupons semi-annually on January 1 and
July 1 at the rate of 20% per year. Mike Brian, FRM, purchases the bond on March 1,
2018, intending to keep it until maturity. The bond is scheduled to mature on July 1,
2021. Compute the dirty price of the bond, given that the required annual yield is
10%.
1. $1,310.25
2. $502.50
3. $400.25
4. $1,100
convention. Under this convention, the number of days between the settlement date
(March 1, 2018) and the next coupon date (July 1, 2018) is 120 (= 4 months at 30
20%
Each coupon payment is valued at × $1, 000 = $100
2
C C C C+F
Price = + + +⋯
k k+1 k+2
(1 + y) (1 + y) (1 + y) (1 + y)(k+n−1)
Where:
P = price
C = semi-annual coupon
number of days in the coupon period, determined as per the relevant day-count
convention.
191
© 2014-2020 AnalystPrep.
y = periodic required yield
Question 2
An analyst has been asked to check for arbitrage opportunities in the Treasury bond
market by comparing the cash flows of selected bonds with the cash flows of
combinations of other bonds. If a 1-year zero-coupon bond is priced at USD 97.25 and
a 1-year bond paying a 20% coupon semi-annually, is priced at USD 114.50, what
should be the price of a 1-year Treasury bond that pays a coupon of 10% semi-
annually?
1. $105.88
2. $100
3. $103.35
4. $105
The secret here is to replicate the 1-year 10% bond using the other two treasury
bonds whose price we already know. To do this, you could solve a system of equations
to determine the weight factors, F1 and F2, which correspond to the proportion of
At every coupon date, the cash flow from the 10% bond should match cash flows from
192
© 2014-2020 AnalystPrep.
At t=1, the 10% bond pays 105, and both the zero-bond and the 20% also have got
At t=0.5, the 10% bond pays 5, the zero bond pays 0, and the 20% bond pays 10
5 = F1 × 0 + F2 × 10 … … … … equation2
Solving equation 2,
F2 = 510 = 0.5
Solving equation 1,
10% bond = 0.5 × price of zero bond + 0.5 × price of 20% bond
= 0.5 × 97.25 + 0.5 × 114.5 = $105.88
Note: You should assume the prices are given as per $100 face value
193
© 2014-2020 AnalystPrep.
Reading 54: Interest Rates
value.
Define spot rate and compute spot rates given discount factors.
Interpret the forward rate and compute forward rates given spot rates.
Define the par rate and describe the equation for the par rate of a bond.
Assess the impact of maturity on the price of a bond and the returns generated by
bonds.
Define the “flattening” and “steepening” of rate curves and describe a trade to reflect
Describe a swap transaction and explain how a swap market defines par rates.
Describe overnight indexed swap (OIS) and distinguish OIS rates from LIBOR swap
rates.
Besides annual interest payments, most securities on today's market have much shorter accrual
periods. For Example, interest may be payable monthly, quarterly (every three months), or semi-
annually resulting in different present values or future values depending on the frequency of
compounding employed. Here's how to calculate the PV and the FV of an investment with
The most important thing is to ensure that the interest rate used corresponds to the number of
194
© 2014-2020 AnalystPrep.
compounding periods present per year.
Future value
m ×n
rq
FV = PV{(1 + )}
m
Where:
Present value
Suppose you make PV the subject of the above formula. You should find that
-m∗n
rq
PV = FV{(1 + )}
m
Suppose you wish to have $20,000 in your savings account at the end of the next four years.
Assume that the account offers a return of 10 percent per year, compounded monthly. How much
would you need to invest now to have the specified amount after the three years?
Solution
-m∗n
rq
PV = FV{(1 + )}
m
195
© 2014-2020 AnalystPrep.
rq=0.10
m=12
n=4 years
−12×4
0.1
PV = 20000{1 + } = $13 , 429
12
Therefore, you will need to invest at least $13,429 in your account to ensure that you have
We can do this with the financial calculator with the following inputs:
I 10
N = 12 ∗ 4 = 48; = = 0.833; PMT = 0; FV = 20, 000;
Y 12
CPT ⇒ PV = −13, 429
Here, we can see that the PV on the financial calculator is a negative value since it’s a cash
We can also rearrange the future value formula to obtain the holding period return (HPR) as
follows:
1
⎡ FV mn ⎤
r q = m ⎢( ) − 1⎥
⎣ PV ⎦
If we have a series of interest swap rates, it is possible to derive discount factors. The notional
amount, which is technically never exchanged between counterparties, determines the size of
If we could exchange the notional amount, the fixed leg of the swap would resemble a fixed
196
© 2014-2020 AnalystPrep.
coupon-paying bond, with fixed leg payments acting like semi-annual, fixed coupons, and the
notional amount acting like the principal payment. Floating rate payments would act like coupon
We denote the discount factor for t-years as d(t). The methodology used to come up with discount
factors when dealing with interest rate swaps is similar to that used to find discount factors
Compute the discount factors for maturities ranging from six months to two years, given a
The four discount factors d(0.5), d(1.0), d(1.5) and d(2.0) can be calculated as follows:
0.75
(100 + )d (0.5) = 100
2
100
d (0.5) = = 0.9963
100.375
…… … …… … …… … …… …
0.85 0.85
d (0.5) + (100 + ) d (1.0) = 100
2 2
0.425 × 0.9963 + 100.425 d(1.0) = 100
100 − 0.4234
d (1.0) + = 0.9916
100.425
…… … …… … …… … …… …
197
© 2014-2020 AnalystPrep.
0.98 0.98 0.98
d (0.5) + d (1.0) + (100 + ) d (1.5) = 100
2 2 2
0.49 × 0.9963 + 0.49 × 0.9916 + 100.49 d (1.5) = 100
100 − 0.4882 − 0.4859
d (1.5) = = 0.9854
100.49
…… … …… … …… … …… …
A t-period spot rate is a yield to maturity on a zero-coupon bond that matures in t years,
Spot rates and discount factors are related as shown in the following formula, assuming semi-
annual coupons:
1
⎡ 1 2t ⎤
z (t) = 2 ⎢( ) − 1⎥
⎣ d (t) ⎦
A spot interest rate gives you the price of a financial contract on the spot date. The spot date is
the day when the funds involved in a business transaction are transferred between the parties
involved. It could be two days after a trade, or even on the same day, we complete the deal. A
spot rate of 5% is the agreed-upon market price of the transaction based on current buyer and
seller action.
198
© 2014-2020 AnalystPrep.
In theory, forward rates are prices of financial transactions that might take place at some future
point. The spot rate tells you "how much it would cost to execute a financial transaction today".
The forward rate, on the other hand, tells you “how much would it cost to execute a financial
We agree on spot and forward rates in the present. The only difference comes in the timing of
execution.
Compute the six-month forward rate in six months, given the following spot rates:
Z (0.5) =1.6%
Z (1.0) =2.2%
Solution
The six-month forward rate, f(1.0), on an investment that matures in one year, must solve the
following equation:
199
© 2014-2020 AnalystPrep.
2 1
0.022 0.016 f(1.0)1
(1 + ) = (1 + ) × 1+
2 2 2
1
f(1.0)
1.0221 = 1.008 × (1 + )
2
f(1.0)
1.01399 − 1 =
2
f (1.0) = 0.02797 = 2.8%
Par Rate
The par rate is the rate at which the present value of a bond equals its par value. It’s the rate
you’d use to discount of all a bond’s cash flows so that the price of the bond is 100 (par). For a
100-par value, the two-year bond that pays semi-annual coupons, and we can easily calculate the
2-year par rate provided we have the discount factor for each period
Par rate
= [d (0.5) + d (1.0) + d (1.5) + d (2.0)] + 100 d (2.0) = 100
2
CT 2T t
∑ d( ) + d (T) = 1
2 t=1 2
The sum of the discount factors is called the annuity factor, A 1, and is given by:
2T t
∑ d( )
t=1 2
Therefore,
CT
× AT + d(T ) =
2
200
© 2014-2020 AnalystPrep.
In general terms, bond prices will tend to increase with maturity whenever the coupon rate is
above the forward rate throughout maturity extension. The opposite holds: bond prices will tend
to decrease with maturity whenever the coupon rate is below the forward rate for a maturity
extension.
To help you understand just how this happens, assume we have two investors with the
opportunity to invest in either STRIPS or a five-year bond. The investor who opts for the 5-year
bond (which utilizes forward rates) will have a simple task: they will invest at the onset and then
wait to receive regularly scheduled coupon payments, plus the principal amount at maturity. The
investor who opts for STRIPS (which utilizes spot rates) will roll them over as they mature
throughout the five years. Rolling over implies that when one STRIP expires, the investor will use
the proceeds to invest in the next six-month contract, and so on for five years.
In market conditions where short-term rates are above the forward rates utilized by bond prices,
the investor who rolls over the STRIPS will tend to outperform the investor in the 5-year bond.
The opposite is exact: In market conditions where short-term rates are below the forward rates
utilized by bond prices, the investors who roll over the STRIPS will tend to underperform the
If the term structure is constant, and that all spot rates are equal, then all par rates
201
© 2014-2020 AnalystPrep.
and all forward rates equal the spot rate.
If the term structure is increasing, the par rate for a specific maturity is below the spot
If the term structure is decreasing, the par rate for a specific maturity is above the spot
If the term structure is increasing, forward rates for a period starting at time T are
A yield curve represents the yield of each bond along a maturity spectrum that's plotted on a
graph. The most and widely accepted yield curves pit the three-month versus two-year T-bonds
or the five-year versus ten-year T-bonds. On occasion, we may use the Federal Funds Rate versus
The yield curve typically slopes upwards indicating that the rate of interest on long-term bonds
is higher than the rate on short-term bonds which reflects the investors' demands to be
compensated for taking on more risk by investing long-term. Such a curve is said to be normal.
202
© 2014-2020 AnalystPrep.
Other than a standard yield curve, we could have a flattening yield curve or a steepening yield
curve.
A flat yield curve indicates that little difference exists between short-term and long-term rates
for similarly rated bonds. It may manifest as a result of long-term interest rates falling more than
short-term interest rates or short-term rates increasing more than long-term rates.
203
© 2014-2020 AnalystPrep.
A flattening curve reflects the expectations of investors about the macroeconomic outlook. It may
decrease; or
An impending increase in the Federal Funds Rate. (An increase in the FFR could cause
an increase in short term rates while long-term rates remain relatively stable, causing
A trader who anticipates a flattening of the yield curve can buy a long-term rate and
sell short a short-term rate because they expect bond prices to rise in the long-term.
204
© 2014-2020 AnalystPrep.
A steepening yield curve indicates a widening gap between the yields on short-term bonds and
long-term bonds. A steepening curve could occur when long-term rates rise faster than short-
term rates. Sometimes, short-term rates can also show some defiance by decreasing even as
For example, assume that a two-year note was at 2.3% on July 15, and the 10-year was at 3.3%.
By August 30, the two-year note could have risen to 2.38% and the 10-year to 3.5%. The
difference would effectively go from 1 percentage point to 1.12 percentage points, resulting in a
A steepening curve reflects the expectations of investors about the macroeconomic outlook. It
205
© 2014-2020 AnalystPrep.
an increase in expected inflation, causing the premium loaded on long-term rates by
investors to increase; or
A trader who anticipates a steepening of the yield curve can sell short a long-term rate and buy a
short-term rate because they expect bond prices to fall in the long-term (bond prices fall as rates
increase).
LIBOR (London Interbank Offered Rate) is a rate of interest at which banks across the world
borrow from each other. LIBOR is derived from average estimates of different borrowing rates
provided by different AA-rated banks globally. These rates are quoted for different currencies,
and the borrowing periods are between one day and one year.
Although we use LIBOR rates as a benchmark borrowing rate, we should note that the estimates
used to come up with the average borrowing rate are subject to manipulation by the different
banks. In efforts to curb these issues, the market has come up with other benchmarks, and an
Overnight rate
While the borrowing period for LIBOR ranges from one day to one year, overnight rates apply for
a day only. Banks with insufficient funds may borrow from those banks with surplus funds to
Brokers always accompany these transactions, and different countries have systems that monitor
the activities of these brokers. For example, we have the Federal Reserve in the US, the Sterling
Overnight Index Average (SONIA) in the UK, and the Euro Overnight Index Average (EONIA) in
the Eurozone.
Swaps
206
© 2014-2020 AnalystPrep.
A swap is a contract in which two parties agree to exchange a sequence of cash flows for a given
period. A contract in which one party agrees to swap a fixed rate for Libor is the most frequently
A swap has two legs: a fixed leg in which the interest rate is fixed and a floating one in which the
rates are varying, and the interest is compound. Each day, banks will come up with a new
Overnight Index; hence the rates keep on fluctuating, i.e., swaps are associated with a fixed
An OIS is a swap in which an investment compounded daily for three months is exchanged with a
(1 + r1 d 1) (1 + r2 2) … . (1 + ri di ) … (1 + rn dn ) − 1
Where:
While Libor swaps define par bond that can be used to determine the Libor term structure, OIS
rates describe par bonds that can be used to determine an OIS term structure.
207
© 2014-2020 AnalystPrep.
Question 1
Given the following spot rates, compute the 6-month forward rate in 1 year.
z (1.0) = 3.25%
z (1.5) = 3.60%
A. 4.2%
B. 1.5%
C. 4.3%
D. 5.1%
Solution:
The 6-month forward rate on an investment that matures in 1.5 years must solve the
following equation:
1
0.036 3 0.03252 f (1.5)
(1 + ) = (1 + ) × (1 + )
2 2 2
1
f (1.5)
1.05498 = 1.03276 × (1 + )
2
f (1.5)
1.02152 − 1 =
2
f (1.5) = 0.04303 = 4.3%
Question 2
Consider a bond with par value of USD 1,000 and maturity in four years. The bond
208
© 2014-2020 AnalystPrep.
n − year S pot rate
1 − year 5%
2 − year 6%
3 − year 7%
4 − year 8%
A. $1,160
B. $500
C. $870.78
D. $850
The value of the bond is the present value of all future cash flows (coupons plus
40 40 40 40 + 1000
PV = + + +
1 2 3
(1 + 0.05) (1 + 0.06) (1 + 0.07) (1 + 0.08)4
= 38.10 + 35.60 + 32.65 + 764.43 = $870.78
209
© 2014-2020 AnalystPrep.
Reading 55: Bond Yields and Return Calculations
Distinguish between gross, and net realized returns and calculate the realized return
Define and interpret the spread of a bond and explain how to derive a spread from a
Define the coupon effect and explain the relationship between the coupon rate, YTM,
Explain the decomposition of P&L for a bond into separate factors, including carry roll-
Explain the following four common assumptions in carry roll-down scenarios: realized
short-term rates; and calculate carry roll down under these assumptions.
The gross realized return on investment has two components: Any increase in the price of the
asset plus income received while holding the investment. When dealing with bonds,
210
© 2014-2020 AnalystPrep.
Example: A bond’s gross realized return over six months
What is the gross realized return for a bond that is currently selling for $1,060 if it was
purchased exactly six-months ago for $1,000 and paid a $20 coupon today?
Solution
Now let's assume the investor financed the purchase of the bond by borrowing cash to invest at a
rate of 1% for six months. The gross realized return less per period financing costs gives the net
realized return.
When calculating the gross/net realized return for multiple periods, it's essential to consider
whether coupons received are reinvested. If the coupons are reinvested, they will grow at the
reinvestment rate from the time they are received up to the end of the period. The risk that
reinvested cash flows will increase by a rate that's lower than the expected rate is known as
reinvestment risk.
Bond Spread
The spread of a bond is the difference between its market price and the price computed
according to spot rates or forward rates – the term structure of interest rates.
As a relative measure, a bond's spread helps us determine whether the bond is trading cheap or
rich relative to the yield curve. We incorporate spread in the bond price formula as follows:
Recall that given a 2-year bond with a face value of P, paying annual coupons each of amount C,
211
© 2014-2020 AnalystPrep.
its price is given by:
C C+P
Market bond price = +
(1 + f (1.0)) (1 + f (1.0)) × (1 + f (2.0))
To incorporate the spread s, we assume that the bond is trading at a premium or discount to this
computed price. We can find the bond’s spread using the following formula:
C C+P
= +
(1 + f (1.0) + s) (1 + f (1.0) + s) × (1 + f (2.0) + s)
Yield to Maturity
Yield to maturity (YTM) of fixed income security is the total return anticipated if we hold the
security until it matures. Yield to maturity is considered a long-term bond yield, but we express
as an annual rate. In other words, it's the security's internal rate of return as long as the investor
holds it up to maturity. To compute a bond’s yield to maturity, we use the following formula:
C1 C2 C3 CN
p= + + ⋯+
1 2 3
(1 + y) (1 + y) (1 + y) (1 + y)N
Where:
When cash flows are received multiple times every year, we can slightly modify the above
C1 C2 C3 Cn
p= + + ⋯+
(1 + y) 1
(1 + y) 2
(1 + y) 3
(1 + y)n
212
© 2014-2020 AnalystPrep.
Where:
Provided all cash flows received are reinvested at the YTM; the yield to maturity is equal to the
For zero-coupon bonds that are not accompanied by recurring coupon payments, the yield to
maturity is equal to the normal rate of return of the bond. We use the formula below to
1
Face value year to maturity
Yield to maturity = ( ) −1
Current price of bond
Exam tip: The yield to maturity assumes cash flows will be reinvested at the YTM and assumes
An annuity is a series of annual payments of PMT until the final time T. The value of an ordinary
1 − (1 + r)−T
PVannuity = PMT
r
Where:
r=discount rate
Perpetuity is a type of annuity whose cash flows continue for an infinite amount of time. The
PMT
213
© 2014-2020 AnalystPrep.
PMT
PVperpetuity =
r
We can use both the spot rate and the yield to maturity to determine the fair market price of a
bond. However, while the yield to maturity is constant, the spot rate varies from one period to
The spot rate is a more accurate measure of the fair market price when interest rates are
Given a bond's cash flows and the applicable spot rates, you can easily calculate the price of a
bond. You can then determine the bond's YTM by equating the price to the present values of cash
The coupon effect describes the fact that reasonably priced bonds of the same maturity but
different coupons have different yields to maturity, which implies that yield is not a reliable
measure of relative value. Even if fixed-income security A has a higher yield than fixed security
It also follows that if two bonds have identical features save for the coupon, the bond with the
smaller coupon is more sensitive to interest rate changes. In other words, given a change in
yield, the lower coupon bond will experience a higher percentage change in price compared to
the bond with larger coupons. The most sensitive bonds are zero-coupon bonds, which do not
Exam tips:
The lower the coupon rate, the higher the interest-rate risk. The greater the coupon
214
© 2014-2020 AnalystPrep.
If coupon rate > YTM, the bond will sell for more than par value or at a premium.
If the coupon rate < YTM, the bond will sell for less than par value, or at a discount.
If coupon rate= YTM, the bond will sell for par value.
Over time, the price of premium bonds will gradually fall until they trade at par value at maturity.
Similarly, the price of discount bonds will gradually rise to par value as maturity gets closer. This
We generate the bond's profitability or loss through price appreciation and explicit cash flows.
changes that emanate from a deviation of term structure from the original
II. Rate changes component: The rate changes component accounts for price
changes due to interest rate movements from an expected term structure to the
III. Spread change component: As the words suggest, the spread change component
215
© 2014-2020 AnalystPrep.
accounts for price changes emanating from changes in the bond’s spread from
time t to t+1.
I. Realized forwards: The return to a bond held to maturity is the same as rolling the
investment one period at a time at the forward rates. However, in reality, some forwards
216
© 2014-2020 AnalystPrep.
Question 1
A bond currently selling for $1,060 was purchased exactly 12 months ago for $1,000
and paid a $20 coupon six months ago. Today, the bond paid a $20 coupon. The
coupon received six months ago was reinvested at an annual rate of 2%. Given that
the purchase price was entirely financed at a yearly rate of 1%, the net realized
A. 8.9%
B. 9.0%
C. 10.0%
D. 9.5%
Solution:
BVt + C t − BV t−1
Netrealised return t−1,t = − financing costs
BVt−1
2%
1, 060 + 20 + 20 (1 + ) − 1000
2
= − 1.0%
1000
= 10.02% − 1% = 8.9%
Question 2
On Jan 1 2017, Commercial Bank of India issued a six-year bond paying an annual
interest rates remain unchanged. Holding all other factors constant, and assuming a
flat term structure of interest rates, how was the bond's price affected? The price:
A. Remained constant
B. Decreased
217
© 2014-2020 AnalystPrep.
C. Increased
From the data given, it's clear that the bond's coupon is higher than the yield. As
such, the bond must have traded at a premium – implying the price must have been
higher than the face value. Provided the yield doesn't change; a bond's price will
always converge to its face value. Since the price starts higher, it must decrease. This
218
© 2014-2020 AnalystPrep.
Reading 56: Applying Duration, Convexity, and DV01
Describe a one-factor interest rate model and identify common examples of interest
rate factors.
Define and compute the DV01 of fixed income security, given a change in yield and the
Calculate the face amount of bonds required to hedge an option position given the
DV01 of each.
Define, compute, and interpret the effective duration of fixed income security given a
Compare and contrast DV01 and effective duration as measures of price sensitivity.
Define, compute, and interpret the convexity of fixed income security, given a change in
Explain the process of calculating the effective duration and convexity of a portfolio of
Construct a barbell portfolio to match the cost and duration of a given bullet
investment, and explain the advantages and disadvantages of a bullet versus barbell
portfolios.
This chapter discusses one-factor risk metrics, which include DV01, duration, and convexity, as
used in the analysis of fixed-income portfolios. We consider these measures to quantify the
consequence of a parallel shift in the interest rate term structure: DV01 and duration consider a
small parallel shift in the term structure while convexity extends the duration to accommodate
larger parallel shifts. Hedging can, in accordance with these risk metrics, be efficient for a
219
© 2014-2020 AnalystPrep.
One-factor Interests Rate Models
The one-factor assumption states that when the rates are driven by one factor, the change of one
interest rate can be used to determine the change in all other interest rates over a short period
of time. For instance, a one-factor model assumes that all interest rates changes by the same
amount. As such, the shape of the term structure never changes. That is, if a one-year spot rate
increases by two basis points, all other spot rates increase by two basis points. DV01, duration,
It is worth noting that shifts in the term structure are not always parallel. For instance, a one-
factor model might predict that if one-year spot rate increases by five basis points over a short
period, then the two year increases by three basis points and the ten-year rate increases by one
basis point.
220
© 2014-2020 AnalystPrep.
DVO1
DV01 reflects the effect of a one-basis movement in interest rates on the value of a portfolio.
ΔP
DV01 = −
Δr
Where
Δr = the size of a parallel shift in the interest rate term structure measured in basis points
Note that for a long position in bonds, the DV01 is positive due to a negative correlation between
221
© 2014-2020 AnalystPrep.
DV01 is defined in three different ways:
i. Year-based DV01: defined as the change in the price from a one-basis point increase in
ii. DVDZ or DPDZ: defined as the change in price from a one-point increase in all spot
rates.
iii. DVDF or DPDF: defined as a change in price from a one-basis point increase in the
forward rate.
A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%
Solution
i. The price of a bond with no spread is USD 101,003.01 calculated using the formula:
C 2T ⎛ 1 ⎞ 100, 000
P= ∑ ⎜⎜⎜ ⎟⎟ +
2 i=1 1 + y ⎟ y
2T
⎝ ⎠ (1 + )
2 2
In this case,
222
© 2014-2020 AnalystPrep.
c = 100, 000 × 0.08 = 8, 000, T = 2
⎛ ⎞
8000⎜ 1 1 1 1 ⎟⎟ 100, 000
⇒p= ⎜⎜ + + + ⎟⎟ +
⎜ 0.06 2 3
2 ⎜
1 + 0.065 0.07 0.075 4 ⎟ 0.075 4
⎝ 2 (1 + ) (1 + ) (1 + ) ⎠ (1 + )
2 2 2 2
1 1 1 1 100, 000
= 4000 ( + 2
+ 3
+ )4 +
1.03 (1.0325) (1.035) (1.0375) (1.0375)4
= 101, 003.01
If the spot rates are each increased by 5 basis points (0.05%) so that the six-month spot rate is
6.05%, the one-year spot rate is 6.55%, and so on, the price of the bond is USD 100,911.18,
calculated as:
We know that:
Note that the rise of the spot rates by 5 basis points decreases the bond price by 91.83
(=100,911.18 − 101,003.01), and the DV01, in this case, measure the decline of a bond price for
ii. Assume now that the spot rates decrease by 5 basis points so that the six-month spot rate is
5.95% (=6 − 0.05), the one-year spot rate is 6.45% (=6.5 − 0.05) and so on. Under decreased
1 1 1 1 100, 000
= 4, 000( + + + ) + = 101, 094.96
1.02975 (1.02975)2 (1.03475)3 (1.03725)4 (1.03725)4
And thus,
223
© 2014-2020 AnalystPrep.
ΔP 101, 094.96 − 101, 003.01
DV01 = − =−
Δr −5
91.95
= = 18.39
−5
As such, an increase of spot rates by five basis points causes the price of the bond to increases
by 91.95 (=101,094.96 − 101,003.01), and thus DV01, in this case, measures the increase of
It is worth noting that the DV01 for the decrease and increases of the basis points are slightly
different because the bond price is not a linear function of interest rates. We estimate the DV01
18.366 + 18.39
DV01 = = 18.38
2
Assume that now bond yield increases by one basis point. Consider the spot rates, as in the
example above.
A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%
224
© 2014-2020 AnalystPrep.
ii. Decreases by the 5 basis points?
Solution
Recall that, using the spot rates, we calculated the bond price to be 101,003.01. To find the yield
Using a financial calculator with the variables N=4, PMT=4000, PV=−101,003.01, we get:
y
= 3.7255% ⇒ y = 7.45%
2
i. Now, if the bond yield increases by 5 basis point to 7.50% (=7.45 + 0.05), the price of the bond
is 100,912.846 (using the financial calculator) and thus for an increase of 5 basis points in bond
yield, the bond price decreases by 90.164 (=100,912.85 − 101,003.01). We can, therefore,
ii. Similarly, if the bond yield decreases by 5 basis points to 7.40% (=7.45 − 0.05), the bond price
is 101,096.71. A decrease of bond yield by 5 basis points increases the bond price by 93.70
Similarly, if the bond yield decreases by 5 basis points to 7.40% (=7.45 − 0.05), the bond price is
101,096.71. A decrease of bond yield by 5 basis points increases the bond price by 93.702
18.0328 + 18.632
≈ 18.33
2
In the cases of the forward rates, the analogy is similar to that of the spot rates. However, this
chapter will primarily emphasize on DV01 computed from a one-basis-point parallel shift in the
In any position that depends on the interest, DV01 can be computed efficiently. DV01 can,
therefore, be used to hedge a position. For example, assume that a bank has a position whose
DV01 is −40. By the definition of DV01, the banks will gain from their position if interest rates
increases and will undoubtedly lose value if interest rates decrease. More specifically, if all the
interest increases by the one-basis point, the value of the bank’s position increases by USD 40.
On the contrary, if all interest rates decrease by one basis point, the value of the bank’s position
Now assume that this bank wants to hedge its position with a coupon bond that pays an annual
coupon rate of 8% payable semiannually, has a face value of USD 100,000 and bond yield of
7.50%. If we can determine a position in the coupon bond that is exactly 40, the bank’s position
A hedge ratio determines the amount of par of the hedge position that needs to be bought or sold
for every $1 par value of the original position. The goal of hedging is to lock in the value of a
position even in the face of small changes in yield. The hedge ratio is given by:
DV01
226
© 2014-2020 AnalystPrep.
DV01Initial Position
HR =
DV01Hedge Position
Assume that DV01 of the coupon bond is 18.33. We use the hedge ratio and the bond’s face value
to increase the DV01 of the bond to 40. In this case, we need to increase the value of the position
by:
40
100, 000 × = $218, 221.50
18.33
Thus, adding the USD 218,221.50 of coupon bond’s position to the bank’s portfolio protects
against small changes in the term structure. In other words, a downward (upward) movement of
the term structure will result in profit (loss), which will be offset by the gain (loss) from the
bond’s position.
Effective duration measures the percentage change in the price of a bond (or other instruments)
caused by small changes in all rates. Note that effective duration is different from the DV01
because DV01 measures actual price changes against small changes in all rates.
ΔP
P ΔP
D=− =
Δr P × Δr
ΔP = −D × P × Δr
When the change in all rates is measured basis points, the effective duration is equivalent to
Consider the DV01 example on the spot rates where we had calculated the price of the bond USD
227
© 2014-2020 AnalystPrep.
101,003.01 and the DV01 of 18.38 so that the duration is:
18.33
= 0.000182 = 0.0182%
101003.01
Effective duration gives the proportional change in the price of an instrument corresponding to
Typically, the effective duration is stated as a percentage change in the price of an instrument
for a 100-basis-change in all rates by multiplying the effect of one-basis-point change by 100.
Therefore, our example above would be stated as 1.82% per 100 basis points.
However, for the sake of clarity, the duration in this chapter will be reported as a decimal. In the
decimal reporting system, one basis point is equivalent to 0.0001, and thus, we measure duration
per 10,000 basis points. Therefore, the duration calculated above will be:
A callable bond is one that an issuing party has the right to purchase back the bond at a pre-
determined price at a particular time in the future before the maturity period of the bond. The
call feature of the bond should not be ignored as it reduces the duration of the bond.
A practical approach to address the callable feature of a bond while calculating duration is
outlined as follows:
Compute the bond value if the all interest rates increase by one basis point while
incorporating the effect of the one-basis point increase on the bond’s callable
Compute the effective duration from the percentage change in the price.
228
© 2014-2020 AnalystPrep.
A puttable bond is one that a holder has a right to ask for early repayment. Calculating the
effective duration of a puttable bond is similar to that of a callable bond. That is, the probability
of the put options increases with the increase of the interest rates.
DV01 is useful in measuring the effect of all rate changes on the value of a position. DV01 is also
Effective duration is appropriate in measuring the effect of rate changes on the value of a
DV01 increases with an increase in the position size, but effective duration does not. In other
words, if the value of a position is doubled, DV01 doubles, but effective duration does not.
Effective Convexity
Convexity measures the sensitivity of duration measure to movement in the interest rates.
1 P+ + P − − 2P
C= [ ]
P (Δr)2
Where:
A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%
229
© 2014-2020 AnalystPrep.
Maturity Period (Years) Spot Rate (%)
0.5 6.0
1.0 6.5
1.5 7.0
2.0 7.5
What is the value of the convexity if all spot rates change by 5 basis points?
Solution
As computed earlier, the price of a bond with no interest spread is USD 101,003.01 calculated
as:
⎛ ⎞
8000⎜ 1 1 1 1 ⎟⎟ 100, 000
p= ⎜⎜ + + + ⎟⎟ +
2 ⎜⎜ 0.06 0.065 2 0.07 3 0.075 4 ⎟ 0.075 4
1 + (1 + ) (1 + ) (1 + ) (1 + )
⎝ 2 ⎠
2 2 2 2
= 101, 003.01
If the spot rates are each increased by 5 basis points (0.05%) so that the six-month spot rate is
6.05%, the one-year spot rate is 6.55%, and so on, the price of the bond is USD 100,911.18
calculated as:
1 1 1 1 100, 000
= 4000 ( + +
2
+
3
)4 + = 101, 094.96
1.02975 (1.03225) (1.03475) (1.03725) (1.03725)4
So,
1 P+ + P− − 2P
C= [ ]
P (Δr)2
1 100, 911.18 + 101, 094.96 − 2 × 101, 003.01
= [ ]
101, 003.01 (0.0005)2
= 0.4752
230
© 2014-2020 AnalystPrep.
Recall that the effective duration can be written
ΔP = −D × P × Δr
The duration of a bond is a linear relationship between the bond price and interest rates where,
as interest rates increase, bond price decreases. It is a good measure of sensitivity to interest
rates for small/sudden changes in yield. However, for much larger changes in yield, the duration
is not so desirable since the relationship between price and interest rates tends to be non-linear.
ΔP
The relationship between and Δr is non-linear. However, effective duration approximates this
P
ΔP
as a linear relationship because it is the gradient of the curve linking and Δr. As seen on the
P
graph, the linear approximation by effective duration becomes less efficient as the rate changes
231
© 2014-2020 AnalystPrep.
become large.
The combination of the duration and convexity is a quadratic function which provides a better
1
ΔP = −DPΔr + CP(Δr)2
2
By combining the duration and the convexity, we can reach a convenient estimate, even for large
A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%
Note that we had calculated the DV01 for the 5 basis point movement of all the spot rates to be
18.38, the price of the bond with no spread is USD 101,003.01, the duration to be 1.82 and
convexity to be 0.4752.
i. Estimate actual bond price change due to an increase in all the spot rates by 30 basis
points;
iii. Estimate bond price change using a combination of the effective duration.
232
© 2014-2020 AnalystPrep.
Solutions
Because all the spot rates increase by 30 basis points (0.3%), the six-month spot rate is 6.3%, the
one-year spot rate is 6.8%, and so on. So the price due to new rates is given by the formula:
i
⎛ 1 ⎞
C 2T 100, 000
P= ∑ ⎜⎜⎜ ⎟⎟ +
⎟
2 i=1 1 + y y 2T
⎝ 2⎠ (1 + )
2
In this case,
⎛ ⎞
8000⎜ 1 1 1 1 ⎟⎟ 100, 000
= ⎜⎜ + + + ⎟⎟ +
2 ⎜⎜ 1 + 0.063 0.068 2 0.073 3 0.078 4 ⎟ 0.078 4
⎝ 2 (1 + ) (1 + ) (1 + ) ⎠ (1 + )
2 2 2 2
1 1 1 1 100, 000
= 4000 ( + + + ) +
1.0315 (1.034)2 (1.0365)3 (1.039)4 (1.039)4
= 100, 453.63
⇒ ΔP = 100, 453.63 − 101, 003.01 = −549.38
ΔP = −DP Δr
= −1.82 × 101, 003.01 × 0.003 = −551.48
The estimated price change provided by the duration is a good approximation of the price
233
© 2014-2020 AnalystPrep.
change, but we can improve it by combining duration and the convexity.
The price change estimation due to a combination of duration and convexity is given by:
1 2
ΔP = −DPΔr + CP(Δr)
2
1 2
= −1.82 × 101, 003.01 × 0.003 + × 0.4752 × 101,003.01 × (0.003)
2
= −551.48 + 0.21598 = −551.26
Hedging using effective duration analogous to that of DV01. Now denote the duration of
investment by DV, and its investment value by V. On the other hand, denote the effective duration
ΔV = −V DV Δr
and
ΔV = −V DB Δr
Where Δr is the size of a small parallel shift in the term structure. These small parallel shifts are
−VDVΔr − PDBΔr = 0
⇒ Δr (−VDV − PDB ) = 0
∴ −VDV − PDB = 0
VDV
P=
DB
234
© 2014-2020 AnalystPrep.
A bank has a position of USD 12 million with an effective duration of 5 and a bond with an
effective duration of 6. How will the bank hedge against its position?
Solution
VDV
P=−
DB
5 × 12
=− = −10
6
Therefore, the bank should short bonds worth 10 million to hedge against its position. This is
true because:
ΔV = −12 × 5 × Δr = −60Δr
and
ΔP = −10 × 6 × Δr = 60Δr
⇒ ΔV + ΔP = −60Δr + 60Δr = 0
Hedging based on both duration and convexity is a complex undertaking where we are trying to
make both effective duration and convexity. We need two bonds. Define:
235
© 2014-2020 AnalystPrep.
Also, define
Now by the definition of approximating the price change using a combination of the duration and
convexity we have:
1
ΔV = −VDV Δr + VC V(Δr)2
2
1
ΔP 1 = −P 1D1 Δr + P1 C 1 (Δr)2
2
and
1
ΔP1 = −P2 D2 Δr + P 2 C 2 (Δr)2
2
−VDV − P1 D1 − P 2 D2 = 0
Also,
VCV − P1 C 1 − P 2 C2 = 0
A bank has a position of USD 12 million with an effective duration of 5 and a convexity of 4. The
bank wishes to hedge its position with two bonds where the first bond has an effective duration
of 6 and a convexity of 9. The second bond has a duration of 4 and a convexity of 7. How will the
236
© 2014-2020 AnalystPrep.
Solution
We know that a position is hedged against if the following equations are satisfied:
VDV − P 1 D1 − P 2D2 = 0
VC V − P 1 C1 − P 2 C2 = 0
Therefore, we have
60 − 6P 1 − 4P 2 = 0
108 − 9P 1 − 7P 2 = 0
Therefore, for the bank to hedge its position, it must take a position of USD 8.77 in the first bond
and short USD 4.15 million in the second bond. In other words, by combining these positions in
bonds, there is no duration or convexity exposure, and thus the bank is hedged against large
parallel shifts in the term structure though it will be exposed to non-parallel shifts.
Yield-Based Duration
Consider a bond whose price is P and yield is y. If the cash flows from the bond are c1 , c 2, … , cn at
n
P = ∑ C i e−yti
i=1
dp
237
© 2014-2020 AnalystPrep.
dp n
= − ∑ Ci e−yti
dy i=1
ΔP n
⇒ = − ∑ tiC i e−yti
Δy i=1
Now, the yield-based duration is defined as the proportional change in the bond price due to a
1 ΔP 1 n
D=− =− (− ∑ ti Ci e−yti)
P Δy P i =1
n Ci e−yti
⇒ D = − ∑ ti
i=1 P
C ie−yti
The expression in the formula denotes the proportion of the value of the bond received at
P
the time ti . Therefore, the duration can be computed by taking the average of the times when the
bond’s cash flows were received weighted by the bond’s value at each time. As such, the
duration will measure the time an investor must wait to receive cash flows.
Typically, the yield on the bond is given as a semi-annual compounding rate rather than
1
continuous compounding. As such, the expression for the duration is divide by y to get:
1+
2
⎛ ⎞
1 ⎟ n C i e−yti
D = ⎜⎜⎜ ⎟ ∑ ti
y ⎟ i=1 P
⎝1 + 2 ⎠
The resulting equation gives the formula for the modified duration while the expression
1
without dividing by y is termed as Macaulay Duration. Modified duration is slightly
1+ 2
different from an effective duration.
Yield-Based Convexity
Recall convexity measures the sensitivity of duration measure to movement in the interest rates.
In the calculus context, yield based convexity is the derivative if the duration. Therefore,
−yt
238
© 2014-2020 AnalystPrep.
1 n n C i e−yti
C= ∑ t2i C ie−yti =∑ t2i
P i =1 i=1 P
Intuitively, convexity can be defined as the weighted (by bond’s value) average of the squared
time to maturity. When yields are given as semi-annual compounding, the convexity expression is
1
multiplied by so that:
y 2
(1+ )
2
1 n C i e−yti
C= ∑ t2i
y 2 i=1 P
(1 + )
2
The last expression is termed as Modified convexity, and it is slightly different from the
effective convexity.
We now need to consider a group of instruments (such bonds) where we need to compute the
The DV01 of a portfolio is simply the sum of individual DV01 of instruments in the portfolio.
For instance, if a bank has four positions whose DV01s are 9, 10, 12, and 13 (in USD millions),
Portfolio duration can be calculated as the weighted sum of the individual durations. The
weight attached to each security is equal to its value as a percentage of total portfolio value.
k
Portfolio duration = ∑ wj Dj
j=1
239
© 2014-2020 AnalystPrep.
Where:
For instance, consider a portfolio consisting of four instruments with values (in USD millions) of
8, 13, 15 and 18, and respective effective durations are eight, seven, six, and four. The effective
8 13 15 18
DEff = × 8+ × 7+ × 6+
8 + 13 + 15 + 18 8 + 13 + 15 + 18 8 + 13 + 15 + 18 8 + 13 + 15 + 18
= 1.1852 + 1.6852 + 1.6667 + 1.3333 = 5.87
The effective portfolio duration is used to determine the impact of the small parallel shift of the
term structure of the interest rates. Similar computations can be used in yield-based durations.
Portfolio convexity is computed using a similar approach. It’s defined as the value-weighted
k
Portfolio convexity = ∑ w jC j
j=1
portfolio is made up of long-term bonds and the other half of short-term bonds. On the other
hand, a bullet strategy is an investment strategy where the investor buys bonds concentrated
Given the prices, coupon rates, maturities, yields, durations, and convexities of a set of bonds, it
240
© 2014-2020 AnalystPrep.
is possible to construct a barbell portfolio with the same cost and duration as the bullet portfolio.
This involves determining the proportion of each security in the Barbell that should be bought,
Assume that the term structure of the interest rates is flat at 5%, compounded semiannually.
Note that this means that the yield on all instruments is 5%.
Now, assume that an investor wants a portfolio with a duration of 7.767. The investor can either
buy a 5-year, 5% coupon (bullet investment) or build a portfolio of other two bonds to have the
If the investor wishes to construct a bond from the other two bonds, denote:
Therefore, the investor can create a bond with a duration of 7.767 by either investing all his
money in a 5-year coupon bond, or invest 63.89% of his funds into the 2-coupon bond, and
Note that the portfolios have equal duration but different convexities. The convexity of the 5-
year, 5% coupon bond is 78.90. For the portfolio consisting of 2-year, 3% coupon, and 10-year,
241
© 2014-2020 AnalystPrep.
7% coupon bonds, the convexity is given by 78.8088 (=0.6389×21.3+0.3611×180.56) which is
An arbitrage opportunity can occur if one invests an amount in the barbell portfolio and short the
same amount of the bullet portfolio. The arbitrage opportunity is possible if the movement in the
term structure is parallel (which is not always true). Typically, the bullet investment is profitable
for many non-parallel movements of the term structure as compared to the barbell investment.
Most of the models are constructed in such a way there are no arbitrage opportunities to the
investors.
Advantages:
Spreading out bond purchases ensures higher yields when rates are rising; and
The investor need not have a “war chest” at the onset since the portfolio is built
gradually.
Disadvantages:
When the yield curve flattens (short rates go up; long rates go down), the Barbell
Note: Sometimes, the bullet and Barbell have the same duration, but they will have different
convexities.
242
© 2014-2020 AnalystPrep.
Question
The price of a 3-year bond is USD 10,000. The DV01 of the bond is 50. What is the
estimated bond price change if all rates increase by five basis points?
A. 15
B. 20
C. 25
D. 40
Solution
Recall that when the change in all rates is measured in basis points, the effective
duration is equivalent to DV01 divided by the bond price. Therefore, in this case, the
50
D= = 0.005 = 50
10, 000
ΔP = −DPΔr
= −50 × 10,000 × 0.0005 = USD 25
Note: We measure the effective is decimal, i.e., per 10,000 basis points and thus:
243
© 2014-2020 AnalystPrep.
Reading 57: Modeling and Hedging Non-Parallel Term Structure Shifts
Describe and assess the major weakness attributable to single-factor approaches when
Describe the principal components analysis and explain its use in understanding term
structure movements.
Define key rate exposures and know the characteristics of key rate exposure factors
Define, calculate, and interpret key rate ‘01 and key rate duration.
Calculate the key rate exposures for a given security, and compute the appropriate
Relate key rates, partial ‘01s, and forward-bucket ‘01s, and calculate the forward-
The main weakness attributable to single-factor approaches to portfolio hedging has much to do
with the assumption that movements in the entire term structure can be exhaustively described
by one interest rate factor. In other words, the single-factor approach erroneously assumes that
all rate changes within the term structure of interest rates are driven by a single factor.
244
© 2014-2020 AnalystPrep.
From a practical point of view, rates in different regions of the term structure are not always
correlated. As an example, the single-factor approach tells us that the 6-month rate can perfectly
predict the change in the 30-year rate. This in turn informs the hedging of the 30-year bond with
a 6-month bill. Such a move is unlikely to hedge the total risk inherent in the 30-year bond.
Predicted changes in the 30-year rate based purely on changes in the 6-month rate can be quite
misleading. That’s because rates in different regions of the term structure (yield curve) are not
always correlated. The risk of such non-parallel shifts along the yield curve is known as yield
curve risk.
Key rate exposures help to describe the risk distribution along the term structure given a bond
portfolio. They help describe how to execute the perfect hedge using highly liquid benchmark
bonds. Bonds used for this purpose are normally government bonds issued in the recent past,
Partial ‘01s are used to measure and hedge the risk of portfolios of swaps or portfolios that
combine both bonds and swaps in terms of the most liquid money market and swap instruments.
Forward bucket ‘01s are also used to measure and the hedge risk of portfolios of swaps/bond
combinations, but the difference here is that instead of measuring risk based on other
comparable securities on the market, they measure risk based on changes in the shape of the
yield curve. Forward bucket ‘01s present an intuitive way to understand the yield curve risk of a
portfolio, but they are otherwise not efficient at recommending the perfect hedges to neutralize
such risks. To compute forward ‘01s, the yield curve is divided into several defined regions.
The assumption behind the key rate shift analysis is that the entire spectrum of rates can be
considered a function of a few select rates at specified points along the yield curve. Thus, to
measure risk and predict interest rate movements, a small number of key rates are used, usually
245
© 2014-2020 AnalystPrep.
those of highly liquid government bonds.
The rates most commonly used are the U.S. Treasury 2-, 5-, 10-, and 30-year par yields. As the
words suggest, a “key rate shift” occurs when any of these rates shifts by one basis point. The
key rate technique indicates that changes in each key rate will affect rates from the term of the
The key rate shift approach enables analysts to estimate changes in all rates based on a few
select rates.
By definition, a Key rate '01 (key rate DV01) is the effect of a dollar change of a one basis point
The Key rate '01 is computed using the same logic as the DV01 formula used in the single-factor
approach.
ΔBV
Key rate '01 = −
10, 000 × Δy
Where:
The change in bond value here is measured in reference to the initial bond value.
Example: Key Rate DV01s and Durations of the May 15, 2045, C-STRIP as
of May 28, 2015
In the table below, column (1) gives the initial price of a C-STRIP and its present value after
246
© 2014-2020 AnalystPrep.
application of key rate one basis point shifts.
The key rate ’01 with respect to the 10-year shift is calculated as:
A key rate of -0.024 implies that the C-STRIP increases in price by 0.024 per $100 face value for
Exam tip: Just like the DV01, a negative key rate ’01 implies an increase in value after a given
shift, relative to the initial value. A positive key rate ’01 implies a decrease in value after a given
The effective duration calculates expected changes in price for a bond or portfolio of bonds given
a basis point change in yield, but it is only valid for parallel shifts in the yield curve. The key rate
duration presents an improvement to the effective duration because it gives the expected
changes in price when the yield curve shifts in a manner that is not perfectly parallel.
1 ∂P
key rate duration = − ( )
P ∂y
247
© 2014-2020 AnalystPrep.
Thus, the key rate duration with respect to the 10-year shift is calculated as:
1 26.13885 − 26.11485
key rate duration = − ( ) ×( )
26.11485 0.01%
= −9.19
Thus,
DV01
duration =
0.0001 × bond value
−0.024
=
0.0001 × 26.11485
= −9.19
Interpreting each key rate duration in isolation can be quite difficult. That’s because, in practice,
it’s highly unlikely that a single point on the yield curve will exhibit an upwards or downwards
shift while all other points remain constant. For this reason, analysts tend to compare key rate
Exam tip: The sum of all the key rate durations along a portfolio yield curve is equal to the
Before looking at some examples, let’s try to map out a typical problem investors often find
themselves in.
Let’s say an investor holds a (long) bond position and is afraid that the bond will lose some value
in the future. We call such a position the “underlying exposure.” How exactly can the position
lose value? Suppose the position has a 5-year key rate exposure of $0.05. This implies that the
248
© 2014-2020 AnalystPrep.
position will drop in value by $0.05 if there’s a one basis point shock to the 5-year key rate. If the
bond is trading at, say, $94 per $100 face value, then the new price will be $93.95 per $100 face
value following a one basis point increase in the 5-year key rate (Remember that bond prices fall
when interest rates rise). To avoid the loss, the investor must identify and sell short another bond
Selling short is all about selling high and buying low. The investor will borrow the bond and sell
it, anticipating an increase in the 5-year key rate which will result in a decrease in the bond’s
price. The investor will still be obligated to “return” the bond to its owner but he will buy it at a
lower price and get to keep the difference, which will offset the loss incurred on the long position
(underlying exposure). This is the argument behind key rate exposure hedging.
Note that the hedging security need not have a 5-year key rate exposure of exactly $0.05. It
could have an exposure of, say, 0.045, implying that the investor will not sell short exactly $100
In key rate exposure hedging, therefore, the secret lies in determining the face amount “F” that’s
An underlying exposure (bond position) has a ten-year key-rate '01 of +$880. If this key rate
exposure can be hedged by trading a ten-year bond that itself has a 10-year KR01 of $0.0520 per
Solution
A positive key rate ’01 implies a decrease in value after a given shift, relative to the initial value.
Thus, a ten-year key-rate '01 of +$880 implies that the bond position stands to lose $880 if there
happens to be a one basis point shock to the ten-year key rate. To avoid this scenario, we must
determine the face amount F(10) of the ten-year bond that must be sold short to neutralize the
0.0520
249
© 2014-2020 AnalystPrep.
0.0520
× F (10) = 880
100
880
F (10) = = $1 , 692, 308
0.00052
Note that since the key rate 01s are reported per 100 face value, they need to be divided by 100
in the hedging equation. However, the key rate 01 of the initial bond position (underlying
The hedge trade requires us to short $1.692 million face amount of the ten-year bond so as to
neutralize the exposure to the ten-year key rate. If there’s a one basis point shock to the ten-year
key rate, the long position will lose $880, while the short position will gain approximately $880
(= 0.052/100 × 1,692,000).
Important note:
Before looking at the second example, it is important to understand exactly what key rates stand
for. When we say that, for example, the 5-year key rate changes, what we mean is that the 5-
year par rate changes; all other par rates are unchanged. It is easy to think of the 5-year key
rate as the 5-year spot rate, but it is not; it’s the par rate. Key rates are not spot rates. (Par rate
denotes the coupon rate for which the price of a bond is equal to its nominal value (or par value).
This leads us to a very important observation: a bond priced at par (i.e., purchase price = par
value = $100) only has price sensitivity to key rates at the same tenor as its maturity. For
instance, a 5-year coupon-paying par bond has zero sensitivity to a change in the 2-year key
rate. However, a 5-year premium/discount bond will have some sensitivity to the 2-year key
rate.
The reason, as we have seen above is that the 5-year par rate doesn’t change. We compute the
price of a bond by discounting all its cash flows by its YTM. If the 5-year par rate doesn’t
change, then the YTM on a 5-year par bond doesn’t change, and therefore the price of a 5-year
250
© 2014-2020 AnalystPrep.
Let’s use an example to illustrate how to pull off the perfect hedge under multi-factor hedging:
Suppose we have a 30-year option-free bond paying semiannual coupons of $5,000 in a flat rate
environment of 5% across all maturities. Using the concepts learnt in the preceding learning
outcome statements, we can compute the following key rate ‘01s and key rate durations,
For example,
The key rate ’01 with respect to the 5-year shift is calculated as:
ΔBV
Key rate '01 = −
10, 000 × Δy
145 , 050.68 − 145066.45
=− = 15.77
10 , 000 × 0.01%
DV01
duration =
0.0001 × bond value
15.77
=
0.0001 × 145066.45
= 1.09
A key rate '01 of 15.77 implies that the bond decreases in value by $15.77 for a one basis point
We can easily come up with the other key rate 01’s and key rate durations by performing similar
calculations.
Now to illustrate how hedging is carried out in this scenario, assume we have four other
251
© 2014-2020 AnalystPrep.
To illustrate how hedging is carried out based on key rates, assume we have four other different
Note: In the table above, we assume that the 2-year bond and the 30-year bond are trading at
par, in which case they are only exposed to the key rate corresponding to their maturity dates (2
years and 30 years, respectively). On the other hand, the 5-year and 10-year securities are
trading at a premium.
For the hedge to work, we must neutralize the key rate exposure at each key rate.
Let F2, F5 , F10 , and F30 be the face amounts of the bonds in the hedging portfolio to be sold.
2-year key rate exposure: Three bonds, namely the two-year, five-year, and 10-year, have an
exposure to the two-year key rate. Therefore, for the two-year key rate exposure of the hedging
portfolio to equal that of the underlying position, it must be the case that
5-year key rate exposure: Only two bonds, namely the five-year and 10-year, have an exposure
to the five-year key rate. Therefore, for the five-year key rate exposure of the hedging portfolio to
0.045 0.001
5-year key rate exposure : × F5 + × F10 = 15.77
100 100
10-year key rate exposure: Only the ten-year bond has an exposure to the ten-year key rate.
Therefore, for the ten-year key rate exposure of the hedging portfolio to equal that of the
0.1
252
© 2014-2020 AnalystPrep.
0.1
10-year key rate exposure : × F10 = 77.43
100
30-year key rate exposure: Only the ten-year bond has an exposure to the ten-year key rate.
Therefore, for the ten-year key rate exposure of the hedging portfolio to equal that of the
0.2
30-year key rate exposure : × F30 = 65.50
100
Solving equations (1) through (4) simultaneously gives the following solution for the face value of
F30 = 32,750
F10 = 77,430
F5 = 33,324
F2 = 317, 150
The investor needs to short $317,150 face amount of the 2-year security, short $33,324 face
amount of the 5-year security, short $77,430 face amount of the 10-year security, and finally
short $32,750 face amount of the 30-year security. Only then would the initial bond position be
However, such a hedge portfolio is not perfect, and the hedged position is actually only
As is the case whenever derivatives are used for hedging purposes, the quality of hedge
The hedge will work only if the par yields between key rates move as assumed (linearly).
Hedging implies more instruments and more transaction costs which may eat up the
scooped gains;
Under the key rate model, the number of key rate durations to be used and the
253
© 2014-2020 AnalystPrep.
corresponding choice of key rates remain quite arbitrary
Partial ‘01s
Key rate shifts make use of a few key rates to determine risk exposures and execute hedging
strategies. For, example, in this reading, we’ve used the 2-year, 5-year, 10-year, and 30-year par
However, when the securities involved contain swaps, we need more to assess the effect of
interest rates at more points along the yield curve. There’s a need to measure more frequently.
When swaps are taken as the benchmark for interest rates in complex portfolios, risk along the
curve is usually measured with Partial ’01s or Partial PV01s, rather than with key-rate ’01s.
Swap market participants fit a swap rate at least once every day from a set of observable par
swap rates or futures rates. Using the fitted swap rate curve, the sensitivity of a portfolio can be
It follows that by definition, partial ’01 (PV01) is the change in value of a portfolio after a
one-basis-point decline in that fitted rate and a refitting of the curve. All other fitted rates are
unchanged. With partial ‘01s, yield curve shifts are able to be fitted more precisely because we
If a curve-fitting algorithm fits the three-month London Interbank Offered Rate (LIBOR) rate and
par rates at 2-, 5-, 10-, and 30-year maturities, then, the two-year partial ’01 would be the
change in the value of a portfolio for a one-basis-point decline in the two-year par rate and
refitting of the curve, where the three-month LIBOR and the par 5-, 10-, and 30-year rates are
254
© 2014-2020 AnalystPrep.
Forward-bucket ’01s
While key rates and partial ’01s do a fantastic job expressing the exposures of a position in terms
of hedging securities, forward-bucket ’01s present a far more direct and intuitive way to convey
Forward-bucket ’01s are computed by shifting the forward rate over each of several defined
regions of the term structure on the region at a time. They, however, aren’t the quickest way to
The first step under this methodology is to subdivide the term structure into buckets. The 5 most
common buckets are 0-2 years, 2-5 years, 5- 10 years, 10-15 years, and 20-30 years. After that,
each forward-bucket ’01 is computed by shifting the forward rates in that bucket by one basis
point. In so doing, the analyst may have to shift all of a bucket’s semiannual forward rates,
The table below lists the cash flows of the fixed side of the 100 notional amount of a swap, the
current forward rates (marked “current”) as of the pricing date, and the three shifted forward
curves.
255
© 2014-2020 AnalystPrep.
Forward Rates
Term Cash flow Current 0-2 shift 2-5 shift shift-all
0.5 1.06 1.012 1.022 1.012 1.022
1 1.06 1.248 1.258 1.248 1.258
1.5 1.06 1.412 1.422 1.412 1.422
2 1.06 1.652 1.662 1.652 1.662
2.5 1.06 1.945 1.945 1.955 1.955
3 1.06 2.288 2.288 2.298 2.298
3.5 1.06 2.614 2.614 2.624 2.624
4 1.06 2.846 2.846 2.856 2.856
4.5 1.06 3.121 3.121 3.131 3.131
5 101.06 3.321 3.321 3.331 3.331
Credit: Bruce Tuckman and Angel Serrat, Fixed Income Securities: Tools for Today’s Markets,
3rd Edition
For the “0-2 Shift,” forward rates of term 0.5 to 2.0 years are shifted up by one basis
For the “2-5 Shift,” forward rates of term 2.0 to 5.0 years are shifted up by one basis
Lastly, for “Shift All,” the forward rates in the curve are shifted
The row labeled “Present Value” gives the present value of the cash flows first under the initial
forward rate curve and then under each of the shifted curves.
The forward-bucket ’01 for each shift can then be computed as the negative of the difference
The ’01 of the “Shift All” scenario is analogous to a DV01. The forward bucket analysis
decomposes this total ’01 into .0196 due to the 0-2-year part of the curve and .0276 due to the 2-
256
© 2014-2020 AnalystPrep.
5-year part of the curve.
Referring to GARP-assigned Tuckman reading, let us say we have a counterparty enters into a
euro 5x10 payer swaption with a strike of 4.044% on May 28, 2010.
This payer swaption gives the buyer the right to pay a fixed rate of 4.044% on a 10-year euro
swap in five years. The underlying is a 10-years swap for settlement on May 31, 2015.
The figure below gives the forward-bucket ‘01s of this swaption for four different buckets, along
Since the overall forward-bucket ’01 of the year swaption is negative (-0.0380), as rates rise, the
value of the option to pay a fixed rate of 4.044% in exchange for a floating rate worth par also
rises.
The figure below shows forward-bucket exposures of three different ways to hedge this payer
swaption (as of May 28, 2010) using securities presented in the previous figure:
As is apparent, the third hedge is the best option since this hedge best neutralizes risk in each of
the buckets (the lowest net position indicates when risk is best neutralized).
Credit: Bruce Tuckman and Angel Serrat, Fixed Income Securities: Tools for Today’s Markets,
3rd Edition
257
© 2014-2020 AnalystPrep.
Security/Portfolio 0-2 2-5 5-10 10-15 All
5*10 payer swaption 0.001 0.0016 −0.0218 −0.0188 −0.0380
Hedge #1:Long 44.34% of 10-year swaps 0.0086 0.0119 0.0175 0.038
Net position 0.0096 0.0135 −0.0043 −0.0188 0.000
Hedge #2:Long 46.66% of 5*10 swaps 0.0209 0.0171 0.038
Net position 0.001 0.0016 −0.0009 0.0017 0.000
Hedge #3:
Long 57.55% of 15-year swaps 0.0112 0.0153 0.022 0.0186 0.067
Short 61.55% of 15-year swaps −0.012 −0.017 −0.029
Net position 0.0002 −0.0001 0.0002 −0.0002 0.000
Although we have studied at length the term-structure of interest rates, we are yet to look at
volatility. Just like there is a term-structure for interest rates, there is also a term-structure for
volatility. In fact, the volatility term structure typically slopes downwards when plotted against
maturity. This implies that the shorter the maturity of the par-rate, the more volatile it tends to
be. The 10-year par rate, for example, is usually more volatile than the 30-year par rate.
In general, portfolios are exposed to interest rates all along the curve but changes in these rates
are not perfectly correlated. How can we go about estimating volatilities for the key-rates?
Step 1: Estimate the volatility for each key rate as well as the correlation for each pair of key
rates.
ΔP = KR011 × ΔC 1 + KR01 2 × ΔC 2
258
© 2014-2020 AnalystPrep.
Furthermore, let σP2 , σ12 and σ22 denote the variances of the portfolio and of the key rates and let
ρ denote the correlation of the key rates. By applying the usual formula for finding the variance
Note that this methodology can be applied equally well to partial ’01s or forward-bucket ’01s.
259
© 2014-2020 AnalystPrep.
Questions
Question 1
recommends portfolio hedging using the key rates of 2-year, 5-year, 7-year, and 20-
year exposures. According to the manager, the 2-year rate has increased by 10 basis
The key rate technique indicates that changes in each key rate will affect rates from
the term of the previous key rate to the term of the subsequent key rate. In this case,
the 2-year key rate will affect all rates from 0 to 5 years; the 5-year key rate affects
all rates from 2 to 7 years; the 7-year key rate affects all rates from 5 to 20 years; and
the 20-year key rate affects all rates from 7 years to the end of the curve.
Question 2
a flat rate environment of 5% across all maturities. The following table provides the
initial price of the bond and its present value after application of a one basis point
260
© 2014-2020 AnalystPrep.
shift in four key rates:
Suppose further that there are four other different bonds with the following key rate
exposures:
If we wish to fully hedge our initial position using these four securities, determine the
face amount of the 5-year security we need to short (assume that the 2-year bond and
A. 27,640
B. 10,150
C. 28,764
D. 30,000
If we assume that the 2-year bond and the 30-year bond are trading at par, they are
only exposed to the key rate corresponding to their maturity dates (2 years and 30
years, respectively). The face amount we need for each security is given by Fi.
⋯ ⋯ ⋯⋯ ⋯ ⋯⋯ ⋯ ⋯⋯
262
© 2014-2020 AnalystPrep.
Reading 58: Binomial Trees
Calculate the value of an American and a European call or put option using a one-step
Describe how the value calculated using a binomial model converges as time periods
are added.
Explain how the binomial model can be altered to price options on stocks with
The binomial option pricing model is a simple approximation of returns which, upon refining,
converges to the analytic pricing formula for vanilla options. The model is also useful for valuing
P S0 u
╱
S0
╲
1− P S0 d
263
© 2014-2020 AnalystPrep.
u=The factor by which the price rises
Over a small time interval Δt, the price today to one of only two potential future values: S0 u , and
S0 d.
Risk-neutral Valuation
The following formula are used to price options in the binomial model:
1 1
D=size of the down move factor= e−σ√ t = =
σ U
e √t
σ is the annual volatility of the underlying asset’s returns and t is the length of the step in the
binomial model.
ert −D
π u = probability of an up move= U−D
The price of an exchange-quoted zero-dividend share is $30. Over the past year, the stock has
exhibited a standard deviation of 17%. The continuously compounded risk-free rate is 5% per
annum. Compute the value of a 1-year European call option with a strike price of $30 using a
0.05 ×1
264
© 2014-2020 AnalystPrep.
(e0.05 ×1 ) − D
1.0513 − 0.8437
πu = = = 0.61
U−D 1.1853 − 0.8437
π d = 1 − 0.61 = 0.39
╱
$30
╲
0.8437 × $30 = $25.30
╱
C0
╲
Max (0, $25.3 − $30) = $0
In the two-period model, the tree is expanded to create room for a greater number of potential
265
© 2014-2020 AnalystPrep.
SUU
╱
SU
╲
S0 ╱ SUD or SDU
╲
╱
SD
╲
SDD
The two-step model uses the same formulae used in the one-step version to calculate the value of
an option.
Note: The value of a put can be calculated once the value of the call has been determined, using
Where “stock” represents the stock price, X represents the strike price, r is the rate of return,
Binomial models with one or two steps are unrealistically simple. Assuming only one or two steps
would yield a very rough approximation of the option price. In practice, the life of an option is
divided into 30 or more time steps. In each step, there is a binomial stock price movement.
As the number of time steps is increased, the binomial tree model makes the same assumptions
about stock price behavior as the Black– Scholes–Merton model. When the binomial tree is used
to price a European option, the price converges to the Black–Scholes–Merton price as the
266
© 2014-2020 AnalystPrep.
Delta
The delta, Δ, of a stock option, is the ratio of the change in the price of the stock option to
the change in the price of the underlying stock. It is the number of units of the stock an
investor/trader should hold for each option shorted in order to create a riskless portfolio. This
The delta of a call option is always between 0 and 1 because as the underlying asset increases in
price, call options increase in price. The delta of a put option, on the other hand, is always
between -1 and 0 because as the underlying security increases, the value of put options
decrease.
For instance, suppose that when the price of a stock change from $20 to $22, the call option
price changes from $1 to $2. We can calculate the value of delta of the call as:
2 −1
= 0.5
22 − 20
This means that if the underlying stock increases in price by $1 per share, the option on it will
Suppose that an investor is long one call option on the stock above (with a delta of 0.5, or 50
since options have a multiplier of 100). The investor could delta hedge the call option by
shorting 50 shares of the underlying stock. Conversely, if the investor is long one put on the
stock (with a delta of -0.5, or -50), they would maintain a delta neutral position by purchasing
As the standard deviation increases, so does the divide (dispersion) between stock prices in up
and down states ( SU and SD, respectively). Suppose there was no deviation at all. Would we have
With zero standard deviation, (SU would be equal to SD, and instead of a tree, we would have a
267
© 2014-2020 AnalystPrep.
straight line. But provided there’s some deviation, the gap between stock prices in the upstate
and stock prices in the downstate increasingly widens as the deviation increases.
To capture volatility, therefore, it would be paramount to evaluate stock prices at each time
Given a stock that pays a continuous dividend yield q , the following formula can be used to price
e(r−q)t − D
Probability of an up move = π u =
U−D
Probability of a down move = 1 − πu
1 1
D=size of the down move factor= e−σ√ t = =
eσ√t U
Note: The sizes of the up move factor and down move factor are the same as in the zero-dividend
model.
Sometimes it may also be necessary to price options constructed with a stock index as the
underlying, for instance, an option on the S&P 500 index. Such an option would be valued in a
manner similar to that of the dividend-paying stock. It’s assumed that the stocks forming part of
The binomial model can also be modified to incorporate the unique characteristics of options on
futures. Of note is the fact that futures contracts are largely considered cost-free to initiate, and
therefore in a risk-neutral environment, they are zero-growth instruments. The only formula that
1 −D
πu =
U−D
268
© 2014-2020 AnalystPrep.
When dealing with options on currencies, a plausible assumption is that the return earned on a
foreign currency asset is equal to the foreign risk-free rate of interest. As such, the probability of
e(r DC −r FC )t − D
πu =
U−D
American Options
To value an American option, we check for early exercise at each node. If the value of the option
is greater when exercised, we assign that value to the node. If that’s not the case, we assign the
value of the option unexercised. We then work backward through the tree as usual.
The binomial model is essentially a discrete-time model where we evaluate option values at
discrete times, say, intervals of one year, intervals of six months, intervals of three months, etc.
However, if we were to shrink the length of time intervals to arbitrarily small values, we’d end up
with a continuous-time model where the price can move at non-discrete times. The binomial
model converges to the continuous-time model when time periods are made arbitrarily small.
269
© 2014-2020 AnalystPrep.
Questions
Question 1
Suppose we have a 6-month European call option with K = $23. Suppose the stock
price is currently $22 and in two-time steps of three months, the stock can go up or
down by 10%. The up move factor is 1.1 while the down move factor is 0.9. The risk-
A. $2
B. $1.54
C. $1.45
D. $0
Suu = $26.62
╱
Su = $24.2
╲
S0 = $22
╱ Su d = $21.78
╲ Sd u = $21.78
╱
Sd = $19.8
╲
Sdd = $17.82
Su = 22 × 1.1 = 24.2,
270
© 2014-2020 AnalystPrep.
Other values at other nodes are calculated using the relevant up/down factors.
The expected value of the call six months from now is given by:
$1.54
The value of the call today = = $1.45
e0.12×0.5
Question 2
A 1-year $50 strike European call option exists on ABC stock currently trading at $49.
risk-free rate is 4%. Assuming an annual standard deviation of 3%, compute the value
A. $0.31
B. $0.30
C. $0.47
D. $0
271
© 2014-2020 AnalystPrep.
Note that the stock is dividend-paying, and therefore the formula for the probability
Let S represent the price of the stock and f represent the value of the call. This is a
Value of the call option one year from today = ($0.47 × 0.67 + $0 × 0.33) = $0.31
$0.31
Value of the call today = = $0.30
e0.04
272
© 2014-2020 AnalystPrep.
Reading 59: The Black-Scholes-Merton Model
Explain the lognormal property of stock prices, the distribution of rates of return, and
non-dividend-paying stock.
Define implied volatilities and describe how to compute implied volatilities from market
Explain how dividends affect the decision to exercise early for American call and put
options.
dividend-paying stock.
Describe warrants, calculate the value of a warrant, and calculate the dilution cost of
Suppose we have a random variable X. This variable will have a lognormal distribution if its
natural log (ln X) is normally distributed. In other words, when the natural logarithm of a random
variable is normally distributed, then the variable itself will have a lognormal distribution.
The two most essential characteristics of the lognormal distribution are as follows:
It has a lower bound of zero, i.e., a lognormal variable cannot take on negative values.
The distribution is skewed to the right, i.e., it has a long right tail.
These characteristics are in direct contrast to those of the normal distribution, which is
symmetrical (zero-skew) and can take on both negative and positive values. As a result, the
273
© 2014-2020 AnalystPrep.
normal distribution cannot be used to model stock prices because stock prices cannot fall below
A crucial part of the BSM model is that it assumes stock prices are log-normally distributed.
Precisely,
σ2
lnST ∼ N (lnS0 + (μ − ) T, σ 2T)
2
Where:
Note: The above relationship holds because mathematically, if the natural logarithm of a random
variable ?,(?? ?) is normally distributed, then ? has a lognormal distribution. It's also imperative
to note that the BSM model assumes stock prices are lognormally distributed, with stock returns
being normally distributed. Specifically, continuously compounded annual returns are normally
274
© 2014-2020 AnalystPrep.
distributed with:
σ2 σ2
a mean of [μ − ] and a variance of
2 T
ABC stock has an initial price of $60, an expected annual return of 10%, and annual volatility of
15%. Calculate the mean and the standard deviation of the distribution of the stock price in six
months.
σ2
lnST ∼ N (lnS0 + (μ − ) T, σ√T )
2
0.152
=N [ln60 + (0.10 − ) 0.5, 0.152 × 0.5]
2
lnST ∼ N [4.139, 0.011]
(In this case,we have μ = √ 0.011 = 0.1049)
Sometimes, the Global Association of Risk Professionals (GARP) may want to test your
understanding of the lognormal concept by involving confidence intervals. Since lnST is log-
normally distributed, 95% of values will fall within 1.96 standard deviations of the mean.
Similarly, 99% of the values will fall within 2.58 standard deviations of the mean. For example, to
obtain the 99% confidence interval for stock prices using the above data, we will proceed as
follows:
Using the properties of a lognormal distribution, we can show that the expected value of
275
© 2014-2020 AnalystPrep.
ST , E(ST ), is:
E(ST ) = S0 eμT
The current price of a stock is $40, with an expected annual return of 15%. What is the expected
Solution
1
R pr = [r1 × r2 × r3 × ⋯ × rn ] n − 1
The continuously compounded return realized over some time of length T is given by:
1 ST
ln ( )
T S0
The realized return of a stock initially priced at $50 growing, with volatility, to $87 over five
1 $87
Realized return = ln ( ) = 11.08%
5 $50
276
© 2014-2020 AnalystPrep.
We can calculate historical volatility from daily price data of stock. We simply need to calculate
continuously compounded returns per day and then determine the standard deviation.
Si
ln ( )
Si−1
The volatility of short periods can be scaled to give the volatility of more extended periods.
For example,
Note that this formula is useful throughout the whole FRM part 1 and FRM part 2 exams in
estimating volatility.
Conversely,
annual volatility
daily volatility =
√ no.of trading days in a year
Black-Scholes-Merton Model
The Black-Scholes-Merton model is used to price European options and is undoubtedly the most
critical tool for the analysis of derivatives. It is a product of Fischer Black, Myron Scholes, and
Robert Merton.
The model takes into account the fact that the investor has the option of investing in an asset
earning the risk-free interest rate. The overriding argument is that the option price is purely a
function of the volatility of the stock's price (option premium increases as volatility increases).
i. There is no arbitrage.
277
© 2014-2020 AnalystPrep.
ii. The price of the underlying asset follows a lognormal distribution.
iii. The continuous risk-free rate of interest is constant and known with certainty.
v. The underlying asset has no cash flow, such as dividends or interest payments.
vi. Markets are frictionless – no transaction costs, taxes, or restrictions on short sales.
vii. Options can only be exercised at maturity, i.e., they are European-style. The model
C 0 = S0 × N (d1 ) − Ke−rT × N (d 2 )
Where:
S0 σ2
ln ( ) + [r + ( )] T
K 2
d1 =
σ√ T
d 2 = d 1 − (σ√T)
S0 =asset price
K =exercise price
278
© 2014-2020 AnalystPrep.
Example: Valuing a call option using the BSM model
S0 σ2
ln ( ) + [r + ( )] T
K 2
d1 =
σ√ T
100 0.252
ln ( ) + [0.10 + ( )] 0.5
90 2 0.1053 + 0.0656
= = = 0.9672
0.25 √0.5 0.1768
d 2 = d 1 − (σ√ T) = 0.9672 − (0.25√0.5) = 0.7904
From a standard normal probability table, look up N(0.97) = 0.8333 and N(0.79) = 0.7852.
Note that the intrinsic value of the option is $10—our answer must be at least that amount.
Exam tips
Tip 1: Given one of either the put value or the call value, you can use the put-call parity to find
279
© 2014-2020 AnalystPrep.
(−Rc T)
C 0 = P 0 + S0 − (Ke f )
(−Rc T)
P 0 = C 0 − S0 + (Ke f )
Tip 3: As S0 becomes very large, calls (puts) are extremely in-the-money (out-of-the-money)
Tip 4: As S0 becomes very small, calls (puts) are extremely out-of-the-money (in-the-money)
Tip 5: Although N(−d1 )and N(−d2 ) can easily be identified from statistical tables; sometimes
Assume that we have a known dividend d distributed a time T1 , T1 < T where T is the maturity
date. To value calls and puts when there are such dividends, we modify the BSM model by
S = S0 − D
D is the sum of the PV(discounted at R cf ) of the dividend payments during the life of the option.
m a division factor in bringing the Δt to a full year. e.g. Δt = 2Δt = 2 months, m=12 months, so
Δt1 2
= 0.1667 years .
m 12 =
After this, everything else in the computational formulas remains the same, i.e.,
280
© 2014-2020 AnalystPrep.
∣ (−R c T) ∣
C 0 = [S0 × N (d1 )] − ∣K × e f × × N (d 2 )∣
∣ ∣
(−Rc ×T)
P 0 = [K × e f × (1 − N (d 2 ))] − [S × (1 − N (d 1 ))]
S
ln ( ) + [R cf + (0.5 × σ2 )] T
K
d1 =
σ√ T
d2 = d1 − (σ√T )
The underlying argument here is that on the ex-dividend dates, the stock prices are expected to
Exam tip: Sometimes, GARP will give you not a dollar amount "d" of the dividend, but a dividend
yield q. For example, you may be told that the dividend yield is 2%, continuously compounded. In
S = e−qT × S0
How Dividends affect the Early Exercise for American Calls and Puts
Call option holders have the right but not the obligation to buy shares as per the terms of the
contract, but they do not hold shares. As such, they cannot benefit from the rights of
shareholders, such as the right to receive dividends – as long as the call options have not been
exercised.
When the underlying stock pays dividends, a call option holder will not receive it unless they
exercise the contract before the dividend is paid. Whoever owns the stock as of the ex-dividend
date receives the cash dividend, so an investor who owns in-the-money call options may exercise
early to capture the dividend. In summary, a call option should only be exercised early to take
281
© 2014-2020 AnalystPrep.
ii. The time value of the option needs to be less than the value of the dividend.
It wouldn't make sense to exercise an out-of-the-money call option and pay an above-market
Provided these conditions have been met, the holder of an American put option can exercise
early, but only after the dividend has been paid. It would make a whole lot more sense to exercise
the put option the day after the dividend is paid to collect the dividend, instead of exercising the
Black's approximation sets the value of an American call option as the maximum of two European
prices:
I. A European call with the same maturity as the American call being valued, but with the
stock price reduced by the present value of the dividend. This implies that S0 is reduced
by the present value of the dividends payable, but all other variables remain the same.
Where
Δt 1 Δt2
−(r) −(r)
PV = D1e m + D2 e m
282
© 2014-2020 AnalystPrep.
m is a division factor in bringing the Δt to a full year. If Δt = 2 months, m=12 months, so
Δt 2
= = 0.1667 years
m 12
II. A European option is maturing just before the final ex-dividend date of the American-
option. This implies that time to maturity is trimmed down to just before the final
dividend is paid. The PV of dividends other than the final one must be deducted from S0
The largest of the two values (I) and (II) above is the desired Black's approximation for the
American call.
Exam tips:
We can extend the BSM result to valuing other assets such as stock indices, currencies, and
futures. For a European option on a stock paying a continuous dividend yield at a rate of q, the
C 0 = S0 e−qT × N (d 1 ) − Ke−rt × N (d 2 )
Where:
σ2
283
© 2014-2020 AnalystPrep.
S0 σ2
ln ( ) + [r − q + ( )] T
K 2
d1 =
σ√T
d 2 = d 1 − (σ√T)
Note that we can also use these formulas to value a European option on a stock index paying
When dealing with an option on foreign currency, we take note that it behaves like a stock paying
a dividend yield at the risk-free foreign rate (rf ). We, therefore, set q= (rf ), and we have the
Suppose the current exchange rate for a currency is 1.100 and the volatility of the exchange rate
is rate is 20%. Calculate the value of a call option to buy 1000 units of the currency in 3 years at
an exchange rate of 2.200. The domestic and foreign risk-free interest rates are 2% and 3%,
respectively.
Solution
S0 σ2
ln ( ) + [r − rf + ( )] T
K 2
d1 =
σ√ T
1.100 0.22
ln ( ) + [0.02 − 0.03 + ( )] 3
2.200 2
=
0.2√ 3
d2 = d1 − σ√T = −1.9143 − 0.2√3 = −2.2607
284
© 2014-2020 AnalystPrep.
From standard normal tables,
This is the value of the option to buy one unit of the currency. The value of an option to buy 1000
units is 0.0036×1000=$3.60
When we are considering an option on futures, we realize that the futures price F is typical to a
stock paying a dividend yield at the risk-free domestic rate (r). We, therefore, set q=r and
Warrants are securities issued by a company on its own stock, which give their owners the right
to purchase shares in the company at a specific price at a future date. They are much like
options, the only difference being that while options are traded on an exchange, warrants are
issued by a company directly to investors in bonds, rights issues, preference shares, and other
securities. They are basically used as sweeteners to make offers more attractive.
When warrants are exercised, the company issues more shares, and the warrant holder buys the
shares from the company at the strike price. An option traded by an exchange does not change
the number of shares issued by the company. However, a warrant allows new shares to be
285
© 2014-2020 AnalystPrep.
purchased at a price lower than the current market price, which dilutes the value of the existing
In an efficient market, the share price reflects the potential dilution from outstanding warrants.
We are not necessarily required to consider these when valuing the outstanding warrants. This
For detachable warrants, their value can be estimated as the difference between the market
price of bonds with the warrants and the market price of the bonds without the warrants.
The Black-Scholes-Merton Model can also be used to value warrants using the BSM call/put
c
C 0 = [S0 × N (d 1 )] − ∣∣K × e−R f ×T × N (d2 )∣∣
1. The stock price S0 s replaced by an "adjusted" stock price. Suppose a company has N
outstanding shares worth S0 . This means that the value of the company's equity is NS 0 .
Further, assume that the company has decided to issue M number of warrants with each
warrant giving the holder the right to buy one share for K. If the stock prices change to
ST at time T, the (adjusted) stock price which accounts for the dilution effect of the issued
warrants, is:
(NS0 + MK)
Sadjusted =
N+M
2. The volatility input is calculated on equity (volatility of the value of the shares plus the
Implied Volatility
The volatility of the stock price is the only unobservable parameter in the BSM pricing formula.
286
© 2014-2020 AnalystPrep.
The implied volatility of an option is the volatility for which the BSM option price equals the
market price.
Implied volatility represents the expected volatility of a stock over the life of the option. It is
influenced by market expectations of the share price as well as by supply and demand of the
underlying options. As expectations rise, and the demand for options increases, the implied
If we use the observable parameters in the BSM formula (S0 , K, r, and T ) and set the BSM
formula equal to the market price, then it's possible to solve for volatility that satisfies the
equation. However, there is no closed-form solution for the volatility, and the only way to find it is
through iteration.
287
© 2014-2020 AnalystPrep.
Practice Questions
Question 1
ABC stock is currently trading at $70 per share. A dividend of $1 is expected after
three months and another one of $1 after six months. An American option on ABC
stock has a strike price of $65 and 8 months to maturity. Given that the risk-free rate
is 10% and the volatility is 32%, compute the price of the option:
A. $9.85
B. $12.5
C. $10
D. $10.94
The current price of the share must be adjusted to take into account the expected
dividends.
S0 =68.0735
K=65
σ=0.32
r=0.1
T=0.6667
0.322
288
© 2014-2020 AnalystPrep.
68.0735 0.322
ln ( ) + [0.1 + ( )] 0.6667
65 2
d1 = = 0.5626
0.32 √0.6667
d2 = d1 − 0.32√0.6667 = 0.3013
N(d1 )=0.7131
N(d2 )=0.6184
Question 2
The stock price is currently $100. Assume that the expected return from the stock is
35% per annum, and its volatility is 20% per annum. Calculate the mean and
standard deviation of the distribution, and determine the 95% confidence interval for
In this case,
S0 =100
μ=0.35 and,
σ=0.20
The probability distribution of the stock price in two years is lognormal and is given
by:
289
© 2014-2020 AnalystPrep.
0.22
lnST ∼ {ln100 + (0.35 − ) 2, 0.22 × 2} ∼ (5.27 , 0.28 2 )
2
The mean stock price = S0 eμT = 100e(0.35 ×2) = 201.38
2×2
= 100e(0.35×2)√(e0.2 − 1) = 58.12
290
© 2014-2020 AnalystPrep.
Reading 60: Option Sensitivity Measures: The “Greeks”
Describe and assess the risks associated with naked and covered option positions.
Describe the use of a stop-loss hedging strategy, including its advantages and
disadvantages, and explain how this strategy can generate naked and covered option
positions.
Describe the dynamic aspects of delta hedging and distinguish between dynamic
Define and describe theta, gamma, vega, and rho for options positions and calculate
Describe how portfolio insurance can be created through option instruments and stock
index futures.
A naked option position occurs when a trader sells a call option without insurance in the form of
a holding of the underlying shares. If the option position is backed by ownership of the
Selling naked call options is laden with risks on the part of the trader. If the market price is
below the pre-agreed strike price on the expiration date, the seller makes a gain equal to the
291
© 2014-2020 AnalystPrep.
premium received. However, if the market price soars above the strike price as at expiration, the
buyer exercises the option. When that happens, the seller has to deliver the agreed number of
shares to the buyer, even if it means buying from the market. Depending on the extent of the
price increase, the entire loss arising from a naked option can be absorbed by the premium
received. In other cases, the price increase may be so high that the seller is left with a net loss.
On the contrary, the trader may insure themselves by selling covered options. This may be a
safer strategy but one that’s also laden with downside risk. If the stock falls, the seller will get to
keep the entire premium, but the shares under their ownership will now be worthless.
Sometimes the price fall may be too high such that the total value lost in the long position
Example:
Suppose a firm sells 10,000 naked call options on a stock on a stock currently going for $30 a
292
© 2014-2020 AnalystPrep.
Scenario 1: Price at expiry = $29
The buyer will not exercise the premium. Total income generated = 10 , 000 × $4 = $40, 000
The buyer will exercise the options, and the seller is obliged to honor the contract. As such, the
seller buys 10,000 shares from the market at a cost of $37 per share and hands them to the
buyer.
In this scenario, the loss is absorbed by the premium, resulting in a net loss of $0
A stop-loss trading strategy is a strategy where the trader initially gets into a naked option
position but later on seeks cover when the option moves in-the-money. In other words, protection
is sought only when market conditions are such that the call writer stands to lose.
With a naked call position, this strategy requires the purchase of the underlying asset
immediately the market price rises above the option’s strike price. But as soon as the market
price returns to a position that’s below the strike price, the trader sells the underlying asset.
Although this sounds like a simple, executable plan on paper, it’s a lot more complicated in
practice thanks to transaction costs and price uncertainty. In practice, buy/sell costs increase as
fluctuations in the strike price increase. It also becomes even more difficult to predict whether
293
© 2014-2020 AnalystPrep.
the option will be in-the-money or out-of-the-money at expiration.
Delta of an Option
Delta is a measure of the degree to which an option is exposed to changes in the price of the
underlying asset. It’s the ratio of the change in the price of the call option to the change in the
For example, if we have a delta value of 0.5, it means that when the price of the underlying
moves by a point, the price of the corresponding call option will change by half a point. If delta =
0.5, a $1 increase in the underlying’s price triggers a $0.5 increase in the price of the call option.
Call Option
Delta of a Call option is closely related to N (d1 ) in the Black-Scholes Pricing model. Precisely,
Δc = e−qT N (d1 )
Where
σ2
ln S +( r+ )T
K 2
d1 =
σ√T
294
© 2014-2020 AnalystPrep.
σ=volatility of the underlying
Exam tips:
I. Delta of a call option is always positive (between 0 and 1). The delta of an at-the-money
call option is close to 0.5. Delta moves to 1 as the call goes deep in-the-money (ITM). It
II. If the underlying does not pay dividends the delta of a call option simplifies to:
(q is zero in this case, and any number raised to power zero is equal to 1)
Put option
Δp = e−qT [N (d 1 ) − 1]
It behaves similar to the call delta, except for the sign (between 0 and -1). As with the call delta,
Δ p = e0 [N (d1 ) − 1] = [N (d 1 ) − 1]
Exam tip: The delta of an at-the-money put option is close to -0.5. Delta moves to -1 as the put
goes deep in-the-money. It moves to zero as the put goes deep out-of-the-money.
Delta of a Forward
Δf = e−qT
295
© 2014-2020 AnalystPrep.
Where q is the dividend yield and T is time to expiry.
By definition, all forward positions have a delta of approximately 1. What does that imply?
It means the underlying asset and the corresponding forward contract have a one-to-one
relationship. As a result, a forward sale position can always be perfectly hedged by buying the
Unlike forward contracts, the delta value of a futures contract is not ordinarily equal to 1. This is
because futures and spot prices move in lockstep, but are not exactly identical.
Δf utures = erT
Δf utures = e(r−q)T
Delta hedging is an attempt to reduce (hedge) the risk associated with price movements in the
underlying, by offsetting long and short positions. For instance, a long call position could be
offset by shorting the underlying stock. Since delta is actually a function of the price of the
When delta changes, the initially option-hedged position is, again, thrust into a state of
imbalance. In other words, the number of stocks is no longer matched with the right number of
296
© 2014-2020 AnalystPrep.
The overall goal of delta-hedging (a delta-neutral position) is to combine a position in the
underlying with another position in an option such that the value of the portfolio remains fixed
even in the face of constant changes in the value of the underlying asset.
An options position can be hedged using shares of the underlying. A share of the underlying has
a delta equal to 1 because of the value changes by $1 for a $1 change in the stock. For instance,
suppose an investor is long one call option on a stock whose delta is 0.6. Because options are
usually held in multiples of 100, we could say that the delta is 60. In such a scenario, the investor
could delta hedge the call option by shorting 60 shares of the underlying. The converse is true: If
the investor is long one put option, he would delta hedge the position by going long 60 shares of
the underlying.
Sometimes an options position can be delta hedged using another options position that has a
delta that’s opposite to that of the current position. This effectively results in a delta-neutral
position. For instance, suppose an investor holds a one call option position with a delta of 0.5. A
call with a delta of 0.5 means it is at-the-money. To maintain a delta neutral position, the trader
can purchase an at-the-money put option with a delta of -0.5, so that the two cancel out.
Delta of a Portfolio
Suppose we want to determine the delta of a portfolio of options, all on a single underlying. The
portfolio delta is equivalent to the weighted average of the deltas of individual options.
n
portfolio delta = Δ portfolio = ∑ w iΔ i
i =1
w i represents the weight of each option position while Δ i represents the delta of each option
position.
Portfolio delta gives the change in the overall option position caused by a change in the price of
297
© 2014-2020 AnalystPrep.
the underlying.
Theta
Theta, θ , tells us how sensitive an option is to a decrease in time to expiration. It gives us the
change in the price of an option for a one-day decrease in its time to expiration.
Options lose value as expiration approaches. Theta estimates the value lost per day if all other
factors are held constant. Time value erosion is nonlinear, and this has implications on theta. As
a matter of fact, the theta of in-the-money, at-the-money, and slightly out-of-the-money options
generally increases as expiration nears. On the other hand, the theta of far out-of-the-money
∂c
θ=
∂t
Where:
∂ t=change in time
For European call options that have zero dividends, the Black-Scholes Merton model can be used
S0 N ′ (d1 ) σ
θcall =− − rXe−rT N (d 2 )
2√T
S0 N ′ (d1 ) σ
θput = − + rXe−rT N (d −2 )
2√T
Where:
298
© 2014-2020 AnalystPrep.
y2
′ 1 −(
2
)
N (y) = e , y = d 1 , d2
√2π
In the above equations, the resulting value for theta is measured in years because T is also
measured in years. To covert theta into a daily value, divide by 252, assuming 252 trading days
in a year.
Gamma
Gamma, Γ, measures the rate of change in an option's Delta per $1 change in the price of the
underlying stock. It tells us how much the option’s delta should change as the price of the
underlying stock or index increases or decreases. Options with the highest gamma are the most
Mathematically,
∂ 2c
Γ=
∂ 2s
Where the numerator and denominator are the partial derivatives of the call and stock prices,
respectively.
N ′ (d1 )
Γ=
S0 σ√T
While delta neutral positions hedge against small changes in stock price, gamma-neutral
positions guard against relatively large stock price moves. As such, a delta-neutral position is
important, but even more important is one that’s also gamma-neutral, because it will be
The number of options that must be added to an existing portfolio to generate a gamma-neutral
Γ
299
© 2014-2020 AnalystPrep.
Γp
−( )
ΓT
Where:
A trader has a short option position that’s delta-neutral but has a gamma of -800. In the market,
there’s a tradable option with a delta of 0.8 and a gamma of 2. To maintain the position gamma-
The number of options that must be added to an existing portfolio to generate a gamma-neutral
?? −800
− =− = 400
?? 2
Buying 400 calls, however, increases delta from zero to 320 (=400×0.8). Therefore, the trader
has to sell 320 shares to restore the delta to zero. Positions in shares always have zero gamma.
The relationship between the three Greeks can best be expressed in the following equation:
rP = θ + rS Δ + 0.5σ2 S 2 Γ
Where:
θ=option theta
300
© 2014-2020 AnalystPrep.
S =price of the underlying stock
Δ=option delta
Γ=option Gamma
If a position is delta-neutral, then Δ = 0, and the above equation narrows down to:
rP = θ + 0.5σ2 S 2 Γ
Vega
Vega measures the rate of change in an option's price per 1% change in the implied volatility of
the underlying stock. And while Vega is not a real Greek letter, it tells us how much an option’s
As an example, a Vega of 6 indicates that for a 1% increase in volatility, the option’s price will
increase by 0.06. For a given exercise price, risk-free rate, and maturity, the Vega of a call equals
Mathematically,
∂c
V ega =
∂σ
Where:
∂σ=change in volatility
V ega = S0 N ′ (d1 ) √T
A drop in Vega will typically cause both calls and puts to lose value. An increase in Vega will
301
© 2014-2020 AnalystPrep.
typically cause both calls and puts to gain value.
Vega decreases with maturity, unlike gamma which increases with maturity. Vega is highest for
at-the-money options.
Rho
Rho measures the expected change in an option's price per 1% change in interest rates. It tells
us how much the price of an option should fall or rise in response to an increase or decrease in
As interest rates increase, the value of call options will generally increase. On the other hand, as
interest rates increase, the value of put options will usually decrease. Although rho is not a
dominant factor in the price of an option, it takes center stage when interest rates are expected
to change significantly.
Long-term options are far more sensitive to changes in interest rates than are short-term
options. Furthermore, in-the-money calls and puts are more sensitive to interest rate changes
Mathematically,
∂c
rho =
∂r
Where:
For European calls and puts on stocks that do not pay dividends,
302
© 2014-2020 AnalystPrep.
Be Used to Formulate Expected Gains and Losses With Option
Positions
On paper, attaining neutrality to all the Greeks might appear a straight forward task but in
practice, this is hardly the case. Although delta-neutral positions are easy to create and maintain,
it’s quite difficult to find securities at reasonable prices that can help tame the negative effects
Traders usually concentrate on maintaining delta neutrality and then purpose to continuously
Sometimes traders may use different values of a portfolio value determinant in order to assess
how sensitive the portfolio is to that determinant. For example, the traders may work with
different values of volatility to estimate the impact on portfolio value. This is called scenario
analysis.
Under scenario analysis, a single parameter may be varied at a time, but two or more
parameters can also be varied simultaneously to estimate their overall effect on the portfolio.
Portfolio insurance is the combination of (1) an underlying instrument and (2) either cash or a
derivative that generates a minimum value for the portfolio in the event that markets crash and
values decline, while still allowing the trader to make a gain in the event that market values rise.
The most common insurance strategy involves using put options to lock in the value of an asset.
This way, the trader is able to maintain a limit on the portfolio value – even if the underlying’s
price tumbles, the trader is insulated from prices below the put’s strike.
To hedge a portfolio with index options, the trader selects an index with a high correlation to
their portfolio. For instance, if the portfolio consists of mainly technology stocks, the Nasdaq
Composite Index might be a good fit. If the portfolio is made up of mainly blue-chip companies,
303
© 2014-2020 AnalystPrep.
then the Dow Jones Industrial Index could be used.
Alternatively, a trader can use stock index futures with a similar end goal. Traders who want to
hedge their portfolios need to calculate the amount of capital they want to hedge and find a
representative index. Assuming an investor wants to hedge a $500,000 stock portfolio, she would
sell $500,000 worth of a specific futures index, such as the S&P 500.
304
© 2014-2020 AnalystPrep.
Questions
Question 1
The current stock price of a company is USD 100. A risk manager is monitoring call
and put options on the stock with exercise prices of USD 70 and 6 days to maturity.
Which of these scenarios is most likely to occur if the stock price falls by USD 1?
The call option is deep-in-the-money and therefore must have a delta close to one.
The put option is deep out-of-the-money and will, therefore, have a delta close to zero.
Therefore, the value of the in-the-money call will decrease by close to USD 1, and the
value of the out-of-the-money put will increase by a much smaller amount close to 0.
Among the four choices, it’s A that is closest to satisfying both conditions.
Question 2
XY Z Inc., a non-dividend-paying stock, has a current price of $200 per share. Eric
Rich, FRM, has just sold a six-month European call option contract on 200 shares of
this stock at a strike price of $202 per share. He wants to implement a dynamic delta
hedging scheme to hedge the risk of having sold the option. The option has a delta of
0.50. He believes that the delta would fall to 0.40 if the stock price falls to $195 per
share.
Identify what action he should take NOW (i.e., when he has just written the option
305
© 2014-2020 AnalystPrep.
contract) to make his position delta-neutral
After the option is written, if the stock price falls to $195 per share, identify the
action Mr. Rich should take at that time, i.e. LATER, to rebalance his delta-hedged
position
A. NOW: buy 200 shares of stock, LATER: buy 100 shares of stock
C. NOW: sell 100 shares of stock, LATER: buy 100 shares of stock
NOW: Eric sold a call on 200 shares, that means he’s short delta of 0.50 × 200, which
is delta = -100. To be delta neutral, he must long (i.e. buy) 100 shares of stock.
LATER: As price falls to $195, the delta moves to −80 = −0.40 × 200 . To be delta
neutral, Eric’s portfolio needs to have 80 shares of stock. He purchased 100 shares at
306
© 2014-2020 AnalystPrep.