0% found this document useful (0 votes)
32 views306 pages

CH 4 Valuation and Risk Models

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
32 views306 pages

CH 4 Valuation and Risk Models

Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 306

FRM Part I Exam

By AnalystPrep

Study Notes - Valuation and Risk Models

Last Updated: 16-10-2020

1
© 2014-2020 AnalystPrep.
Table of Contents

45 - Measures of Financial Risk 3


46 - Calculating and Applying VaR 26
47 - Measuring and Monitoring Volatility 45
48 - External and Internal Credit Ratings 72
49 - Country Risk 82
50 - Measuring Credit Risk 94
51 - Operational Risk 120
52 - Stress Testing 141
53 - Pricing Conventions, Discounting, and Arbitrage 164
54 - Interest Rates 177
55 - Bond Yields and Return Calculations 196
56 - Applying Duration, Convexity, and DV01 204
57 - Modeling and Hedging Non-Parallel Term Structure Shifts 228
58 - Binomial Trees 248
59 - The Black-Scholes-Merton Model 258
60 - Option Sensitivity Measures: The “Greeks” 75

2
© 2014-2020 AnalystPrep.
Reading 45: Measures of Financial Risk

After completing this reading, you should be able to:

Describe the mean-variance framework and an efficient frontier.

Explain the limitations of the mean-variance framework with respect to assumptions

about return distributions.

Compare the normal distribution with the typical distribution of returns of risky

financial assets such as equities.

Define the VaR measure of risk, describe assumptions about return distributions and

holding period, and explain the limitations of VaR.

Explain and calculate Expected Shortfall (ES), and compare and contrast VaR and ES.

Define the properties of a coherent risk measure and explain the meaning of each

property.

Explain why VaR is not a coherent risk measure.

Describe spectral risk measures and explain how VaR and ES are special cases of

spectral risk measures.

The Mean-Variance Framework

The mean-variance framework uses the expected mean and standard deviation to measure the

financial risk of portfolios. Under this framework, it is necessary to assume that returns follow a

specified distribution, usually the normal distribution.

The normal distribution is particularly common because it concentrates most of the data around

the mean return. 66.7% of returns occur within plus or minus one of the standard deviations of

the mean. A whopping 95% of the returns occur within plus or minus two standard deviations of

the mean.

3
© 2014-2020 AnalystPrep.
Investors are normally concerned with downside risk and are therefore interested in

probabilities that lie to the left of the expected mean.

Note the expected does not imply the anticipated return but rather the average returns. On the

other hand, the risk is measured using the standard deviation of returns.

Example: Calculating Expected Return and Standard Deviation

The expected returns for an asset with corresponding probabilities are given below:

Return (%) Probability


10% 0.25
−20% 0.09
15% 0.40
7% 0.06
30% 0.20

Calculate the expected return and standard deviation of the asset return.

4
© 2014-2020 AnalystPrep.
Solution

To calculate the expected return, we weight the expected return by their corresponding

probability. That is:

n
R̄ = ∑ p iR
i=1

So for this case,

R̄ = (0.10 × 0.25) + (−0.20 × 0.09) + (0.15 × 0.40) + (0.07 × 0.06) + (0.30 × 0.20)
= 0.1312 = 13.12%

Recall that the variance for a variable X is given by:

Var (X) = E (X 2) − [E (X)]2

Then the variance of the return R is given by:

Var (R) = E (R 2 ) − [E (R)]2

The standard deviation is equal to the square root of the variance

σR = √E (R2 ) − [E (R)]2

Therefore,

E (R 2 ) =(0.102 × 0.25) + ((−0.20)2 × 0.09) + (0.152 × 0.40) + (0.072 × 0.06)


+(0.302 × 0.20)
=0.033394
⇒ σR =√0.033394 − [0.1312]2 = 0.1272 = 12.72%

Combinations of Investments

Consider two investments with respective means μ1 and μ 2 . Suppose that an investor wishes to

5
© 2014-2020 AnalystPrep.
invest in both investments with a proportion of w 1 in the first investment and w2 in the second

investment. It is safe to state that w2 = 1 − w1 .

The portfolio expected return is equivalent to weighted returns from individual investments. That

is:

μp = w 1 μ1 + w 2 μ2

The variance of the portfolio expected return is given by:

σP2 = w 21 σ12 + w22 σ22 + 2ρw 1 w 2σ1 σ2

Where

σ1 : standard deviation of the first investment

σ2 : standard deviation of the second investment

ρ: correlation between investment the first and the second investment

Therefore, the standard deviation of the portfolio is given by:

σP = √w 21 σ12 + w 22 σ22 + 2ρw1 w 2 σ

Note that the variance of the portfolio can be written as:

σP2 = w 21 σ12 + w22 σ22 + 2w1 w 2 Cov (R 1 , R 2 )

This is true from the fact that:

Cov (R 1 ,R 2 )
Cov (R 1 , R2 ) = ρ =
σ1 σ2
⇒ Cov (R 1 , R2 ) = ρσ1 σ2

Example: Calculating the expected return and the standard deviation of a


portfolio

6
© 2014-2020 AnalystPrep.
An investor invests in two assets X and Y, with an expected return of 10% and 15%. The investor

invests 45% of his funds in asset X and the rest in asset Y. The correlation coefficient is 0.45.

Given that the standard deviation of asset X is 15% and Y is 30%, what are the expected return

and standard deviations of the portfolio?

Solution

The portfolio expected return is given by:

μp = wX μX + wY μ Y
= 0.10 × 0.45 + 0.15 × 0.55
= 0.1275 = 12.75%

The portfolio standard deviation is given by:

σP = √ w21 σ12 + w 22 σ22 + 2ρw 1 w2 σ1 σ2

= √0.452 × 0.152 + 0.552 × 0.32 + 2 × 0.45 × 0.45 × 0.55 × 0.15 × 0.3


= √0.041805 = 0.2045 = 20.45%

Calculating the portfolio expected return and standard deviation can be extended to a portfolio

with n investments. The portfolio expected return for n returns is given by:

n
μ p = ∑ w i μi
i=1

Where μ i and wi are the mean return and weight of ith investment

And then the standard deviation of the portfolio is given by:

n n
σP = ∑ ∑ ρij w i wj σiσj
i=1 j=1

where ρij is the correlation coefficient between investments i and j. Other variables are intuitively

definitive.

Efficient Frontier

7
© 2014-2020 AnalystPrep.
The efficient frontier represents the set of optimal portfolios that offers the highest expected

return for a defined level of risk or the lowest risk for a given level of expected return. This

concept can be represented on a graph pitting the expected return (Y-axis) against the standard

deviation (X-axis).

For every point on the efficient frontier, at least one portfolio can be constructed from all

available investments that have the expected risk and return corresponding to that point.

Portfolios that do not lie on the efficient frontier are suboptimal: those that lie below the line do

not provide enough return for the level of risk, and those that lie on the right of the line have a

higher level of risk for the defined rate of return.

8
© 2014-2020 AnalystPrep.
The choice between optimal portfolios A, B, and D above will depend on an individual investor's

appetite for risk. A very risk-averse investor will choose portfolio A because it offers an optimal

return at the lowest risk, whereas an investor with room for more risk might pick D. After all, it

offers the highest optimal return at the highest risk.

The efficient frontier above considers only the risky assets. Now, consider when we introduce a

risk-free investment with a return of RF . It can be shown that the efficient frontier is a straight

line. That is, there is a linear relationship between expected return and the standard deviation of

return.

9
© 2014-2020 AnalystPrep.
F represents the risk-free return on the diagram above. Note that risk-free asset sits on the

efficient frontier because you cannot get a higher return with no risk, and you cannot have less

risk than zero. Consider the tangent line FM. By proportioning our investment between the risk-

free asset F and the risky asset M (market portfolio), we can obtain a risk-return that lies on the

line FM combination. In other words, the risk-return tradeoff is a linear function.

Denote the risk-free return by R F (with a standard deviation of 0). Also, let the market portfolio

return be RM , and its standard deviation is σM . Let the proportion of funds in a risky portfolio be

β and that in risk-free assets, be 1 − β . Now using the formula

μp = w 1 μ1 + w 2 μ2

10
© 2014-2020 AnalystPrep.
We have w1 = 1 − β, w2 = β, μ1 = R F, μ2 = R M so that return from the portfolio is given by

μ p = R F (1 − β) + βR M
μp − RF
⇒β=
RM − RF

Also, the standard deviation of a portfolio with two components is given by

σ = √w21 σF2 + w 22 σM
2 + 2ρw w σ σ
1 2 F M

But σF = 0

⇒ σ = √ 0 + w 22σM
2 + 0 = w 2 σM = βσ M

Therefore,

μP − R F
σ = σM ( )
RM − RF
σM σMR F
⇒ σ = μp ( )−
RM − RF RM − RF

The efficient frontier involving risk-free asset also shows that the investor should invest in risky

assets (in this case, M) by borrowing and lending at a risk-free rate rF . For instance, we assume

that an investor borrows at the rate rF so that now we are considering the efficient frontier

beyond M. If this is the case, then β > 1 and the proportion of amount borrowed will be β − 1,

and the total amount available is β multiplied by available funds. Assume now that we invest in

risky asset M. Then the expected return is:

βrM − (β − 1) RF = (1 − β) rF + βrM

The standard deviation can be shown to be βσM , which is similar to arguments for the points

below point M.

Therefore, it is safe to say risk-averse investors will invest in points on line FM and close to F,

11
© 2014-2020 AnalystPrep.
and those investors that are risk-seeking will invest on points close to M or even points beyond M

on line FM.

Limitations of the Mean-Variance Framework with Respect to


Assumptions about the Return Distributions

The mean-variance framework is unreliable when the assumption of normality is not met. If the

return distribution is not symmetric, the standard deviation is not a reliable or relevant measure

of probabilities at the extreme ends of the distribution.

The normality assumption is only strictly appropriate in the presence of a zero-skew (symmetric)

distribution. If the distribution leans to the left or the right – something that often happens with

financial returns – the mean-variance framework churns out misleading estimates of risk.

12
© 2014-2020 AnalystPrep.
The Normal Distribution

The normal distribution, also called Gaussian distribution, is a widely used continuous

distribution with two parameters: mean denoted by μ and the standard deviation denoted by σ .

The density function of the normal distribution is given by

(x − μ)2
1 −
f (x) = e 2σ2
√2πσ 2

The shape of the standard curve is as shown below.

The height of the normal distribution is equivalent to the probability that a value X has occurred.

This is commonly stated as X ∼ N(μ, σ 2). Values close to the center of the distribution are most

likely to occur while those values at the tails of the distribution are less likely to occur.

Note that, similar to other probability distributions, the probability that value lies between the

value a and b is equivalent to the area under the curve between the values a and b. This can be

13
© 2014-2020 AnalystPrep.
thought of as the cumulative distribution up to a point B less the cumulative distribution up to

point A.

Standard Normal Distribution

A standard normal distribution has a mean of 0 and a standard deviation of 1. In other words, μ

=0 and σ=1. As such, the normal distribution density function reduces to:

x2
1 −
f (x) = f (x) = e 2σ 2
√2π

The normal tables give a cumulative distribution of the standard normal distribution. For normal

distribution with the mean μ and the standard deviation σ , it can be transformed into z-scores,

which gives cumulative probability up to a value x for standard normal. The z-score is defined by:

x −μ
z=
σ

Where z ∼N(0,1).

For example, consider a normal distribution with a mean of 4 and a standard deviation of 5. What

is the probability that a value X is less than 7? Using standard normal transformation,

X −μ 7− 4
Pr(X < 7) = P r( < ) = 0.6
σ 5
= P r(z < 0.6) = Φ(0.6) = 1 − 0.2743 = 0.7257

Note that this is taken from the reference table you will be given on your exam:

14
© 2014-2020 AnalystPrep.
A normal distribution is usually assumed to apply to financial data because financial analysts are

mostly concerned with the mean and standard deviation. However, financial variables have fatter

15
© 2014-2020 AnalystPrep.
tails than the normal distribution. For instance, the means created can have fatter tails. Consider

the diagram below.

The diagram shows that the normal distribution and the actual distribution have equal standard

deviations, but the actual distribution has more peaked and has fatter tails than the normal

distribution. In other words, the actual distribution suggests that small and large changes

frequently occur while intermediate events occur less frequently.

Values at Risk (VaR)

Assuming a normal distribution for financial variables (by use of mean and standard deviation)

may likely underestimate the probability of the adverse events. The standard deviation can be a

perfect measure of risk, but it does not capture the tails of the probability distribution.

16
© 2014-2020 AnalystPrep.
VaR is a risk measure that is concerned with the occurrence of adverse events and their

corresponding probability. VaR is built from two parameters: the time horizon and the

confidence level. Therefore, we can say that VaR is the loss that we do not anticipate to be

exceeded over a given time period at a specified confidence level.

For example, consider a time horizon of 30 days and a confidence interval of 98%. Therefore 98%

VaR of USD 5 million implies that we are 98% certain that over the next 30 days, the loss will be

less than USD 5 million. Similarly, we can say that we are 2% certain that over the next 30 days,

the loss will be greater than USD 5 million.

Consider the following loss distribution density function curve:

Therefore, calculating X% VaR involves finding the loss that has an X% chance of being

exceeded, or the area under the curve that equal to (100-X)%.

Consider the following examples:

Example 1: Computing the VaR for normally distributed investment


returns

17
© 2014-2020 AnalystPrep.
The investment return over a period of time has a normal loss distribution with a mean of -200

and a variance of 300. What is 99% VaR of the loss distribution?

Solution

Denote the VaR level by t, then we need:

P (X < t) = 0.99

(Note we can also use P(X > t) =0.01)

Standardizing the normal distribution with a given mean and standard deviation, we have:

P (X < t) = 0.99

(Note we can also use P (X < t)=0.01.)

Standardizing the normal distribution with a given mean and standard deviation, we have:

t − −200
P (z < ) = 0.99
300
t + 200
⇒ Φ( ) = (0.99)
300
t + 200 −1
⇒ = Φ (0.99)
300

Now, Φ −1(0.99) is the inverse of standard normal cumulative probability. To do this using a

standard table, look for 0.99 (or closest value) in the table and read the corresponding vertical

and horizontal values and add them. In other words, we are reversing the reading of the

standard normal table.

In our cases, consider the following table:

18
© 2014-2020 AnalystPrep.
And thus:

19
© 2014-2020 AnalystPrep.
Φ−1 (0.99) = 2.33 + 0.02 = 2.32
t + 200
∴ = 2.33 ⇒ t = 499
300

The VaR level is 499 at a 99% confidence level.

Example 2: Calculating the VaR for a discrete loss distribution

The loss distribution of investment is as shown below:

Amount of Loss Probability


USD 10 Million 75%
USD 13 Million 22%
US 17 million 3%

What is the value of the 99% VaR?

Solution

To find the 99%, we need to find the cumulative probability distribution and locate 99%:

Amount of Loss Probability Cumulative Probability Range


USD 10 Million 75% 0 to 75%
USD 13 Million 22% 75% to 97%
US 17 million 3% 97% to 100%

Therefore, with the confidence level of 99%, the VaR value is USD 17 million because 99% falls

the range of 97% and 100% cumulative probability range.

Note that if we reduce our confidence level to 95%, VaR will change to USD 13 million because

95% is in the range of 75% to 97% cumulative probability range.

However, if the confidence level is 97%, then we could have two VaR values: USD 13 million and

USD 17 million. This will be ambiguous, and so the best estimate is the average of the values,

which is USD 15 million.

Limitations of VaR

20
© 2014-2020 AnalystPrep.
I. It does not describe the worst possible loss. Indeed, as seen from the example above,

we would expect the $100 million loss mark to be breached 5 times out of a hundred for

a 95% confidence level.

II. VaR does not describe the losses in the left tail. It indicates the probability of a value

occurring but stops short of describing the distribution of losses in the left tail.

III. Two arbitrary parameters are used in its calculation – the confidence level and the

holding period. The confidence level indicates the probability of obtaining a value greater

than or equal to VaR. The holding period is the time span during which we expect the

loss to be incurred, say, a week, month, day, or year. VaR increases at an increasing rate

as the confidence level increases. VaR also increases with increases in the holding

period.

IV. VaR estimates are subject to both model risk and implementation risk. Model risks arise

from incorrect assumptions, while implementation risk is the risk of errors from the

implementation process.

Expected Shortfall (ES)

Recall the VaR does not describe the worst possible loss. For instance, if 99% VaR is USD 10

million, we know that we are 1% certain that the loss will exceed USD 10 million. From the VaR

level, we cannot say that the loss is greater than 20 million or USD 50 million. Therefore, VaR

sets a risk measure equal to a certain percentile of the loss distribution and does not consider

the possible losses beyond the VaR level.

Expected shortfall (ES) is a risk measure that considers the expected losses beyond the VaR

level. In other words, ES is the expected loss conditional that the loss is greater than the VaR

level.

Exam tip: Expected shortfall is also called conditional value at risk (CVaR), average value at risk

(AVaR), or expected tail loss (ETL). Think about this as the average loss beyond the VaR.

When the losses are normally distributed with the mean μ and standard deviation σ, then the ES

is given by:

U2
21
© 2014-2020 AnalystPrep.
U2
⎛ − ⎞
⎜⎜ e 2 ⎟⎟
ES = μ + σ ⎜
⎜⎜ (1 − X) √2π ⎟⎟⎟
⎝ ⎠

Where

X = the confidence level; and

U = the point in the standard normal distribution that has a probability of X% of being exceeded.

Example: Calculating expected shortfall

The investment return over a period of time has a normal loss distribution with a mean of -200

and a variance of 300. What is 99% Expected of the loss distribution?

Solution

We know that:

U2
⎛ e− 2 ⎞
ES = μ + σ ⎜ ⎟
⎝ (1 − X) √2π ⎠

Now, U=Φ−1 (0.01) = −2.32

(2.32) 2
⎛ e− 2 ⎞

ES = −200 + 300 ⎜ ⎟⎟ = 611.44
⎝ (1 − 0.99) √ 2π⎠

ES should always be greater than the VaR level because the ES gives us the average of the

values that are in tail exceeding VaR.

Example: Calculating the ES for a discrete loss distribution

The loss distribution of investment is as shown below:

22
© 2014-2020 AnalystPrep.
Amount of Loss Probability
USD 20 Million 2%
USD 17 Million 8%
US 13 million 12%
US 10 million 78%

What is the 95% expected shortfall (ES)?

Solution

At a 95% confidence level, we need to answer the question, "given that we are at 5% of the loss

distribution, what is the value of the expected loss?".Now looking at the probability column, it is

clear to see that 5% tail distribution consists of a 2% probability that the loss is USD 20 million

and 3% that the loss is USD 17 million. Conditioned that we are dealing with tail distribution,

then there is a 2/5 chance that the loss is USD 20 million and 3/5 chance that the loss is USD 17

million. Therefore, the expected shortfall is given by:

2 3
× 20 + × 17 = 18.20
5 5

Again, note that the 97% VaR is USD 10 million, which is less than ES.

Properties of a Coherent Risk measure

A risk measure summarizes the entire distribution of dollar returns X by one number, ρ (X). There

are four desirable properties every risk measure should possess. These are:

I. Monotonicity: If X1 ≤ X2 , ρ (X1 ) ≥ ρ (X 2 )

Interpretation: If a portfolio has systematically lower values than another, it must have a

greater risk in each state of the world. In other words, if a portfolio gives undesirables

results than the others, then it must be riskier.

II. Subadditivity: ρ (X 1 + X2 ) ≤ ρ (X1 ) + ρ (X2 )

23
© 2014-2020 AnalystPrep.
Interpretation: When two portfolios are combined, their total risk should be less than (or

equal to) the sum of their risks. Merging of portfolios ought to reduce risk. This property

captures the implications of diversification. If two portfolios are perfectly correlated,

then the overall risk is the sum of their risk when considered separately. However, if the

two portfolios are not perfecdtly correlated, their overall risk should decrease due to

diversification benefits.

III. Homogeneity: ρ (kX) = kρ (X)

Interpretation: Increasing the size of a portfolio by a factor k should result in a

proportionate scale in its risk measure. For instance, if we increase the portfolio size by a

quarter, then the risk should be increased by a quarter.

IV. Translation invariance: ρ (X + h) = ρ (X) − h

Interpretation: Adding cash h to a portfolio should reduce its risk by h. Like X, h is

measured in dollars. This property reflects that more cash acts as a "loss absorber" and

can be taken as a replacement for capital.

If a risk measure satisfies all four properties, then it is a coherent risk measure. Expected

shortfall is a coherent risk, but VaR is not.

Why VaR is Not a Coherent Risk Measure

Value at risk is not a coherent risk measure because it fails the subadditivity test. Here's an

illustration:

Suppose we want to calculate the VaR of a portfolio at 95% confidence over the next year of two

zero-coupon bonds (A and B) scheduled to mature in one year. Assume that:

The current yield on each of the two bonds is 0;

The bonds have different issuers;

Each bond has a probability of 4% of defaulting over the next year;

The event of default in either bond is independent of the other; and

24
© 2014-2020 AnalystPrep.
The recovery rate upon default is 30%.

Given these conditions, the 95% VaR for holding either of the bonds is 0 because the probability

of default is less than 5%. Now, what is the probability 'P' that at least one bond defaults?

P = 0.04 × 0.96 + 0.96 × 0.04 + 0.04 × 0.04 = 7.84

The probability of at least one default is 7.84%, which exceeds 5%.

So if we held a portfolio that consisted of 50% A and 50% B, then the 95%

VaR = 0.7 × 0.5 + 0 × 0.5 = 35

This violates the subadditivity principle, and VaR is, therefore, not a coherent risk measure.

Spectral Risk Measures

A spectral risk measure is a risk measure given as a weighted average of return quantiles from

the loss distribution. It's given as a weighted average of outcomes where bad outcomes typically

have more weight. A spectral risk measure is always a coherent risk measure.

A risk measure is a coherent risk measure if and only if the weights are a non-decreasing

function of the percentile of the loss distribution. To demonstrate this, let w (p) be a weighted

given to the percentile, then a risk measure is coherent if:

w (p1 ) ≥ w (p 2 ) if p 1 > p2

VaR is also a particular case of risk spectrum measurement, but it places no weight on tail losses.

For instance, given a VaR confidence level of 95%, all the weight is concentrated on the 95th

percentile. If we denote p2 =95 % and p 1=98%, w (p2 ) = 1 and w (p 2 )=0 implying that rule above

has been isolated.

The expected shortfall is a particular case of risk spectrum measurement where the weighting

function is set to [1/(1−confidence level)] for tail losses, with all the other quantiles weighting

zero. In other words, no weight is assigned to losses that are less than the VaR level and that all

losses greater than VaR loss are given the same weight. Conclusively, ES is a coherent risk

25
© 2014-2020 AnalystPrep.
measure where the weights assigned to percentiles increase in a way that risk aversion is

captured. One method is to make the weights proportional to:

1− p

e λ

Where:

p is the percentile in decimal form; and

λ is a constant that reflects the degree of aversion for a user (decreases with an increase in risk

aversion).

26
© 2014-2020 AnalystPrep.
Practice Question

Ann Conway, FRM, has spent the last several months trying to develop a new risk

measure to appraise a set of defaultable zero-coupon bonds owned by her employer.

Prior to its use, her supervisor has asked her to demonstrate that it is a coherent risk

measure. The results are listed below:

Given:

X and y are state-contingent payoffs of two different bond portfolios.

P (x) and P (y) are risk measures for portfolio x and portfolio y.

K and l are arbitrary constants, with k > 0

Which of the following equations shows that Conway's risk measure is not coherent?

A. P (kx) = kP (x)

B. P (x) + P (y) ≥ P (x+y)

C. P (x) ≤ P (y) if x ≤ y

D. P (x+l) = P (x) − l

The correct answer is C.

Option C, as represented above, shows that the risk measure does not satisfy the

monotonicity property. Monotonicity requires that P (x) ≥ P (y) if x ≤ y. (If a portfolio

has systematically lower values than another, in each state of the world, it must have

a greater risk.)

Option A demonstrates that the measure satisfies the homogeneity property.

Option B demonstrates that the measure satisfies the subadditivity property.

Option D demonstrates that the measure satisfies the translation invariance

property.

27
© 2014-2020 AnalystPrep.
Reading 46: Calculating and Applying VaR

After completing this reading, you should be able to:

Explain and give examples of linear and nonlinear derivatives.

Describe and calculate VaR for linear derivatives.

Describe and explain the historical simulation approach for computing VaR and ES.

Describe the delta-normal approach for calculating VaR for nonlinear derivatives.

Describe the limitations of the delta-normal method.

Explain the full revaluation method for computing VaR.

Compare delta-normal and full revaluation approaches for computing VaR.

Explain structured Monte Carlo and stress testing methods for computing VaR, and

identify strengths and weaknesses of each approach.

Describe the implications of correlation breakdown for scenario analysis.

Describe the worst-case scenario (WCS) analysis and compare WCS to VaR.

Linear and Nonlinear Portfolios

A linear portfolio linearly depends on the changes in the values of its corresponding variables

(risk factors). For instance, consider a portfolio consisting of 100 shares, each valued at USD

100. Therefore, the change in portfolio value (ΔP) is attributed to change in stock (share price)

which can be denoted by ΔS, and thus the change in portfolio value is given by :

ΔP = 100ΔS

The value of the portfolio is USD 10,000 (=100×100). Now, if we introduce the effect of the

interest rate, the change in the value of the portfolio will be given by:

28
© 2014-2020 AnalystPrep.
ΔP = 100Δr

Generally, consider a portfolio consisting of long and short positions in stocks. The change in a

linear portfolio is given by:

ΔP = ∑ n iΔSi
i

Where:

ni = number of shares of stock i in the portfolio;

Si = the price of stock i.

Intuitively, the amount invested in stock i is ni Si , and the price change is Δn i .

Si
Now, if we multiply the above equation by , we have:
Si

Si ΔSi
ΔP = ∑ n iΔSi × = ∑ ni ΔSi
i Si i Si

ΔSi
Let qi = ni Si and Δri = , then
Si

ΔP = ∑ q iΔri
i

ΔS i
Note that q i is the amount invested in stock i, and Δri = is the return on stock i.
Si

Therefore, we can say the portfolio change is a linear function of change in stock price or change

in stock returns.

Nonlinear portfolios contain complex securities that are not linear. For instance, consider a

portfolio made of call options. The payoff from the call option is nonlinear because the payoff is

zero if the stock price at maturity is less than the strike price and S-K if the stock price is higher

than the strike price K.

29
© 2014-2020 AnalystPrep.
However, a forward contract is an example of a derivative whose value is a linear function of the

asset because even if the contracts do give a payoff. The holder is obligated to buy the asset at a

future time T at agreed price K. As such, the forward contract's value is given by:

S − PV (K)

Where S is the current asset price, and PV (K) denoted the present value of the future price K.

30
© 2014-2020 AnalystPrep.
VaR for Linear Derivatives

In general terms, the VaR of a linear derivative can be expressed as:

VaRlinear derivative = Δ × VaRUnderlying factor

Where Δ represents the sensitivity of the derivative's price to the price of the underlying asset. It

is usually expressed as a percentage.

Example: VaR for linear derivatives

Suppose the permitted lot size of S&P 500 futures contracts is 300 and multiples thereof. What is

31
© 2014-2020 AnalystPrep.
the VaR of an S&P 500 futures contract?

VaRS&P500futures c. = 300 × VaRS&P500 index

Calculating VaR and Expected Shortfall using Historical


Simulation

Historical simulation is used to calculate one-day VaR and ES. However, for longer periods T it is

assumed that,

VaR (T , X) = √T × VaR (1, X)


ES (T , X) = √T × ES (1, X)

VaR (T, X) = value at risk for a time horizon of T days and confidence level X.

ES (T, X) = expected shortfall for a time horizon of T days and confidence level X.

The above estimates assume that the portfolios' changes are normally distributed with a mean of

zero and independent of each other.

In a historical simulation, market variables (risk factors) on which the portfolio value depends.

Examples of such variables include commodity prices, equity prices, and volatilities. After this,

the data on the movement of these risk factors in the past is collected. After data collection,

scenarios are built by assuming that each risk factor's change over the next day corresponds to a

change observed during one of the past days.

Risk factors are broadly classified into:

i. Those whose past percentage change is to define the future percentage, for example,

stock prices and exchange rates; and

ii. Those whose past actual change is used to define an actual change in the future, for

example, interest rates and credit spreads.

Example: Illustration of Historical Simulation

32
© 2014-2020 AnalystPrep.
A portfolio is assumed to depend on many risk factors. For simplicity, let us assume three risk

factors (exchange rates, interest rates, and stock price) over the past 300 days (longer periods

are usually considered, such as 500 days). The most recent 301 days of historical data is as

follows:

Day Stock Price Exchange Interest Portfolio


(USD) rates rate (%) Value
(USD/CAD) (USD
millions)
0 30 1.3901 3.51 60.0
1 34 1.4000 2.64 62.5
2 40 1.3921 2.52 61.25
.. .. .. .. ..
298 42 1.3876 2.40 65.0
299 40 1.3910 2.45 60.25
300 60 1.4021 2.50 71.25

Assuming that today is the 300th day, we need to know what will happen between today and

tomorrow (301st day). To achieve this, we use the above data to create 300 scenarios (that is why

we have 301-day historical data).

In the first scenario, we will assume that the risk factors behave between the days 300 and 301

in a similar manner as they did between days 0 and 1. For instance, in the first scenario, the
34
stock price increased by 13% (= − 1), and thus the stock price on day 101 is USD 68
30

[= ( 34 1.400
− 1) × 60] . For the interest rate, it increased by 0.7% (= − 1) , and thus, we expect the
30 1.3901

exchange rate for the 101st day to be 1.4121 (= 1.4021 [ 1.400 − 1]). For the interest rate, it
1.3901

decreased by 0.87% (2.64%-3.51%) and thus 101st interest rate is 1.63% (2.50%-0.87%).

The values for the second and subsequent scenarios' risk factors are calculated similarly as the

first scenario. For the second scenario, assume that the risk factors behave in a similar manner

as they did between days 1 and 2. This will create the following table.

33
© 2014-2020 AnalystPrep.
Scenario Stock Price Exchange Interest Portfolio Loss
(USD) rates rate (%) Value
(USD/CAD) (USD
millions)
1 68 1.4121 3.37 70.25 1.0
2 71 1.3922 2.38 72.15 0.9
.. .. .. .. ..
299 57.14 1.3910 2.55 71.25 0
300 90 1.4021 2.55 73.25 2.0

The risk factor values for the scenario table are directly calculated from the historical data table.

For the scenario portfolio values, we must generate based on the risk factors. We assume that

the current portfolio value is USD 71.25 million (300th day's value). After generating the

portfolio values, we then calculate the losses while attaching a negative to create a loss

distribution.

Assume that the first scenario's portfolio value is 70.25, 72.15 for the second scenario, and so on.

Therefore, the loss for the first portfolio is 1.00 (=71.25-70.25), and the second scenario is 0.9

(71.25-72.15), and so on.

In order to calculate the VaR and the expected shortfall, we ought to arrange the scenario losses

from the largest to the smallest. Assume that in our example, we wish to calculate one day VaR

and ES at a 99% confidence interval. The sorted losses are as follows:

Scenario Loss
200 3.9
10 3.0
25 2.5
100 2.0
.. . .. .
.. . .. .

In this case, VaR is equivalent to third-worst loss since the third-worst loss is the first percentile
3
point of the distribution, i.e., = 0.01. Therefore, VaR=2.5 million.
100

By definition, the expected shortfall is calculated as the average of the losses that is worse than

the VaR. In this case,

1
34
© 2014-2020 AnalystPrep.
1
ES = (3.9 + 3.0) = 3.45 million
2

Example: Calculating VaR using the historical simulation method

The following are hypothetical ten worst returns for an asset B from 120 days of data for 6

months. Find the 1-day 5% VaR and the ES for B.

-3.45%, -14.12%, -15.72%, -10.92%, -5.50%, -3.56%, -6.90%, -2.50%, -5.30%, -4.31%.

Solution

First, we rearrange starting with the worst day, to the least bad day, as shown below:

-15.72%, -14.12%, -10.92%, -6.90%, -5.50%, -5.30%, -4.31%, -3.56%, -3.45%, -2.50%.

The VaR corresponds to the (5% × 120)=6th worst day = -5.30%. However, recall that VaR need

not be represented as a negative.

This implies that there is a 95% probability of getting at most 5.3% loss.

The expected shortfall (ES) is calculated as the average of the losses that is worse than the VaR.

In this case,

(15.72% + 14.12% + 10.92% + 6.90% + 5.50%)


ES = = 10.63%
5

Full Revaluation VaR

Under the full revaluation approach, the VaR of a portfolio is established by fully repricing the

portfolio under a set of scenarios over a period of time.

Revaluation of a portfolio is a time-consuming activity. One approach to address this challenge is

to use Greek letters. The Greek letters are the hedging parameters used by the analysts to

quantify and manage risks.

35
© 2014-2020 AnalystPrep.
One of the crucial Greek letter deltas (δ) which is defined as:

ΔP
δ=
ΔS

Where ΔS is a small change in risk factors such as stock price and ΔP the corresponding change

in the portfolio value. Therefore, delta can be defined as a change of the portfolio value with

respect to the change in the risk factor.

For instance, consider stock price as a risk factor. If the delta of a portfolio with respect to the

stock price is USD 100, it implies that the portfolio value changes by USD 100 if the stock price

changes by 1 USD.

From the delta formula, we express the change in the portfolio value as :

ΔP = δΔS

Generally, if we have multiple risk factors, we find each risk factor's effect and sum it up. That is,

if we have i risk factors, then change in the portfolio would be:

ΔP = ∑ δi ΔSi
i

However, the delta concept gives relatively accurate estimates in linear portfolios as compared

to nonlinear portfolios.

The accuracy of nonlinear portfolios can be enhanced by including another Greek letter gamma

(γ) so that the delta-gamma formula is given by:

1
ΔP = δΔS + γ(ΔS)2
2

In case of multiple risk factors, the above equation changes to:

1 2
ΔP = ∑ δi ΔSi + ∑ γi (ΔSi)
i 2 i

36
© 2014-2020 AnalystPrep.
Risk Factors Expressed as Term Structures

We have considered risk factors such as stock and prices, which are expressed with a single

number and hence easy to manage. However, some of the risk factors, such as interest rates,

credit spreads, and at-the-money implied volatility are described using term structures.

Interest rates such US treasury rate are usually given in points such as a one-month rate, three-

month rate, and so on. Linear interpolation is usually used to approximate the rates between

these points.

Credit spreads are excess of the borrowing rate of a company over the set limit (benchmark

level). Credit spreads are stated in terms of basis points. As such, the credit spread for a

company might be stated as 100-basis points for a 2-year maturity and 200-basis points for a 6-

year maturity. Respectively, this implies that a firm will pay an extra 1% more than the risk-free

rate in case it borrows for two years and 2% more for six years.

At-the-money implied volatility is also expressed by the term structure. At the money implied

volatility is applied in the Black-Scholes Merton formula for option valuation.

Using Stressed Measures

We usually use recent data to develop scenarios to calculate VaR and ES. Nevertheless, bank

regulators have suggested the stressed VaR and stressed ES to evaluate the capital requirements

of a bank.

Stressed VaR and stressed ES are calculated from one-year data that is particularly stressful to

the bank's current portfolio. For instance, a stressful period that might be considered in the

2007-2008 financial crisis.

The scenarios are created from the percentage change or actual changes from a stressful period.

Therefore, 250 scenarios are generated from the one-year data. All the calculations in computing

traditional VaR and ES remain the same.

The Delta-Normal Model

37
© 2014-2020 AnalystPrep.
The delta-normal model is based on the equation (as seen earlier):

ΔP = ∑ δi ΔSi
i

Recall this equation gives an exact value in a linear portfolio and an approximate value in

nonlinear portfolios.

Also, recall that we have two types of risk factors:

1. Those whose past percentage change is to define the future percentage; and

2. Those whose past actual change is used to define an actual change in the future.

In order to accommodate both types of risk factors, the equation above is written as:

38
© 2014-2020 AnalystPrep.
ΔP = ∑ aix i
i

ΔS i
For risk factors where percentage changes are used, ai = and x i = δiSi .
Si

And for the risk factors where actual changes are considered ai = ΔSi and x i = δi.

From the resulting equations, the mean and the standard deviation of the change in portfolio

value can be calculated as:

n
μ P = ∑ aiμ i
i=1
n n
σP2 = ∑ ∑ ai aj ρij σi σj
i=1 j=1

Where:

μi and σi are the mean and standard deviation of x i , respectively; and

ρij is the coefficient of correlation between xi and x j.

The formula for the standard deviation can be written as:

σP2 = a2i σi2 + 2 ∑ ai aj ρij σiσj


i>j

Assuming that the portfolio changes are normally distributed, then we can comfortably compute

VaR and ES. Recall that the VaR is given by:

VaR = μ P + σP U
U2
⎛ e− 2 ⎞
ES = μ P + σP ⎜ ⎟
⎝ (1 − X) √2π ⎠

Where:

X = confidence level

U = point on the normal distribution where X is exceeded

39
© 2014-2020 AnalystPrep.
For instance, if X=95% then U=Φ−1 (0.05) = −1.645.

At this point, you might guess where the name "delta-normal" name comes from: the model uses

deltas of the risk factors and assumes that the portfolio changes are normally distributed.

A typical assumption is that the mean change in the risk factor is zero. This assumption is

sometimes not reasonable but is useful when dealing with short time periods because the mean

is less than the standard deviation for short periods when dealing with portfolio value changes.

As such, VaR and ES are given by:

VaR = σP U
U2

⎛ e 2 ⎞
ES = σP ⎜ ⎟
⎝ (1 − X) √2π ⎠

Limitations of Delta-Normal Model

The method has several disadvantages, chief among them being that:

It is computationally easy but quite inaccurate compared to other VaR measurement

methods. Put more precisely, it may underestimate the occurrence of extreme losses

because it relies on the normal distribution.

This method is accurate for small moves of the underlying, but quite inaccurate for

large moves. As we can see from the following graph, the slope of the green line is

entirely different from the slope of the blue line:

40
© 2014-2020 AnalystPrep.
For large changes in a nonlinear derivative, we must use the delta + gamma approximation, or

full revaluation.

Using Monte Carlo Simulation to Calculate VaR and ES

Monte Carlo approach is similar to that of the historical simulation, but Monte Carlo simulation

produces scenarios by randomly selecting samples from the distribution assumed for the risk

factors instead of historical data. Monte Carlo simulations work for both linear and nonlinear

portfolios.

Now, if for instance, we assume that risk factor changes have a multivariate normal distribution

(as in delta-normal), Monte Carlo procedure is as follows:

41
© 2014-2020 AnalystPrep.
Step 1: Calculate the value of the portfolio today using the current values of the risk factors.

Step 2: Sample once from the multivariate normal probability distribution of Δxi . Sampling

should be consistent with the assumed standard deviations and correlations, which are usually

approximated from historical data.

Step 3: Using the sample values of Δx i , determine the values of the risk factors at the end of the

period being considered (such as one day).

Step 4: Revalue the portfolio using these new risk factor values.

Step 5: Subtract the revaluated portfolio from the current portfolio value to determine the loss.

Step 6: Repeat step 2 to step 5 multiple times to come up with a probability distribution for the

loss.

For instance, a total of 500 trials are conducted in a Monte Carlo simulation, then 99% VaR for

the period under consideration will be the fifth-worst loss, and thus the expected shortfall will be

the average of the four losses worse than VaR.

Like other approaches, Monte Carlo simulation computes one-day VaR, and thus the following

equations apply when we want to compute T-day time horizon VaR and ES:

VaR (T,X) = √ T × VaR (1, X)


ES (T,X) = √ T × ES (1, X)

Monte Carlo simulation is slow because it is computationally intensive. This can be explained by

the fact that portfolios considered are usually huge, and evaluating each one of them in each

trial is quite time-consuming. To address this challenge, the delta-gamma approach can be used

(as discussed earlier) to determine the change in the portfolio value. This is called partial

simulation.

Delta-normal model assumes normal for the risk factors. However, Monte Carlo simulation uses

any distribution for the risk factors only if the correlation between the risk factors can be

defined.

Estimating the Parameter Values

42
© 2014-2020 AnalystPrep.
To execute the Monte Carlo simulation or delta-normal model, an approximation of the standard

deviations and correlations of either percentage or actual changes of the risk factors is

necessary. The approximation of these parameters is made using recent historical data. More

weights can be applied to more recent data using models such as GARCH (1,1), which we will

see in the following chapter.

In the case of stressed VaR and ES, standard deviations and correlations should be estimated

from the past period, which would be considered stressful to the current portfolio.

Correlation Breakdown

During the stressed market conditions, standard deviations increase as well as correlations. This

phenomenon was witnessed during the 2007-2008 financial crisis, where default rates of

mortgages increased all over the US.

Therefore, correlations in a high volatility period are quite different from those of normal market

conditions. Thus, when calculating VaR or ES, risk managers might need to determine what will

happen in extreme market conditions.

Stress Testing

In finance, contagion is the spread of an economic shock in an economy or region, so that price

movements in one market are gradually but increasingly noticed in other markets. During a

contagion, both volatility and correlations increase, rendering diversification less effective as a

risk mitigation strategy. Stress testing is an attempt to model the contagion effect that would

occur in the event of a major crisis.

Some of the historical events that have been used to stress test by various firms include the

Mexican crisis of 1994, the Gulf war of 1990, and the near-collapse of LTCM in 1998. The

underlying question among analysts while stress testing is quite simple: If a similar event

occurred right now, how would it affect the firm's current position?.

43
© 2014-2020 AnalystPrep.
The other approach to stress testing may not necessarily turn to historical events. It can also be

done using predetermined stress scenarios. For example, the firm could seek to find out the

effect on its current positions if interest rates short up by, say, 300 bps.

Strength:

Assumptions regarding the underlying distribution are not needed.

Weakness:

Stress testing based on history is limited to a few past events; therefore, there is a cap

on the usefulness of such a strategy. In fact, a constant historical event problem is that

obtaining accurate data on the exact happenings can be a tough ask. Furthermore, as

is the case with all past events, there can be no guarantees that what happened back

then will happen in the future.

Worst-case Scenario Analysis

Worst-case scenario analysis focuses on extreme losses at the tail end of the distribution. First,

firms assume that an unfavorable event is certain to occur. They then attempt to establish the

worst possible outcomes that could come out of it.

WCS analysis dissects the tail further to establish the range of worst-case losses that could be

incurred. For example, within the lowest 5% of returns, we can construct a "secondary"

distribution that specifies the 1% WCS return.

WCS analysis complements the VaR, and here is how. Recall that the VaR specifies the minimum

loss for a given percentage, but it stops short of establishing the severity of losses in the tail.

WCS analysis goes a step further to describe the distribution of extreme losses more precisely.

44
© 2014-2020 AnalystPrep.
Practice Questions

Question 1

Which of the following statements about stress testing are correct?

I. A weakness of stress testing is that it is highly subjective

II. Stress testing can complement VaR estimation by helping managers establish

crucial vulnerabilities in a portfolio.

III. The larger the number of test scenarios, the better the understanding of the

risk exposure of a portfolio

IV. Stress testing allows users to include scenarios that did not occur in the

lookback horizon of the VAR data but are nonetheless possible.

A. All of the above

B. I, II, and IV

C. I, II, and III

D. Only II and IV

The correct answer is B.

In stress testing, the fewer the scenarios under consideration, the easier it is to

understand and interpret results. Too many scenarios make it difficult to interpret the

risk exposure.

Question 2

A risk manager wishes to calculate the VaR for a Nikkei futures contract using the

historical simulation approach. The current price of the contract is 955, and the

multiplier is 250. For the last 300 days, the following return data have been recorded:

-7.8%, -7.0%, -6.2%, -5.2%, -4.6%, -3.2%, -2.0%, …, 3.8%, 4.2%, 4.8%, 5.1%, 6.3%,

6.8%, 7.0%

45
© 2014-2020 AnalystPrep.
What is the VaR of the position at 99% using the historical simulation methodology?

A. $12,415

B. $16,713

C. $18,623

D. $14,803

The correct answer is D.

The 99% return among 300 observations would be the third-worst observation among

the returns (0.99 × 301 = 297).

Among the returns given above, the third-worst return is −6.2%. As such,

VaR(99%) = 6.2% × 955 × 250 = 14 , 802.50

A is incorrect. This answer incorrectly uses the fourth-worst observation as the 99%

return among 300 observations.

B is incorrect. This answer incorrectly uses the second-worst observation as the 99%

return among 300 observations.

C is incorrect. This answer incorrectly uses the worst observation as the 99% return

among 300 observations.

Question 3

Bank X and Bank Y are two competing investment banks that are calculating the 1-

day 99% VaR for an at-the-money call on a non-dividend-paying stock with the

following information:

Current stock price: USD 100

Estimated annual stock return volatility: 20%

Current Black-Scholes-Merton option value: USD 4.80

46
© 2014-2020 AnalystPrep.
Option delta: 0.7

To compute VaR, Bank X uses the linear approximation method, while Bank Y uses a

Monte Carlo simulation method for full revaluation. Which bank will estimate a

higher value for the 1-day 99% VaR?

A. Bank X

B. Bank Y

C. Both will have the same VaR estimate

D. Insufficient information to determine

The correct answer is B.

The option’s return function is convex with respect to the value of the underlying;

therefore the linear approximation method will always underestimate the true value

of the option for any potential change in price. Therefore the VaR will always be

higher under the linear approximation method than a full revaluation conducted by

Monte Carlo simulation analysis. The difference is the bias resulting from the linear

approximation, and this bias increases in size with the change in the option price and

with the holding period.

47
© 2014-2020 AnalystPrep.
48
© 2014-2020 AnalystPrep.
Reading 47: Measuring and Monitoring Volatility

After completing this reading, you should be able to:

Explain how asset return distributions tend to deviate from the normal distribution.

Explain reasons for fat tails in a return distribution and describe their implications.

Distinguish between conditional and unconditional distributions.

Describe the implications of regime-switching on quantifying volatility.

Evaluate the various approaches for estimating VaR.

Compare and contrast different parametric and non-parametric approaches for

estimating conditional volatility.

Calculate conditional volatility using parametric and non-parametric approaches.

Evaluate implied volatility as a predictor of future volatility and its shortcomings.

Apply the exponentially weighted moving average (EWMA) approach and the GARCH

(1,1) model to estimate volatility.

Explain and apply approaches to estimate long-horizon volatility/VaR and describe the

process of mean reversion according to a GARCH (1,1) model.

Calculate conditional volatility with and without mean reversion.

Describe the impact of mean reversion on long-horizon conditional volatility estimation.

Describe an example of updating correlation estimates.

Constant volatility is easily approximated from historical data. However, volatility varies through

time. Therefore, an alternative to constant normality of asset returns is to assume that asset

returns are normally distributed conditioned on a known volatility. More specifically, high

volatility indicates that the daily asset return is normally distributed with high standard

deviation, and when the volatility is low, the daily returns are normally distributed with low

49
© 2014-2020 AnalystPrep.
standard deviation.

Monitoring volatility is made possible through two methodologies (as will be covered later in this

chapter): the exponentially weighted moving average (EMWA) model and he GARCH (1,1) model.

We will also apply the estimation of the volatility and application of volatility monitoring to

correlation.

Deviation of Asset Returns from Normal Distribution

There are several reasons as to why the normality framework has been adopted for most risk

estimation attempts. For starters, the normal distribution is relatively easy to implement. We

need two parameters – the mean, μ, and the standard deviation, σ, of returns. With these two, it

is possible to characterize the distribution fully and even come up with Value at Risk measures at

specified levels of confidence. In the real world, however, asset returns are not normally

distributed, and this can be observed empirically. Asset return distributions exhibit several

attributes of non-normality. These are:

1. Fat Tails (Negative Skewness): A defining characteristic of the normal distribution is

that most data points are concentrated around the center (mean), with very few points at

the tails. The distribution is symmetrical (has equal and opposite halves) with outliers

that consistently diminish, resulting in smooth, narrow tails.

50
© 2014-2020 AnalystPrep.
2. Instability of Parameter Values: Unlike the normal distribution, asset return

distributions have unstable parameters (mean and standard deviation). This tendency

arises from market conditions that are continually changing. This instability is evident in

asset volatility where asset returns turn out to be way more volatile than predictions

using the normal distribution would suggest.

3. The Asset Returns can be Non-symmetrical: As seen in the figure below, the equity

returns lean further to the right than the normal distribution. This phenomenon is known

as negative skewness. A direct consequence of negative skewness is that the left slope of

equity returns is longer than the left slope of the normal distribution, indicating a greater

magnitude of extreme negative events.

51
© 2014-2020 AnalystPrep.
Fat Tails and their Implications

For this reading, it’s essential to keep in mind that the term ‘fat tails’ is relative: it refers to the

tails of one distribution relative to the normal distribution. If a distribution has fatter tails

relative to the normal distribution, it has a similar mean and variance, but probabilities at the

extremes are significantly different.

When modeling asset returns, analysts focus on extreme events – those that have a low

probability of occurrence but resulting in catastrophic losses. There’s no point concentrating on

non-extreme events that would reasonably not be expected to cause severe losses.

If we were to assume that the normal distribution holds as far as asset returns are concerned, we

would not expect not even a single daily move of 4 standard deviations or more per year. The

normal distribution predicts that about 99.7% of outcomes lie within 3 standard deviations from

the mean.

52
© 2014-2020 AnalystPrep.
In reality, every financial market experience one or more daily price moves of 4 standard

deviations or more every year. There’s at least one market that experiences a daily price move of

10 or more standard deviations per year. What then does this suggest?

The overriding argument is that the true distribution of asset returns has fatter tails.

Implications of Fatter Tails

There is a higher probability of extreme events than the normal distribution would suggest. This

means that reliance on normality results in inaccurately low estimates of VaR. As a result, firms

tend to be unprepared for such tumultuous events and are heavily affected, sometimes leading to

foreclosure and/or large-scale ripple effects that reverberate and destabilize the entire financial

system.

Unconditional and Conditional Normality

Unconditional normal distribution manifests when the mean and standard deviation of asset

returns in a model is the same for any given day, regardless of market and economic conditions.

In other words, even if there is information about the distribution of asset returns that suggests

the existence of different parameters, we ignore that information and assume that the same

distribution exists on any given day.

Conditionally normal distribution occurs when a model contains asset returns that are

normal every day, but the standard deviation of the returns varies over time. That is the standard

deviation high at one in times and low at some other time.

The underpinning idea here is that if we collect daily return data, we will observe the

unconditional distribution and that by monitoring volatility, we can estimate the conditional

distribution for daily return. For instance, consider that from the (historical) data we have

collected, we estimate volatility to be 0.5% per day, but from volatility monitoring, we estimate

the volatility to be 1.5% per day. Therefore, it will be more accurate to assume that asset returns

are normally distributed with a 1.5% standard deviation and thus use this 1.5% measure to

calculate VaR and the Expected Shortfall (ES). In other words, results from monitoring volatility

are better than using the results from the fat-tailed distribution or by assuming a normal

53
© 2014-2020 AnalystPrep.
distribution with volatility from the observed data.

Various Approaches for Estimating VaR

Approaches for estimating the value at risk can be broadly categorized as either historical-based

or implied-volatility-based.

Historical-based Approaches

Historical-based approaches can be further subdivided into three categories: parametric, non-

parametric, and hybrid.

1. The parametric approach: The parametric approach is characterized by specific

distribution assumptions on conditional asset returns. In most cases, it is assumed that

returns are normally or lognormally distributed with volatility that varies with time.

2. The non-parametric approach: The non-parametric approach differs from the

parametric approach in that it has no assumptions regarding the distribution of asset

returns. It uses historical data directly. A good example of non-parametric models is a

historical simulation.

3. Hybrid approach: A hybrid approach borrows from both parametric and non-parametric

models. A result is a wholesome approach that still makes use of historical data.

The Implied-volatility-based Approach

Whereas historical approaches use historical data to gauge the volatility of an asset, the implied

volatility approach is based on current market data, i.e., it is forward-looking, which is its most

significant advantage. It is informed by the idea that historical events are not always very

accurate predictors of future events.

That said, it is essential to note just like historical approaches; this method does not guarantee

certain results. Implied volatility is a product of probabilities and other factors such as current

investor sentiments, and therefore there can be no assurances that market prices will follow the

54
© 2014-2020 AnalystPrep.
predicted pattern. Another disadvantage of the model is that it is model-dependent.

Nonparametric vs. Parametric VaR methods

The advantages of non-parametric models over parametric models are:

They do not require any assumptions regarding the distribution of returns to estimate

VaR.

Multivariate density estimation allows for weights to vary to reflect data relevance of

the data in light of current market conditions, regardless of the timing of the data.

Problems posed by Fat tails, skewness, and other deviations from the assumed

distribution are avoided.

Multivariate density estimation exhibits much flexibility in introducing dependence on

economic variables.

The disadvantages of non-parametric approaches compared to parametric models are:

Subdividing the full sample data into different market regimes reduces the amount of

data available for historical simulations.

Their use of data is less efficient compared to parametric models.

Multivariate density estimation requires a large amount of data that has a direct

relationship with the number of conditioning variables incorporated in the model.

Multivariate density estimation may lead to data snooping or overfitting when working

to come up with weighting scheme assumptions, as well as the number of observations

required to produce reliable volatility estimates.

The Implications of Regime Switching on Quantifying


Volatility

In an attempt to model asset returns better, analysts may subdivide a specified time period into

55
© 2014-2020 AnalystPrep.
regimes, where each regime has a clearly noticeable set of parameters that are markedly

different from those of other regimes. For example, volatility could rise sharply during a market

crash only to stabilize when conditions improve. This is the idea behind the regime-switching

volatility model.

For illustration, let us use real interest rates of a developed nation between 1990 and 2015:

Given the graph above, it would be difficult to identify different states of the economy. Now,

suppose we make use of an econometric model to try and identify the different economic states.

From the econometric model above, we can identify three distinct states of the economy. These

are precisely what we would call regimes. As we switch from one regime to another, at least one

56
© 2014-2020 AnalystPrep.
parameter has to change. So, how exactly does the regime-switching model help to better

measure volatility?

The conditional distribution of returns is always normal with either low or high volatility but a

constant mean. The regime-switching model captures the conditional normality and, in so doing,

helps resolve the fat tails problem.

Measuring Volatility

Conventionally volatility is defined as a change of a variable value over a period of time. In the

context risk management, volatility is defined as the standard deviation of an asset return in one

day.

The assets return in a day i is defined as:

Si − Si −1
ri =
Si−1

Where

ri: return on a day i.

Si : value of an asset on a close of trading i.

Si−1 : value of an asset on a close of trading i-1.

For a day n, the variance (σn2 ) from the previous m days is given by

1 m
σn2 = ∑ (rn−i − r̄)2
m − 1 i=1

And thus the standard deviation is given by:

1 m
2
σn = ⎷( ∑ (rn−i − r̄) )
m − 1 i=1

57
© 2014-2020 AnalystPrep.
Where:

r̄: average return from the previous m-days defined as:

1 m
r̄ = ∑ rn−i
m i=1

The risk management simplifies the above formulas by:

Substituting m-1 by m: This is justified in that, by using m-1, we obtain an unbiased

estimate of volatility, but when we use m, we get the maximum likelihood estimator,

which is most likely given any data.

Assuming that r̄ = 0: For a short period data, the standard deviation is far more

significant than the mean return, and we need expected return and not the historical

mean.

Therefore, defining the variables as before, the formula for the standard deviation changes to:

1 m
σn2 = ∑ r2n−i
m i =1

And thus volatility is given by:

1 m
σn = ⎷( ∑ r2n−i )
m i =1

The square of the volatility is defined as the variance rate, which is typically the mean of

squared returns.

Example: Calculating the volatility of asset returns

The asset returns over five days are 10%, -5%. 6%, -3%, and 12%. What is the volatility of the

asset returns?

58
© 2014-2020 AnalystPrep.
Solution

The volatility of the asset returns is given by:

1 m
σn = ⎷( ∑ r2n−i )
m i =1

Now,

5
∑ r2n−i = 0.12 + (−0.05)2 + 0.062 + (−0.03)2 + 0.12 2 = 0.0314
i=1

1
⇒ σn = ⎷ (0.0314) = 0.07925 = 7.925%
5

An alternative method of calculating volatility is to use absolute returns rather than squared

returns when calculating volatility. This method is suitable when dealing with non-normal data

because it provides an appropriate prediction for fat-tailed distribution. However, the commonly

used method uses squared returns, which we have considered and continue to do.

Approximating the Current Volatility

For the effective use of the conditional normal model, it is essential to estimate current volatility.

Current volatility can be estimated using the formula

1 m
σn = ⎷( ∑ r2n−i )
m i =1

For large values of ?, it would not capture current volatility due to variations of the volatility over

the period when data was collected. We can argue that the small data sample might be relatively

reliable, but the resulting estimate may not be accurate due to small data.

Alternatively, we can incorporate the effect of the standard error of the estimate. Recall that the

standard error of an estimate is the difference between the volatility estimate and the actual

59
© 2014-2020 AnalystPrep.
value. The standard error is estimated volatility divided by 2(m-1) where m is the number of

observations. That is:

1 1 m
S.E.Eσn = ( ∑ r2 )
2 (m − 1)⎷ m i=1 n−i

For instance in our example above,

1
S.E.Eσn = (0.07925) ≈ 0.01
2 (4)

We can then improve the accuracy of our volatility estimate by computing the confidence

intervals. Typically, confidence intervals are usually two standard errors from the estimated

value. In our case, the confidence interval is 5.925% to 9.925% (=0.07925± 2×0.01) per day. We

can reduce the confidence interval’s width by increasing the size of the observation, but we will

still face the disadvantage of the variability of the asset return volatility.

To solve this problem, we use a technique called exponential smoothing, also called an

exponentially weighted moving average (EWMA) used by RiskMetrics to estimate volatilities for

a wide range of market variables. Also, we use GARCH (1,1) as an exponential smoothing

technique.

Exponentially Weighted Moving Average (EWMA)

1 1
Recall that from the formula σn2 = m ∑ m 2
i=1 r n−i , equal weight ( m ) is applied to the squared returns.

However, in EWMA, the weights given to the squared returns are not equal and must sum up to

1.

Weight in EWMA is defined as weight applied to squared return from the k days ago denoted by

λ multiplied by weight applied to squared return from k-1 days ago. λ is a constant positive and

must be less than 1.

For example, denote the weight from the recent squared return, that is, squared return from day

60
© 2014-2020 AnalystPrep.
n-1 by w0 . Therefore, the weight on day n-2 is λw0 , and for the squared return for day n-3 is λ2

w 0 and so on. Consider the following table for easy understanding.

Day Squared Return Weight


n-1 r2n−1 w0
n-2 r2n−2 λw0
n-3 r2n−3 λ2 w 0
⋯ ⋯ ⋯

The suitable value of λ is the one that leads to an estimate that gives the lowest error. Now

assume that we have K days of data. Then, the sum of the weight applied is given by:

w 0 + w 0 λ + w 0λ 2 + ⋯ + w 0 λk−1
2 k−1
=w 0 (1 + λ + λ + ⋯ + λ )

Practically, EWMA can be applied to data spanning one to two years, but data from a long time

ago carries less weight; thus, we can assume that data that spans to infinite past is the result of

the total weight :

w0 + w 0 λ + w0 λ 2 + w 0 λ3 + ⋯

The above expression can be compressed to:


w 0 ∑ λk
i=0

The summation above in an infinite series so that:

∞ 1
∑ λk =
i =0 1 −λ
∞ 1
⇒ w 0 ∑ λk = w0 ( )
i=0 1 −λ

However, as stated earlier, the sum of weights in EWMA must be one. Therefore,

61
© 2014-2020 AnalystPrep.
1
w0 ( )=1
1−λ
∴ w0 = 1 − λ

By definition of the EWMA model, the estimated volatility on day n is computed by applying the

weights to past squared returns. More specifically,

σn2 = w0 r2n−1 + w 0 r2n−2 + w 0 r2n−3 + ⋯ (Eq1)

Intuitively, the estimated volatility on day n-1 is given by:

2
σn−1 = w0 r2n−2 + w 0 r2n−3 + w 0 r2n−4 + ⋯ (Eq2)

If we substitute Eq2 in Eq1 we get:

σn2 = w0 r2n−1 + λσn−1


2

Bur, we know that, w 0 =1-λ. Therefore,

σn2 = (1 − λ) r2n−1 + λσn−1


2

The last formula results from the EWMA model, which gives the estimate of the variance rate on

day n is the weighted average of the estimated variance rate for the previous day (n-1) and the

most recent (n-1) observation of the squared return. The equation

σn2 = (1 − λ) r2n−1 + λσn−1


2

is termed as adaptive volatility because it incorporates prior information about the volatility into

the new information. One merit of the EWMA model is that less data is required once the EWMA

model has been built for a particular market variable. Only recent volatility estimates need to be

remembered. Moreover, no return history is required.

Example: Calculating volatility using the EWMA model

Assume that you estimate recent volatility to be 3%, with a corresponding return of 2%. Given

62
© 2014-2020 AnalystPrep.
that λ=0.84, what is the value of new volatility?

Solution

Based on the EWMA model, the variable rate is given by:

σn2 = (1 − λ) r2n−1 + λσn−1


2

= (1 − 0.84) × 0.022 + 0.84 × 0.032


= 0.00082
⇒ σn = √0.00082 = 0.02864 = 2.864%

Determining the Value of λ

RiskMetrics determined the value of λ to be 0.94. This estimate, however, may not be

appropriate for today’s use, nor values higher than 0.94. Choosing a higher value of λ makes

EWMA less responsive to new data. For instance, let λ=0.99. From the equation

σn2 = (1 − 0.99) r2n−1 + 0.99σn−1


2

it implies that the new volatility is assigned a weight of 99%, and the new squared return is

assigned a weight of 1%. Clearly, for a number of days of high volatility will not change the value

of volatility significantly. Moreover, using a small value of λ, such 2% will make the volatility

estimate to overreact to new information.

There are two methods of determining the value of λ. One of them is to compute realized

volatility for a given day using 20 or 30 days of subsequent returns and then determine the value

of λ that minimizes the difference between the estimated volatility and the realized volatility.

The second method is the maximum likelihood method, where the λ that maximizes the

probability of observed data occurring is determined. For instance, if a trial leads to a value of λ

that results in predicting low volatility for a particular day, but the observed data is large, then

the estimated value of λ is the best estimate.

Historical Simulation

63
© 2014-2020 AnalystPrep.
Similar to EWMA, exponentially declining weights are applied when using historical simulation.

Exponentially declining weights are appropriate for determining VaR or expected shortfalls from

recent preceding observations because scenarios from the immediate data are more relevant

than those from a long time ago.

While weighting in historical simulation, the scenarios are arranged from the largest to the

smallest loss, after which the weights are accumulated to determine the VaR.

Other Weighting Schemes

Recall that in the EWMA model, the weights were exponentially declining. An alternative

methodology is multivariate density estimation (MDE). In MDE, an evaluation is carried out to

determine which periods in the past bear the same feature as the current period, after which

weights are assigned to the day’s historical data depending on the level of similarity of that day

to the present day.

For instance, the volatility of an interest rate varies depending on the level of the interest rates:

decreases when the interest rates increases and vice versa. We can calculate the volatility of an

interest rate by assigning the weights to interest data that is similar to the present interest rates

and decreasing the weight as the difference between past and today’s interest rate decreases.

Determining the level of similarity between one period and another is done using conditioning

variables. For example, while determining the interest rate weights, we could use GDP growth

rates as conditioning variables for the interest rate volatility. For conditional variables

X1 , X2 , … , Xn the similarity between today and the previous period is determined by calculating

the measure:

n 2
^i − X* )
∑ ai ( X i
i =1

Where

X*i : the value of Xi today;

^ i: the historical value of Xi ; and


X

64
© 2014-2020 AnalystPrep.
ai : a constant variable reflecting the importance of the ith variable.

The measure given by the formula becomes smaller as the similarity between current and

historical periods increases.

The GARCH Model

The generalized autoregressive conditional heteroscedasticity (GARCH) model is an extension of

the EWMA model, where we apply a weight to the recent variance rate estimate and the latest

squared return. According to the GARCH(1,1) model, the updated model for the variance rate is

given by:

σn2 = αr2n−1 + βr2n−1 + γVL

Where:

VL : long-run average variance rate;

α: weight given to the most recent squared returns;

β: weight given to the previous variance rate estimate; and

γ: weight given to long-run average variance rate.

The sum of the weights must be one so that:

α+β≤1

And,

γ = 1− α −β

Note that α and β are positive.

It is easy to see that EWMA is a particular case of GARCH (1,1) where γ=0, α =1-γ , and β = γ .

Both GARCH (1,1) and EWMA are called first-order autoregressive (AR(1)) models since the

65
© 2014-2020 AnalystPrep.
forecast for the variance rate depends on the immediately preceding variable.

Similar to the EWMA model, the weights in GARCH (1,1) decline exponentially. Now, by

definition of the GARCH (1,1) model, we have:

σn2 = αr2n−1 + βr2n−1 + γVL

And,

2
σn-1 = αr2n−2 + βr2n−2 + γ VL
2
σn-2 = αr2n−3 + βr2n−3 + γ VL
2
σn-3 = αr2n−4 + βr2n−4 + γ VL

2
Starting from the equation defining GARCH (1,1), substitute σn-1 2 2
, σn-2 , σn-3, … , we define:

w = γ VL
ω
⇒ VL =
γ

So,

σn2 = ω + αr2n−1 + βr2n−1

ω
Now since α + β + γ = 1 and VL = , then
γ

ω
VL =
1 − α− β

Example: Calculating implied volatility using GARCH(1,1)

Consider a GARCH (1,1) model where ω=0.00005, α=0.15, and β=0.75. Given that the current

volatility is 3% and the new return is -5%, (a) what is the long-run variance rate?

Solution

We know that:

66
© 2014-2020 AnalystPrep.
α+ β + γ = 1
⇒ γ = 1− α − β
= 1 − 0.15 − 0.75 = 0.10

Long-run average variance rate is given by:

ω 0.00005
VL = = = 0.0005
1 − α− β 0.10

(b) Assuming the GARCH (1,1) model, what is the new volatility?

Solution

This corresponds to the volatility of 2.24% (=√ 0.0005). Now according to GARCH (1,1) model,

σn2 = ω + αr2n−1 + βr2n−1


2
= 0.00005 + 0.15 × (−0.05) + 0.75 × 0.03 2 = 0.0011

Therefore, the new volatility is 3.32% (=√0.0011).

Mean Reversion According to an AR(1) Model

Mean reversion is simply the assumption that a stock's price (or any other variable) will tend to

move to its average over time.

67
© 2014-2020 AnalystPrep.
The simplest form of mean reversion can be illustrated by a first-order autoregressive process

[AR(1)], where the current value is based on the immediately preceding value.

Xt+1 = a + bX t + et+1

The expected value of Xt as a function of period t information is:

E [Xt+1 ] = a + bXt

We can restate this as:

a
E [Xt+1 ] = (1 − b) × + bXt
1− b

a
The long-run mean of this model is evaluated as [ (1−b) ]. The parameter of utmost interest in this

long-run mean equation is b, often termed “the speed of reversion” parameter. There are two

scenarios:

If b = 1, the long-run mean is infinite (i.e., the process is a random walk, nonstationary

process with an undefined long-run mean. That implies that the next period’s expected

value is equal to today’s value.

68
© 2014-2020 AnalystPrep.
If b is less than 1, then the process is mean-reverting (i.e., the time series will trend

toward its long-run mean). That implies that when Xt is above the LRM, it will be

expected to decline, and if it is below the long-run mean, it is expected to increase in

value.

Conditional Volatility With and Without Mean Reversion

Let’s consider our time series model above:

Xt+1 = a + bX t + et+1

Without mean reversion (b = 1),

The single period volatility is σ

The two-period volatility follows the square root rule and is σ × √2

With mean reversion (b < 1),

The single-period volatility is σ

The two-period volatility is √1 + b 2 ) × σ (which is less than σ × √ 2

Mean Reversion According to GARCH (1,1)

In GARCH (1,1), the long-run variance rate (VL ) provides a “pull” mechanism towards the long-

run average mean. Not that this is absent in the EWMA model due to a lack of the VL factor.

69
© 2014-2020 AnalystPrep.
Consider the figure above. When the volatility is above the long-run average mean, VL “pulls”

down toward it. On the other hand, when the volatility is below the long-run mean, VL pulls it

towards the long-run mean. This tendency of the “pull” mechanism is termed as the mean

reversion.

Therefore, GARCH (1,1) exhibits mean reversion, while EWMA does not. Mean reversion applies

to market variables. More importantly, market variables that are traded should not exhibit

predictable mean reversion to avoid arbitrage opportunities.

Notably, volatility cannot be traded, and thus it exhibits mean reversion. In this case, mean

reversion implies that when the volatility is high, we do not expect to remain in that position

forever but rather return to normal levels.

Long Horizon Volatility

Our discussion has been based on approximating one-day volatility. We might want to know what

will happen over a longer period, such as one year. To achieve this, we assume that the variance

rate over T days is equivalent to T days times variance over one day. That is,

70
© 2014-2020 AnalystPrep.
T − days variance rate = one-day variance rate × T

This means that volatility over days is equivalent to the volatility over one day multiplied by the

square root of T days. That is

T − day = one-day volatility × √T

Recall that this was true also for the VaR.

Mean Reversion and Long Horizon Volatility

Mean reversion provides an improvement to the square root rule of calculating long-horizon

volatility. Note that according to mean reversion, if the current daily volatility is high, we expect

it to decline. Therefore, using square root rule, in this case, overstates VaR, and if the volatility is

low, we expect it to rise a thus, the square root rule understates VaR. This applies to an expected

shortfall (ES).

According to GARCH (1,1), the expected variance rate on day t can be shown to be:

σn+t
2 = VL + (α + β)t (σn2 − VL )

We can use this formula to compute the average variance rate over the next T-days and

consequently expected volatility.

Example: Calculating long horizon volatility

Assume that the daily variance rate is 0.000025. What is the 30-day volatility?

T− day = one-day volatility × √T


= √0.000025 × √30 = 0.0274 = 2.74%

Implied Volatility

Implied volatility is an alternative measure of volatility that is constructed using the option

71
© 2014-2020 AnalystPrep.
valuation. The options (both put and call) have payouts that are non-linear functions of the price

of the underlying asset. For instance, the payout from the put option is given by:

max(K − PT )

Where:

P T is the price of the underlying asset,

K is the strike price, and

T is the maturity period.

Therefore, the price payout from an option is sensitive to the variance in the asset’s return.

The Black-Scholes-Merton model is commonly used for option pricing valuation. The model

relates the price of an option to the risk-free rate of interest, the current price of the underlying

asset, the strike price, time to maturity, and the variance of return. For instance, the price of the

call option can be denoted by:

C t = f (rf , T , Pt , σ 2 )

Where:

rf= Risk-free rate of interest

T=Time to maturity

P t=Current price of the underlying asset

σ 2=Variance of the return

The implied volatility σ relates the price of an option with the other three parameters. The

implied volatility is an annualized value but can be converted by dividing by 252, which is an

estimated number of trading days in a year. For instance, if annual volatility is 30%, then the

30%
daily implied volatility is 1.89% (= ).
√ 252

72
© 2014-2020 AnalystPrep.
The difference between the implied volatility and historical volatility (such as the one estimated

by GARCH(1,1) and EWMA models) is that implied volatility is forward-looking while historical

volatility is backward-looking. For example, a one-month implied volatility indicates average

volatilities over the next month.

Options are not actively traded, and thus finding reliable volatility is an uphill task. However, risk

managers evaluate both implied and historical volatility.

The volatility index (VIX) measures the volatility on the S&P 500 over the coming 30 calendar

days. VIX is constructed from a variety of options with different strike prices. VIX applies to a

large variety of assets such as gold, but it is only applicable to highly liquid derivative markets

and thus not applicable to most financial assets.

Advantages and Disadvantages of Implied Volatility

There are two main advantages of implied volatility over historical volatility:

It is a forward-looking, predictive measure that reflects the market’s consensus; and

Historical distribution patterns do not restrain it.

On the other hand, implied volatility has its shortcomings:

It is model dependent;

Options on the same underlying asset may trade at different implied volatilities. For

example, deep out of the money and deep in the money options trade at higher

volatility than at the money options;

It has limited availability because it can only be deduced when there are current

market prices; and

It assumes volatility will remain constant over a period of time, but volatility will

change over that same period.

Monitoring Correlation

73
© 2014-2020 AnalystPrep.
Recall that in a delta-normal model, correlations between the daily asset returns are needed to

compute the VaR or expected shortfall for a linear portfolio. Therefore, it is crucial to monitor

correlations.

Updating the correlations is analogous to that of volatilities. Recall that while updating the

volatilities, we use the variances. In the case of relationships, we use covariances. Now,

assuming that mean of daily returns is zero, then the covariance is defined as the expectation of

the product of returns.

According to the EWMA model, updating the correlation between returns X and Y is given by:

covn = λcovn−1 + (1 − λ) x n−1 y n−1

Where:

covn : covariance for day n

x n−1 and y n−1 : values of X and Y at day n.

Recall that for variables X and Y,

Cov (X, Y)
Corr (X, Y) =
σX σY

Where:

Corr (X, Y) : correlation between variables X and Y;

Cov (X, Y) : covariance between variables X and Y;

σX : standard deviation of X; and

σY : standard deviation of Y.

Therefore, if the EWMA model has been used to estimate the standard deviations of the returns,

we can comfortably estimate the correlation coefficient. However, the same value of λ is used

while updating both variances and covariances for consistency.

74
© 2014-2020 AnalystPrep.
Example: Updating the correlation coefficient

Assume that recent volatilities for returns on X and Y are 2% and 5%, respectively. Their

corresponding correlation coefficient is 0.4. Moreover, their recent returns are 3% and 4%,

respectively.

Assuming λ=0.92, calculate the value of:

i. updated volatilities; and

ii. updated correlation coefficient.

Solution

i. Updated volatilities

Denote the “recent” period by n-1. Now using the formula:

σn2 = (1 − λ) r2n−1 + λσn−1


2

The updated volatility for return X is given by:

σXn = √(1 − 0.92) × 0.032 + 0.92 × 0.022 = 0.02098 = 2.098%

And for return Y we have:

σYn = √(1 − 0.92) × 0.04 2 + 0.92 × 0.05 2 = 0.04927 = 4.927%

ii. Updated correlation coefficient

Using the formula:

Cov (X, Y)
Corr (X, Y) =
σX σY

It intuitively means that:

75
© 2014-2020 AnalystPrep.
covn−1 = σXn−1 × σY n−1 × Corr (Xn−1, Y n−1 )
= 0.02 × 0.05 × 0.4
= 0.0004

Now using:

cov n = λcovn−1 + (1 − λ) Xn−1 Yn−1

we have:

covn = 0.92 × 0.0004 + (1 − 0.92) × 0.03 × 0.04 = 0.000464

And thus, the updated correlation coefficient is given by:

cov n
corr (x n ,y n ) =
σXn σY n
0.000464
= = 0.4489
0.02098 × 0.04927

76
© 2014-2020 AnalystPrep.
Practice Question

The current estimate of daily volatility is 2.3 percent. The closing price of an asset

yesterday was CAD 46. Today, the asset ended the day at CAD 47.20. Using log-

returns and the exponentially weighted moving average model with λ = 0.94,

determine the updated estimate of volatility.

A. 2.319%

B. 0.0537%

C. 2.317%

D. 2.315%

The correct answer is C.

The updated variance is given by:

h t = λ(current volatility)2 + (1 − λ) (current return)2

Current volatility = 0.023

Current log-return = ln47.2– ln46 = 3.85439– 3.82864 = 0.02575


ht = 0.94 (0.0232 ) + (0.06 × 0.025752 ) = 0.000537
Updated estimate of volatility = √0.000537 = 0.02317

77
© 2014-2020 AnalystPrep.
Reading 48: External and Internal Credit Ratings

After completing this reading you should be able to:

Describe external rating scales, the rating process, and the link between ratings and

default.

Describe the impact of time horizon, economic cycle, industry, and geography on

external ratings.

Define and use the hazard rate to calculate unconditional default probability of a credit

asset.

Define the recovery rate and calculate the expected loss from a loan.

Explain and compare the through-the-cycle and at-the-point internal ratings

approaches.

Describe alternative methods to credit ratings produced by rating agencies.

Compare external and internal rating approaches.

Describe a rating transition matrix and explain its uses.

Explain the potential impact of rating changes on bond and stock prices.

Explain historical failures and potential challenges to the use of credit ratings in

making investment decisions.

A Description of External Rating Scales and the Rating


Process

An external rating scale is a scale used as an ordinal measure of risk. The highest grade on the

scale represents the least risky investments, but as we move down the scale, the amount of risk

gradually increases (safety decreases).

An issue-specific credit rating conveys information about a specific instrument, such as a

78
© 2014-2020 AnalystPrep.
zero-coupon bond issued by a corporate entity. An issuer-specific credit rating, on the other

hand, conveys information about the entity behind an issue. The latter usually incorporates a lot

more information about the issuer.

Here are S&P’s and Moody’s credit rating scores for long-term obligations:

79
© 2014-2020 AnalystPrep.
successive move down the scale represents an increase in risk. In the case of Moody’s

80
© 2014-2020 AnalystPrep.
ratings, Baa and above are said to be investment-grade while those below this level are said

to be non-investment-grade.

In the case of S&P’s, ratings BBB and above are investment-grade. All the others are non-

investment-grade.

The Rating Process

The process leading up to the issuance of a credit rating follows certain steps. These are:

I. A qualitative analysis of the company, including assessments of the quality of

management and competitive aspects

II. A quantitative analysis of financials such as ratio analysis

III. A meeting with the firm’s management

IV. A meeting of the rating agency committee assigned to rating the firm

V. A notification is sent to the rated firm detailing the assigned rating

VI. A fee is paid to the rating agency.

VII. The rated firm has a window to appeal the assigned rating or offer new information

VIII. The assigned rating is published

Outlooks and watchlists

Apart from the ratings themselves, the rating agencies also provide outlooks which shows the

changes likely to be experienced over the medium term.

A positive outlook indicates that a rating is likely to be raised.

A negative outlook indicates that a rating is likely to be lowered.

A stable outlook shows that the rating is stationary.

A developing outlook is an evolving one in which we can’t tell the direction of the

change.

When a rating is placed on a watchlist, it shows that a very small short-term change is expected.

81
© 2014-2020 AnalystPrep.
Rating stability

Rating stability is necessary since ratings are majorly used by bond traders. If the ratings were

to change, then the bond traders are required to trade more frequently and, in this case, they are

likely to incur a lot of transaction costs.

Rating stability is important because ratings are also used in financial contracts, and if the

ratings vary for different bonds, it would be difficult to administer the underlying contracts.

The Impact of Time Horizon, Economic Cycle, Industry, and Geography


on External Ratings

Time Horizon

The probability of default given any rating at the beginning of a cycle increases with the time

horizon. Non-investment bonds are the worst hit. Their default probabilities can dramatically

increase within a short time.

Economic Cycle

Since ratings are generally produced with an eye on a long-term period, they must take into

account any economic/industrial cycle on the horizon. Rating agencies make efforts to

incorporate the effects associated with an economic cycle in their ratings. Although this practice

is generally valid, it can lead to underestimation or overestimation of default if the predicted

economic cycle doesn’t play out exactly as expected. Put precisely, the probability of default can

be underestimated if an economic recession occurs, or overestimated if an economic boom

occurs. In addition, the default rate of lower-grade bonds is correlated with the economic cycle,

while the default rate of high-grade bonds is fairly stable.

Industry

Two firms in different industries – say, banking and manufacturing – could have the same rating,

but the probability of default may be higher for one of the firms than for the other. What does

that mean? The implication here is that for a given rating category, default rates can vary from

82
© 2014-2020 AnalystPrep.
industry to industry. However, there’s little evidence to support the notion that geographic

location has a similar effect.

Hazard rates

Consider a firm defaulting in a very short time, that is, δt.

The task is to answer the question, “What is the conditional probability of a firm defaulting

between time t and time t+δ t given that there is no default before time t?” We can denote this by

hδ t, where h is the rate at which defaults are happening at time t.

Unconditional default probabilities can be calculated using the hazard rates.

Suppose that h is the average hazard rate between time 0 and time t.

Then, the unconditional probability between time 0 and t is

1 − exp (−ht)

and the survival probability to time t is therefore given by

exp (−ht)

and the unconditional probability between time t1 and t2 is given by the expression;

exp (−ht 1 ) − exp (−ht 2 )

Example: Calculating default probabilities given hazard rates

Suppose you have been given a constant hazard of 0.05,

Calculate:

a. The probability of default by the end of 2 years.

b. The unconditional probability during the 3rd year.

Solution

83
© 2014-2020 AnalystPrep.
a. The probability of default at the end of the 2nd year is given by:

1 − exp (−ht)
= 1 − exp (−0.05 × 2) = 0.09516

b. The unconditional probability during the 3 rd year.

exp (−0.05 × 2) − exp (−0.05 × 3) = 0.04413

Recovery Rates

In the event that a firm runs bankruptcy or defaults, it may pay part of the amount of the total

loan to the lender. This amount that is repaid, expressed as a percentage, is known as the

recovery rate.

Since the loan is not fully repaid, then we can calculate the expected loss from the loan over a

given period of time as;

Expected loss = Probability of default × Loss given default


EL = PD × LGD

But since Loss given default = 1 − Recovery rate

Then, the expected loss from a loan is also calculated as

Expected loss = Probability of default × (1 − Recovery Rate)

For example, if the recovery rate is 70%, then

Loss given default = 100% − 70% = 30%.

Suppose the debt instrument has a notional value of $100 million, then the expected loss when

the loan defaults is $30 million.

Credit Spreads and Risk Premiums

The interest rate on a given risky bond and the yield on an equivalent risk-free security will

84
© 2014-2020 AnalystPrep.
always. This extra interest charged over the risk-free rate is known as a credit spread. This extra

interest acts as compensation to investors for bearing the risk of default.

From the following image, we can see the spread between the high yield index (risky bonds) and

the treasury notes (risk-free rate) over time.

Since the default rates keep fluctuating, giving rise to a non-diversifiable risk, bond investors, in

addition to the credit spread, will require a risk premium. Difficulties are also likely to be

experienced in selling off the bonds. Since risky bonds sometimes have low liquidity, the

investors will demand additional compensation in the form of a liquidity premium.

The Impact of Rating Changes on Bond and Stock Prices

Bonds

There’s overwhelming evidence that a rating downgrade triggers a decrease in bond prices. In

85
© 2014-2020 AnalystPrep.
fact, bond prices sometimes decrease just because there’s a strong possibility of a downgrade.

Anxious investors tend to sell bonds whose credit quality is declining.

A rating upgrade triggers an increase in bond prices, although there’s relatively less market

evidence to support this conclusion.

Therefore, the underperformance of bonds whose credit quality has been downgraded is more

statistically significant compared to the over-performance of bonds recently upgraded.

Stocks

There’s moderate evidence to support the view that a rating downgrade will lead to a stock price

decrease. A ratings upgrade, on the other hand, is somewhat likely to trigger an increase in bond

prices.

In practice, the relationship between changes in rating and stock prices can be quite complex

and will usually be heavily impacted by the reason behind the changes. Furthermore,

downgrades tend to have more impact on the stock price compared to upgrades.

Comparing External and Internal Ratings Approaches

External ratings are produced by independent rating agencies and aim at revealing the financial

stability of both lenders and borrowers. For example, Moody’s periodically releases ratings for

big banks around the globe. Such ratings are important because banks usually rely on customer

deposits and money raised through the issuance of various assets such as bonds to sustain

lending. The funds raised this way to create a pool of money that is then loaned to borrowers in

smaller chunks. Thus, depositors and bond owners use such ratings to assess the riskiness of

giving their money to the bank.

Sometimes, however, banks also need their own ratings so as to undertake an independent

assessment of the creditworthiness of a specific borrower – either an individual or a corporate.

That’s where internal credit ratings come in.

In modern times, internal credit ratings are usually developed based on the techniques used to

86
© 2014-2020 AnalystPrep.
develop external credit ratings. The same indicators are used, albeit with a few adjustments

depending on whether the borrower is an individual or a corporate.

One way of carrying out an internal rating is by use of a statistical technique known as the

Altman’s Z-score. The following ratios need to be provided when using this technique:

i. X1 ∶Working capital to total assets

ii. X2 ∶Retained earnings to total assets

iii. X3 ∶Earnings before interest and taxes to total assets

iv. X4 ∶Market value of equity to book value of total liabilities

v. X5 ∶Sales to total asset

Using the discriminant analysis, the Z-score is given by:

Z = 1.2X1 + 1.4X2 + 3.3X3 + 0.6X4 + 0.999X5 .

A Z-score above 3 means that the firm is not likely to default and when the Z-score is below 3,

then the firm is likely to default.

Nowadays, machine learning algorithms use more than five input variables as compared to

Altman’s Z-score. Also, the functions used in machine learning algorithms can be non-linear.

Some of the factors that have contributed to the increased sophistication of modern internal

credit ratings are:

I. The ever-growing use of external credit rating agency language in financial markets

II. Enforcement of capital requirements such as Basel II

Alternative to Rating

Apart from the commonly known rating agencies, that is, Moody’s, S&P, and Fitch, we have some

organizations such as KMV and Kamakura which use some models to come up with default

probabilities and hence can then use probabilities to provide important information to clients.

Factors considered includes:

87
© 2014-2020 AnalystPrep.
The amount of debt the firm has in its capital structure.

The market value of the firm’s equity.

The volatility of the firm’s equity.

In the underlying model, a company defaults if the value of its debt exceeds the value of its

assets.

Suppose v is the value of the asset and d is the value of the debt, the firm defaults when v < d.

The value of the equity, at a future point in time, is:

Equity = max (v-d , 0)

This implies that equity in a company is a call option on the assets of the firm with a strike price

equal to the face value of the debt. The firm defaults if the option is not exercised.

Comparing the Through-the-cycle and At-the-point Internal Ratings


Approaches

At-the-point Internal Ratings

At-the-point internal ratings, also called point-in-time ratings, evaluate the current situation

of a customer by taking into account both cyclical and permanent effects. As such, they are

known to react promptly to changes in the customer’s current financial situation.

At-the-point ratings try to assess the customer’s quantitative financial data (e.g. balance sheet

information), qualitative factors (e.g. quality of management), and information about the state of

the economic cycle. Using statistical procedures such as scoring models, all that information is

transformed into rating categories.

At-the-point internal ratings are only valid for the short-term or medium term, and that’s

largely because they take into account cyclic information. They are usually valid for a period not

exceeding one year.

88
© 2014-2020 AnalystPrep.
Through-the-cycle Internal Ratings

Through-the-cycle (ttc) internal ratings try to evaluate the permanent component of default risk.

Unlike at-the-point ratings, they are said to be nearly independent of cyclical changes in the

creditworthiness of the borrower. They are not affected by credit cycles, i.e. they are through-

the-cycle. As a result, they are less volatile than at-the-point ratings and are valid for a much

longer period (exceeding one year).

Advantages of ttc ratings include:

I. They are much more stable over time compared to at-the-point ratings

II. Because of their low volatility, ttc ratings help financial institutions to better manage

customers. Too many rating changes necessitate changes in the way a bank handles a

customer, including the products the bank is ready to offer.

One of the disadvantages of ttc ratings over at-the-point ratings is that they can at times be too

conservative if the stress scenarios used to develop the rating are frequently materially different

from the firm’s current condition. If the firm’s current condition is worse than the stress

scenarios simulated, then the ratings may be too optimistic. In fact, ttc ratings have very low

default prediction in the short-term.

Ratings Transition Matrices and Their Uses

A rating transition matrix gives the probability of a firm ending up in a certain rating category at

some point in the future, given a specific starting point. The matrix, which is basically a table,

uses historical data to show exactly how bonds that begin, say, a 5-year period with an Aa rating,

change their rating status from one year to the next. Most matrices show one-year transition

probabilities.

Transition matrices demonstrate that the higher the credit rating, the lower the probability of

default.

The table below presents an example of a rating transition matrix according to S&P’s rating

categories:

89
© 2014-2020 AnalystPrep.
One−year transition matrix

Exam tips:

Each row corresponds to an initial rating

Each column corresponds to a rating at the end of 1 year. For example, a bond initially

rated BB has an 8.84% chance of moving to a B rating by the end of the year.

The sum of the probabilities of all possible destinations, given an initial rating, is equal

to 1 (100%)

You will need to recall the rules of probability from mathematics to come up with n-year

transition probabilities, where n>1.

Credit ratings are their most stable over a one-year horizon. Stability decreases with

longer horizons.

Building, Calibrating, and Backtesting an Internal Rating System

Building

To build an internal rating system, banks try to replicate the methodology used by rating agency

analysts. Such a methodology consists of identifying the most meaningful financial ratios and

risk factors. After that, these ratios and factors are assigned weights such that the final rating

90
© 2014-2020 AnalystPrep.
estimate is close to what a rating agency analyst would come up with. Weights attached to

financial ratios or risk factors are either defined qualitatively following consultations with an

agency analyst, or extracted using statistical techniques.

Calibrating

Internal ratings have two main uses:

I. Assessing the creditworthiness of a customer during the loan application process

II. To determine the value of inputs used in the modeling of capital required as per the

existing regulations, e.g. Basel II

For these reasons, internal ratings have to be calibrated. This involves establishing a link

between the internal rating scale and tables displaying the cumulative probabilities of default.

The timeline of such tables must capture all maturities, from, say, 1 year to 30 years. Sometimes,

it may be necessary to build different transition matrices that are specific to the asset classes

owned by the bank.

Backtesting

Before linking default probabilities to internal ratings, backtesting of the current internal rating

system is vital. The question is: Just how many years are needed to pull this off?

A historical sample of between 11 and 18 years is considered sufficient to test the validity of

ratings.

Biases that May Affect a Rating System

91
© 2014-2020 AnalystPrep.
Bias Description
Time horizon bias Using a combination of at-the-point and through-the-cycle
approaches to score a company.
Information bias Assigning rating based on insufficient information
Homogeneity bias Inability to maintain consistent rating methods.
Principal-agent bias Rating developers fail to act in the best interest of the
management.
Backtesting bias Incorrectly linking rating systems to default rates.
Distribution bias Modeling the probability of a default using an inappropriate
distribution.
Scale bias Producing ratings that are not stable with the passage of
time

92
© 2014-2020 AnalystPrep.
Questions

Question 1

You have been given the following one-year transition matrix:

Rating From Rating To


A B CCC Default
A 80% 10% 10% 0%
B 5% 85% 5% 5%
CCC 0% 10% 70% 20%

Determine the probability that a B –rated firm will default over a two-year period.

A. 5%

B. 4.25%

C. 1%

D. 10.25%

The correct answer is D.

Required probability = Sum of probabilities of all possible paths that could lead to a

rating of D (default) after two years.

In other words, in how many ways can a B-rated firm default over a two-year period?

The following are the possible paths:

Path Probability
B→ default 0.05
B→ B → default 0.85 x 0.05= 0.0425
B→ CCC → default 0.05 x 0.20= 0.01
Total 0.1025

Question 2

ABC Co., currently rated BBB, has an outstanding bond trading in the market.

93
© 2014-2020 AnalystPrep.
Suppose the company is upgraded to A. What will be the most likely effect on the

bond’s price?

A. Positive and stronger than the negative effect triggered by a bond downgrade

B. Negative and stronger than the positive effect triggered by a bond

downgrade

C. Positive and weaker than the negative effect triggered by a bond downgrade

D. Positive and as strong as the negative effect triggered by a bond downgrade

The correct answer is C.

Rating downgrades tend to have more impact on the stock price compared to

upgrades. This can be explained by the fact that firms tend to release good news a lot

more often than bad news, and thus the expectations among investors are generally

positive. Negative news is usually unexpected and unanticipated, triggering a

stronger downward effect.

94
© 2014-2020 AnalystPrep.
Reading 49: Country Risk

After completing this reading you should be able to:

Identify sources of country risk.

Explain how a country’s position in the economic growth life cycle, political risk, legal

risk, and economic structure affect its risk exposure.

Evaluate composite measures of risk that incorporate all types of country risk and

explain limitations of the risk services.

Compare instances of sovereign default in both foreign currency debt and local

currency debt, and explain common causes of sovereign defaults.

Describe the consequences of sovereign default.

Describe factors that influence the level of sovereign default risk; explain and assess

how rating agencies measure sovereign default risks.

Describe characteristics of sovereign credit spreads and sovereign credit default swap

(CDS) and compare the use of sovereign spreads to credit ratings.

Sources of Country Risk

Country risk is the risk associated with investing in a given country. In most cases, an investor is

exposed to more risk by investing in some countries than others. When Apple, for example,

pushes for a bigger market presence in Latin America, they are exposed to the political and

economic turmoil that plagues these markets.

Country risk could be attributed to the following:

A country’s position in the economic growth life cycle

Differences in political risk

The legal system

95
© 2014-2020 AnalystPrep.
Disproportionate dependence on certain products or services.

A Country’s Position in the Economic Growth Cycle

The value of all goods and services produced within a country’s borders is known as GDP. When

we look at the economic growth rate with regard to an increase in the price of goods and

services, then this is called real GDP. The GDP, therefore, gives us a clear snapshot of the

economic performance of a given country.

Countries in early growth are more exposed to risk than larger, more mature countries. For

example, a global recession hits small, emerging markets harder than it does mature markets.

For instance, while countries like Japan and the U.K. experienced a 1-2% dip in GDP following a

recession, early growth economies like Kenya and Panama can record a dip as high as 5%.

Emerging markets are also hit hard in case of an economic shock. Even in the face of a robust

legal framework and good governance, there’s an upper cap on the powers that countries have

over their risk exposure. Some risks may simply be unavoidable. This is why it’s important to

thoroughly analyze a country’s risk profile before making critical investment decisions.

A country’s competitive advantage depends on:

Factors of production;

Demand for the products of given industries;

The availability of other related industries that produce high-quality products; and

The structure and management of the industries.

Differences in Political Risk

The political environment in a country can have a major bearing on its risk exposure. This can

particularly be explained in four main ways:

I. Continuous vs. Discontinuous risk: Countries deeply rooted in democracy and free

96
© 2014-2020 AnalystPrep.
speech have continuous risk in the sense that rules and regulations governing are

continuously being challenged and amended. As such, regulations enacted for the long-

term may remain in place only for a while before new laws and bills are introduced and

signed into law. This means that an investor may have to contend with new capital

requirements, for example, at a time when they are faced with a crippling cash crunch.

In authoritarian states, on the other hand, dictatorial policies create discontinuous

risk. That means it may be difficult to amend rules and regulations however good or bad

they might be from the perspective of the investor. In practice, the strength and

reliability associated with stable policies is usually undermined by issues such as

corruption and ineffective legal systems.

II. Corruption and Side costs: The rules and regulations governing business and

operations in a country are only as good as the systems put in place to enforce them.

High levels of corruption make it easy to circumvent regulations or ignore them outright.

An investor defending a certain move or arguing against an unfavorable regulation may

have to part with huge sums of money to get the job done.

III. Physical violence: Internal conflicts or civil war expose investors to both physical harm

and heavy operational costs, including high insurance costs and depreciation of physical

assets.

IV. Expropriation risk: Expropriation risk is the risk that a government may seize

ownership of a firm’s assets or impose certain rights that collectively reduce the firm’s

value. This could also happen when firms are subjected to specific taxes either by virtue

of their presence in a country or the nature of their core business. Compensation for

expropriation may be well below the value of rights or assets relinquished. Mining firms

would be a good example of firms constantly under expropriation risk.

Legal Risk

Legal risk has much to do with the enforcement of property/contractual rights and fidelity to the

rule of law. Laxity toward enforcement of rules and regulations not only disadvantages current

investors; it also serves to discourage potential investors from coming in. At best, those who

97
© 2014-2020 AnalystPrep.
decide to invest in the face of a failure of the legal system have to incorporate similar outcomes

in their expectations for the future.

Legal risk is also a function of how efficiently the system operates. For example, if enforcing a

contractual right takes years in a given country, investors will most likely shun that country.

Businesses and individual investors would not tolerate legal limbo for too long.

Some countries are known to have (and enforce) very robust property rights, especially in North

America. Others have very weak enforcement, especially African and South American countries.

Disproportionate Dependence on Certain Products or Services

A country that depends too much on one product or service exposes investors to additional risk.

A decline in the price or demand of the product or service can create severe economic shocks

that may reverberate well beyond the companies immediately affected. For example, a country

whose oil proceeds amount to, say, 50% of the GDP, exposes all investors within the country to

economic pain if the price of oil tumbles. In most cases, prices of other commodities shoot up.

Composite Measures of Risk: Risk Services

A good measure of country risk should incorporate all the dimensions of such risk, be they

political, financial, or economic. Collectively these risks make up what we call total risk. There

are several professional organizations around the world that offer country risk measurement

services. These include:

I. Political Risk Services (PRS)-PRS is a subscription-based service that offers members

extensive risk analysis of over 100 countries based on three core dimensions: political,

financial, and economic. These three dimensions are made up of 22 distinct variables.

PRS gives both dimensional and composite scores. The maximum score is 100, while the

minimum score is zero. For example, in its report dated July 2015, Switzerland was rated

the least risky country with a score of 88.5. Syria took the bottom spot with a score of

35.3

II. Euromoney- Euromoney is an organization that uses surveys of 400 prominent

98
© 2014-2020 AnalystPrep.
economists to come up with a composite risk score of a country or region. The score

ranges from 0 to 100.

III. The Economist- The popular media house develops in-house country risk scores built

upon currency risk, sovereign debt risk, and banking risk.

IV. The World Bank- The World Bank develops country risk scores based on six key

indicators. These are corruption, government effectiveness, political stability, regulatory

quality, rule of law, and accountability. The WB’s scores are scaled around zero, with

negative numbers indicating more risk and positive numbers less risk.

Limitations of Risk Services

Risk services greatly help to understand and keep track of country risk, but they are not devoid

of limitations. These include:

I. Some of the models used to churn out the final risk score incorporate risks that have

very little impact on business. The final score could be appropriate for policy-making or

economic reading but otherwise irrelevant for investors.

II. There’s no standardization of scores, and each service uses its own protocol and

calibration techniques.

III. The scores can be quite misleading when used to measure relative risk. For example, if a

country has a PRS score of 80, that doesn’t imply it’s twice as safe as another country

with a score of 40.

Sovereign Default

A sovereign default is the failure or refusal of the government of a sovereign state to pay back its

debt in full. Sovereign debt can be denominated in either local or foreign currency.

Foreign Currency Defaults

Foreign currency defaults are defaults that occur on the sovereign currency that’s denominated

in foreign currency. In most cases, governments find themselves unable to raise the amount of

foreign currency required to meet contractual obligations. And the worst thing about foreign

99
© 2014-2020 AnalystPrep.
currency debt is that governments cannot print foreign currency to make up for the shortfall.

In the last few decades, a majority of sovereign defaults have been based on foreign currency. In

addition, countries have been more likely to default on bank debt owed than on sovereign bonds

issued. And in dollar value terms, Latin American countries have accounted for much of

sovereign defaulted debt in the last 50 years.

Local Currency Defaults

Some countries have previously defaulted on debt denominated in local currency. Examples

include Argentina (2002-2004) and Russia (1998-1999). Local currency defaults could be traced

down to three main reasons:

I. Mandatory gold backups: In the years prior to 1971, the printed currency had to be

backed up with gold reserves. Without enough reserves, countries could not print enough

cash to pay up debts.

II. Presence of shared currencies: When a country shares a currency with other

countries, it lacks the freedom to print cash to prevent a sovereign default. This scenario

played out in 2015 whereby the Greek government could not print more Euros even in

the face of a crippling economic meltdown punctuated by a huge sovereign debt. The

country defaulted on a USD1.7 billion IMF payment – becoming the first country to do so

since Zimbabwe in 2001.

III. A reluctance to print cash purely for debt repayment: Printing cash to meet debt

obligations is fraught with dangers, including reputation risk, political instability, and the

very real possibility of an economic recession. The local currency can dramatically lose

value, forcing investors to shun financial investments in favour of real assets such as real

estate and precious stones.

Consequences of Sovereign Default

I. Reputation loss: Sovereign default can lead to a dramatic deterioration of both

diplomatic and economic ties between states. The defaulting government suffers a loss of

trust, making it incredibly difficult to negotiate debt in the future.

100
© 2014-2020 AnalystPrep.
II. Political instability: Following a sovereign default event, the populace may lose its

confidence in their leadership, leading to wave after wave of demos, unrest, as well as

coups in extreme cases.

III. Real output declines:? As investors increasingly shun financial assets in favour of real

assets, domestic consumption may also decrease. This, in turn, may lead to a drop-in

production

IV. Capital markets are thrust into a state of chaos and turmoil: Following a default

event, investors increasingly grow reluctant to commit their funds in long-term

investments. The demand for long-term securities such as bonds declines, making it

difficult for firms to raise funds for expansion and other operations.

Factors that Influence the Level of Sovereign Default Risk

I. Degree of indebtedness: The larger the debt a sovereign state has, the more likely it is

to default on part of the debt.

II. The size of revenue: A high amount of revenue reduces the chances of a sovereign

default event occurring. If a government has regular cash inflows in the form of taxes

and other revenue streams such as income earned from the sale of state-owned natural

resources, the less likely it is to default on a payment. The opposite is true.

III. Stability of revenue: Countries with more stable revenue streams have less default

risk. Normally, stability in the revenue collected increases as a country’s economic

activities become more diversified. Countries that depend too much on a specific source

of income have higher exposure to default risk.

IV. Political risk: If the leadership of a country is somewhat immune from public pressure –

something common in autocracies like Iran and other Middle East nations, default events

can easily occur. In such situations, the leadership enjoys the stability of tenure, giving

them a free hand to dictate the nation’s financial priorities. Democracies, on the other

hand, are less likely to default because their leadership is constantly under pressure to

deliver.

V. The level of backing/support a country enjoys courtesy of other sovereign states:

Countries that form part of a regional economic block usually offer each other some

101
© 2014-2020 AnalystPrep.
implicit backing on matters debt. For example, an EU member state is unlikely to default

on a sovereign debt because other EU countries are likely to chip in. In fact, when

countries like Spain and Portugal joined the EU, renowned credit rating agencies

reduced their assessment of default risk in these countries. However, such support is not

guaranteed. That’s why investors don’t read too much from this kind of backing.

How Rating Agencies Measure Sovereign Default Risks

Rating agencies offer opinions on a firm or country’s ability to incur and/or pay back the debt.

While assessing sovereign risk, agencies consider:

Political and social risks – including issues like public participation in political decision

making, respect for the rule of law, transparency in government, and long-term stability

of political institutions.

International security – including the safety and security of a country’s borders.

Regime legitimacy

Power and governance structures in place

Existing government obligations –including both local and foreign debt

Shortcomings of Ratings and Rating Agencies

I. Ratings are upward biased: In the aftermath of the 2007/2008 financial crisis, rating

agencies were widely criticized for having awarded unduly high credit ratings to banks

and other financial institutions. In general, rating agencies have been accused of being

far too optimistic in their assessment of sovereign ratings. An upward bias on corporate

ratings could be explained by the fact that the same corporates double up as the credit

agencies’ remunerators. This argument, however, does not hold when it comes to

sovereign ratings because individual governments are not required to pay the rating

agencies.

II. They are at times reactive rather than proactive: Rather than updating their ratings

102
© 2014-2020 AnalystPrep.
before an actual credit event occurs, rating agencies sometimes downgrade countries

after a problem has become evident. This does little to protect investors.

III. Too much interdependence: Although rating agencies claim to work independently

albeit using similar risk indicators, they have been accused of exhibiting herd behaviour,

so that an upgrade/downgrade by one agency is soon replicated by other agencies. As

such, independence is lost.

IV. Vicious cycle: Sometimes rating agencies have been accused of worsening a crisis by

unduly downgrading ratings – painting a situation as being worse than it actually is.

Sovereign Default Spread

Sovereign default spread describes the difference between the rate of interest on a sovereign

bond denominated in a foreign currency and the rate of interest on a riskless investment in that

currency. For example, suppose that country A has a 10-year dollar-denominated bond with a

market interest rate of 8%. At the same time, the 10-year U.S. Treasury bond is trading at 2%.

This implies that the sovereign default spread for country A is 6%. The sovereign default spread

represents the market’s assessment of a government’s credit risk.

Here is a table representing the sovereign default spread in some specific countries as compared

to the 10-years US Treasury bonds:

Country Moody’s rating $10-Year Bond U.S. T. Bond Default Spread


Rate Rate
Argentina B3 6.96% 2.85% 4.11%
Turkey Baa2 6.15% 2.85% 3.30%
Peru 3.55% 3.55% 2.85% 0.70%
Chile 3.35% 3.35% 2.85% 0.50%

Advantages of Using the Sovereign Default Spread as a Predictor of


Defaults

Compared to rating agencies, the market differentiation for risk is more granular. In

other words, the market has an even more refined understanding of country risk

compared to credit ratings. For example, countries A and could have the same Moody’s

rating, but the sovereign spread for A may be greater than the spread for B. That could

103
© 2014-2020 AnalystPrep.
mean the market sees more default risk in A than in B.

Market-based spreads reflect changes in real-time, unlike credit ratings which can be

reactive rather than proactive. As such, market-based spreads can be more effective at

predicting imminent danger.

Disadvantages of Using the Sovereign Default Spread as a Predictor of


Defaults

Market-based spreads tend to be far more volatile than credit ratings and can be affected by

variables with little or no correlation to default. For example, liquidity and market forces of

supply and demand can trigger shifts in spreads that have nothing to do with default.

Conclusion

Both credit ratings and market spreads are useful measures of default. Sovereign bond markets

usually price bonds guided by credit ratings. In addition, credit rating agencies also leverage

market data to make changes in ratings.

104
© 2014-2020 AnalystPrep.
Questions

Question 1

In the context of sovereign default spread, which of the following statements is least

accurate?

A. It describes the difference between the rate of interest on a sovereign bond

denominated in a foreign currency and the rate of interest on a riskless

investment in the local currency

B. A higher spread represents greater sovereign risk

C. Market-based spreads tend to be far more volatile than credit ratings

D. There is a strong correlation between sovereign ratings and market default

spreads

The correct answer is A.

Sovereign default spread describes the difference between the rate of interest on a

sovereign bond denominated in a foreign currency and the rate of interest on a

riskless investment in that currency. For example, suppose that country A has a 20-

year dollar-denominated bond with a market interest rate of 8%. At the same time,

the 20-year U.S. Treasury bond is trading at 2%. This implies that the sovereign

default spread for country A is 6%.

Question 2

Which of the following statements is most accurate in reference to the effect of a

country’s position in the economic cycle on its risk exposure?

A. Mature growth countries are insured from economic shocks

B. Young, growth companies are more exposed to risk partly because they have

limited resources to overcome setbacks

C. In markets, a shock to global markets will travel across the world, but mature

105
© 2014-2020 AnalystPrep.
market equities will often show much greater reactions, both positive and

negative to the same news

D. A country that is still in the early stages of economic growth will generally

have less exposure than a mature country, provided it is well-governed and

has a solid legal system.

The correct answer is B.

Young, growth companies are more exposed to risk partly because they have limited

resources to overcome setbacks and also because they rely too much on a stable

macroeconomic environment so as to register success in terms of GDP.

Choice A is inaccurate. Mature growth countries are not insured from economic

shocks. They are still exposed to risk, but on a lower scale compared to young,

emerging markets.

Choice C is inaccurate. In markets, a shock to global markets will travel across the

world, but early growth market equities will often show much greater reactions,

both positive and negative to the same news

Choice D is also inaccurate. A country that is still in the early stages of economic

growth will generally have more risk exposure than a mature country, even if it has

good governance backed up by a solid legal system.

106
© 2014-2020 AnalystPrep.
Reading 50: Measuring Credit Risk

After completing this reading, you should be able to:

Evaluate a bank’s economic capital relative to its level of credit risk.

Explain the distinctions between economic capital and regulatory capital, and describe

how economic capital is derived. Identify and describe important factors used to

calculate economic capital for credit risk: the probability of default, exposure, and loss

rate.

Define and calculate the expected loss (EL).

Define and explain unexpected loss (UL).

Estimate the mean and standard deviation of credit losses assuming a binomial

distribution.

Describe the Gaussian copula model and its application.

Describe and apply the Vasicek model to estimate default rate and credit risk capital

for a bank.

Describe the CreditMetrics model and explain how it is applied in estimating economic

capital.

Describe and use the Euler’s theorem to determine the contribution of a loan to the

overall risk of a portfolio.

Explain why it is more difficult to calculate credit risk capital for derivatives than for

loans.

Describe challenges to quantifying credit risk.

Economic Capital, Regulatory Capital, and Debt Capital

Economic capital of a bank is the approximate amount of capital that the bank requires to

107
© 2014-2020 AnalystPrep.
absorb the losses from the loan portfolios. In other words, it is the cushion that a bank estimates

it will need in order to remain solvent.

Regulatory capital is the amount of capital the regulators require the banks to maintain. For

instance, the Global bank requirements are determined by the Basel Committee on Banking

Supervision (BCBS) based in Switzerland. Supervisors then implement the BCBS requirements in

each country.

Banks also have what is called debt capital, funded by bondholders. In case the incurred losses

deplete the equity capital, the debt holders should incur losses before the depositors.

The equity capital is described as “going concern capital” because the bank is solvent if its

capital is positive. On the other hand, the debt capital is described as the “gone concern capital”

because it acts as a cushion to the depositors when the bank becomes insolvent (no longer a

going concern).

Banks face many risks through their transactions, which need to be quantified. Credit risk is a

primary risk that the banks have concentrated since the inception of the banking industry. It is

majorly explained by the fact that the banks’ activities mainly involve taking deposits and making

loans. The loans made are predisposed to some level of default of risk and hence some level of

credit loss (credit risk).

Credit risk is quantified using different models. This chapter discusses three models:

i. The mean and standard deviation of the loss from a loan portfolio determined from the

features of individual loans.

ii. The Vasicek model – used by the bank regulators to approximate the extreme percentile

of the loss distribution.

iii. CreditMetrics – used by banks to estimate economic capital.

Evaluating a Bank’s Economic Capital Relative to its Level of


Credit Risk

The term credit risk describes the risk that arises from nonpayment or rescheduling of any

108
© 2014-2020 AnalystPrep.
promised payment. It can also arise from credit migration – events related to changes in the

credit quality of the borrower. These events have the potential to cause economic loss to the

bank.

The expected loss is the amount a bank can expect to lose, on average, over a predetermined

period when extending credits to its customers. Unexpected loss is the volatility of credit losses

around its expected loss.

Once a bank determines its expected loss, it sets aside credit reserves in preparation. However,

for unexpected loss, the bank must estimate the excess capital reserves needed subject to a

predetermined confidence level. This excess capital needed to match the bank’s estimate of

unexpected loss is known as economic capital.

To safeguard its long-term financial health, a lender must match its capital reserves with the

amount of credit risk borne. Economic capital is primarily determined by (i) level of confidence,

and (ii) level of risk. An increase in any of these two parameters causes the economic capital also

to increase.

Important Factors Used to Calculate Economic Capital for Credit Risk

Probability of Default

The probability of default (PD), describes the probability that a borrower will default on

contractual payments before the end of a predetermined period. This probability is in and of

itself not the most significant concern to a lender because a borrower may default but then

bounce back and settle the missed payments soon afterward, including any imposed penalties. It

is expressed as a percentage.

Exposure Amount

Exposure amount (EA), also known as exposure at default (EAD), is the loss exposure of a bank

at the time of a loan’s default, expressed as a dollar amount. It is the predicted amount of loss in

the event the borrower defaults.

109
© 2014-2020 AnalystPrep.
EAD is a dynamic amount that keeps on changing as the borrower continues to make payments.

Loss Rate

The loss rate, also known as the loss given default (LGD), is the percentage loss incurred if the

borrower defaults. It can also be described as the expected loss expressed as a percentage. The

loss rate is the amount that’s impossible to recover after selling (salvaging) the underlying asset

following a default event.

The LGD can also be expressed as:

LGD = 1– Recovery Rate

Calculation of the Expected Loss

The expected loss, EL, is the average credit loss that we would expect from an exposure or a

portfolio over a given period. It is the anticipated deterioration in the value of a risky asset. In

mathematical terms,

EL = EA × PD × LGD

Credit loss levels are not constant but rather fluctuate from year to year. The expected loss

represents the anticipated average loss that can be statistically determined. Businesses will

typically have a budget for the EL and try to bear the losses as part of the standard operating

cash flows.

Exam tip: The expected loss of a portfolio is equal to the summation of expected losses of

personal losses.

ELP = ∑ EA i × PD i × LGD i

Example: Calculating expected loss (EL) and unexpected loss (UL)

A Canadian bank recently disbursed a CAD 2 million loan, of which CAD 1.6 million is currently

outstanding. According to the bank’s internal rating model, the beneficiary has a 1% chance of

110
© 2014-2020 AnalystPrep.
defaulting over the next year. In case that happens, the estimated loss rate is 30%. The

probability of default and the loss rate have standard deviations of 6% and 20%, respectively.

Determine the expected loss figures for the bank.

EL = EA × P D × LR

Where:

EA = CAD 1,600,000

P D = 1%

LR = 30%

Thus,

EL = 1, 600, 000 × 0.01 × 0.3 = CAD 4, 800

Unexpected Loss

Unexpected loss is the average total loss over and above the expected loss.

111
© 2014-2020 AnalystPrep.
It is the variation in the expected loss. It is calculated as the standard deviation from the mean at

a certain confidence level.

Let UHL denote the unexpected loss at the horizon for asset value VH . Then,

UHL ≡ √(var (VH))

You will usually apply the following formula to determine the value of the unexpected loss:

2 + LR2 × σ 2
UL = EA × √ PD × σLR PD

Where

2
σPD = PD × (1 − PD)

112
© 2014-2020 AnalystPrep.
The Mean and Standard Deviation of Credit Losses

Given that a bank has n loans, define the following quantities:

L i – the amount borrowed in the ith assumed to be constant throughout the year.

pi – the probability of the default for the ith loan

R i – the recovery rate in the event of the default by the ith loan in the event of the default by the

ith loan

pij – the correlation between losses on the ith and jth loan

σi – the standard deviation of loss on the ith and jth loan

σp – the standard deviation of loss from the portfolio

a – the standard deviation of portfolio loss as a fraction of the size of the portfolio

If the ith loan defaults, the loss is given by:

L i (1 − Ri )

Intuitively, the probability distribution for the loss from the ith loan is made of a probability p i

that there will be a loss of this amount and the probability 1 − pi that there is no loss, which is

typically a binomial distribution. Therefore, we can present the mean and the standard deviation

of the loss.

The mean of the loss is given by:

p i × L i (1 − R i) + (1 − p i) × 0 = p iL i (1 − Ri )

Now, recall that for a random variable X, the variance is defined as:

σx = E (x2 ) − [E (x)]2

113
© 2014-2020 AnalystPrep.
Where E denotes the expectation. Intuitively,

σi2 = E (Loss2 ) − [E (Loss)]2

This follows immediately that:

σi2 = p i[L i (1 − R i)]2 − [pi L i (1 − R i)]2 = (pi − p2i ) [Li (1 − Ri )]2

So that the standard deviation of the loss is given by:

σi = √p i − p 2i [L i (1 − R i )]

Note that we can also calculate the standard deviation of the loan portfolio from the losses on

the individual loans. It is given by:

n n
σP2 = ∑ ∑ p ij σi σj … … … … … Eq1
i=1 j=1

Now, the standard deviation expressed as the percentage of the size of the portfolio is

n n
√∑i=1 ∑j=1 pij σi σj
α= … … … … … . . Eq2
∑ni=1 L i

Assume that all loans have the same principal L, all recovery rate R are equal, and all default

probabilities are equal and denoted by p, and the correlation coefficient is defined as:

1 when i = j
ρij = {
ρ when i ≠ j

Where ρ is constant.

Therefore, the standard deviation of the loss from loan i is the same for all i so that the common

standard deviation denoted by σ is given by:

σ = √p − p 2 [L (1 − R)]

114
© 2014-2020 AnalystPrep.
Therefore, Eq1 reduces to:

σP2 = nσ 2 + n (n − 1) ρσ 2

Also, the standard deviation of the loss from the loan portfolio as a percentage of its size (Eq2)

reduces to:

σP σ√1 + (n − 1) ρ
α= =
nL L√n

Example: Loss from a loan Portfolio as a percentage of its size

The bank of Baroda has a portfolio consisting of 10,000 loans each loan amounting to $2 million

and has a 1% probability of default over the following year. The recovery rate is 40%, and the

correlation coefficient is 0.1. Calculate α, the standard deviation of the loss from the loan

portfolio as a percentage of its size?

Solution

The α parameter is given by:

σ√1 + (n − 1) ρ
α=
L √n

For this case,

L = $2 million

ρ = 0.1

n = 10,000

R = 0.4

σ = √p − p2 [L (1 − R)]

= √ 0.01 − 0.012 [2 (1 − 0.4)] = 0.1194

115
© 2014-2020 AnalystPrep.
Therefore,

0.1194√1 + (10,000 − 1) × 0.1


α= = 0.01889
2 × √10,000

The Gaussian Copula Model

A Gaussian copula maps the marginal distribution of each variable to the standard normal

distribution, which, by definition, has a mean of zero and a standard deviation of one. Copula

correlation models create a joint probability distribution for two or more variables while still

preserving their marginal distributions. The joint probability of the variables of interest is

implicitly defined by mapping them to other variables whose distribution properties are known.

We can demonstrate how Gaussian copulas work using a simple example:

Let us define two variables V1 and V 2 that have unknown distributions and unique marginal

distributions. V1 and V2 are mapped into new variables U1 and U 1 that have standard normal

distributions. The mapping is done on a percentile-to-percentile basis to create a Gaussian

copula.

116
© 2014-2020 AnalystPrep.
For example, the one-percentile point of the V 1 distribution is mapped to the one-percentile point

of the U 1 distribution. Similarly, the 50-percentile point of the V1 distribution is mapped to the

50-percentile point of the U 1 distribution.

Before mapping variables V 1 and V2 to the normal distribution, it is very difficult to define a

relationship between them since their marginal distributions are unknown and are pretty much

incomprehensible structures. Once they have been mapped to the standard normal distribution

as new variables U1 and U 2, respectively, we can now define a relationship between them since

the standard normal distribution has a known structure. The Gaussian copula, therefore, helps us

to define a correlation between variables when it is not possible to define a correlation directly.

A One-Factor Correlation Model

Assume now that we have many variables Vi for all i=1,2,…,n for which each of the Vi can be

117
© 2014-2020 AnalystPrep.
mapped to a standard normal distribution Ui . The main challenge that remains is to define the

correlation between U i distributions because the presence of many unique distributions means

that we have to specify numerous correlation parameters. To address this issue, we use the one-

factor model.

The one-factor model is defined as:

U i = ai F + √1 − a2i Zi

Where F is a common factor for all Ui and Zi is a component of Ui that is unrelated to the factor

F and uncorrelated to each other. The ai are the parameter values that lie between -1 and +1,

that is ai ∈ [−1, +1].

The variables F and Zi have the standard normal distributions, that is, F ∼ N (0 , 1) and

Zi ∼ N(0, 1). Therefore, U i is a sum of two independent normal distributions, and it is, therefore,

a normal variable with a mean of 0 and a standard deviation of 1. The variance of Ui is 1 since F

and Zi are uncorrelated so that:

Var (U i ) = Var [ai F + √ 1 − a2i Zi ]


2
= a2i Var (F) + (√ 1 − a2i ) Var (Zi )
2
= a2i .1 + (√1 − a2i ) .1 = a2i + 1 − a2i = 1

Note that we are summing the variances of two uncorrelated variables.

So in a nutshell, the one-factor model takes one standard normally distributed variable Ui and

defines it in relation to two other variables, which are both are standard normally distributed.

The factor F affects across all Ui but factor Zi affects only Ui .

The coefficient of correlation between Ui and Uj comes in as a result of the shared factor F and

this ai aj. The correlation coefficient between Ui and Uj is defined from the basic statistics as

E (Ui U j) − E (Ui ) E(U j)


ρUi, Uj =
σUi σUj

118
© 2014-2020 AnalystPrep.
But E (Ui )=E (U j)=0 and σUi = σUj = 1 and thus the above equation reduces to

E (U i Uj ) − 0
ρUi ,Uj = = E (U iU j)
1

Now,

E (Ui Uj ) = E [(aiF + √1 − a2i Zi ) (ajF + √1 − a2j Zj )]

But F is uncorrelated with all Zi and Zi is uncorrelated with Zj so that

E (FZi ) = E (FZj) = E (Zi Zj ) = 0

And the above equation reduces to:

E(U i Uj ) = E (ai , ajF2 ) = ai , ajE (F2 ) = ai, aj .1 = ai , aj


∴ ρ U i, U j = ai aj

Note that the immediate result stems from the fact that E(F2 ) = 1 because F is a standard normal

variable, F ∼ N(0, 1) so that:

E (F2 ) − [E (F)]2 = Var (F) = 1


⇒ E (F2 ) − [0]2 = 1
∴ E (F2 ) = 1

A notable example of a one-factor model is the capital asset pricing model (CAPM). In CAPM, the

correlation coefficient between two assets is assumed to arise from the dependence on the

common factor – the return from the market index. However, CAPM is significantly palatable

since it specifies the correlation between the returns of the different assets.

The Vasicek Model

The Vasicek model uses the Gaussian copula model to define the correlation between the

119
© 2014-2020 AnalystPrep.
defaults when determining the capital for loan portfolios. The Vasicek model has an advantage

over the standard deviation of the loss from the loan portfolio as a percentage of its size, α in

that the unexpected loss can be determined analytically.

Now, assume that the probability of default (PD) is the same for all firms in a huge portfolio so

that the PD for a company i for one year is mapped to a standard normal U i as described earlier.

Consider the following figure below:

The values at the further left (shaded region) tail of this standard normal distribution is the

distribution of the default, while the rest is the distribution of no default.

Mathematically, a firm i defaults if:

Ui ≤ N−1 (PD)

Where N −1 is the inverse of the inverse cumulative normal distribution. That is if the probability

of default for the loan portfolio of a bank is 1%, then the bank defaults if:

120
© 2014-2020 AnalystPrep.
U i ≤ N −1 (0.01) = −2.326

This can be stated as a bank defaults if:

Ui ∈ (−∞, −2.326)

And no default if:

Ui ∈ (−2.326 , ∞)

Recall that the one-factor model is given by:

121
© 2014-2020 AnalystPrep.
U i = ai F + √1 − a2i Zi

To make the model more palatable, ai ’s are assumed to be equal for all i so that ai = a and thus,

the one-factor model is rewritten as:

Ui = aF + √ 1 − a2 Zi

Thus the correlation between each pair od Ui ’s is:

ρUi,Uj = ai aj = a. a = a2

Now, the factor F can be taken as an index of the recent economic health. That is, if the F is high

the doing economy is healthy implying that U i is high, making the default relatively low.

Otherwise, if F is relatively low, then Ui is also relatively low, and thus making default most likely.

For a large portfolio, the default rate defined as the probability that:

Ui ≤ N−1 (PD)

which was discussed earlier. Now using the properties of a normal distribution:

N −1 (PD) − aF
Default Rate as a Function of F = N ( )
√1 − a2

We anticipate that the default rate is not exceeded with a 99.9% likelihood and thus given by the

low value of F. Moreover, we require that the probability of the true value of F will be worse than

other F' and it will be 0.1%. Now, recall that F is normally distributed so that:


F = N−1 (0.001)

and thus,

N −1 (PD) − aN−1 (0.001)


99.9 percentile for default rate = N ( )
√ 1 − a2

122
© 2014-2020 AnalystPrep.
Now, ρ is defined as the correlation between each pair of Ui given by:

ρ = a2 ⇒ a = √ρ

Therefore,

N −1 (PD) + √ ρN−1 (0.001)


99.9 percentile for default rate = N ( )
√1 − ρ

By the property of the normal distribution, we know that:

N (0.001) = −N (0.999)

The formula above can be rewritten as:

N −1 (PD) + √ ρN−1 (0.999)


99.9 percentile for default rate = N ( )
√1 − ρ

Now consider a loan portfolio with the same default probability. The last equation allows us to

convert the average default rate (PD) into a portfolio default rate, which is interestingly only

once in 1000 years!

Moreover, when ρ = 0,

N −1 (PD) + √0N−1 (0.999)


99.9 percentile for default rate = N ( )
√1 − 0
= N (N −1 (PD)) = PD

This makes much sense because if the firms default independently, the “law of large numbers”

makes sure that the default rate in a large portfolio is always the same.

Example: Calculating the 99.9 percentile for default rate under the
Vasicek model

The Bank of Baroda has a loan portfolio that has a default rate of 2% and a correlation coefficient

123
© 2014-2020 AnalystPrep.
of 0.3. Under the Vasicek Model, what is the 99.9 percentile for the portfolio default rate?

Solution

The 99.9 percentile under the Vasicek model is given by:

N −1 (PD) + √ ρN−1 (0.999)


99.9 percentile for default rate = N ( )
√1 − ρ

So for the information provided in the question,

N −1 (0.02) + √0.3N −1 (0.999)


= N( ) = N (−0.42) = 34%
√1 − 0.3

124
© 2014-2020 AnalystPrep.
If a loan has the same PD, the same ρ and the same loss given default (LGD) and the same

principle, the Basel II capital requirement for banks under the IRB approach is given by:

125
© 2014-2020 AnalystPrep.
(WCDR − PD) × LGD × EAD

Where WCDR is defined as the worst-case default rate, and it is 99.9 percentile of the default

rate distribution defined as in Vasicek model. LGD is the loss given the default, which is defined

as one minus the recovery rate. EAD is the total exposure at default, which is the sum of the

principal of all the loans.

The Basel II equation provides a way of calculating the unexpected loss with a 99.9% confidence

level. This can be seen by the fact that WCDR×LGD is the percentile point of loss rate

distribution, while WCDR×LGD×EAD is the loss at the 99.9 percentile. Therefore, the expected

loss is given by:

PD × LGD × EAD

For a non-homogeneous loan portfolio, the one-factor model equation can be rewritten so that for

loan i:

∑ (WCDRi − PD i) × LGD i × EAD i


i

Where the variables are as defined earlier for each loan i.

The last equation gives a way of calculating the capital for each loan separately, and then the

results are added. Moreover, the equation can be adjusted to include the maturity adjustment

factor.

Basel II defines correlation ρ in that the banks must assume different conditions. Based on the

IRB approach, the banks estimate PD while EAD and LGD estimates are approximated in

accordance with Basel II rules or using the bank’s internal model but subject to circumstances

and regulatory requirements.

CreditMetrics Model

The CreditMetrics model is used by the banks to calculate economic capital where each

126
© 2014-2020 AnalystPrep.
borrower is given an external or internal credit rating. A one-year transition table is then utilized

to define the changes in the credit ratings.

The loan portfolio of the bank is determined at the beginning of the one year, and the Monte

Carlo simulation is used to define how the rating changes over the year. In each of the simulation

experiments, the ratings of all borrowers are determined at the end of the year, which allows the

bank to reevaluate the portfolio.

Reevaluation of the portfolio involves calculating the credit loss of the portfolio, which is defined

as the value of the portfolio at the beginning of the year, less the value of the portfolio at the end

of the year. Through numerous simulation trials, complete credit loss distribution is produced.

Example: Applications of the CreditMetrics model

To demonstrate the CreditMetrics model, consider a bank X whose credit ratings are A, B, C, and

default. The probabilities of the rating transition are as follows: a B-rated can transition to rating

A with a probability of 5%, 85% chance of staying in B, 14% of transitioning to C, and 3% chance

of defaulting as shown in the table below.

Rating Transition Movement Probability of Transition


B to A 6%
Remains in B 85%
B to C 14%
Defaults 3%

In each Monte Carlo simulation experiment, a number is sampled from a standard normal

distribution to ascertain what will happen to a borrower. Now given that

N −1 (0.06) = −1.555, N −1 (0.90) = 1.282, N −1 (0.97) = 1.881. From these values, the corresponding

“rule table” is as shown below:

Standard Normal Sample Rating Transition


Less than -1.555 B to A
Between -1.555 and 1.282 Remains in B
Between 1.282 and 1.881 B to C
Greater than 1.881 Defaults

127
© 2014-2020 AnalystPrep.
However, since the bank borrowers do not default independently, sampling should be done in a

way that reflects the correlation between the samples. As such, a factor model (such as the one-

factor model) is used to capture the relationship between the normal distributions. In other

words, the Monte Carlo simulation should be based on the Gaussian model such that

probabilities of the rating transitions for each borrower are converted to a normal distribution,

and the correlations are defined on those distributions and not rating transitions themselves.

Notably, the correlations between the traded equities are usually used in CreditMetrics, which

can be justified by the Monte Carlo simulation, where the company defaults if its market value

becomes less than the book value of its debt.

Lastly, as can be seen in the CreditMetrics model, as opposed to the Vasicek model, it includes

both the effects of rating variations and defaults.

Euler's Theorem on Risk Allocation

Leonhard Euler developed a model that can be used to divide risk measures. The Euler Theorem

is based on the homogenous functions F of a set of variables x1 , x 2 ,… , x nwhich a feature defined

as

F (λx 1 , λx2 , . . . , λx n ) = λF (x 1 , x2 , . . . , x n )

Where λ is a constant.

Now, define:

ΔFi
ΔFi Δx i
Q i = xi =
Δxi xi

Where

Δx i = small change in x i

ΔFi = small change in Fi

Δx
128
© 2014-2020 AnalystPrep.
Δx i
Qi = ratio of ΔFi to a proportional change in x i
xi

Euler showed that as Δx i tends to 0,

n
F = ∑ Qi
i=1

So, many risk measures are homogeneous functions, which is a property of a coherent risk

measure.

Now, if a portfolio is adjusted such that each position is multiplied by some constant λ, a risk

measure is typically multiplied by λ. So the Euler’s theorem provides a way to allocate a risk

measure F, which is defined as a function of many trades into its components.

As an application of the Euler’s theorem on the credit risk, we can determine the contribution of

each loan in a portfolio to the overall risk measure. Consider the following example.

Example: Application of the Euler’s Theorem

The Basley bank has three loans, A, B, and C. Losses from the loans are 1.0, 1.2, and 1.3,

respectively. The correlations between the losses are as shown by the table below:

Loan A Loan B Loan C


Loan A 1 0 0
Loan B 0 1 0.8
Loan C 0 0.8 1

The standard deviation of the total loss is given by:

σP = √σA2 + σB2 + σC2 + 2ρAB σAσB + 2ρAC σA σC + 2ρBC σB σC

= √1.02 + 1.22 + 1.32 + 0 + 0 + (2 × 0.8 × 1.2 × 1.3) ≈ 2.57

Now assume that the size of the loan A is increased by 1% so that its new standard deviation is

now 1 × 1.01 = 1.01. The increase in the standard deviation of the loan portfolio is:

129
© 2014-2020 AnalystPrep.
= (√1.012 + 1.22 + 1.32 + (2 × 0.8 × 1.2 × 1.3) − (√1.02 + 1.22 + 1.32 + (2 × 0.8 × 1.2 × 1.3)) = 0.0039

Now using the equation:

ΔFi
ΔFi Δx i
Q i = xi =
Δxi xi

It implies that,

0.0039
QA = = 0.39
0.01

Now if the size of loan B is also increased by 1%, so that its standard deviation of loss is now

1.2 × 1.01 = 1.2120 and thus the increase in the loan portfolio is given by:

= (√1.02 + 1.21202 + 1.32 + 2 × 0.8 × 1.2120 × 1.3) − (√ 1.02 + 1.22 + 1.32 + 2 × 0.8 × 1.2 × 1.3) = 0.01045

And thus,

0.01115
QB = = 1.045
0.01

Again if the loan size of loan C is increased by 1% then,

QC = 1.142

So,

QA + QB + QC = 0.39 + 1.045 + 1.142 ≈ 2.57

As per the context of Euler’s theorem, we have divided the total loss of 2.57 into the loan

contributions from loans A, B, and C. It is easy to see that that the contribution of loan A is low

because it is uncorrelated with loans B and C, hence contributes less risk to the entire portfolio.

130
© 2014-2020 AnalystPrep.
Credit Risk Capital for Derivatives

Similar to loans, derivatives (such as options and swaps) generate credit risk. For instance, if

company X purchases an option from company Y, the credit risk to company X arises in the sense

that the company Y might default and thus fail to honor its obligation. In the case of a swap, if

company X enters into an interest rate swap with company Y, the credit risk to company X occurs

if company Y defaults when the value of the swap is positive to company X.

The credit risk capital for derivatives can be similarly calculated using the equation:

∑ (WCDRi − PD i) × LGD i × EAD i


i

However, it is challenging to compute EAD for the derivatives transactions. Note that in the case

of loans, EAD is the amount that has been advanced or expected to be advanced to the borrower.

Obviously, in derivatives, exposures varies with the value of the derivative. The solution this

challenge is addressed by the Basel Committee’s rules of computing EAD – setting the exposure

at default for the derivatives equal the current exposure plus add-on amount. The current

exposure is the maximum amount of capital that might be lost in case of default today. The add-

on amount is an additional amount for the possibility of the exposure worsening by the time

default occurs.

An additional challenge for the derivatives involves netting agreements so that all the

outstanding derivatives with a given counterparty may be considered a single derivative in case

of a default. As such, the equation above cannot be utilized on a transaction-by-transaction

grounds, yet it must be implemented on a counterparty-by-counterparty basis.

Challenges to Quantifying Credit Risk

The analysis of the credit risk needs many estimates, such as PD. Just like through-the-cycle and

point-in-time for credit ratings, we can differentiate between the through-the-cycle PD (mean of

the PD over the economic cycle) and point-in-time PD (indicates the current economic times).

The banks are required to estimate the through-the-cycle PD for regulatory reasons, but point-in-

131
© 2014-2020 AnalystPrep.
time PD is estimated for internal uses.

The accounting standards, such as IFRS 9, requires that the loans should be valued for

accounting reasons. As such, the expected losses over one year or loan’s life must be computed

and subtracted from the loan amount, at which point-in-time estimate is required. Therefore, the

bank is faced with the difficulty of estimating both through-the-cycle estimates for regulators and

point-time estimators to satisfy the auditors.

The recovery rate or loss given default is usually negatively correlated with default rate so that

in adverse economic conditions contribute to credit risk is twofold: the default rate increases

while recovery decreases.

The exposure to default (the amount the borrower owes at the time of default) is such that in

case of an overdraft facility or line of credit, it can be conservatively approximated as the

borrowing limit given to the customer. Usually, for a term loan, the exposure to default is the

expected principal during a given year. However, for a portfolio of derivatives, the exposure to

default is complex to be calculated during a year, which might lead to wrong-way risk. Wrong-

way risk occurs where a counterparty to an institution is most likely to default if the value of the

derivative is negative to the bank and positive to the counterparty.

The correlations are challenging to approximate. Despite the fact the Gaussian model is easy to

use, there is no surety that it will reflect how bad the amount lost by a loan in one-in-thousand

days occurrence or the case of economic capital, which is an extreme event.

Apart from the credit risk, a bank must also get concerned about other risks it faces. Such risks

include market, operational, liquidity, and strategic risks. Typically, these risks are assigned to

different arms in a bank but not necessarily independent of each other. These risks interact and

consequently impacting both economic and regulatory capital requirements of the bank.

132
© 2014-2020 AnalystPrep.
Questions

Question 1

Big Data Inc., a U.S. based cloud technology and computing firm, has been offered a

USD 10 million term loan fully repayable in exactly two years. The bank behind the

offer estimates that it will be able to recover 65% of its exposure if the borrower

defaults, and the probability of that happening is 0.8%. The bank’s expected loss one

year from today is closest to:

A. USD 52,000

B. USD 26,000

C. USD 14,000

D. USD 28,000

The correct answer is D.

EL = EA × P D × LR

EA = USD 10, 000, 000

P D = 0.8%

LR = 35%

EL = 10, 000,000 × 0.008 × 0.35 = USD 28, 000

Maturity is irrelevant since the loan is fully repayable in two years.

Option A is incorrect. The loss given default is taken to be 65%. Note that

LGD = 1– Recovery rate .

Option B is incorrect. The loss given default is taken to be 65% and the final result

dividend by two.

Option C is incorrect. The final result is incorrectly divided by 2.

133
© 2014-2020 AnalystPrep.
134
© 2014-2020 AnalystPrep.
Reading 51: Operational Risk

After completing this reading, you should be able to:

Describe the different categories of operational risk and explain how each type of risk

can arise.

Compare the basic indicator approach, the standardized approach, and the advanced

measurement approach for calculating operational risk regulatory capital.

Describe the standardized measurement approach and explain the reasons for its

introduction by the Basel committee.

Explain how a loss distribution is derived from an appropriate loss frequency

distribution and loss severity distribution using Monte Carlo simulations.

Describe the common data issues that can introduce inaccuracies and biases in the

estimation of loss frequency and severity distributions.

Describe how to use scenario analysis in instances when data is scarce.

Describe how to identify causal relationships and how to use Risk and Control Self-

Assessment (RCSA) and Key Risk Indicators (KRIs) to measure and manage operational

risks.

Describe the allocation of operational risk capital to business units.

Explain how to use the power-law to measure operational risk.

Explain the risks of moral hazard and adverse selection when using insurance to

mitigate operational risks.

According to the Basel Committee, operational risk is “the risk of direct and indirect loss

resulting from inadequate or failed internal processes, people, and systems or from external

events.”

The International Association of Insurance Supervisors describes the operational risk as to the

135
© 2014-2020 AnalystPrep.
risk of adverse change in the value of the capital resource as a result of the operational

occurrences such as inadequacy or failure of internal systems, personnel, procedures, and

controls as well as the external events.

Operational risk emanates from internal functions or processes, systems, infrastructural flaws,

human factors, and outside events. It includes legal risk but leaves out reputational and strategic

risks in part because they can be difficult to measure quantitatively.

This chapter primarily discusses the methods of computing the regulatory and economic capital

for operational risk and how the firms can reduce the likelihood of adverse occurrence and

severity.

The Basel Committee’s Seven Categories of Operational Risk

1. Internal fraud: Internal fraud encompasses acts committed internally that diverge from

a firm’s interests. These include forgery, bribes, tax non-compliance, mismanagement of

assets, and theft.

2. External fraud: External fraud encompasses acts committed by third parties. Commonly

encountered practices include theft, cheque fraud, hacking, and unauthorized access to

information.

3. Clients, products, and business practices: This category has much to do with

intentional and unintentional practices that fail to meet a professional obligation to

clients. That includes issues such as fiduciary breaches, improper trading, misuse of

confidential client data, and money laundering.

4. Employment practices and work safety: These are acts that go against laws put in

place to safeguard the health, safety, and general well-being of both employees and

customers. Issues covered include unethical termination, discrimination, and the coerced

use of defective protective material.

5. Damage to physical assets: These are losses incurred to either natural phenomena like

earthquakes or human-made events like terrorism and vandalism

6. Business disruption and system failures: This included supply-chain disruptions and

system failures like power outages, software crashes, and hardware malfunctions.

136
© 2014-2020 AnalystPrep.
7. Execution, delivery, and process management: This describes the failure to execute

transactions and manage processes correctly. Issues such as data entry errors and

unfinished legal documents can cause unprecedented losses.

Large Operational Risks

The three large operational risks faced by financial institutions include cyber risk, compliance

risks, and rogue trader risk.

Cyber Risk

The banking industry has developed technologically. This development is evident through online

banking, mobile banking, credit and debit cards, and many other advanced banking technologies.

Technological advancement is beneficial to both the banks and their clients, but it can also be an

opportunity for cybercriminals. Cybercriminals can either be individual hackers, organized

crime, nation-states, or insiders.

The cyber-attack can lead to the destruction of data, theft of money, intellectual property and

personal and financial data, embezzlement, and many other effects. Therefore, the financial

institutions have developed defenses mechanisms such as account controls and cryptography.

However, financial institutions should be aware that they are vulnerable to attacks in the future;

thus, they should have a plan that can be executed on short notice upon the attack.

Compliance Risks

Compliance risks occur when an institution incurs fines due to knowingly or unknowingly

ignoring the industry’s set of rules and regulations, internal policies, or best practices. Some

examples of compliance risks include money laundering, financing terrorism activities, and

helping clients to evade taxes.

Compliance risks not only lead to hefty fines but also reputational damage. Therefore, financial

institutions should put in place structures to ensure that the applicable laws and regulations are

adhered to. For example, some banks have developed a system where suspicious activities are

137
© 2014-2020 AnalystPrep.
detected as early as possible.

Rogue Trader Risk

Rogue trader risk occurs when an employee engages in unauthorized activities that consequently

lead to large losses for the institutions. For instance, an employee can trade in highly risky assets

while hiding losses from the relevant authorities.

To protect itself from rogue trader risk, a bank should make the front office and back office

independent of each other. The front office is the one that is responsible for trading, and the

back office is the one responsible for record-keeping and the verifications of transactions.

Moreover, the treatment of the rogue trader upon discovery matters. Typically, if the

unauthorized trading occurs and leads to losses, the trader will most likely be disciplined (such

as lawful prosecution). On the other hand, if the trader makes a profit from the unauthorized

trading, this violation should not be ignored because it breeds the culture of risk ignorance,

which can lead to adverse financial drawback.

Comparing the Three Approaches for Calculating Regulatory


Capital

The Basel Committee on Banking Supervision develops the global regulations which are

instituted by the supervisors of each of the banks in each country. Basel II, which was drafted in

1999, revised the methods of computing the credit risk capital. Basel II regulation includes the

approaches to determine the operational risk capital.

The Basel Committee recommends three approaches that could be adopted by firms to build a

capital buffer that can protect against operational risk losses. These are:

I. Basic indicator approach

II. Standardized approach

III. Advanced measurement approach (AMA)

138
© 2014-2020 AnalystPrep.
Basic Indicator Approach

Under the basic indicator approach, the amount of capital required to protect against

operational risk losses is set equal to 15% of annual gross income over the previous three years.

Gross income is defined as:

Gross income = Interest earned − Interest paid + Noninterest income

Standardized Approach

To determine the total capital required under the standardized approach is similar to the

primary indicator method, but a bank’s activities are classified into eight distinct business lines,

with each of the lines having a beta factor. The average gross income for each business line is

then multiplied by the line’s beta factor. After that, the capital results from all eight business

lines are summed up. In other words, the percentage applied to gross income varies business

lines.

Below are the eight business lines and their beta factors:

Business Line Beta Factor


Corporate Finance 18%
Retail Banking 12%
Trading and Sales 18%
Commercial Banking 15%
Agency Services 15%
Retail Brokerage 12%
Asset Management 12%
Payment and Settlement 18%

To use the standardized approach, a bank has to satisfy several requirements. The bank must:

I. Have an operational risk management function tasked with the identification,

assessment, monitoring, and control of operational risk

II. Consistently keep records of losses incurred in each business line.

III. Regularly report operational risk losses incurred in all business lines.

IV. Install an operational risk management system that’s well documented.

V. Regularly subject its operational risk management processes to independent reviews by

139
© 2014-2020 AnalystPrep.
both internal and external auditors.

Advanced Measurement Approach (AMA)

The AMA approach is much more complicated compared to other approaches. Under this

method, the banks are required to treat operational risk as credit risk and set the capital equal

to the 99.9 percentile of the loss distribution less than the expected operational loss, as shown by

the figure below.

Moreover, under the AMA approach, the banks are required to take into consideration every

combination of the eight business lines mentioned in the standardized approach. Combining the

seven categories of operational risk with the eight business lines gives a total of (7 × 8 =) 56

potential sources of operational risk. The bank must then estimate the 99.9 percentile of one-

year loss for each combination and then aggregate each combination together to determine the

total capital requirement.

To use the AMA method, a bank has to satisfy all the requirements under the standardized

140
© 2014-2020 AnalystPrep.
approach, but the bank must also:

I. Be able to estimate unexpected losses, guided by the use of both external and internal

data.

II. Have a system capable of allocating economic capital for operational risk across all

business lines in a way that creates incentives for these business lines to manage

operational risk better.

Currently, the Basel Committee has replaced the three approaches with a new standardized

measurement (SMA) approach. Despite being abandoned by the Basel Committee, some aspects

of the AMA approach are still being used by some of the banks to determine economic capital.

The AMA approach opened the eyes of risk managers to operational risk. However, bank

regulators found flaws in the AMA approach in that there is a considerable level of variation in

the calculation done by different banks. In other words, if different banks are provided with the

same data, there is a high chance that each will come up with different capital requirements

under the AMA.

Standardized Measurement Approach (SMA)

The Basel Committee announced in March 2016 to substitute all three approaches for

determining operational risk capital with a new approach called standardized measurement

approach (SMA).

The SMA approach first defines a quantity termed as Business Indicator (BI). BI is similar to

gross income, but it is structured to reflect the size of a bank. For instance, trading losses and

operating expenses are treated separately so that they lead to an increase in BI.

The BI Component for a bank is computed from the BI employing a piecewise linear relationship.

The loss component is calculated as:

7X + 7Y + 5Z

Where X, Y, and Z are the approximations of the average losses from the operational risk over

141
© 2014-2020 AnalystPrep.
the past ten years defined as:

X – all losses

Y – losses greater than EUR 10 million

Z – losses greater than EUR 100 million

The computations are structured so that the losses component and the BI component are equal

for a given bank. The Basel provides formula used the required capital for the loss and BI

components.

Determination of Operational Risk Loss Distribution

Computations of the economic capital require the distributions for various categories of

operational risk losses and the aggregated results. The operational risk distribution is

determined by the average loss frequency and loss severity.

Average Loss Frequency

The term “average loss frequency” is defined as the average number of times that large losses

occur in a given year. The loss frequency distribution shows just how the losses are distributed

over the one year, specifying the mean and variance.

If the average losses in a given year are λ, then the probability of n losses in a given year is given

by Poisson distribution defined as

e−λ λn
Pr (n) =
n!

Example: Calculating the Probability of Loss

The average number of losses of a given bank is 6. What is the probability that there are

precisely 15 losses in a year?

142
© 2014-2020 AnalystPrep.
Solution

From the question, λ = 6 . Therefore,

e−6 615
Pr (15) = ≈ 0.001
15!

Loss Severity

Loss severity is defined as a probability distribution of the size of each loss. The mean and

variance of the loss severity are modeled using the lognormal distribution. That is, suppose the

standard deviation of the loss size is σ, and the mean is μ then the mean of the logarithm of the

loss size is given by

μσ
ln ( )
√1 + w

The variance is given by:

ln (1 + w)

Where:

2
σ
w =( )
μ

Example: Calculating the Logarithm of Mean and Variance Loss Severity.

The estimated mean and standard deviation of the loss size is 50 and 20, respectively. What is

the mean and standard deviation of the logarithm of the loss size?

Solution

We start by calculating w,

143
© 2014-2020 AnalystPrep.
2
20
w= ( ) = 0.16
50

So the mean of the logarithm of the loss size is given by:

μσ 20 × 50
ln ( ) = ln ( ) = 6.8335
√1 + w √1.16

The variance of the logarithm of the log size is given by:

ln (1 + w) = ln (1.16) = 0.1484

Monte Carlo Simulation for Operational Risk

After estimating λ , μ, and σ, Monte Carlo simulation can be utilized to determine the probability

distribution of the loss as illustrated below:

144
© 2014-2020 AnalystPrep.
The necessary steps of the Monte Carlo Simulations are as follows:

I. Sampling is done from the Poisson distribution to determine the number of loss events

(=n) in a year. For instance, we can sample Percentile of Poisson distribution as a random

number between 0 and 1

II. Sample n times from the lognormal distribution of the loss size for each of the n loss

145
© 2014-2020 AnalystPrep.
occurrences.

III. Sum the n loss sizes to determine the total loss.

IV. Repeat the process (steps I to III) many times.

Example: Demonstration of the Monte Carlo Simulation

Step 1: Sampling is done from the Poisson distribution

Assume the average loss frequency is six, and a number sampled in step I is 0.29. Therefore,

0.29 corresponds to three loss events in a year because using the formula.

e−λ λn
Pr (n) =
n!
Pr (0) + Pr (1) + Pr (2) = 0.284 and Pr (0) + Pr (1) + Pr (2) + Pr (3) = 0.4660

Therefore 0.29 lies in between two cumulative probabilities.

Also, assume that loss size it has a mean of 60 and a standard deviation of 5 using the formulas
μσ
ln ( ) and ln (1 + w).
√1 + w

Step 2: Sample n times from the lognormal distribution

In step II we will sample three times from the lognormal distribution using the mean

⎛ ⎞
2
⎜⎜ 60 × 5 ⎟⎟ ⎛ 5 ⎞
ln ⎜⎜ ⎟⎟ = 5.70 and standard deviation ln 1 + ( ) = 0.0069
⎜⎜ ⎟⎟ ⎝ 60 ⎠
5 2
√ 1 + ( )
⎝ 60 ⎠

Now, assume the sampled numbers are 4.12, 4.70, and 5.5. Note that the lognormal distribution

gives the logarithm of the loss size. Therefore we need to exponentiate the sampled numbers to

get the actual losses. As such, the three losses are e4.12=61.56, e4.70 =109.95 and e5.5 =244.69.

Step 3: Sum the n loss sizes to determine the total loss

146
© 2014-2020 AnalystPrep.
This gives the total loss of 416.20 (61.56+109.95+244.69) in the trial herein.

Step 4: Repeat the process (steps I to III) many times

Step 4 requires that the same process be repeated many times to generate the probability

distribution for the total loss, from which the desired percentile can be computed.

Estimation Procedures for Loss Frequency and Loss Severity

The estimation of the loss frequency and loss severity involves the use of data and subjective

judgment. Loss frequency is estimated from the banks’ data or subjectively estimated by

operational risk professionals after considering the controls in place.

In the case that the loss severity cannot be approximated from the bank’s data, the loss incurred

by other financial institutions may be used as a guide. The methods in which the banks share

information have been laid out. Moreover, there exist data vendor services (such as Factiva),

which is useful at supplying data on publicly reported losses incurred by other banks.

Common Data Issues in the Estimation of Loss Frequency and Severity


Distributions

I. Inadequate historical records: The data available for operational risk losses –

including loss frequency and loss amounts – is grossly inadequate, especially when

compared to credit risk data. This inadequacy creates problems when trying to model the

loss distribution of expected losses.

II. Inflation: When modeling the loss distribution using both external and internal data, an

adjustment must be made for inflation. The purchasing power of money keeps on

changing so that a $10,000 loss recorded today would not have the same effect as a

similar loss recorded, say, ten years ago.

III. Firm-specific adjustments: No two firms are the same in terms of size, financial

structure, and operational risk management. As such, when using external data, it is

essential to make adjustments to the data in cognizance of the different characteristics of

the source and your bank. A simple proportional adjustment can either underestimate or

147
© 2014-2020 AnalystPrep.
overestimate the potential loss.

The generally accepted scale adjustment for firm size is as follows:

Estimated Loss for Bank A


0.23
Bank A Revenue
= Observed Loss for Bank B × ( )
Bank B Revenue

Example: Calculating the Estimated Loss based on its Size

Suppose that Bank A has revenues of USD 20 billion and incurs a loss of USD 500 million.

Another bank B has revenues of USD 30 billion and experiences a loss of USD 300 million. What

is the estimated loss for Bank A?

Solution

0.23
Bank A Revenue
Estimated Losss for Bank A = Observed Loss for Bank B × ( )
Bank B Revenue
0.23
20
= 300 × ( ) = USD 273.29 million.
30

Scenario Analysis in Instances when Data is Scarce

Scenario analysis aims at estimating how a firm would get along in a range of scenarios, some of

which have not occurred in the past. It’s particularly essential when modeling low-frequency

high-severity losses, which are essential to determine the extreme tails of the loss distribution.

The objective of the scenario analysis is to list events and create a scenario for each one.

Scenarios considered come from:

The firms’ own experience;

The experience of other firms;

Market analysts and consultants; or

148
© 2014-2020 AnalystPrep.
The risk management unit in liaison with senior management.

For each scenario, loss frequency and loss severity are approximated. Monte Carlo simulations

are used to determine a probability distribution for total loss across diverse types of losses. The

loss frequency estimates should capture the existing controls at the financial institution and the

type of business.

Estimation of the probability of rare events is challenging. One method is to state several

categories and ask operational risk experts to put each loss to a category. For instance, some of

the categories might be a scenario that happens once every 1,000 years on average, which is

equivalent to λ = 0.001. The bank could also use a scenario happening once every 100 years on

average, which is equivalent to λ = 0.01 , and so on.

Operational risk experts estimate the loss severity, but rather than in the form of the mean and

standard deviation, it is more suitable to estimate the 1 percentile to 99 percentile range of the

loss distribution. These estimated percentiles can be fitted with the lognormal distribution. That

is, if 1 percentile and 99 percentiles of the loss are 50 and 100 respectively, then 3.91 (ln(50))

and 4.61 (ln(100)) are 1 and 99 percentiles for the logarithm of the loss distribution respectively.

The concluding point in scenario analysis is that it takes into consideration the losses that have

never been incurred by a financial institution but can occur in the future. Managerial judgment

is used to analyze the loss frequency and loss severity which can give hints on how such loss

event may appear, which in turn assist the firms in setting up plans to respond to loss event or

reduce the likelihood of it happening.

Allocation of Operational Risk Capital to Business Units

Typically, economic capital is allocated to business units, after which the return on capital is

computed. Similar to credit risk, the same principles are used in the allocation of operational risk

capital. The provision of operational risk capital to business units acts as an incentive to the

business unit manager to reduce the operational risk because if the manager reduces the loss

frequency and severity, less operational capital will be allocated. Consequently, the profit from

the business unit will improve,

149
© 2014-2020 AnalystPrep.
In a nutshell, the allocation of operational risk capital should sensitize the manager on the

benefits of operational risk. Operational risk reduction does not necessarily reach an optimal

point because there exists operational risk in a firm that cannot be avoided. Therefore, cost-

benefit analysis is carried out when operational risk is reduced by increasing the operational

cost.

Use of the Power-law to Measure Operational Risk

The power-law states that, if v is a random variable and another x is a higher value of v (v>x),

then it is true that:

Pr(v > x) = Kx−α

Where K and α are the parameters.

The power holds for some probability distributions, and it describes the fatness of the right tail of

the probability distribution of v. K is a scale factor, and α depends on the fatness of the right tail

of the distribution. That is, the fatness of the right tail increases with a decrease in α.

According to the mathematician G.V Gnedenko, the power for many distributions increases as x

tends to infinity. Practically, the power-law is usually taken to be true for the values of x at the

top 5% of the distribution. Some of the distributions in which the power-law holds to be true are

the magnitude of earthquakes, trading volume of the stocks, income of individuals, and the sizes

of the corporations.

Generally, the power-law holds for the probability distributions of random variables resulting

from aggregating numerous independent random variables. Adding up the independent

variables, we usually get a normal distribution, and fat tails arise when the distribution is a

result of many multiplicative effects.

According to Fontnouvelle (2003), the power-law holds for the operational risk losses, which

turns to be crucial.

150
© 2014-2020 AnalystPrep.
Example: Measuring Operational Risk Using the Power-law

A risk manager has established that there is a 95% probability that losses over the next year will

not exceed $50 million. Given that the power-law parameter is 0.7, calculate the probability of

the loss exceeding (a) 20 million, (b) 70 million, and (c) 100 million.

Solution

According to the power-law

Pr(v > x) = Kx−α

This implies from the question that,

0.05 = K50−0.7 ⇒ K = 0.7731

Thus,

P(v > x) = 0.7731V −0.7

Now,

when x=20

p(v > 20) = 0.7731 (20)−0.7 = 0.09495

when x=70

p(v > 20) = 0.7731 (70)−0.7 = 0.03951

when x=100

p(v > 100) = 0.7731 (100)−0.7 = 0.03071

Managing Operational Risk

151
© 2014-2020 AnalystPrep.
It is crucial to measure operational risk and to compute the required amount of operational risk

capital. However, it is also imperative to reduce the likelihood of significant losses and severity in

case an event occurs. More often, financial institutions learn from each other. That is, if

significant losses are incurred in one of the financial institutions, risk managers other financial

institutions will try and study what happened so that they can make necessary plans to avoid a

similar event.

Some of the methods of reducing operational risk include: reducing the cause of losses, risk

control, and self-assessment, identifying key risk indicators (KRI’s), and employee education.

Causal Relationship

Causal relationships describe the search for a correlation between firm actions and operational

risk losses. It is an attempt to identify firm-specific practices that can be linked to both past and

future operational risk losses. For example, if the use of new computer software coincides with

losses, it is only wise to investigate the matter in a bid to establish whether the two events are

linked in any way.

Once a causal relationship has been identified, the firm should then decide whether or not to act

on it. This should be done by conducting a cost-benefit analysis of such a move.

Risk and Control Self-assessment (RCSA)

Risk and control self-assessment (RCSA) involves asking departmental heads and managers to

single out the operational risks in their jurisdiction. The underlying argument is that unit

managers are the focal point of the flow of information and correspondence within a unit. As

such, they are the persons best placed to understand the risks pertinent to their operations.

Some of the approaches RCSA methods include:

I. Analyzing the historical incidences with line managers

II. Requesting line managers to complete a risk questionnaire

III. Carrying out interviews with line managers and their staff

IV. Utilizing the suggestion boxes and intranet reporting portals

152
© 2014-2020 AnalystPrep.
V. Executing brainstorming in a workshop environment

VI. Analyzing the reports from third parties such as auditors and regulators

RCSA should be done periodically, such as yearly. The problem with this approach is that

managers may not divulge information freely if they feel they are culpable or the risk is out of

control. Also, a manager’s perception of risk and its potential rewards may not conform to the

firm-wide assessment. For these reasons, there is a need for independent review.

Key Risk Indicators (KRIs)

Key risk indicators seek to identify firm-specific conditions that could expose the firm to

operational risk. KRIs are meant to provide firms with a system capable of predicting losses,

giving the firm ample time to make the necessary adjustments. Examples of KRIs include:

Staff turnover

Number of vacant positions

Number of failed transactions over a specified time period

Percentage of employees that take up the maximum leave days on offer.

The hope is that key risk indicators can identify potential problems and allow remedial action to

be taken before losses are incurred.

Education

It is essential to educate the employees on the prohibited business practices and breeding risk

culture where such unacceptable practices might be executed. Moreover, the legal branch of a

financial institution educates the employees to be cautious when writing emails and answering

phone calls. Essentially, employees should be mindful that their emails and recorded calls could

become public knowledge.

Moral Hazard and Adverse Selection when Using Insurance to


Mitigate Operational Risks

153
© 2014-2020 AnalystPrep.
Earlier in the reading, we saw that a bank using the AMA approach could reduce its capital

charge, subject to extensive investment in operational risk management. One of the ways

through which a bank can achieve this is by taking an insurance cover. That way, the firm is

eligible for compensation if it suffers a loss emanating from a covered risk.

For all its advantages, taking an insurance policy comes with two problems:

1. Moral Hazard: Moral hazard describes the observation that an insured firm is likely to

act differently in the presence of an insurance cover. In particular, traders might

increasingly take high-risk positions in the knowledge that they are well protected from

heavy losses. Without such an insurance policy, the traders would be a bit more

cautionary and restricted in their trading behavior.

In a bid to tame the moral hazard problem, insurers use a range of tactics. These may

include deductibles, coinsurance, and policy limits. Stiff penalties may also be imposed in

case there is indisputable evidence of reckless, unrestricted behavior.

A firm can intentionally keep insurance cover private. This ensures that its traders do not

take unduly high-risk positions.

2. Adverse Selection: Adverse selection describes a situation where the risk seller has

more information than the buyer about a product, putting the buyer at a disadvantage.

For example, a company providing life assurance may unknowingly attract heavy

smokers, or even individuals suffering from terminal illnesses. If this happens, the

company effectively takes on many high-risk persons but very few low-risk individuals.

This may result in a claim experience that’s worse than initially anticipated.

On matters trading, firms with poor internal controls are more likely to take up insurance

policies compared to firms with robust risk management frameworks. To combat adverse

selection, an insurer has to go to great lengths toward understanding a firm’s internal

risk controls. The premium payable can then be adjusted to reflect the risk of the policy.

154
© 2014-2020 AnalystPrep.
Question 1

Melissa Roberts, FRM, has observed 12 losses in her portfolio over the last four

years. She believes the frequency of losses follows a Poisson distribution with a

parameter λ. The probability that she will observe a total of 4 losses over the next

year is closest to:

A. 17%

B. 16%

C. 20%

D. 0.53%

The correct answer is A.

We need to find the parameter λ which from the questions we have,

12 losses
λ= = 3 losses per year
4 years

The probability of n losses in a given year is given by Poisson distribution defined as

e−λ λn
Pr (n) =
n!
e−3 34
Pr (n = 4) = = 0.168
4!

Question 2

According to the Basel Committee, a bank has to satisfy certain qualitative standards

to be allowed to use the advanced measurement approach when computing the

economic capital required. Which of the following options is NOT one of the

standards?

A. The bank must have a system capable of allocating economic capital for

155
© 2014-2020 AnalystPrep.
operational risk across all business lines in a way that creates incentives for

these business lines to manage operational risk better.

B. Internal and external auditors must regularly and independently review all

operational risk management processes. The review must include the policy

development process and independent scrutiny of the risk management

function.

C. The bank’s operational risk measurement system should only make use of

internally generated data to avoid the bias associated with external data.

D. The bank must have an operational risk management function tasked with

identification, assessment, monitoring, and control of operational risk.

The correct answer is C.

The Basel committee does not rule out the use of external data by banks. In fact, the

committee recommends the use of a combination of both external and internal data to

estimate the unexpected loss. External data may not conform to a particular firm, but

firms are allowed to scale the data to fit their profiles. In some cases, internal data

may be either insufficient or entirely unavailable, forcing the firm to look elsewhere.

Question 3

Melissa Roberts, FRM, has observed 12 losses in her portfolio over the last four

years. She believes the frequency of losses follows a Poisson distribution with a

parameter λ . The probability that she will observe a total of 4 losses over the next

year is closest to:

A. 17%

B. 16%

C. 20%

D. 0.53%

The correct answer is A.

12 losses
156
© 2014-2020 AnalystPrep.
12 losses
λ (1) = =3 losses per year
4 years

The probability of n losses in T years can be given by:

(λT )n
P r (n) = e−λT
n!

34
Pr (n = 4) = e−3
4!

= 0.168

157
© 2014-2020 AnalystPrep.
Reading 52: Stress Testing

After completing this reading, you should be able to:

Describe the rationale for the use of stress testing as a risk management tool.

Identify key aspects of stress testing governance, including choice of scenarios,

regulatory specifications, model building, stress-testing coverage, capital, and liquidity

stress testing, and reverse stress testing.

Describe the relationship between stress testing and other risk measures, particularly

in enterprise-wide stress testing.

Explain the importance of stressed inputs and their importance in stressed VaR and

stressed ES.

Identify the advantages and disadvantages of stressed risk metrics.

Describe the key elements of effective governance over stress testing.

Describe the responsibilities of the Board of directors and senior management in stress

testing activities.

Identify elements of clear and comprehensive policies, procedures, and documentation

for stress testing.

Identify areas of validation and independent review for stress tests that require

attention from a governance perspective.

Describe the important role of the internal audit in stress testing governance and

control.

Describe the Basel stress testing principles for banks regarding the implementation of

stress testing.

Stress testing is a risk management tool that involves analyzing the impacts of the extreme

scenarios that are unlikely but feasible. The main question for financial institutions is whether

158
© 2014-2020 AnalystPrep.
they have adequate capital and liquid assets to survive stressful times. Stress testing is done for

regulatory purposes or for internal risk management by financial institutions. Stress testing can

be combined with measurement of the risk such as the Vale-at-Risk (VaR) and the Expected

Shortfall (ES) to give a detailed picture of the risks facing a financial institution.

This chapter deals with the internally generated stress testing scenarios, regulatory

requirements of stress testing, governance issues of stress testing, and the Basel stress testing

principles.

Rationale for the Use of Stress Testing as a Risk Management


Tool

Stress testing serves to warn a firm’s management of potential adverse events arising

from the firm’s risk exposure and goes further to give estimates of the amount of

capital needed to absorb losses that may result from such events.

Stress tests help to avoid any form of complacency that may creep in after an extended

period of stability and profitability. It serves to remind management that losses could

still occur, and adequate plans have to be put in place in readiness for every

eventuality. This way, a firm is able to avoid issues like underpricing of products,

something that could prove financially fatal.

Stress testing is a key risk management tool during periods of expansion when a firm

introduces new products into the market. For such products, there may be very limited

loss data or none at all, and hypothetical stress testing helps to come up with reliable

loss estimates.

Under pillar 1 of Basel II, stress testing is a requirement for all banks using the

Internal Models Approach (IMA) to model market risk and the internal ratings-based

approach to model credit risk. These banks have to employ stress testing to determine

the level of capital they are required to have.

Stress testing supplements other risk management tools, helping banks to mitigate

159
© 2014-2020 AnalystPrep.
risks through measures such as hedging and insurance. By itself, stress testing cannot

address all risk management weaknesses, nor can it provide a one-stop solution.

Comparison between Stress Testing and the VaR and ES

Recall that the VaR and ES are estimated from a loss distribution. VaR enables a financial

institution to conclude with X% likelihood that the losses will not exceed the VaR level during

time T. On the other hand, ES enables the financial institutions to conclude whether the losses

exceed the VaR level during a given time T and hence the expected loss will be the ES amount.

VaR and ES are backward-looking. That is, they assume that the future and the past are the

same. On the other hand, stress testing is forward-looking. It asks the question, “what if?”.

While stress testing largely does not involve probabilities, VaR, and ES models are founded on

probability theory. For example, a 99.9% VaR can be viewed as a 1-in-1,000 event.

The backward-looking ES and VaR consider a wide range of scenarios that are potentially good

or bad to the organization. However, stress testing considers a relatively small number of

scenarios that are all bad for the organization.

Specifically, for the market risk, VaR/ES analysis often takes a short period of time, such as a day,

while stress testing takes relatively long periods, such as a decade.

The primary objective of stress testing is to capture the enterprise view of the risks impacting a

financial institution. The scenarios used in the stress testing are often defined based on the

macroeconomic variables such as the unemployment rates and GDP growth rates. The effect of

these variables should be considered in all parts of an institution while considering interactions

between diverse areas of an institution.

Stressed VaR and Stressed ES

Conventional VaR and ES are calculated from data spanning from one to five years, where a daily

variation of the risk factors during this period is used to compute the potential future

movements.

160
© 2014-2020 AnalystPrep.
However, in the case of the stressed VaR and stressed ES, the data is obtained from specifically

stressed periods (12-month stressed period on current portfolios according to Basel rules). In

other words, stressed VaR and stressed ES generates conditional distributions and conditional

risk measures. As such, they are conditioned to a recurrence of a given stressed period and thus

can be taken as a historical stress testing.

Though stressed VaR and stressed ES might be objectively similar, they are different. Typically

the time horizon for the stressed VaR/ES is short (one to ten days), while for the stress testing, it

considers relatively longer periods.

For instance, assume that a stressed period is the year 2007. The stressed VaR would conclude

that if there was a repeat of 2007, then there is an X% likelihood that losses over a period of T

days will not surpass the stressed VaR level. On the other hand, stressed ES would conclude that

if the losses over T days do not exceed the stressed VaR level, then the expected loss is the

stressed ES.

However, stress testing would ask the questions “if the following year (2008) is the same as in

2007, will the financial institution survive?” Alternatively, what if the conditions of the next year

are twice as adverse as that of 2007, will the financial institution survive? Therefore, stress

testing does not consider the occurrence of the worst days of 2008 but rather the impact of the

whole year.

There is also a difference between conventional VaR and the stressed VaR. Conventional VaR can

be back-tested while stressed VaR cannot. That is, if we can compute one-day VaR with 95%

confidence, we can go back and determine how effective it would have worked in the past. We

are not able to back-test the stressed VaR output and its results because it only considers the

adverse conditions which are generally infrequent.

Types of Scenarios in Stress Testing

The basis of choosing a stress testing scenario is the selection of a time horizon. The time

horizon should be long enough to accommodate the full analysis of the impacts of scenarios.

Long time horizons are required in some situations. One-day to one-week scenarios can be

161
© 2014-2020 AnalystPrep.
considered, but three months to two-year scenarios are typically preferred.

The regulators recommend some scenarios, but in this section, we will discuss internally chosen

scenarios. They include using historical scenarios, stressing key variables, and developing ad hoc

scenarios that capture the current conditions of the business.

Historical Scenarios

Historical scenarios are generated by the use of historical data whose all relevant variables will

behave in the same manner as in the past. For instance, variables such as interest rates and the

credit rate spreads are known to repeat past changes. As such, actual changes in the stressed

period will be assumed to repeat themselves while proportional variations will be assumed for

others. A good example of a historical scenario is the 2007-2008 US housing recession, which

affected a lot of financial institutions.

In some cases, a moderately adverse scenario is made worse by multiplying variations of all risk

factors by a certain amount. For instance, we could multiply what happened in the loss-making

one-month period and increase the frequency of movement of all relevant risk movements by ten.

As a result, the scenario becomes more severe to financial institutions. However, this approach

assumes linear relationships between the movements in risk factors, which is not always the case

due to correlations between the risk factors.

Other historical scenarios are based on one-day or one-week occurrences of all market risk

factors. Such events include terrorist attacks (such as 9/11 terrorist attacks) and one-day

massive movement of interest rates (such as on April 10, 1992, when ten-year bond yields

changed by 8.7 standard deviations).

Stressing Key Variables

A scenario could be built by assuming that a significant change occurs in one or more key

variables. Such changes include:

A 2% decline in the GDP

162
© 2014-2020 AnalystPrep.
A 25% decrease in equity prices

A 100% increase in all volatilities

A 4% increase in the unemployment rate

A 200-basis point increase in all interest rates

Some other significant variations could occur in factors such as money exchange rates, prices of

commodities, and default rates.

In the case of the market risk, small changes in measured using the Greek letters (such as delta

and gamma). The Greek letters cannot be used in stress testing because the changes are usually

large. Moreover, Greeks are used to measure risk from a unit market variable over a short period

of time, while stress testing incorporates the interaction of the different market variables over a

long period of time.

Ad Hoc Stress Tests

The stress testing scenarios we have been discussing above are performed regularly, after which

the results are used to test the stability of the financial structure of a financial institution in case

of extreme conditions. However, the financial institutions need to develop ad hoc scenarios that

capture the current economic conditions, specific exposures facing the firm, and update analysis

of potential future extreme events. The firms either generate new scenarios or modify the

existing scenarios based on previous data.

An example of an event that will prompt the firms to develop an ad hoc scenario is the change in

the government policy on an important aspect that impacts the financial institutions or change in

Basel regulation that requires increment of the capital within short periods of time.

The boards, senior management, and economic experts use their knowledge in markets, global

politics, and current global instabilities to come with adverse scenarios. The senior management

carries out a brain-storming event, after which they recommend necessary actions to avoid

unabsorbable risks.

163
© 2014-2020 AnalystPrep.
Using the Stress Testing Results

While stress testing, it is vital to involve the senior management for it to be taken seriously and

thus used for decision making. The stress-testing results are not only to satisfy the “what if”

question, but also the Board and management should analyze the results and decide whether a

certain class of risk mitigation is necessary. Stress testing makes sure that the senior

management and the Board do not base their decision making on what is most likely to happen,

but also consider other alternatives less likely to happen that could have a dramatic result on the

firm.

Model Building

It is possible to see how the majority of the relevant risk factors behave in a stressed period

while building a scenario, after which the impact of the scenario on the firm is analyzed in an

almost direct manner. However, scenarios generated by stressing key variables and ad hoc

scenarios capture the variations of a few key risk factors or economic variables. Therefore, in

order to exhaust the scenarios, it is necessary to build a model to determine how the “left out”

variables are expected to behave in a stressed market condition. The variables stated in the

context of the stress testing are termed as core variables, while the remaining variables are

termed as peripheral variables.

One method is performing analysis, such as regression analysis, to relate the peripheral

variables to the core variables. Note that the variables are based on the stressed economic

conditions. Using the data of the past stressed periods is most efficient in determining

appropriate relationships.

For example, in the case of the credit risk losses, data from the rating agencies, such as default

rates, can be linked to an economic variable such as GDP growth rate. Afterward, general default

rates expected in various stressed periods are determined. The results can be modified (scaled

up or down) to determine the default rate for different loans or financial institutions. Note that

the same analysis can be done to the recovery rates to determine loss rates.

The Knock-On Effects

164
© 2014-2020 AnalystPrep.
Apart from the immediate impacts of a scenario, there are also knock on-effects that reflect how

financial institutions respond to extreme scenarios. In its response, a financial institution can

make decisions that can further worsen already extreme conditions.

For instance, during the 2005-2006 US housing price bubble, banks were concerned with the

credit quality of other banks and were not ready to engage in interbank lending, which made

funding costs for banks rise.

Reverse Stress Testing

Recall that stress testing involves generating scenarios and then analyzing their effects. Reverse

stress testing, as the name suggests, takes the opposite direction by trying to identify

combinations of circumstances that might lead financial institutions to fail.

By using historical scenarios, a financial institution identifies past extreme conditions. Then, the

bank determines the level at which the scenario has to be worse than the historical observation

to cause the financial institution to fail. For instance, a financial institution might conclude that

twice the 2005-2006 US housing bubble will make the financial institution to fail. However, this

kind of reverse stress testing is an approximation. Typically, a financial institution will use

complicated models that take into consideration correlations between different variables to make

the market conditions more stressed.

Finding an appropriate combination of risk factors that lead the financial institution to fail is a

challenging feat. However, an effective method is to identify some of the critical factors such as

GDP growth rate, unemployment rates, and interest rate variations, then build a model that

relates all other appropriate variables to these key variables. After that, possible factor

combinations that can lead to failure are searched iteratively.

Regulatory Stress Tests

US, UK, and EU regulators require banks and insurance companies to perform specified stress

tests. In the United States, the Federal Reserve performs stress tests of all the banks whose

consolidated assets are over USD 50 billion. This type of stress test is termed as Comprehensive

Capital Analysis and Review (CCAR). Under CCAR, the banks are required to consider four

165
© 2014-2020 AnalystPrep.
scenarios:

I. Baseline Scenario

II. Adverse Scenario

III. Scenario

IV. An internal Scenario

The baseline scenario is based on the average projections from the surveys of the economic

predictors but does not represent the projection of the Federal Reserve.

The adverse and the severely adverse scenarios describe hypothetical sets of events which are

structured to test the strength of banking organization and their resilience. Each of the above

scenarios consists of the 28 variables (such as the unemployment rate, stock market prices, and

interest rates) which captures domestic and international economic activity accompanied by the

Board explanation on the overall economic conditions and variations in the scenarios from the

past year.

Banks are required to submit a capital plan, justification of the models used, and the outcomes of

their stress testing. If a bank fails to stress test due to insufficient capital, the bank is required to

raise more capital while restricting the dividend payment until the capital has been raised.

Banks with consolidated assets between USD 10 million and USD 50 million are under the Dodd-

Fank Act Stress Test (DFAST). The scenarios in the DFAST is similar to that in the CCAR.

However, in the DFAST, banks are not required to produce a capital plan.

Therefore, through stress tests, regulators can evaluate the banks to determine their ability to

extreme economic conditions consistently. However, they recommend that banks develop their

scenarios.

Stress Testing Governance

Stress testing governance is important for determining the extent to which financial institutions

do the exercise. Moreover, governance ensures that the assumption is a stress testing has been

well evaluated, and the senior management analyzes the stress testing outcomes and the

166
© 2014-2020 AnalystPrep.
necessary actions taken out of relevant results.

Key Elements of Effective Governance

Stress testing is a broad term used to describe a firm’s examination of its risks and the potential

impact of stressful events on its financial condition. Stress testing seeks to lay bare all the

financial circumstances a firm may find itself in. It helps to understand the range of potential risk

exposures.

The importance of stress testing was highlighted in the post-analysis of the 2007/2008 financial

crises. At many firms, stress testing was insufficiently integrated into their risk management

practices. The few that had shown more commitment to stress testing did not examine

sufficiently severe scenarios, and test results had little or no consideration in decision making.

Most importantly, the crisis brought to the fore the key role of governance in stress testing. In

summary, the following points stand out:

Stress testing is the ultimate responsibility of the management body of an institution.

The buck stops with senior management and the Board of directors. The management

should have an in-depth understanding of the institution’s risk profile and the potential

impact of stress events.

The stress testing program should be guided by clear, well-detailed policies that outline

the procedure to follow from start to finish, as well as describing the role played by

various employees.

There should be consistent documentation of stress testing results and methodologies

There should be an independent review of not just the testing procedure but also the

results of the program.

The ensuing learning outcomes seek to break down these key elements even further.

Responsibilities of The Board of Directors and Senior Management in


Stress Testing Activities

167
© 2014-2020 AnalystPrep.
For effective operation of stress testing, the Board of directors and senior management should

have distinct responsibilities. What’s more, there should be some shared responsibilities,

although a few roles can be set aside exclusively for one of the two groups.

Responsibilities of the Board of Directors

1. The buck stops with the Board: The Board of directors is “ultimately” responsible for

a firm’s stress tests. Even if board members do not immerse themselves in the technical

details of stress tests, they should ensure that they stay sufficiently knowledgeable about

stress-testing procedures and interpretation of results.

2. Continuous involvement: Board members should regularly receive summary

information on stress tests, including results from every scenario. Members should then

evaluate these results to ensure they take into account the firm’s risk appetite and

overall strategy.

3. Continuous review: Board members should regularly review stress testing reports with

a view to not just critic key assumptions but also supplement the information with their

views that better reflect the overall goals of the firm.

4. Integrating stress testing results in decision making: The Board should make key

decisions on investment, capital, and liquidity based on stress test results along with

other information. While doing this, the Board should proceed with a certain level of

caution in cognizance of the fact that stress tests are subject to assumptions and a host

of limitations.

5. Formulating stress-testing guidelines: It’s the responsibility of the Board to come up

with guidelines on stress testing, such as the risk tolerance level (risk appetite).

Responsibilities of Senior Management

1. Implementation oversight: Senior management has the mandate to ensure that stress

testing guidelines authorized by the Board are implemented to the letter. This involves

establishing policies and procedures that help to implement the Board’s guidelines.

2. Regularly reporting to the Board: Senior management should keep the Board up-to-

date on all matters to do with stress testing, including test designs, emerging issues, and

168
© 2014-2020 AnalystPrep.
compliance with stress-testing policies.

3. Coordinating and Integrating stress testing across the firm: Members of senior

management have the responsibility to propagate widespread knowledge on stress tests

across the firm, making sure that all departments understand its importance.

4. Identifying grey areas: Senior management should seek to identify inconsistencies,

contradictions, and possible gaps in stress tests to make improvements to the whole

process.

5. Ensuring stress tests have a sufficient range: In consultations with the Board of

directors, senior management has to ensure that stress testing activities are sufficiently

severe to gauge the firm’s preparation for all possible scenarios, including low-frequency

high-impact events.

6. Using stress tests to assess the effectiveness of risk mitigation strategies: Stress

tests should help the management to assess just how effective risk mitigation strategies

are. If such strategies are effective, significantly severe events will not cause significant

financial strain. If the tests predict significant financial turmoil, it could be that the

hedging strategies adopted are ineffective.

7. Updating stress tests to reflect emerging risks: As time goes, an institution will

gradually gain exposure to new risks, either as a result of market-wide trends or its

investment activities. It is the responsibility of senior management to develop new stress-

testing techniques that reflect the institution’s new risk profile.

Policies, Procedures, and Documentations on Stress Testing

Policies and Procedures

A financial institution should set out clearly stated and understandable policies and procedures

governing stress testing, which must be adhered to. The policies and procedures ensure that the

stress testing of parts of a financial institution converges to the same point.

The policies and procedures should be able to:

Explain the purpose of stress testing;

169
© 2014-2020 AnalystPrep.
Describe the procedures of stress testing;

State the frequency at which the stress testing can be done;

Describe the roles and responsibilities of the parties involved in stress testing;

Provide an explanation of the procedures to be followed while choosing the scenarios;

Describe how the independent reviews of the stress testing will be done;

Give clear documentation on stress testing to third parties (e.g., regulators, external

auditors, and rating agencies);

Explain how the results of the stress testing will be used and by whom;

They were amended as the stress testing practices changes as the market conditions

change;

Accommodate tracking of the stress test results as they change through time; and

Document the activities of models and the software acquired from the vendors or other

third parties.

Documentation

A firm should ensure that its stress-testing activities are well documented, including clear and

precise details about key assumptions, methodologies, limitations, uncertainties, and test results.

Documentation has several advantages:

It gives management the chance to identify trends in risk exposures

It promotes transparency by allowing third parties to scrutinize an institution’s stress-

testing practices and even compare firms

It provides for continuity of stress tests even in the face of significant changes in

management. The new leadership easily “learns the ropes.”

It helps stress test developers to further improve the process by forcing them to think

170
© 2014-2020 AnalystPrep.
clearly about their methods, assumptions, and choices.

Validation and Independent Review Governance

The stress testing governance covers the independent review procedures, which are expected to

unbiased and provide assurance to the Board that stress testing is carried out while following the

firm’s policies and procedures. Financial institutions use diverse models that are subject to

independent review to make sure that they serve the intended purpose.

Areas of Validation and Independent Review for Stress Tests that Require Attention from a

Governance Perspective include:

Ensuring that validation and independent review are conducted on an ongoing basis;

Ensuring that subjective or qualitative aspects of a stress test are also validated and

reviewed, even if they cannot be tested in quantitative terms;

Acknowledging limitations in stress testing;

Ensuring that stress-testing standards are upheld;

Acknowledging data weaknesses or limitations, if any;

Ensuring that there is sufficient independence in both validation and review of stress

tests;

Ensuring that third-party models used in stress-testing activities are validated and

reviewed to determine if they are fit for the purpose at hand;

Ensuring that stress tests results are implemented rigorously, and verifying that any

departure from the recommended actions is backed up by solid reasons;

Role of the Internal Audit in Stress Testing Governance and Control

Internal audit should:

Independently evaluate the performance, integrity, and reliability of stress-testing

171
© 2014-2020 AnalystPrep.
activities;

Ensure that stress tests across the organization are conducted in a sound manner and

remain current in terms of the scenarios tested;

Assess the skills and expertise of the staff involved in stress-testing activities;

Check that approved changes to stress-testing policies and procedures are

implemented and appropriately documented;

Evaluate the independent review and validation exercises;

To accomplish all the above, internal audit staff must be well qualified. They should be well-

grounded in stress-testing techniques and technical expertise to be able to differentiate between

excellent and inappropriate practices.

Capital and Liquidity Stress-testing

When stress-testing for capital and liquidity adequacy, the firm’s overall strategy and

annual planning cycle should be taken into account. Results should be refreshed in

light of emerging strategic decisions that have the potential to affect the financial

stability of the firm significantly.

Stress tests for capital and liquidity accuracy should seek to reveal how cash flows –

both income and expenses – would be affected, especially in an environment where

multiple risks materialize simultaneously.

For entities with subsidiaries, tests should assess the likelihood of individual

subsidiaries facing capital/liquidity crises on their own, regardless of the financial

stability of the parent company.

Stress tests should simulate the effects of capital problems and illiquidity pressures

happening at the same time. Also, there should be efforts to show if one of these

problems can exacerbate the other.

Stress tests should help the firm to come up with contingency plans to deal with

172
© 2014-2020 AnalystPrep.
crippling capital/liquidity problems. The tests should make clear the specific steps

taken, as well as those that cannot be taken in the face of the aforementioned

problems.

Basel Stress-Testing Principles

The Basel Committee emphasizes that stress testing is a crucial aspect by requiring that the

market risk calculations are based on the internal VaR and the Expected Shortfall (ES) models,

which should be accompanied by “ rigorous and comprehensive” stress testing. Moreover, Banks

that use the internal rating approach of the Basel II to calculate the credit risk capital should

perform a stress test to evaluate the strength of their assumptions.

Influenced by the 2007-2008 financial crisis, the Basel Committee published the principles of the

stress-testing for the banks and corresponding supervisors. The overarching emphasis of the

Basel committee was the importance of stress testing in determining the amount of the capital

that will cushion banks against losses due to large shocks.

Therefore, the Basel committee recognized the importance of stress testing in:

Giving a forward-looking perspective on the evaluation of risk;

Overcoming the demerits of modes and historical data;

Facilitating the development of risk mitigation, or any other plans to reduce risks in

different stressed conditions;

Assisting internal and external communications;

Supporting the capital and liquidity planning procedures; and

Notifying and setting of risk tolerance.

When the Basel committee considered the stress tests done before 2007-2008, they concluded

that:

It is crucial to involve the Board and the senior management in stress testing. The

173
© 2014-2020 AnalystPrep.
Board and the senior management should be involved in stress testing aspects such as

choosing scenarios, setting stress testing objectives, analysis of the stress testing

results, determining the potential actions, and strategic decision making. During the

crisis, banks that have had senior management that had an interest in developing a

stress test, which eventually affected their decision making, performed fairly well.

The approaches of the stress-testing did not give room for the aggregation of different

exposures in different parts of a bank. That is, experts from different parts of the bank

did not cooperate to produce an enterprise-wide risk view.

The scenarios chosen in the stress tests were too moderate and were based on a short

period of time. The possible correlations between different risk types, products, and

markets were ignored. As such, the stress test relied on the historical scenarios and left

out risks from new products and positions taken by the banks.

Some of the risks were not considered comprehensively in the chosen scenarios. For

example, counterparty credit risk, risks related to structured products, and product

awaiting securitizations were partially considered. Moreover, the effect of the stressed

scenario on the liquidity was underrated.

Basel Committee Stress Testing Principles

According to the Basel Committee on Banking Supervision’s “Stress Testing Principles”

published in December 2017:

1. Stress testing frameworks should incorporate an effective


governance structure.

The stress testing frameworks should involve a governance structure that is clear,

documented, and comprehensive. The roles and responsibilities of senior management,

oversight bodies, and those concerned with stress testing operation should be clearly

stated.

The stress testing framework should incorporate a collaboration of all required

174
© 2014-2020 AnalystPrep.
stakeholders and the appropriate communication to stakeholders of the stress testing

methodologies, assumptions scenarios, and results.

2. Stress testing frameworks should have clearly articulated and


formally adopted objectives.

The stress testing frameworks should satisfy the objectives that are documented and

approved by the Board of an organization or any other senior governance. The objective

should be able to meet the requirements and expectations of the framework of the bank

and its general governance structure. The staff mandated to carry out stress testing

should know the stress testing framework’s objectives.

3. Stress testing frameworks should capture material and relevant


risks and apply sufficiently severe stresses.

Stress testing should reflect the material and relevant risk determined by a robust risk

identification process and key variables within each scenario that is internally consistent.

A narrative should be developed explaining a scenario that captures risks, and those

risks that are excluded by the scenario should be described clearly and well documented.

4. Stress testing should be utilized as a risk management tool and to


convey business decisions.

Stress testing is typically a forward-looking risk management tool that potentially helps a

bank in identifying and monitoring risk. Therefore, stress testing plays a role in the

formulation and implementation of strategic and policy objectives. When using the

results of stress testing, banks and authorities should comprehend crucial assumptions

and limitations such as the relevance of the scenario, model risks, and risk coverage.

Lastly, stress testing as a risk management tool should be done regularly in accordance

with a well-developed schedule (except ad hoc stress tests). The frequency of a stress

test depends on:

The objective of the stress testing framework;

175
© 2014-2020 AnalystPrep.
The size and complexity of the financial institution; and

Changes in the macroeconomic environment.

5. Resources and organizational structures should be adequate to


meet the objectives of the stress testing framework.

Stress testing frameworks should have adequate organizational structures that meet the

objectives of the stress test. The governance processes should ensure that the resources

for stress testing are adequate, such that these resources have relevant skill sets to

implement the framework.

6. Stress tests should be supported by accurate and sufficiently


granular data and robust IT systems.

Stress tests identify risks and produce reliable results if the data used is accurate and

complete, and available at an adequately granular level and on time. Banks and

authorities should establish a sound data infrastructure which is capable of retrieving,

processing, and reporting of information used in stress tests. The data infrastructure

should be able to provide adequate quality information to satisfy and objectives of the

stress testing framework. Moreover, structures should put place to cover any material

information deficiencies.

7. Models and methodologies to assess the impacts of scenarios and


sensitivities should be fit for purpose.

The models and methodologies utilized in stress testing should serve the intended

purpose. Therefore,

There should be an adequate definition of coverage, segmentation, and

granularity of the data and the types of risks based on the objectives of the

stress test framework. All is done at the modeling stage;

The complexity of the models should be relevant to both the objectives of the

176
© 2014-2020 AnalystPrep.
stress testing and target portfolios being assessed using the models; and

The models and the methodologies in a stress test should be adequately

justified and documented.

The model building should be a collaborative task between the different experts. As such,

the model builders engage with stakeholders to gain knowledge on the type of risks

being modeled and understand the business goals, business catalysts, risk factors, and

other business information relevant to the objectives of the stress testing framework.

8. Stress testing models, results, and frameworks should be subject


to challenge and regular review.

Periodic review and challenge of stress testing for the financial institutions and the

authorities is important in improving the reliability of the stress testing results,

understanding of results’ limitations, identifying the areas that need improvement and

ensuring that the results are utilized in accordance with the objectives of the stress

testing framework.

9. Stress testing practices and findings should be communicated


within and across jurisdictions.

Communicating the stress testing results to appropriate internal and external

stakeholders provides essential perspectives on risks that would be unavailable to an

individual institution or authority. Furthermore, disclosure of the stress test results by

banks or authorities improves the market discipline and motivate the resilience of the

banking sector towards identified stress.

Banks and authorities who choose to disclose stress testing results should ensure that

the method of delivery should make the results understandable while including the

limitations and assumptions to which the stress test is based. Clear conveyance of stress

test results prevents inappropriate conclusions on the resilience of the banks with

different results.

177
© 2014-2020 AnalystPrep.
Question 1

Hardik and Simriti compare and contrast stress testing with economic capital and

value at risk measures. Which of the following statements regarding differences

between the two types of risk measures is most accurate?

A. Stress tests tend to calculate losses from the perspective of the market, while

EC/VaR methods compute losses based on an accounting point of view

B. While stress tests focus on unconditional scenarios, EC/VaR methods focus on

conditional scenarios

C. While stress tests examine a long period, typically spanning several years, EC

models focus on losses at a given point in time, say, the loss in value at the

end of year tt.

D. Stress tests tend to use cardinal probabilities while EC/VaR methods use

ordinal arrangements

The correct answer is C.

A is inaccurate. Stress tests tend to calculate losses from the perspective of

accounting, while EC/VaR methods compute losses based on a market point of view.

B is inaccurate. While stress tests focus on conditional scenarios, EC/VaR methods

focus on unconditional scenarios.

D is also inaccurate. Stress tests do not focus on probabilities. Instead, they focus on

ordinal arrangements like “severe,” “more severe,” and “extremely severe.” EC/VaR

methods, on the other hand, focus on cardinal probabilities. For instance, a 95% VaR

loss could be interpreted as 5-in-100 events.

Question 2

One of the approaches used to incorporate stress testing in VaR involves the use of

stressed inputs. Which of the following statements most accurately represents a

178
© 2014-2020 AnalystPrep.
genuine disadvantage of relying on risk metrics that incorporate stressed inputs?

A. The metrics are usually more conservative (less aggressive)

B. The metrics are usually less conservative (more aggressive)

C. The capital set aside, as informed by the risk metrics, is likely to be

insufficient

D. The risk metrics primarily depend on portfolio composition and are not

responsive to emerging risks or current market conditions.

The correct answer is D.

The most common disadvantage of using stressed risk metrics is that they do not

respond to current issues in the market. As such, significant shocks in the market can

“catch the firm unawares” and result in extensive financial turmoil.

Question 3

Sarah Wayne, FRM, works at Capital Bank, based in U.S. The bank owns a portfolio of

corporate bonds and also has significant equity stakes in several medium-size

companies across the United States. She was recently requested to head a risk

management department subcommittee tasked with stress testing. The aim is to

establish how well prepared the bank is for destabilizing events. Which of the

following scenario analysis options would be best for the purpose at hand?

A. Hypothetical scenario analysis

B. Historical scenario analysis

C. Forward-looking hypothetical scenario analysis and historical scenario

analysis

D. Cannot tell based on the given information

The correct answer is C.

Scenario analyses should be dynamic and forward-looking. This implies that historical

scenario analysis and forward-looking hypothetical scenario analysis should be

179
© 2014-2020 AnalystPrep.
combined. Pure historical scenarios can give valuable insights into impact but can

underestimate the confluence of events that are yet to occur. What’s more, historical

scenario analyses are backward-looking and hence neglect recent developments (risk

exposures) and current vulnerabilities of an institution. As such, scenario design

should take into account both specific and systematic changes in the present and

near future.

Question 4

Senior management should be responsible for which of the following tasks?

I. Ensuring that stress testing policies and procedures are followed to the letter

II. Assessing the skills and expertise of the staff involved in stress-testing

activities

III. Evaluating the independent review and validation exercises

IV. Making key decisions on investment, capital, and liquidity based on stress

test results along with any other information available.

V. Propagating widespread knowledge on stress tests across the firm, and

making sure that all departments understand its importance

A. I, II, and IV

B. I and V

C. III and IV

D. V only

The correct answer is B.

Roles II and III belong to internal audit. Role IV belongs to the board of directors.

180
© 2014-2020 AnalystPrep.
Reading 53: Pricing Conventions, Discounting, and Arbitrage

After completing this reading, you should be able to:

Define discount factor and use a discount function to compute present and future

values.

Define the “law of one price,” explain it using an arbitrage argument, and describe how

it can be applied to bond pricing.

Identify arbitrage opportunities for fixed income securities with certain cash flows.

Identify the components of a US Treasury coupon bond, and compare and contrast the

structure to Treasury STRIPS, including the difference between P-STRIPS and C-

STRIPS.

Construct a replicating portfolio using multiple fixed income securities to match the

cash flows of a given fixed-income security.

Differentiate between “clean” and “dirty” bond pricing and explain the implications of

accrued interest with respect to bond pricing.

Describe the common day-count conventions used in bond pricing.

Discount Factors and Their Use to Compute Present and


Future Values

A discount factor for a particular term gives the value today of one unit of currency due at the

end of that term. It's essentially a discount rate. The discount factor for t years is denoted as

d(t). For example, if d(1) =0.85, then the present value of, say, $1 to be received a year from

today is given by d(t)×$1=$0.85d(t)×$1=$0.85. Discount factors can easily be extracted from

Treasury bond prices. The discount factor d(t) is the factor which, when multiplied by the total

amount of money to be received (principal + interest), gives the price (present value) of the

bond. However, when performing these calculations, it’s important to note that cash flows with

different timings have different discount factors, in line with the time value of money. For

181
© 2014-2020 AnalystPrep.
example, the discount factor that applies to interest due in six months will be different from the

discount factor for interest due in a year, i.e., d(0.5) ≠ d(1), and d(1) < d(0.5).

Bond Price Quotations

A bond quote is the last price at which a bond is traded, expressed as a percentage of par value

(100). Those bonds sold at a discount are priced at less than 100, and another group, although

fewer, are sold at a premium and are priced at more than 100.

US T-bonds are quoted in dollars and fractions of a dollar – paving the way for the so-called

"32nds" convention. And as the wording suggests, 32 portions of a dollar are considered. For

example, if we have a T-bond quoted at 98–16, this means 98 "full" dollars plus 16/32 of a dollar,

i.e., 0.5 dollars. Hence, the quote represents a price of $98.50.

Corporate or municipal bonds, on the other hand, use dollars and eight fractions of a dollar.

A “+” sign at the end of a quote represents half a tick. For example,

16.5
98 − 16 + implies 98 +
32

Treasury Bills

Treasury bills are short term debt obligation issued by the government, which usually last for

one year or less and do not pay coupons. They are usually quoted at a discount to the face value

of 100. The cash price is the face value minus the quoted discount rate.

Let Q be the quoted price of a T-bill and C as the cash price. If the cash price is 97 and there are

90 days until the maturity of the Treasury bill, then:

360
Q= (100 − C)
n
360
Q= (100 − 97)
n
Q = 12

182
© 2014-2020 AnalystPrep.
This means that we would pay 97 today to get the face value of 100 in 90 days. Our discount is
12
12 for every 100 of face value, which means our annual discount rate is around = 12%.
100

Treasury Bonds

Treasury bonds last for more than one year and usually pay coupons. The accrued interest is the

amount of coupon payment accrued between two coupon dates. When we are talking of treasury

bonds,

Cash price = Quoted price + Accrued interest

The Law of One Price and How it Applies to Bond Pricing

The Law of one price states that the price of a security, commodity, or asset should be the same

in two different markets, say, A and B. In other words, if two securities have the same cash flows,

they must have the same price. Otherwise, a trader can generate a risk-free profit by buying on

market A and selling on market B in one risk-free move. Such a possibility is called an arbitrage

opportunity.

The Law of one price describes security price quite well because, in case of an arbitrage

opportunity, traders rush en masse to take advantage of it. Within no time, market forces of

supply and demand adjust the price to eliminate any deviations.

Let’s look at an example.

Consider a 1-year maturity bond with a face value of $100, a coupon rate of 10%, paying coupon

semi-annually. Assume that the borrowing (bank) interest rate is 5% per annum.

The present value of the cash flows from this bond is:

5 105
PV = + = $104.82
1.025 1.025 2

If the bond has a price of $100, an investor can borrow $100 from a bank and buy the bond. After

183
© 2014-2020 AnalystPrep.
six months, they will be able to repay $5 after receiving the first coupon. At this point, the debt

outstanding will be equal to:

debt = $100 + $100 × 2.5% − $5 = $97.5

At the end of the year, the investor will pocket the principal ($100) as well as the second coupon

of $5, making a total of $105.

Debt at this point = $97.5(1.025) = $99.94

Thus, after fully repaying the debt, they will be left with $105 − $99.94 = $5.06, which would

effectively be a risk-free profit.

To exploit this situation, eagle-eyed investors in an efficient market would attempt to buy this

bond by borrowing funds from banks. Increased demand would drive the price up so that at the

end of the day, there would be no arbitrage opportunity.

Liquidity

Liquidity refers to how easily an asset converts to cash. It affects the price of a bond since it

determines how easily the bond could be sold in the future.

Liquidity issues have, at times, causing a violation of the Law of one price, and therefore we can

not always conclude that arbitrage opportunities do not exist at all.

Components of US Treasury Coupon Bond and the Structure of


Treasury Strips

A US Treasury coupon can have strips in two distinct securities: The principal security, also

known as the P-STRIP, and the detached coupons, also called C-STRIPS. The two types of

securities can then be traded separately via a broker.

For instance, suppose we have a 10-year bond with a $100,000 face value and a 10% annual

interest rate. Assuming it initially pays coupons semi-annually, 21 zero-coupon bonds can be

184
© 2014-2020 AnalystPrep.
created. That's the 20 C-STRIPS plus the principal strip (P-STRIP). Each of the C-STRIPS has a

$%5,000 face value, which is the amount of each coupon. The P-STRIP to be received at maturity

has a face value of $100,000.

Why are STRIPS so popular?

They have a very high credit quality because US Treasury securities back them.

Because they are sold at a discount, investors do not need a large stash of money to

purchase them

The payout is known in advance as long as the investor holds them to maturity

They offer a range of maturity dates and can, therefore, be used to match liabilities due

at specific points in the future.

STRIPS are eligible for inclusion in tax-deferred retirement plans and non-taxable

accounts such as pension funds, in which their value would grow tax-free until your

retirement.

Disadvantages of STRIPS

Shorter-term STRIPS tend to trade rich while longer-term STRIPS tend to trade cheap

Sometimes they can be quite illiquid

They typically trade very close to the fair value, thus potential profits are small

How to Construct a Replicating Portfolio Using Multiple Fixed


Income Securities to Match the Cash Flows of a Given Fixed
Income Security

Here’s an example of how a replicating portfolio can be created from multiple fixed income

securities:

Assume we have a 2-year fixed-income security with $100 face value and a 20% coupon rate,

185
© 2014-2020 AnalystPrep.
paid on a semi-annual basis. Assume further that the security has a yield to maturity of 5%.

The present value of the security would be:

10 10 10 10
PBB1 = + + + = $128.21
1 2 3
1.025 1.025 1.025 1.0254

If this bond is determined to be trading cheap, a trader can carry out an arbitrage trade by

purchasing the undervalued bond and shorting a portfolio that mimics (replicates) the bond’s

cash flows. Assume that in addition to our bond above, which we shall call bond 1, we have four

Bond Coupon PV FV Time to


maturity
Bond 2 14% $106.35 $100 Six months
Bond 3 24% $122.58 $100 12 months
Bond 4 10% $113.07 $100 18 months
Bond 5 12% $120.94 $100 24 months

fixed income securities with the following characteristics:

Note that these bonds also pay semi-annual coupons.

Using the above multiple fixed-income securities, we can create a replicating portfolio. However,

we must first determine the percentage face amounts of each bond to purchase,Fi , where

i=1,2,3,4, which match bond one cash flows in every semi-annual period.

14% 24% 10% 12%


Bond1 CF t = F2 × + F3 × + F4 × + F5 ×
2 2 2 2

In these types of calculations, the most straightforward approach to obtaining the values of Fi

involves starting from the end and then working backward. The logic here is simple. At 24

months, only bond 5 makes a payment. Hence at this point, all other values are equal to zero.

12
$110 = F2 × 0 + F3 × 0 + F4 × 0 + F5 × (100 + )%
2

Solving this gives

110
186
© 2014-2020 AnalystPrep.
110
F5 = = 103.77%
106%

Thus, we have to purchase 103.77%×100=$103.77 face value of bond 5

At 18 months, only bonds 4 and 5 make a payment. We can, therefore, obtain the value of F4 as

follows:

10 12%
$10 = F2 × 0 + F3 × 0 + F4 × (100 + ) % + 103.77 ×
2 2

Solving this gives

10 − 103.77 × 0.06
F4 =
1.05

Thus, we have to purchase $3.59 face value of bond 4

To solve for F3,

24 10% 12%
$10 = F2 × 0 + F3 × (100 + ) % + 3.59 × + 103.77% ×
2 2 2
10 − 0.18 − 6.23
F3 = = 3.21%
1.12

Similarly,

14 24% 10% 12%


$10 = F2 × (100 + ) % + 3.21 × + 3.59 × + 103.77 ×
2 2 2 2
10 − 0.39 − 0.18 − 6.23
F2 = = 2.99%
1.07

We can create Cash flows from the replicating portfolio as the product of each bond's initial cash

flows and the face amount percentage. For example, the cash flow from bond five at 12 months is

equal to:

(12%)
× $100 × 103.77% = $6.22
2

187
© 2014-2020 AnalystPrep.
Similarly, the cash flow from bond two at six months

(14%)
× $100 × 2.99% = $0.21
2

Bond Coupon Face Amount CF(t=6) CF(t=12) CF(t=18) CF(t=24)


Bond 2 14% 2.99% 3.2
Bond 3 24% 3.21% 0.39 3.59
Bond 4 10% 3.59% 0.18 0.18 3.77 110
Bond 5 12% 103.77% 6.23 6.23 6.23
Total cashflows 10 10 10 110
Bond 1 cashflows 10 10 10 110

Where t= time in months

As can be seen above, the cash flows from the four bonds replicate bond one cash flows.

Clean vs. Dirty Bond Pricing

The dirty price of a bond is a bond pricing quote that's equal to the present value of all future

cash flows, including interest accruing on the next coupon payment date. Bonds do trade in the

secondary-market before paying any coupon, or after clearing several coupons. In other words,

the day a trader buys or sells the bond could be in between coupon payment dates.

In line with the principle of the time value of money, it's only fair to compensate the seller of a

bond for the number of days they have held the bond between coupon payment dates. We call

this compensation the accrued interest – the interest earned in between any two coupon dates.

(number of days that have elapsed since the last coupon was paid
Accrued interest = c ( )
number of days in the coupon period)

For example, suppose a $1,000 par value bond pays semi-annual coupons at a rate of 20%, and

we've had 120 days since the last coupon was paid. Assuming that there are 30 days in a month,

120
Accrued interest = × $100 = $66.70
180

188
© 2014-2020 AnalystPrep.
The seller would be compensated to the tune of $67.70, while the buyer would see out the

coupon period and receive the remaining $33.30.

The clean price of a bond is the price that doesn’t include any coupon payments.

The dirty and clean prices are also known as the full and quoted prices, respectively.

Day-Count Conventions

When computing the accrued interest, we use one of several day-count conventions. These

include:

Actual/actual

Actual/360

Actual/365

30/360

30E/360 (E stands for Europe)

Interpretation of these conventions is relatively straightforward. For example, the actual/actual

convention considers the actual number of days between two coupon dates. The 30/360

convention assumes there are 30 days in any given month and 360 days in a year.

For purposes of the exam, note the following:

I. We use the actual/actual convention for US government bonds

II. US corporate and municipal bonds use the 30/360 convention

III. The actual/360 convention is common in money markets

Exam tips:

If coupons are paid semi-annually, the denominator should be 180 in both actual/360

and 30/360 conventions. Similarly, the denominator would be 90 for quarterly coupons.

Almost all US Treasury trades settle T + 1, which means that the exchange of bonds for

189
© 2014-2020 AnalystPrep.
cash happens one business day after the trade date.

Clean price = dirty price – accrued interest

Dirty Price Formula

C C C C+F
Price = + + +⋯
k k+1 k+2
(1 + y) (1 + y) (1 + y) (1 + y)(k+n−1)

Where:

P = price

C = semi-annual coupon

k = number of days until the next coupon payment divided by the

the number of days in the coupon period, determined as per the relevant day-count convention.

y = periodic required yield

n = number of periods remaining, including the present one.

F = face value (par value) of the bond

190
© 2014-2020 AnalystPrep.
Question 1

A $1,000 par value US corporate bond pays coupons semi-annually on January 1 and

July 1 at the rate of 20% per year. Mike Brian, FRM, purchases the bond on March 1,

2018, intending to keep it until maturity. The bond is scheduled to mature on July 1,

2021. Compute the dirty price of the bond, given that the required annual yield is

10%.

1. $1,310.25

2. $502.50

3. $400.25

4. $1,100

The correct answer is A.

As a US corporate issue, this bond is valued based on the 30/360 day-count

convention. Under this convention, the number of days between the settlement date

(March 1, 2018) and the next coupon date (July 1, 2018) is 120 (= 4 months at 30

days per month).

20%
Each coupon payment is valued at × $1, 000 = $100
2

C C C C+F
Price = + + +⋯
k k+1 k+2
(1 + y) (1 + y) (1 + y) (1 + y)(k+n−1)

Where:

P = price

C = semi-annual coupon

k = number of days until the next coupon payment divided by the

number of days in the coupon period, determined as per the relevant day-count

convention.

191
© 2014-2020 AnalystPrep.
y = periodic required yield

n = number of periods remaining, including the present one.

F = face value (par value) of the bond

In this case, n=7

100 100 100 100 100 100 100 + 1000


Price = + + + + + +
(1.05)0.67 (1.05)1.67 (1.05)2.67 (1.05)3.67 (1.05)4.67 (1.05)5.67 (1.05)6.67
= 96.78 + 92.18 + 87.79 + 83.61 + 79.62 + 75.83 + 794.44 = 1, 310.25

Question 2

An analyst has been asked to check for arbitrage opportunities in the Treasury bond

market by comparing the cash flows of selected bonds with the cash flows of

combinations of other bonds. If a 1-year zero-coupon bond is priced at USD 97.25 and

a 1-year bond paying a 20% coupon semi-annually, is priced at USD 114.50, what

should be the price of a 1-year Treasury bond that pays a coupon of 10% semi-

annually?

1. $105.88

2. $100

3. $103.35

4. $105

The correct answer is A.

The secret here is to replicate the 1-year 10% bond using the other two treasury

bonds whose price we already know. To do this, you could solve a system of equations

to determine the weight factors, F1 and F2, which correspond to the proportion of

the zero and the 20% bond to be held, respectively.

At every coupon date, the cash flow from the 10% bond should match cash flows from

the zero-bond and the 20% bond.

192
© 2014-2020 AnalystPrep.
At t=1, the 10% bond pays 105, and both the zero-bond and the 20% also have got

payouts of 100 and 110, respectively

105 = F1 × (100) + F2 × 110 … … … … equation1

At t=0.5, the 10% bond pays 5, the zero bond pays 0, and the 20% bond pays 10

5 = F1 × 0 + F2 × 10 … … … … equation2

Solving equation 2,

F2 = 510 = 0.5

Solving equation 1,

105 = 100F1 + 0.5 × 110


50 = 100F1
F1 = 0.5

Thus, the price of the

10% bond = 0.5 × price of zero bond + 0.5 × price of 20% bond
= 0.5 × 97.25 + 0.5 × 114.5 = $105.88

Note: You should assume the prices are given as per $100 face value

193
© 2014-2020 AnalystPrep.
Reading 54: Interest Rates

After completing this reading, you should be able to:

Calculate and interpret the impact of different compounding frequencies on a bond’s

value.

Define spot rate and compute spot rates given discount factors.

Interpret the forward rate and compute forward rates given spot rates.

Define the par rate and describe the equation for the par rate of a bond.

Interpret the relationship between spot, forward, and par rates.

Assess the impact of maturity on the price of a bond and the returns generated by

bonds.

Define the “flattening” and “steepening” of rate curves and describe a trade to reflect

expectations that a curve will flatten or steepen.

Describe a swap transaction and explain how a swap market defines par rates.

Describe overnight indexed swap (OIS) and distinguish OIS rates from LIBOR swap

rates.

Different Compounding Frequencies and Their Effect on Bond


Value

Besides annual interest payments, most securities on today's market have much shorter accrual

periods. For Example, interest may be payable monthly, quarterly (every three months), or semi-

annually resulting in different present values or future values depending on the frequency of

compounding employed. Here's how to calculate the PV and the FV of an investment with

multiple compounding periods per year.

The most important thing is to ensure that the interest rate used corresponds to the number of

194
© 2014-2020 AnalystPrep.
compounding periods present per year.

Future value

m ×n
rq
FV = PV{(1 + )}
m

Where:

rq is the quoted annual rate,

m represents the number of compounding periods (per year)

Lastly, n is the number of years

Present value

Suppose you make PV the subject of the above formula. You should find that

-m∗n
rq
PV = FV{(1 + )}
m

Example: Simple time value of money calculation

Suppose you wish to have $20,000 in your savings account at the end of the next four years.

Assume that the account offers a return of 10 percent per year, compounded monthly. How much

would you need to invest now to have the specified amount after the three years?

Solution

First, we write down the formula to use,

-m∗n
rq
PV = FV{(1 + )}
m

Second, we establish the components that we already have:

195
© 2014-2020 AnalystPrep.
rq=0.10

m=12

n=4 years

Then, we factor everything into the equation to find our PV.

−12×4
0.1
PV = 20000{1 + } = $13 , 429
12

Therefore, you will need to invest at least $13,429 in your account to ensure that you have

$20,000 after three years.

We can do this with the financial calculator with the following inputs:

I 10
N = 12 ∗ 4 = 48; = = 0.833; PMT = 0; FV = 20, 000;
Y 12
CPT ⇒ PV = −13, 429

Here, we can see that the PV on the financial calculator is a negative value since it’s a cash

outflow. The FV has a positive sign since it’s a cash inflow.

We can also rearrange the future value formula to obtain the holding period return (HPR) as

follows:

1
⎡ FV mn ⎤
r q = m ⎢( ) − 1⎥
⎣ PV ⎦

Calculating Discount Factors Given Interest Rate Swap Rates

If we have a series of interest swap rates, it is possible to derive discount factors. The notional

amount, which is technically never exchanged between counterparties, determines the size of

both fixed and floating leg payments.

If we could exchange the notional amount, the fixed leg of the swap would resemble a fixed

196
© 2014-2020 AnalystPrep.
coupon-paying bond, with fixed leg payments acting like semi-annual, fixed coupons, and the

notional amount acting like the principal payment. Floating rate payments would act like coupon

payments of the floating-rate bond.

We denote the discount factor for t-years as d(t). The methodology used to come up with discount

factors when dealing with interest rate swaps is similar to that used to find discount factors

when dealing with bonds.

Example: Calculating discount factors

Compute the discount factors for maturities ranging from six months to two years, given a

notional swap amount of $100 and the following swap rates:

Maturity (years) Swap rates


0.5 0.75
1.0 0.85
1.5 0.98
2.0 1.20

The four discount factors d(0.5), d(1.0), d(1.5) and d(2.0) can be calculated as follows:

0.75
(100 + )d (0.5) = 100
2
100
d (0.5) = = 0.9963
100.375

…… … …… … …… … …… …

0.85 0.85
d (0.5) + (100 + ) d (1.0) = 100
2 2
0.425 × 0.9963 + 100.425 d(1.0) = 100
100 − 0.4234
d (1.0) + = 0.9916
100.425

…… … …… … …… … …… …

197
© 2014-2020 AnalystPrep.
0.98 0.98 0.98
d (0.5) + d (1.0) + (100 + ) d (1.5) = 100
2 2 2
0.49 × 0.9963 + 0.49 × 0.9916 + 100.49 d (1.5) = 100
100 − 0.4882 − 0.4859
d (1.5) = = 0.9854
100.49

…… … …… … …… … …… …

1.20 1.20 1.20 1.20


d (0.5) + d (1.0) + d (1.5) (100 + ) d (2.0) = 100
2 2 2 2
100 − 0.5978 − 0.5950 − 0.5912
d (2.0) = = 0.9763
100.6

Maturity (years) Discount factor


0.5 0.9963
1.0 0.9916
1.5 0.9854
2.0 0.9763

Computing Spot Rates Given Discount Factors

A t-period spot rate is a yield to maturity on a zero-coupon bond that matures in t years,

assuming semi-annual compounding. We denote the t-periodic spot rate as z(t).

Spot rates and discount factors are related as shown in the following formula, assuming semi-

annual coupons:

1
⎡ 1 2t ⎤
z (t) = 2 ⎢( ) − 1⎥
⎣ d (t) ⎦

Spot Rate vs. Forward Rates

A spot interest rate gives you the price of a financial contract on the spot date. The spot date is

the day when the funds involved in a business transaction are transferred between the parties

involved. It could be two days after a trade, or even on the same day, we complete the deal. A

spot rate of 5% is the agreed-upon market price of the transaction based on current buyer and

seller action.

198
© 2014-2020 AnalystPrep.
In theory, forward rates are prices of financial transactions that might take place at some future

point. The spot rate tells you "how much it would cost to execute a financial transaction today".

The forward rate, on the other hand, tells you “how much would it cost to execute a financial

transaction at a future date X”.

We agree on spot and forward rates in the present. The only difference comes in the timing of

execution.

Example: Converting spot rates into forward rates

Compute the six-month forward rate in six months, given the following spot rates:

Z (0.5) =1.6%

Z (1.0) =2.2%

Solution

The six-month forward rate, f(1.0), on an investment that matures in one year, must solve the

following equation:

199
© 2014-2020 AnalystPrep.
2 1
0.022 0.016 f(1.0)1
(1 + ) = (1 + ) × 1+
2 2 2
1
f(1.0)
1.0221 = 1.008 × (1 + )
2
f(1.0)
1.01399 − 1 =
2
f (1.0) = 0.02797 = 2.8%

Par Rate

The par rate is the rate at which the present value of a bond equals its par value. It’s the rate

you’d use to discount of all a bond’s cash flows so that the price of the bond is 100 (par). For a

100-par value, the two-year bond that pays semi-annual coupons, and we can easily calculate the

2-year par rate provided we have the discount factor for each period

Par rate
= [d (0.5) + d (1.0) + d (1.5) + d (2.0)] + 100 d (2.0) = 100
2

In general, for any maturity C T, and assuming a par value of $1,

CT 2T t
∑ d( ) + d (T) = 1
2 t=1 2

The sum of the discount factors is called the annuity factor, A 1, and is given by:

2T t
∑ d( )
t=1 2

Therefore,

CT
× AT + d(T ) =
2

Impact of Maturity on the Price of a Bond and the Returns


Generated by Bonds

200
© 2014-2020 AnalystPrep.
In general terms, bond prices will tend to increase with maturity whenever the coupon rate is

above the forward rate throughout maturity extension. The opposite holds: bond prices will tend

to decrease with maturity whenever the coupon rate is below the forward rate for a maturity

extension.

To help you understand just how this happens, assume we have two investors with the

opportunity to invest in either STRIPS or a five-year bond. The investor who opts for the 5-year

bond (which utilizes forward rates) will have a simple task: they will invest at the onset and then

wait to receive regularly scheduled coupon payments, plus the principal amount at maturity. The

investor who opts for STRIPS (which utilizes spot rates) will roll them over as they mature

throughout the five years. Rolling over implies that when one STRIP expires, the investor will use

the proceeds to invest in the next six-month contract, and so on for five years.

In market conditions where short-term rates are above the forward rates utilized by bond prices,

the investor who rolls over the STRIPS will tend to outperform the investor in the 5-year bond.

The opposite is exact: In market conditions where short-term rates are below the forward rates

utilized by bond prices, the investors who roll over the STRIPS will tend to underperform the

investor in the 5-year bond.

Properties of Spot, Forward and Par Rates

If the term structure is constant, and that all spot rates are equal, then all par rates

201
© 2014-2020 AnalystPrep.
and all forward rates equal the spot rate.

If the term structure is increasing, the par rate for a specific maturity is below the spot

rate for the same timeline.

If the term structure is decreasing, the par rate for a specific maturity is above the spot

rate for the same maturity.

If the term structure is increasing, forward rates for a period starting at time T are

higher than the spot rate for maturity T.

Flattening and Steepening of Rate Curves

A yield curve represents the yield of each bond along a maturity spectrum that's plotted on a

graph. The most and widely accepted yield curves pit the three-month versus two-year T-bonds

or the five-year versus ten-year T-bonds. On occasion, we may use the Federal Funds Rate versus

the 10-year Treasury note.

The yield curve typically slopes upwards indicating that the rate of interest on long-term bonds

is higher than the rate on short-term bonds which reflects the investors' demands to be

compensated for taking on more risk by investing long-term. Such a curve is said to be normal.

202
© 2014-2020 AnalystPrep.
Other than a standard yield curve, we could have a flattening yield curve or a steepening yield

curve.

Flattening Yield Curve

A flat yield curve indicates that little difference exists between short-term and long-term rates

for similarly rated bonds. It may manifest as a result of long-term interest rates falling more than

short-term interest rates or short-term rates increasing more than long-term rates.

203
© 2014-2020 AnalystPrep.
A flattening curve reflects the expectations of investors about the macroeconomic outlook. It may

be the result of the following events:

A decrease in expected inflation, causing the premium loaded on long-term rates to

decrease; or

An impending increase in the Federal Funds Rate. (An increase in the FFR could cause

an increase in short term rates while long-term rates remain relatively stable, causing

the curve to flatten.)

A trader who anticipates a flattening of the yield curve can buy a long-term rate and

sell short a short-term rate because they expect bond prices to rise in the long-term.

Steepening Yield Curve

204
© 2014-2020 AnalystPrep.
A steepening yield curve indicates a widening gap between the yields on short-term bonds and

long-term bonds. A steepening curve could occur when long-term rates rise faster than short-

term rates. Sometimes, short-term rates can also show some defiance by decreasing even as

long-term rates rise.

For example, assume that a two-year note was at 2.3% on July 15, and the 10-year was at 3.3%.

By August 30, the two-year note could have risen to 2.38% and the 10-year to 3.5%. The

difference would effectively go from 1 percentage point to 1.12 percentage points, resulting in a

steeper yield curve.

A steepening curve reflects the expectations of investors about the macroeconomic outlook. It

may be the result of the following events:

205
© 2014-2020 AnalystPrep.
an increase in expected inflation, causing the premium loaded on long-term rates by

investors to increase; or

an impending decrease in the Federal Funds Rate.

A trader who anticipates a steepening of the yield curve can sell short a long-term rate and buy a

short-term rate because they expect bond prices to fall in the long-term (bond prices fall as rates

increase).

LIBOR and Overnight Rates

LIBOR (London Interbank Offered Rate) is a rate of interest at which banks across the world

borrow from each other. LIBOR is derived from average estimates of different borrowing rates

provided by different AA-rated banks globally. These rates are quoted for different currencies,

and the borrowing periods are between one day and one year.

Although we use LIBOR rates as a benchmark borrowing rate, we should note that the estimates

used to come up with the average borrowing rate are subject to manipulation by the different

banks. In efforts to curb these issues, the market has come up with other benchmarks, and an

excellent example of this is the overnight rate.

Overnight rate

While the borrowing period for LIBOR ranges from one day to one year, overnight rates apply for

a day only. Banks with insufficient funds may borrow from those banks with surplus funds to

meet their financial needs overnight.

Brokers always accompany these transactions, and different countries have systems that monitor

the activities of these brokers. For example, we have the Federal Reserve in the US, the Sterling

Overnight Index Average (SONIA) in the UK, and the Euro Overnight Index Average (EONIA) in

the Eurozone.

Swaps

206
© 2014-2020 AnalystPrep.
A swap is a contract in which two parties agree to exchange a sequence of cash flows for a given

period. A contract in which one party agrees to swap a fixed rate for Libor is the most frequently

exercised type of swap.

A swap has two legs: a fixed leg in which the interest rate is fixed and a floating one in which the

rates are varying, and the interest is compound. Each day, banks will come up with a new

Overnight Index; hence the rates keep on fluctuating, i.e., swaps are associated with a fixed

interest rate and a variable interest rate.

Overnight Indexed Swap (OIS)

An OIS is a swap in which an investment compounded daily for three months is exchanged with a

fixed rate of interest for five years.

The floating rate per three-month is given by:

(1 + r1 d 1) (1 + r2 2) … . (1 + ri di ) … (1 + rn dn ) − 1

Where:

r is the overnight rate per day on a day i.

d is the number of calendar days where r applies and,

n is the number of trading days in the three months.

OIS is the fixed rate in the overnight indexed swaps.

While Libor swaps define par bond that can be used to determine the Libor term structure, OIS

rates describe par bonds that can be used to determine an OIS term structure.

207
© 2014-2020 AnalystPrep.
Question 1

Given the following spot rates, compute the 6-month forward rate in 1 year.

z (1.0) = 3.25%

z (1.5) = 3.60%

A. 4.2%

B. 1.5%

C. 4.3%

D. 5.1%

The correct answer is C.

Solution:

The 6-month forward rate on an investment that matures in 1.5 years must solve the

following equation:

1
0.036 3 0.03252 f (1.5)
(1 + ) = (1 + ) × (1 + )
2 2 2
1
f (1.5)
1.05498 = 1.03276 × (1 + )
2
f (1.5)
1.02152 − 1 =
2
f (1.5) = 0.04303 = 4.3%

Question 2

Consider a bond with par value of USD 1,000 and maturity in four years. The bond

pays a coupon of 4% annually. The spot rate curve is as follows:

208
© 2014-2020 AnalystPrep.
n − year S pot rate
1 − year 5%
2 − year 6%
3 − year 7%
4 − year 8%

The value of the bond is closest to:

A. $1,160

B. $500

C. $870.78

D. $850

The correct answer is C.

The value of the bond is the present value of all future cash flows (coupons plus

principal), discounted at the various spot rates.

Each coupon = 4% × 1 , 000 = $40

40 40 40 40 + 1000
PV = + + +
1 2 3
(1 + 0.05) (1 + 0.06) (1 + 0.07) (1 + 0.08)4
= 38.10 + 35.60 + 32.65 + 764.43 = $870.78

209
© 2014-2020 AnalystPrep.
Reading 55: Bond Yields and Return Calculations

After completing this reading, you should be able to:

Distinguish between gross, and net realized returns and calculate the realized return

for a bond over a holding period, including reinvestments.

Define and interpret the spread of a bond and explain how to derive a spread from a

bond price and a term structure of rates.

Define, interpret, and apply a bond’s yield-to-maturity (YTM) to bond pricing.

Compute a bond's YTM, given a bond structure and price.

Calculate the price of an annuity and perpetuity.

Explain the relationship between spot rates and YTM.

Define the coupon effect and explain the relationship between the coupon rate, YTM,

and bond prices.

Explain the decomposition of P&L for a bond into separate factors, including carry roll-

down, rate change, and spread change effects.

Explain the following four common assumptions in carry roll-down scenarios: realized

forwards, unchanged term structure, consistent yields, and realized expectations of

short-term rates; and calculate carry roll down under these assumptions.

Gross vs. Net Realized Returns

The gross realized return on investment has two components: Any increase in the price of the

asset plus income received while holding the investment. When dealing with bonds,

Ending value + Coupon - Beginning value


Gross realized return t−1 ,t =
Beginning value

210
© 2014-2020 AnalystPrep.
Example: A bond’s gross realized return over six months

What is the gross realized return for a bond that is currently selling for $1,060 if it was

purchased exactly six-months ago for $1,000 and paid a $20 coupon today?

Solution

Ending value + Coupon-Beginning value


Gross realized return =
Beginning value
1, 060 + 20 − 1000
= = 8%
1000

Now let's assume the investor financed the purchase of the bond by borrowing cash to invest at a

rate of 1% for six months. The gross realized return less per period financing costs gives the net

realized return.

Ending value + Coupon-Beginning value


1%
Net realised return = −
Beginning value 2

When calculating the gross/net realized return for multiple periods, it's essential to consider

whether coupons received are reinvested. If the coupons are reinvested, they will grow at the

reinvestment rate from the time they are received up to the end of the period. The risk that

reinvested cash flows will increase by a rate that's lower than the expected rate is known as

reinvestment risk.

Bond Spread

The spread of a bond is the difference between its market price and the price computed

according to spot rates or forward rates – the term structure of interest rates.

As a relative measure, a bond's spread helps us determine whether the bond is trading cheap or

rich relative to the yield curve. We incorporate spread in the bond price formula as follows:

Recall that given a 2-year bond with a face value of P, paying annual coupons each of amount C,

211
© 2014-2020 AnalystPrep.
its price is given by:

C C+P
Market bond price = +
(1 + f (1.0)) (1 + f (1.0)) × (1 + f (2.0))

To incorporate the spread s, we assume that the bond is trading at a premium or discount to this

computed price. We can find the bond’s spread using the following formula:

C C+P
= +
(1 + f (1.0) + s) (1 + f (1.0) + s) × (1 + f (2.0) + s)

Yield to Maturity

Yield to maturity (YTM) of fixed income security is the total return anticipated if we hold the

security until it matures. Yield to maturity is considered a long-term bond yield, but we express

as an annual rate. In other words, it's the security's internal rate of return as long as the investor

holds it up to maturity. To compute a bond’s yield to maturity, we use the following formula:

C1 C2 C3 CN
p= + + ⋯+
1 2 3
(1 + y) (1 + y) (1 + y) (1 + y)N

Where:

P=price of the bond

C t=annual cash flow in year t

N=time to maturity in years

y=annual yield (YTM to maturity)

When cash flows are received multiple times every year, we can slightly modify the above

formula such that:

C1 C2 C3 Cn
p= + + ⋯+
(1 + y) 1
(1 + y) 2
(1 + y) 3
(1 + y)n

212
© 2014-2020 AnalystPrep.
Where:

P=price of the bond

C t=periodic cash flow in period t

n=N×m=number of periods (=years×number of periods per year)

Provided all cash flows received are reinvested at the YTM; the yield to maturity is equal to the

bond's realized return.

For zero-coupon bonds that are not accompanied by recurring coupon payments, the yield to

maturity is equal to the normal rate of return of the bond. We use the formula below to

determine YTM for zero-coupon bond:

1
Face value year to maturity
Yield to maturity = ( ) −1
Current price of bond

Exam tip: The yield to maturity assumes cash flows will be reinvested at the YTM and assumes

that we hold the bond until maturity.

Prices of Annuities and Perpetuities

An annuity is a series of annual payments of PMT until the final time T. The value of an ordinary

annuity is given by:

1 − (1 + r)−T
PVannuity = PMT
r

Where:

r=discount rate

Perpetuity is a type of annuity whose cash flows continue for an infinite amount of time. The

present value of a perpetuity is given by:

PMT
213
© 2014-2020 AnalystPrep.
PMT
PVperpetuity =
r

The Relationship between Spot Rates and YTM

We can use both the spot rate and the yield to maturity to determine the fair market price of a

bond. However, while the yield to maturity is constant, the spot rate varies from one period to

the next to reflect interest rate expectations as time goes.

The spot rate is a more accurate measure of the fair market price when interest rates are

believed to rise and fall over the coming years.

Given a bond's cash flows and the applicable spot rates, you can easily calculate the price of a

bond. You can then determine the bond's YTM by equating the price to the present values of cash

flows discounted at the YTM.

The Coupon Effect

The coupon effect describes the fact that reasonably priced bonds of the same maturity but

different coupons have different yields to maturity, which implies that yield is not a reliable

measure of relative value. Even if fixed-income security A has a higher yield than fixed security

B, A is necessarily not a better investment.

It also follows that if two bonds have identical features save for the coupon, the bond with the

smaller coupon is more sensitive to interest rate changes. In other words, given a change in

yield, the lower coupon bond will experience a higher percentage change in price compared to

the bond with larger coupons. The most sensitive bonds are zero-coupon bonds, which do not

make any coupon payments.

Exam tips:

The lower the coupon rate, the higher the interest-rate risk. The greater the coupon

rate, the lower the interest rate risk.

214
© 2014-2020 AnalystPrep.
If coupon rate > YTM, the bond will sell for more than par value or at a premium.

If the coupon rate < YTM, the bond will sell for less than par value, or at a discount.

If coupon rate= YTM, the bond will sell for par value.

Over time, the price of premium bonds will gradually fall until they trade at par value at maturity.

Similarly, the price of discount bonds will gradually rise to par value as maturity gets closer. This

phenomenon is known as "pulling to par."

Components of a Bond’s P&L

We generate the bond's profitability or loss through price appreciation and explicit cash flows.

The total price appreciates as follows:

P t+1 (R t+1 s t+1 ) − P t (R tst )

There are three components of price appreciation:

I. Carry-roll-down component: The carry-roll-down component comprises of price

changes that emanate from a deviation of term structure from the original

structure to an expected term structure, denoted as R e, as maturity approaches.

It does not account for spread changes

Carry roll down component = P t+1 (Ret+1 st) − P t (Rt st)

II. Rate changes component: The rate changes component accounts for price

changes due to interest rate movements from an expected term structure to the

term structure that exists at time t+1.

Rate changes component = P t+1 (R t+1 s t) − Pt (R et+1 st )

This component also doesn’t account for spread changes.

III. Spread change component: As the words suggest, the spread change component

215
© 2014-2020 AnalystPrep.
accounts for price changes emanating from changes in the bond’s spread from

time t to t+1.

Spread change component = P t+1 (Rt+1 st+1 ) − P t+1 (R t+1 st )

Assumptions in Carry-Roll-Down Scenarios

I. Realized forwards: The return to a bond held to maturity is the same as rolling the

investment one period at a time at the forward rates. However, in reality, some forwards

are realized above or below the initial forwards.

II. Unchanged term structure

III. Unchanged yields

216
© 2014-2020 AnalystPrep.
Question 1

A bond currently selling for $1,060 was purchased exactly 12 months ago for $1,000

and paid a $20 coupon six months ago. Today, the bond paid a $20 coupon. The

coupon received six months ago was reinvested at an annual rate of 2%. Given that

the purchase price was entirely financed at a yearly rate of 1%, the net realized

return of the bond is closest to:

A. 8.9%

B. 9.0%

C. 10.0%

D. 9.5%

The correct answer is A.

Solution:

BVt + C t − BV t−1
Netrealised return t−1,t = − financing costs
BVt−1
2%
1, 060 + 20 + 20 (1 + ) − 1000
2
= − 1.0%
1000
= 10.02% − 1% = 8.9%

Question 2

On Jan 1 2017, Commercial Bank of India issued a six-year bond paying an annual

coupon of 6% at a price reflecting a yield to maturity of 4%. As of Dec 31, 2017,

interest rates remain unchanged. Holding all other factors constant, and assuming a

flat term structure of interest rates, how was the bond's price affected? The price:

A. Remained constant

B. Decreased

217
© 2014-2020 AnalystPrep.
C. Increased

D. Increased, but only in the second half of the year

The correct answer is B.

From the data given, it's clear that the bond's coupon is higher than the yield. As

such, the bond must have traded at a premium – implying the price must have been

higher than the face value. Provided the yield doesn't change; a bond's price will

always converge to its face value. Since the price starts higher, it must decrease. This

phenomenon is called 'pulling to par."

218
© 2014-2020 AnalystPrep.
Reading 56: Applying Duration, Convexity, and DV01

After completing this reading, you should be able to:

Describe a one-factor interest rate model and identify common examples of interest

rate factors.

Define and compute the DV01 of fixed income security, given a change in yield and the

resulting change in price.

Calculate the face amount of bonds required to hedge an option position given the

DV01 of each.

Define, compute, and interpret the effective duration of fixed income security given a

change in yield and the resulting change in price.

Compare and contrast DV01 and effective duration as measures of price sensitivity.

Define, compute, and interpret the convexity of fixed income security, given a change in

yield and the resulting change in price.

Explain the process of calculating the effective duration and convexity of a portfolio of

fixed income securities.

Describe an example of hedging based on effective duration and convexity.

Construct a barbell portfolio to match the cost and duration of a given bullet

investment, and explain the advantages and disadvantages of a bullet versus barbell

portfolios.

This chapter discusses one-factor risk metrics, which include DV01, duration, and convexity, as

used in the analysis of fixed-income portfolios. We consider these measures to quantify the

consequence of a parallel shift in the interest rate term structure: DV01 and duration consider a

small parallel shift in the term structure while convexity extends the duration to accommodate

larger parallel shifts. Hedging can, in accordance with these risk metrics, be efficient for a

parallel shift of term structure as compared to non-parallel shifts.

219
© 2014-2020 AnalystPrep.
One-factor Interests Rate Models

The one-factor assumption states that when the rates are driven by one factor, the change of one

interest rate can be used to determine the change in all other interest rates over a short period

of time. For instance, a one-factor model assumes that all interest rates changes by the same

amount. As such, the shape of the term structure never changes. That is, if a one-year spot rate

increases by two basis points, all other spot rates increase by two basis points. DV01, duration,

and convexity are examples of such models.

It is worth noting that shifts in the term structure are not always parallel. For instance, a one-

factor model might predict that if one-year spot rate increases by five basis points over a short

period, then the two year increases by three basis points and the ten-year rate increases by one

basis point.

220
© 2014-2020 AnalystPrep.
DVO1

DV01 reflects the effect of a one-basis movement in interest rates on the value of a portfolio.

DV01 is defined as:

ΔP
DV01 = −
Δr

Where

Δr = the size of a parallel shift in the interest rate term structure measured in basis points

ΔP = resultant change in the value of the position being considered

Note that for a long position in bonds, the DV01 is positive due to a negative correlation between

the bond’s price and interest rate changes.

221
© 2014-2020 AnalystPrep.
DV01 is defined in three different ways:

i. Year-based DV01: defined as the change in the price from a one-basis point increase in

the yield of a bond.

ii. DVDZ or DPDZ: defined as the change in price from a one-point increase in all spot

rates.

iii. DVDF or DPDF: defined as a change in price from a one-basis point increase in the

forward rate.

Example: Calculating DV01 from the One-Basis Point Change in Spot


Rates

A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%

paid semiannually. The spot rates, as shown in the table below:

Maturity Period (Years) Spot Rate (%)


0.5 6.0
1.0 6.5
1.5 7.0
2.0 7.5

What is the value of the DV01 if:

i. the spot rates increase by five basis points?

ii. the spot rates decrease by five basis points?

Solution

i. The price of a bond with no spread is USD 101,003.01 calculated using the formula:

C 2T ⎛ 1 ⎞ 100, 000
P= ∑ ⎜⎜⎜ ⎟⎟ +
2 i=1 1 + y ⎟ y
2T
⎝ ⎠ (1 + )
2 2

In this case,

222
© 2014-2020 AnalystPrep.
c = 100, 000 × 0.08 = 8, 000, T = 2

⎛ ⎞
8000⎜ 1 1 1 1 ⎟⎟ 100, 000
⇒p= ⎜⎜ + + + ⎟⎟ +
⎜ 0.06 2 3
2 ⎜
1 + 0.065 0.07 0.075 4 ⎟ 0.075 4
⎝ 2 (1 + ) (1 + ) (1 + ) ⎠ (1 + )
2 2 2 2
1 1 1 1 100, 000
= 4000 ( + 2
+ 3
+ )4 +
1.03 (1.0325) (1.035) (1.0375) (1.0375)4
= 101, 003.01

If the spot rates are each increased by 5 basis points (0.05%) so that the six-month spot rate is

6.05%, the one-year spot rate is 6.55%, and so on, the price of the bond is USD 100,911.18,

calculated as:

8000 1 1 1 1 100, 000


= ( + + + )4 +
2 3
2 1.03025 (1.03275) (1.03525) (1.03775) (1.03775)4
= 100, 911.18

We know that:

ΔP 100, 911.18 − 101, 003.01


DV01 = − =−
Δr 5
91.83
= = 18.366
5

Note that the rise of the spot rates by 5 basis points decreases the bond price by 91.83

(=100,911.18 − 101,003.01), and the DV01, in this case, measure the decline of a bond price for

each one-basis point increase in spot rates.

ii. Assume now that the spot rates decrease by 5 basis points so that the six-month spot rate is

5.95% (=6 − 0.05), the one-year spot rate is 6.45% (=6.5 − 0.05) and so on. Under decreased

spot rates, the price of the bond is USD 101,094.96 calculated as

1 1 1 1 100, 000
= 4, 000( + + + ) + = 101, 094.96
1.02975 (1.02975)2 (1.03475)3 (1.03725)4 (1.03725)4

And thus,

223
© 2014-2020 AnalystPrep.
ΔP 101, 094.96 − 101, 003.01
DV01 = − =−
Δr −5
91.95
= = 18.39
−5

As such, an increase of spot rates by five basis points causes the price of the bond to increases

by 91.95 (=101,094.96 − 101,003.01), and thus DV01, in this case, measures the increase of

bond price for each one-basis point decrease in spot rates.

It is worth noting that the DV01 for the decrease and increases of the basis points are slightly

different because the bond price is not a linear function of interest rates. We estimate the DV01

by averaging the estimates above:

18.366 + 18.39
DV01 = = 18.38
2

Example: Calculating DV01 from the one-basis point change in bond


yield

Assume that now bond yield increases by one basis point. Consider the spot rates, as in the

example above.

Maturity Period (Years) Spot Rate (%)


0.5 6.0
1.0 6.5
1.5 7.0
2.0 7.5

A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%

paid semiannually. The spot rates, as shown in the table above.

What is the value of DV01 if the bond yield

i. Increases by 5 basis points?

224
© 2014-2020 AnalystPrep.
ii. Decreases by the 5 basis points?

Solution

Recall that, using the spot rates, we calculated the bond price to be 101,003.01. To find the yield

to maturity of the bond (bond yield), we solve the equation:

4 , 000 4, 000 4, 000 104 , 000


+ + + = 101, 003.01
y y 2
y 3
y 4
1+ (1 + ) (1 + ) (1 + )
2 2 2 2

Using a financial calculator with the variables N=4, PMT=4000, PV=−101,003.01, we get:

y
= 3.7255% ⇒ y = 7.45%
2

i. Now, if the bond yield increases by 5 basis point to 7.50% (=7.45 + 0.05), the price of the bond

is 100,912.846 (using the financial calculator) and thus for an increase of 5 basis points in bond

yield, the bond price decreases by 90.164 (=100,912.85 − 101,003.01). We can, therefore,

calculate the DV01:

ΔP 100, 912.85 − 101, 003.01


DV01 = − =−
Δr 5
−90.164
= = 18.033
5

ii. Similarly, if the bond yield decreases by 5 basis points to 7.40% (=7.45 − 0.05), the bond price

is 101,096.71. A decrease of bond yield by 5 basis points increases the bond price by 93.70

(=101,096.71 − 101,003.01). Thus,

Similarly, if the bond yield decreases by 5 basis points to 7.40% (=7.45 − 0.05), the bond price is

101,096.71. A decrease of bond yield by 5 basis points increases the bond price by 93.702

(=101,096.71 − 101,003.01). Thus,

ΔP 101 096.71 − 101 003.01


225
© 2014-2020 AnalystPrep.
ΔP 101, 096.71 − 101, 003.01
DV01 = − =−
Δr −5
−93.164
= = 18.632
5

The average of the estimates is 18.33 given as:

18.0328 + 18.632
≈ 18.33
2

In the cases of the forward rates, the analogy is similar to that of the spot rates. However, this

chapter will primarily emphasize on DV01 computed from a one-basis-point parallel shift in the

term structure of spot rates.

Hedging Using DV01 Metric

In any position that depends on the interest, DV01 can be computed efficiently. DV01 can,

therefore, be used to hedge a position. For example, assume that a bank has a position whose

DV01 is −40. By the definition of DV01, the banks will gain from their position if interest rates

increases and will undoubtedly lose value if interest rates decrease. More specifically, if all the

interest increases by the one-basis point, the value of the bank’s position increases by USD 40.

On the contrary, if all interest rates decrease by one basis point, the value of the bank’s position

decreases by USD 40.

Now assume that this bank wants to hedge its position with a coupon bond that pays an annual

coupon rate of 8% payable semiannually, has a face value of USD 100,000 and bond yield of

7.50%. If we can determine a position in the coupon bond that is exactly 40, the bank’s position

would have been hedged.

A hedge ratio determines the amount of par of the hedge position that needs to be bought or sold

for every $1 par value of the original position. The goal of hedging is to lock in the value of a

position even in the face of small changes in yield. The hedge ratio is given by:

DV01
226
© 2014-2020 AnalystPrep.
DV01Initial Position
HR =
DV01Hedge Position

Assume that DV01 of the coupon bond is 18.33. We use the hedge ratio and the bond’s face value

to increase the DV01 of the bond to 40. In this case, we need to increase the value of the position

by:

40
100, 000 × = $218, 221.50
18.33

Thus, adding the USD 218,221.50 of coupon bond’s position to the bank’s portfolio protects

against small changes in the term structure. In other words, a downward (upward) movement of

the term structure will result in profit (loss), which will be offset by the gain (loss) from the

bond’s position.

Effective Duration (D)

Effective duration measures the percentage change in the price of a bond (or other instruments)

caused by small changes in all rates. Note that effective duration is different from the DV01

because DV01 measures actual price changes against small changes in all rates.

Effective duration is defined as:

ΔP
P ΔP
D=− =
Δr P × Δr

The effective duration can also be rewritten as:

ΔP = −D × P × Δr

When the change in all rates is measured basis points, the effective duration is equivalent to

DV01 divided by the bond price.

Consider the DV01 example on the spot rates where we had calculated the price of the bond USD

227
© 2014-2020 AnalystPrep.
101,003.01 and the DV01 of 18.38 so that the duration is:

18.33
= 0.000182 = 0.0182%
101003.01

Effective duration gives the proportional change in the price of an instrument corresponding to

one-basis point change in all interest rates.

Typically, the effective duration is stated as a percentage change in the price of an instrument

for a 100-basis-change in all rates by multiplying the effect of one-basis-point change by 100.

Therefore, our example above would be stated as 1.82% per 100 basis points.

However, for the sake of clarity, the duration in this chapter will be reported as a decimal. In the

decimal reporting system, one basis point is equivalent to 0.0001, and thus, we measure duration

per 10,000 basis points. Therefore, the duration calculated above will be:

0.000182 × 10 , 000 = 1.82

Callable and Puttable Bonds

A callable bond is one that an issuing party has the right to purchase back the bond at a pre-

determined price at a particular time in the future before the maturity period of the bond. The

call feature of the bond should not be ignored as it reduces the duration of the bond.

A practical approach to address the callable feature of a bond while calculating duration is

outlined as follows:

Calculate the value of the bond today;

Compute the bond value if the all interest rates increase by one basis point while

incorporating the effect of the one-basis point increase on the bond’s callable

probability of the bond; and

Compute the effective duration from the percentage change in the price.

228
© 2014-2020 AnalystPrep.
A puttable bond is one that a holder has a right to ask for early repayment. Calculating the

effective duration of a puttable bond is similar to that of a callable bond. That is, the probability

of the put options increases with the increase of the interest rates.

Comparing DV01 and Effective Duration

DV01 is useful in measuring the effect of all rate changes on the value of a position. DV01 is also

appropriate in measuring the changes in swaps and interest rate futures.

Effective duration is appropriate in measuring the effect of rate changes on the value of a

position as a percentage. Effective duration is an appropriate measure in bond valuation.

DV01 increases with an increase in the position size, but effective duration does not. In other

words, if the value of a position is doubled, DV01 doubles, but effective duration does not.

Effective Convexity

Convexity measures the sensitivity of duration measure to movement in the interest rates.

Denoted convexity by C and value of a position by P, convexity is defined as by:

1 P+ + P − − 2P
C= [ ]
P (Δr)2

Where:

P + = the value of the position when all rates increase by Δr

P − = the value of the position corresponding to the decrease of all rates by Δr

Note that Δr is given in decimal.

Example: Calculating Convexity

A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%

paid semiannually. The spot rates, as shown in the table below:

229
© 2014-2020 AnalystPrep.
Maturity Period (Years) Spot Rate (%)
0.5 6.0
1.0 6.5
1.5 7.0
2.0 7.5

What is the value of the convexity if all spot rates change by 5 basis points?

Solution

As computed earlier, the price of a bond with no interest spread is USD 101,003.01 calculated

as:

⎛ ⎞
8000⎜ 1 1 1 1 ⎟⎟ 100, 000
p= ⎜⎜ + + + ⎟⎟ +
2 ⎜⎜ 0.06 0.065 2 0.07 3 0.075 4 ⎟ 0.075 4
1 + (1 + ) (1 + ) (1 + ) (1 + )
⎝ 2 ⎠
2 2 2 2
= 101, 003.01

If the spot rates are each increased by 5 basis points (0.05%) so that the six-month spot rate is

6.05%, the one-year spot rate is 6.55%, and so on, the price of the bond is USD 100,911.18

calculated as:

1 1 1 1 100, 000
= 4000 ( + +
2
+
3
)4 + = 101, 094.96
1.02975 (1.03225) (1.03475) (1.03725) (1.03725)4

So,

1 P+ + P− − 2P
C= [ ]
P (Δr)2
1 100, 911.18 + 101, 094.96 − 2 × 101, 003.01
= [ ]
101, 003.01 (0.0005)2
= 0.4752

The Impact of Parallel Shifts

230
© 2014-2020 AnalystPrep.
Recall that the effective duration can be written

ΔP = −D × P × Δr

The duration of a bond is a linear relationship between the bond price and interest rates where,

as interest rates increase, bond price decreases. It is a good measure of sensitivity to interest

rates for small/sudden changes in yield. However, for much larger changes in yield, the duration

is not so desirable since the relationship between price and interest rates tends to be non-linear.

Consider the following diagram:

ΔP
The relationship between and Δr is non-linear. However, effective duration approximates this
P
ΔP
as a linear relationship because it is the gradient of the curve linking and Δr. As seen on the
P

graph, the linear approximation by effective duration becomes less efficient as the rate changes

231
© 2014-2020 AnalystPrep.
become large.

The combination of the duration and convexity is a quadratic function which provides a better

estimation. As such the price change is given by

1
ΔP = −DPΔr + CP(Δr)2
2

Where Δr is given in decimal.

By combining the duration and the convexity, we can reach a convenient estimate, even for large

changes. Consider the example in the DV01 section.

Example: Price Change Approximation Using Effective Duration and


Convexity

A two-year treasury bond has a face value of USD 100,000, with an annual coupon rate of 8%

paid semiannually. The spot rates, as shown in the table below:

Maturity Period (Years) Spot Rate (%)


0.5 6.0
1.0 6.5
1.5 7.0
2.0 7.5

Note that we had calculated the DV01 for the 5 basis point movement of all the spot rates to be

18.38, the price of the bond with no spread is USD 101,003.01, the duration to be 1.82 and

convexity to be 0.4752.

Now we want to:

i. Estimate actual bond price change due to an increase in all the spot rates by 30 basis

points;

ii. Estimate bond price change using effective duration; and

iii. Estimate bond price change using a combination of the effective duration.

232
© 2014-2020 AnalystPrep.
Solutions

i. Actual Change due to a change in all rates

Because all the spot rates increase by 30 basis points (0.3%), the six-month spot rate is 6.3%, the

one-year spot rate is 6.8%, and so on. So the price due to new rates is given by the formula:

i
⎛ 1 ⎞
C 2T 100, 000
P= ∑ ⎜⎜⎜ ⎟⎟ +

2 i=1 1 + y y 2T
⎝ 2⎠ (1 + )
2

In this case,

c = 100, 000 × 0.08 = 8, 000, T = 2

⎛ ⎞
8000⎜ 1 1 1 1 ⎟⎟ 100, 000
= ⎜⎜ + + + ⎟⎟ +
2 ⎜⎜ 1 + 0.063 0.068 2 0.073 3 0.078 4 ⎟ 0.078 4
⎝ 2 (1 + ) (1 + ) (1 + ) ⎠ (1 + )
2 2 2 2
1 1 1 1 100, 000
= 4000 ( + + + ) +
1.0315 (1.034)2 (1.0365)3 (1.039)4 (1.039)4
= 100, 453.63
⇒ ΔP = 100, 453.63 − 101, 003.01 = −549.38

Therefore the price of the bond decreases by 549.38 (=100,453.63 − 101,003.01).

ii. Estimated price change using effective duration

The estimated price change using effective duration is given by:

ΔP = −DP Δr
= −1.82 × 101, 003.01 × 0.003 = −551.48

iii. Estimated price change using a combination of duration and


convexity

The estimated price change provided by the duration is a good approximation of the price

233
© 2014-2020 AnalystPrep.
change, but we can improve it by combining duration and the convexity.

The price change estimation due to a combination of duration and convexity is given by:

1 2
ΔP = −DPΔr + CP(Δr)
2
1 2
= −1.82 × 101, 003.01 × 0.003 + × 0.4752 × 101,003.01 × (0.003)
2
= −551.48 + 0.21598 = −551.26

Hedging Based on Effective Duration and Convexity

Hedging Based on Duration only

Hedging using effective duration analogous to that of DV01. Now denote the duration of

investment by DV, and its investment value by V. On the other hand, denote the effective duration

of a bond by DB and its value by P. By the definition of effective duration,

ΔV = −V DV Δr

and

ΔV = −V DB Δr

Where Δr is the size of a small parallel shift in the term structure. These small parallel shifts are

hedged against if:

−VDVΔr − PDBΔr = 0
⇒ Δr (−VDV − PDB ) = 0
∴ −VDV − PDB = 0

If we make P the subject, the position required in the bond is:

VDV
P=
DB

Example: Hedging Based on Effective Duration

234
© 2014-2020 AnalystPrep.
A bank has a position of USD 12 million with an effective duration of 5 and a bond with an

effective duration of 6. How will the bank hedge against its position?

Solution

The position required is given by:

VDV
P=−
DB
5 × 12
=− = −10
6

Therefore, the bank should short bonds worth 10 million to hedge against its position. This is

true because:

ΔV = −12 × 5 × Δr = −60Δr

and

ΔP = −10 × 6 × Δr = 60Δr
⇒ ΔV + ΔP = −60Δr + 60Δr = 0

Hedging Based on Duration and Convexity

Hedging based on both duration and convexity is a complex undertaking where we are trying to

make both effective duration and convexity. We need two bonds. Define:

P 1 = a value of the first bond

D1 = duration of the first bond

C 1 = convexity of the first bond

P 2 = a value of the second bond

D2 = duration of the second bond

C 2 = convexity of the second bond

235
© 2014-2020 AnalystPrep.
Also, define

V = a value of the position to hedged

P V = duration of the position to be hedged

C V = convexity of the position to be hedged

Now by the definition of approximating the price change using a combination of the duration and

convexity we have:

1
ΔV = −VDV Δr + VC V(Δr)2
2
1
ΔP 1 = −P 1D1 Δr + P1 C 1 (Δr)2
2

and

1
ΔP1 = −P2 D2 Δr + P 2 C 2 (Δr)2
2

If we make both the duration and convexity to be zero, we have:

−VDV − P1 D1 − P 2 D2 = 0

Also,

VCV − P1 C 1 − P 2 C2 = 0

These equations must be satisfied for a position to be hedged.

Example: Hedging Based on Convexity and Duration

A bank has a position of USD 12 million with an effective duration of 5 and a convexity of 4. The

bank wishes to hedge its position with two bonds where the first bond has an effective duration

of 6 and a convexity of 9. The second bond has a duration of 4 and a convexity of 7. How will the

bank hedge against its position?

236
© 2014-2020 AnalystPrep.
Solution

We know that a position is hedged against if the following equations are satisfied:

VDV − P 1 D1 − P 2D2 = 0
VC V − P 1 C1 − P 2 C2 = 0

Therefore, we have

V = 12 , DV = 5 , and CV = 4.For the bonds we have C 1 = 9, D1 = 6 also C2 = 7 , D2 = 4so that:

60 − 6P 1 − 4P 2 = 0
108 − 9P 1 − 7P 2 = 0

Solving the above equation simultaneously, we get:

P1 = 8.77 and P2 = 4.15

Therefore, for the bank to hedge its position, it must take a position of USD 8.77 in the first bond

and short USD 4.15 million in the second bond. In other words, by combining these positions in

bonds, there is no duration or convexity exposure, and thus the bank is hedged against large

parallel shifts in the term structure though it will be exposed to non-parallel shifts.

Yield-Based Duration and Convexity

Yield-Based Duration

Consider a bond whose price is P and yield is y. If the cash flows from the bond are c1 , c 2, … , cn at

times t1 , t2 ,… , tn respectively, the relationship between P and y (continuously compounded) is:

n
P = ∑ C i e−yti
i=1

If we differentiate with respect to y, we have:

dp
237
© 2014-2020 AnalystPrep.
dp n
= − ∑ Ci e−yti
dy i=1
ΔP n
⇒ = − ∑ tiC i e−yti
Δy i=1

Now, the yield-based duration is defined as the proportional change in the bond price due to a

small change in the yield. More precisely,

1 ΔP 1 n
D=− =− (− ∑ ti Ci e−yti)
P Δy P i =1
n Ci e−yti
⇒ D = − ∑ ti
i=1 P

C ie−yti
The expression in the formula denotes the proportion of the value of the bond received at
P
the time ti . Therefore, the duration can be computed by taking the average of the times when the

bond’s cash flows were received weighted by the bond’s value at each time. As such, the

duration will measure the time an investor must wait to receive cash flows.

Typically, the yield on the bond is given as a semi-annual compounding rate rather than
1
continuous compounding. As such, the expression for the duration is divide by y to get:
1+
2

⎛ ⎞
1 ⎟ n C i e−yti
D = ⎜⎜⎜ ⎟ ∑ ti
y ⎟ i=1 P
⎝1 + 2 ⎠

The resulting equation gives the formula for the modified duration while the expression
1
without dividing by y is termed as Macaulay Duration. Modified duration is slightly
1+ 2
different from an effective duration.

Yield-Based Convexity

Recall convexity measures the sensitivity of duration measure to movement in the interest rates.

In the calculus context, yield based convexity is the derivative if the duration. Therefore,

−yt

238
© 2014-2020 AnalystPrep.
1 n n C i e−yti
C= ∑ t2i C ie−yti =∑ t2i
P i =1 i=1 P

Intuitively, convexity can be defined as the weighted (by bond’s value) average of the squared

time to maturity. When yields are given as semi-annual compounding, the convexity expression is
1
multiplied by so that:
y 2
(1+ )
2

1 n C i e−yti
C= ∑ t2i
y 2 i=1 P
(1 + )
2

The last expression is termed as Modified convexity, and it is slightly different from the

effective convexity.

Portfolio Interest-Rate Sensitivity Measures

We now need to consider a group of instruments (such bonds) where we need to compute the

interest-rate sensitivity measures we have been discussing.

DV01 for a Portfolio

The DV01 of a portfolio is simply the sum of individual DV01 of instruments in the portfolio.

For instance, if a bank has four positions whose DV01s are 9, 10, 12, and 13 (in USD millions),

then the DV01 for the portfolio is 44 (=9+10+12+13).

Duration for a Portfolio

Portfolio duration can be calculated as the weighted sum of the individual durations. The

weight attached to each security is equal to its value as a percentage of total portfolio value.

k
Portfolio duration = ∑ wj Dj
j=1

239
© 2014-2020 AnalystPrep.
Where:

Dj = duration of the bond j

w j = market value of the bond j divided by total portfolio market value

k = number of bonds in the portfolio

For instance, consider a portfolio consisting of four instruments with values (in USD millions) of

8, 13, 15 and 18, and respective effective durations are eight, seven, six, and four. The effective

duration for the portfolio is given by:

8 13 15 18
DEff = × 8+ × 7+ × 6+
8 + 13 + 15 + 18 8 + 13 + 15 + 18 8 + 13 + 15 + 18 8 + 13 + 15 + 18
= 1.1852 + 1.6852 + 1.6667 + 1.3333 = 5.87

The effective portfolio duration is used to determine the impact of the small parallel shift of the

term structure of the interest rates. Similar computations can be used in yield-based durations.

Convexity for a Portfolio

Portfolio convexity is computed using a similar approach. It’s defined as the value-weighted

average of the individual bond convexities making up the portfolio.

k
Portfolio convexity = ∑ w jC j
j=1

Barbell Portfolio vs. Bullet Portfolio

A barbell is an investment strategy applicable to fixed-income securities whereby half the

portfolio is made up of long-term bonds and the other half of short-term bonds. On the other

hand, a bullet strategy is an investment strategy where the investor buys bonds concentrated

only in the intermediate maturity range.

Given the prices, coupon rates, maturities, yields, durations, and convexities of a set of bonds, it

240
© 2014-2020 AnalystPrep.
is possible to construct a barbell portfolio with the same cost and duration as the bullet portfolio.

This involves determining the proportion of each security in the Barbell that should be bought,

such that their total value equals that of the bullet.

Example: Demonstration of Barbell and Bullet Portfolio

Consider the following three bonds:

Bonds Bond Value Effective Duration Effective Convexity


2-year, 3% coupon 88.90 4.567 21.23
5-year, 5% coupon 99.80 7.767 78.90
10-year, 7% coupon 120.45 13.43 180.56

Assume that the term structure of the interest rates is flat at 5%, compounded semiannually.

Note that this means that the yield on all instruments is 5%.

Now, assume that an investor wants a portfolio with a duration of 7.767. The investor can either

buy a 5-year, 5% coupon (bullet investment) or build a portfolio of other two bonds to have the

desired duration of 7.767 (barbell investment).

If the investor wishes to construct a bond from the other two bonds, denote:

The proportion of the 2-year, 3% coupon bond by α, and

The proportion of the 10-year, 7% coupon bond by 1-α.

Thus, the duration of the bond is

4.567α + 13.43(1 − α) = 7.767


⇒ α = 0.6389

Therefore, the investor can create a bond with a duration of 7.767 by either investing all his

money in a 5-year coupon bond, or invest 63.89% of his funds into the 2-coupon bond, and

36.11% of his funds in the 10-year coupon bond.

Note that the portfolios have equal duration but different convexities. The convexity of the 5-

year, 5% coupon bond is 78.90. For the portfolio consisting of 2-year, 3% coupon, and 10-year,

241
© 2014-2020 AnalystPrep.
7% coupon bonds, the convexity is given by 78.8088 (=0.6389×21.3+0.3611×180.56) which is

better than the investing in one bond only.

An arbitrage opportunity can occur if one invests an amount in the barbell portfolio and short the

same amount of the bullet portfolio. The arbitrage opportunity is possible if the movement in the

term structure is parallel (which is not always true). Typically, the bullet investment is profitable

for many non-parallel movements of the term structure as compared to the barbell investment.

Most of the models are constructed in such a way there are no arbitrage opportunities to the

investors.

Advantages and Disadvantages of Barbell Portfolios

Advantages:

It may be more liquid compared to a barbell portfolio;

Spreading out bond purchases ensures higher yields when rates are rising; and

The investor need not have a “war chest” at the onset since the portfolio is built

gradually.

Disadvantages:

When the yield curve flattens (short rates go up; long rates go down), the Barbell

outperforms the bullet.

Note: Sometimes, the bullet and Barbell have the same duration, but they will have different

convexities.

242
© 2014-2020 AnalystPrep.
Question

The price of a 3-year bond is USD 10,000. The DV01 of the bond is 50. What is the

estimated bond price change if all rates increase by five basis points?

A. 15

B. 20

C. 25

D. 40

Solution

Recall that when the change in all rates is measured in basis points, the effective

duration is equivalent to DV01 divided by the bond price. Therefore, in this case, the

effective duration is given by:

50
D= = 0.005 = 50
10, 000

The estimated price change using effective duration is given by:

ΔP = −DPΔr
= −50 × 10,000 × 0.0005 = USD 25

Note: We measure the effective is decimal, i.e., per 10,000 basis points and thus:

D = 0.005 = 0.005 × 10, 000 = 50

Also, 5 basis points = 0.05% = 0.0005.

243
© 2014-2020 AnalystPrep.
Reading 57: Modeling and Hedging Non-Parallel Term Structure Shifts

After completing this reading you should be able to:

Describe and assess the major weakness attributable to single-factor approaches when

hedging portfolios or implementing asset-liability techniques.

Describe the principal components analysis and explain its use in understanding term

structure movements.

Define key rate exposures and know the characteristics of key rate exposure factors

including partial ‘01s and forward-bucket ‘01s.

Describe key-rate shift analysis.

Define, calculate, and interpret key rate ‘01 and key rate duration.

Describe the key rate exposure technique in multi-factor hedging applications;

summarize its advantages and disadvantages.

Calculate the key rate exposures for a given security, and compute the appropriate

hedging positions given a specific key rate exposure profile.

Relate key rates, partial ‘01s, and forward-bucket ‘01s, and calculate the forward-

bucket ‘01 for a shift in rates in one or more buckets.

Apply key rate and multi-factor analysis to estimating portfolio volatility.

Major Weakness Attributable to Single-factor Approaches


When Hedging Portfolios or Implementing Asset Liability
Techniques

The main weakness attributable to single-factor approaches to portfolio hedging has much to do

with the assumption that movements in the entire term structure can be exhaustively described

by one interest rate factor. In other words, the single-factor approach erroneously assumes that

all rate changes within the term structure of interest rates are driven by a single factor.

244
© 2014-2020 AnalystPrep.
From a practical point of view, rates in different regions of the term structure are not always

correlated. As an example, the single-factor approach tells us that the 6-month rate can perfectly

predict the change in the 30-year rate. This in turn informs the hedging of the 30-year bond with

a 6-month bill. Such a move is unlikely to hedge the total risk inherent in the 30-year bond.

Predicted changes in the 30-year rate based purely on changes in the 6-month rate can be quite

misleading. That’s because rates in different regions of the term structure (yield curve) are not

always correlated. The risk of such non-parallel shifts along the yield curve is known as yield

curve risk.

Key Rate Exposures

Key rate exposures help to describe the risk distribution along the term structure given a bond

portfolio. They help describe how to execute the perfect hedge using highly liquid benchmark

bonds. Bonds used for this purpose are normally government bonds issued in the recent past,

which means they are likely to be trading at or near par.

Partial ‘01s are used to measure and hedge the risk of portfolios of swaps or portfolios that

combine both bonds and swaps in terms of the most liquid money market and swap instruments.

Forward bucket ‘01s are also used to measure and the hedge risk of portfolios of swaps/bond

combinations, but the difference here is that instead of measuring risk based on other

comparable securities on the market, they measure risk based on changes in the shape of the

yield curve. Forward bucket ‘01s present an intuitive way to understand the yield curve risk of a

portfolio, but they are otherwise not efficient at recommending the perfect hedges to neutralize

such risks. To compute forward ‘01s, the yield curve is divided into several defined regions.

A Description of the Key-rate Shift Analysis

The assumption behind the key rate shift analysis is that the entire spectrum of rates can be

considered a function of a few select rates at specified points along the yield curve. Thus, to

measure risk and predict interest rate movements, a small number of key rates are used, usually

245
© 2014-2020 AnalystPrep.
those of highly liquid government bonds.

The rates most commonly used are the U.S. Treasury 2-, 5-, 10-, and 30-year par yields. As the

words suggest, a “key rate shift” occurs when any of these rates shifts by one basis point. The

key rate technique indicates that changes in each key rate will affect rates from the term of the

previous key rate to the term of the subsequent key rate.

The key rate shift approach enables analysts to estimate changes in all rates based on a few

select rates.

Definition, Calculation, and Interpretation of Key Rate ’01 and


Key Rate Duration

By definition, a Key rate '01 (key rate DV01) is the effect of a dollar change of a one basis point

shift around each key rate on the value of the security.

The Key rate '01 is computed using the same logic as the DV01 formula used in the single-factor

approach.

ΔBV
Key rate '01 = −
10, 000 × Δy

Where:

ΔBV =change in bond value

Δy =change in yield (0.01%)

Note that yield here implies the yield to maturity.

The change in bond value here is measured in reference to the initial bond value.

Example: Key Rate DV01s and Durations of the May 15, 2045, C-STRIP as
of May 28, 2015

In the table below, column (1) gives the initial price of a C-STRIP and its present value after

246
© 2014-2020 AnalystPrep.
application of key rate one basis point shifts.

(1) Value (2) Key Rate ‘01 (3) Calculation

Initial value 26.11485


2-year shift 26.11582 −0.001 − 26.11582−26.11485
10 000∗0.01%
,
26.13885−26.11885
5-year shift 26.11885 −0.040 − 10 000∗0.01%
,

10-year shift 26.13885 −0.024 − 26.13885−26.11485


10 000∗0.01%
,
26.01192−26.11485
30-year shift 26.01192 0.103 − 10 000∗0.01%
,

The key rate ’01 with respect to the 10-year shift is calculated as:

ΔBV 26.13885 − 26.11485


Key rate '01 = − =− = −0.024
10, 000 × Δy 10 , 000 × 0.01%

How do we interpret the key rate?

A key rate of -0.024 implies that the C-STRIP increases in price by 0.024 per $100 face value for

a one basis point 10-year shift.

Exam tip: Just like the DV01, a negative key rate ’01 implies an increase in value after a given

shift, relative to the initial value. A positive key rate ’01 implies a decrease in value after a given

shift, relative to the initial value.

Key Rate Duration

The effective duration calculates expected changes in price for a bond or portfolio of bonds given

a basis point change in yield, but it is only valid for parallel shifts in the yield curve. The key rate

duration presents an improvement to the effective duration because it gives the expected

changes in price when the yield curve shifts in a manner that is not perfectly parallel.

The key rate duration is actually analogous to duration so that:

1 ∂P
key rate duration = − ( )
P ∂y

247
© 2014-2020 AnalystPrep.
Thus, the key rate duration with respect to the 10-year shift is calculated as:

1 26.13885 − 26.11485
key rate duration = − ( ) ×( )
26.11485 0.01%
= −9.19

Alternatively, recall that:

DV 01 = duratio n × 0.0001 × bond value

Thus,

DV01
duration =
0.0001 × bond value
−0.024
=
0.0001 × 26.11485
= −9.19

How do we interpret the key rate duration?

Interpreting each key rate duration in isolation can be quite difficult. That’s because, in practice,

it’s highly unlikely that a single point on the yield curve will exhibit an upwards or downwards

shift while all other points remain constant. For this reason, analysts tend to compare key rate

durations across the curve.

Exam tip: The sum of all the key rate durations along a portfolio yield curve is equal to the

effective duration of the portfolio.

The Key Rate Exposure Technique in Multi-factor Hedging


Applications

Before looking at some examples, let’s try to map out a typical problem investors often find

themselves in.

Let’s say an investor holds a (long) bond position and is afraid that the bond will lose some value

in the future. We call such a position the “underlying exposure.” How exactly can the position

lose value? Suppose the position has a 5-year key rate exposure of $0.05. This implies that the

248
© 2014-2020 AnalystPrep.
position will drop in value by $0.05 if there’s a one basis point shock to the 5-year key rate. If the

bond is trading at, say, $94 per $100 face value, then the new price will be $93.95 per $100 face

value following a one basis point increase in the 5-year key rate (Remember that bond prices fall

when interest rates rise). To avoid the loss, the investor must identify and sell short another bond

whose key rate exposure matches that of the underlying exposure.

Selling short is all about selling high and buying low. The investor will borrow the bond and sell

it, anticipating an increase in the 5-year key rate which will result in a decrease in the bond’s

price. The investor will still be obligated to “return” the bond to its owner but he will buy it at a

lower price and get to keep the difference, which will offset the loss incurred on the long position

(underlying exposure). This is the argument behind key rate exposure hedging.

Note that the hedging security need not have a 5-year key rate exposure of exactly $0.05. It

could have an exposure of, say, 0.045, implying that the investor will not sell short exactly $100

face value of the bond; the amount will be slightly more.

In key rate exposure hedging, therefore, the secret lies in determining the face amount “F” that’s

needed to neutralize the key rate exposure of the underlying position.

Example 1 on Multi-factor Hedging

An underlying exposure (bond position) has a ten-year key-rate '01 of +$880. If this key rate

exposure can be hedged by trading a ten-year bond that itself has a 10-year KR01 of $0.0520 per

100 face amount, what is the hedge trade?

Solution

A positive key rate ’01 implies a decrease in value after a given shift, relative to the initial value.

Thus, a ten-year key-rate '01 of +$880 implies that the bond position stands to lose $880 if there

happens to be a one basis point shock to the ten-year key rate. To avoid this scenario, we must

determine the face amount F(10) of the ten-year bond that must be sold short to neutralize the

key rate exposure. We proceed as follows:

0.0520
249
© 2014-2020 AnalystPrep.
0.0520
× F (10) = 880
100
880
F (10) = = $1 , 692, 308
0.00052

Note that since the key rate 01s are reported per 100 face value, they need to be divided by 100

in the hedging equation. However, the key rate 01 of the initial bond position (underlying

exposure) is reported for the face amount to be hedged, so it stands as is.

The hedge trade requires us to short $1.692 million face amount of the ten-year bond so as to

neutralize the exposure to the ten-year key rate. If there’s a one basis point shock to the ten-year

key rate, the long position will lose $880, while the short position will gain approximately $880

(= 0.052/100 × 1,692,000).

Important note:

Before looking at the second example, it is important to understand exactly what key rates stand

for. When we say that, for example, the 5-year key rate changes, what we mean is that the 5-

year par rate changes; all other par rates are unchanged. It is easy to think of the 5-year key

rate as the 5-year spot rate, but it is not; it’s the par rate. Key rates are not spot rates. (Par rate

denotes the coupon rate for which the price of a bond is equal to its nominal value (or par value).

This leads us to a very important observation: a bond priced at par (i.e., purchase price = par

value = $100) only has price sensitivity to key rates at the same tenor as its maturity. For

instance, a 5-year coupon-paying par bond has zero sensitivity to a change in the 2-year key

rate. However, a 5-year premium/discount bond will have some sensitivity to the 2-year key

rate.

The reason, as we have seen above is that the 5-year par rate doesn’t change. We compute the

price of a bond by discounting all its cash flows by its YTM. If the 5-year par rate doesn’t

change, then the YTM on a 5-year par bond doesn’t change, and therefore the price of a 5-year

par bond doesn’t change.

Example 2 on Multi-factor Hedging

250
© 2014-2020 AnalystPrep.
Let’s use an example to illustrate how to pull off the perfect hedge under multi-factor hedging:

Suppose we have a 30-year option-free bond paying semiannual coupons of $5,000 in a flat rate

environment of 5% across all maturities. Using the concepts learnt in the preceding learning

outcome statements, we can compute the following key rate ‘01s and key rate durations,

assuming a one-basis point shift in the key rates used:

Value Key Rate ‘01 Key Rate Duration


Initial value 145,066.45
2-year shift 145,061.23 5.22 0.36
5-year shift 145,050.68 15.77 1.09
10-year shift 144,989.02 77.43 5.34
30-year shift 145,000.95 65.50 4.52
Total 163.92 11.31

For example,

The key rate ’01 with respect to the 5-year shift is calculated as:

ΔBV
Key rate '01 = −
10, 000 × Δy
145 , 050.68 − 145066.45
=− = 15.77
10 , 000 × 0.01%

And the corresponding key rate duration is:

DV01
duration =
0.0001 × bond value
15.77
=
0.0001 × 145066.45
= 1.09

A key rate '01 of 15.77 implies that the bond decreases in value by $15.77 for a one basis point

shock to the 5-year key rate.

We can easily come up with the other key rate 01’s and key rate durations by performing similar

calculations.

Now to illustrate how hedging is carried out in this scenario, assume we have four other

different securities, each with the following key rate exposures:

251
© 2014-2020 AnalystPrep.
To illustrate how hedging is carried out based on key rates, assume we have four other different

securities, each with the following key rate exposures:

Security Exposure ( per 100 face value)


2-year key 5-year key 7-year 30-year
rate rate key rate key rate
2-year security 0.001
5-year security 0.0015 0.045
7-year security 0.002 0.001 0.1
30-year security 0.20

Note: In the table above, we assume that the 2-year bond and the 30-year bond are trading at

par, in which case they are only exposed to the key rate corresponding to their maturity dates (2

years and 30 years, respectively). On the other hand, the 5-year and 10-year securities are

trading at a premium.

For the hedge to work, we must neutralize the key rate exposure at each key rate.

Let F2, F5 , F10 , and F30 be the face amounts of the bonds in the hedging portfolio to be sold.

2-year key rate exposure: Three bonds, namely the two-year, five-year, and 10-year, have an

exposure to the two-year key rate. Therefore, for the two-year key rate exposure of the hedging

portfolio to equal that of the underlying position, it must be the case that

0.001 0.0015 0.002


2-year key rate exposure : × F2 + × F5 + × F10 = 5.22
100 100 100

5-year key rate exposure: Only two bonds, namely the five-year and 10-year, have an exposure

to the five-year key rate. Therefore, for the five-year key rate exposure of the hedging portfolio to

equal that of the underlying position, it must be the case that

0.045 0.001
5-year key rate exposure : × F5 + × F10 = 15.77
100 100

10-year key rate exposure: Only the ten-year bond has an exposure to the ten-year key rate.

Therefore, for the ten-year key rate exposure of the hedging portfolio to equal that of the

underlying position, it must be the case that

0.1
252
© 2014-2020 AnalystPrep.
0.1
10-year key rate exposure : × F10 = 77.43
100

30-year key rate exposure: Only the ten-year bond has an exposure to the ten-year key rate.

Therefore, for the ten-year key rate exposure of the hedging portfolio to equal that of the

underlying position, it must be the case that

0.2
30-year key rate exposure : × F30 = 65.50
100

Solving equations (1) through (4) simultaneously gives the following solution for the face value of

the hedging bonds in the hedging portfolio:

F30 = 32,750
F10 = 77,430
F5 = 33,324
F2 = 317, 150

What then, do these figures imply?

The investor needs to short $317,150 face amount of the 2-year security, short $33,324 face

amount of the 5-year security, short $77,430 face amount of the 10-year security, and finally

short $32,750 face amount of the 30-year security. Only then would the initial bond position be

insured from changes in rates close to the key rates used.

However, such a hedge portfolio is not perfect, and the hedged position is actually only

approximately immune due to two main reasons

As is the case whenever derivatives are used for hedging purposes, the quality of hedge

deteriorates as the size of the interest rate change increases.

The hedge will work only if the par yields between key rates move as assumed (linearly).

Other reasons why the hedge may not work include:

Hedging implies more instruments and more transaction costs which may eat up the

scooped gains;

Under the key rate model, the number of key rate durations to be used and the

253
© 2014-2020 AnalystPrep.
corresponding choice of key rates remain quite arbitrary

Partial ‘01s and Forward-bucket ‘01s

Partial ‘01s

Key rate shifts make use of a few key rates to determine risk exposures and execute hedging

strategies. For, example, in this reading, we’ve used the 2-year, 5-year, 10-year, and 30-year par

yields. The hedging strategy must involve all four.

However, when the securities involved contain swaps, we need more to assess the effect of

interest rates at more points along the yield curve. There’s a need to measure more frequently.

This leads us to partial ‘01s and forward-bucket ‘01s.

When swaps are taken as the benchmark for interest rates in complex portfolios, risk along the

curve is usually measured with Partial ’01s or Partial PV01s, rather than with key-rate ’01s.

Swap market participants fit a swap rate at least once every day from a set of observable par

swap rates or futures rates. Using the fitted swap rate curve, the sensitivity of a portfolio can be

measured in terms of changes in the rates of the fitting securities.

It follows that by definition, partial ’01 (PV01) is the change in value of a portfolio after a

one-basis-point decline in that fitted rate and a refitting of the curve. All other fitted rates are

unchanged. With partial ‘01s, yield curve shifts are able to be fitted more precisely because we

are constantly fitting securities.

Example of Partial ‘01s

If a curve-fitting algorithm fits the three-month London Interbank Offered Rate (LIBOR) rate and

par rates at 2-, 5-, 10-, and 30-year maturities, then, the two-year partial ’01 would be the

change in the value of a portfolio for a one-basis-point decline in the two-year par rate and

refitting of the curve, where the three-month LIBOR and the par 5-, 10-, and 30-year rates are

kept the same.

254
© 2014-2020 AnalystPrep.
Forward-bucket ’01s

While key rates and partial ’01s do a fantastic job expressing the exposures of a position in terms

of hedging securities, forward-bucket ’01s present a far more direct and intuitive way to convey

the exposures of a position to different parts of the curve.

A bucket is a jargon for a region of the term structure of interest rates.

Forward-bucket ’01s are computed by shifting the forward rate over each of several defined

regions of the term structure on the region at a time. They, however, aren’t the quickest way to

determine the hedges required to pull off that perfect “immunization”.

The first step under this methodology is to subdivide the term structure into buckets. The 5 most

common buckets are 0-2 years, 2-5 years, 5- 10 years, 10-15 years, and 20-30 years. After that,

each forward-bucket ’01 is computed by shifting the forward rates in that bucket by one basis

point. In so doing, the analyst may have to shift all of a bucket’s semiannual forward rates,

quarterly forward rates, or even shorter-term rates.

Example: Computation of forward-bucket ‘01s of a 5-year swap given a 0-


2 year bucket and a 2-5 year bucket.

The table below lists the cash flows of the fixed side of the 100 notional amount of a swap, the

current forward rates (marked “current”) as of the pricing date, and the three shifted forward

curves.

255
© 2014-2020 AnalystPrep.
Forward Rates
Term Cash flow Current 0-2 shift 2-5 shift shift-all
0.5 1.06 1.012 1.022 1.012 1.022
1 1.06 1.248 1.258 1.248 1.258
1.5 1.06 1.412 1.422 1.412 1.422
2 1.06 1.652 1.662 1.652 1.662
2.5 1.06 1.945 1.945 1.955 1.955
3 1.06 2.288 2.288 2.298 2.298
3.5 1.06 2.614 2.614 2.624 2.624
4 1.06 2.846 2.846 2.856 2.856
4.5 1.06 3.121 3.121 3.131 3.131
5 101.06 3.321 3.321 3.331 3.331

Present value 99.9955 99.976 99.9679 99.9483


01 0.0196 0.0276 0.0472

Credit: Bruce Tuckman and Angel Serrat, Fixed Income Securities: Tools for Today’s Markets,

3rd Edition

For the “0-2 Shift,” forward rates of term 0.5 to 2.0 years are shifted up by one basis

point while holding all other forward rates constant.

For the “2-5 Shift,” forward rates of term 2.0 to 5.0 years are shifted up by one basis

point while, again, holding all other forward rates constant.

Lastly, for “Shift All,” the forward rates in the curve are shifted

The row labeled “Present Value” gives the present value of the cash flows first under the initial

forward rate curve and then under each of the shifted curves.

The forward-bucket ’01 for each shift can then be computed as the negative of the difference

between the shifted and initial present values, i.e.,

Forward bucket ’01 = –(shifted present value – initial present value)


For the 0-2-year shift, for example, the ’01 is −(99.976 − 99.9955), or
.0.0196

The ’01 of the “Shift All” scenario is analogous to a DV01. The forward bucket analysis

decomposes this total ’01 into .0196 due to the 0-2-year part of the curve and .0276 due to the 2-

256
© 2014-2020 AnalystPrep.
5-year part of the curve.

Hedging across Forward-Bucket Exposures

Referring to GARP-assigned Tuckman reading, let us say we have a counterparty enters into a

euro 5x10 payer swaption with a strike of 4.044% on May 28, 2010.

This payer swaption gives the buyer the right to pay a fixed rate of 4.044% on a 10-year euro

swap in five years. The underlying is a 10-years swap for settlement on May 31, 2015.

The figure below gives the forward-bucket ‘01s of this swaption for four different buckets, along

with other swaps for hedging purposes.

Since the overall forward-bucket ’01 of the year swaption is negative (-0.0380), as rates rise, the

value of the option to pay a fixed rate of 4.044% in exchange for a floating rate worth par also

rises.

Security Rate 0-2 2-5 5-10 10-15 All


5×10 payer swaption 4.04% 0.0010 0.0016 −0.0218 −0.0188 -0.0380
5-year swap 2.120% 0.0196 0.0276 0.0000 0.0000 0.0472
10-year swap 2.943% 0.0194 0.0269 0.0394 0.0000 0.0857
15-year swap 3.290% 0.0194 0.0265 0.0383 0.0323 0.1164
5×10 swap 4.044% 0.0000 0.0000 0.0449 0.0366 0.8150

The figure below shows forward-bucket exposures of three different ways to hedge this payer

swaption (as of May 28, 2010) using securities presented in the previous figure:

As is apparent, the third hedge is the best option since this hedge best neutralizes risk in each of

the buckets (the lowest net position indicates when risk is best neutralized).

Credit: Bruce Tuckman and Angel Serrat, Fixed Income Securities: Tools for Today’s Markets,

3rd Edition

Applying Key Rate and Multi-factor Analysis to Estimate


Portfolio Volatility

257
© 2014-2020 AnalystPrep.
Security/Portfolio 0-2 2-5 5-10 10-15 All
5*10 payer swaption 0.001 0.0016 −0.0218 −0.0188 −0.0380
Hedge #1:Long 44.34% of 10-year swaps 0.0086 0.0119 0.0175 0.038
Net position 0.0096 0.0135 −0.0043 −0.0188 0.000
Hedge #2:Long 46.66% of 5*10 swaps 0.0209 0.0171 0.038
Net position 0.001 0.0016 −0.0009 0.0017 0.000
Hedge #3:
Long 57.55% of 15-year swaps 0.0112 0.0153 0.022 0.0186 0.067
Short 61.55% of 15-year swaps −0.012 −0.017 −0.029
Net position 0.0002 −0.0001 0.0002 −0.0002 0.000

Although we have studied at length the term-structure of interest rates, we are yet to look at

volatility. Just like there is a term-structure for interest rates, there is also a term-structure for

volatility. In fact, the volatility term structure typically slopes downwards when plotted against

maturity. This implies that the shorter the maturity of the par-rate, the more volatile it tends to

be. The 10-year par rate, for example, is usually more volatile than the 30-year par rate.

In general, portfolios are exposed to interest rates all along the curve but changes in these rates

are not perfectly correlated. How can we go about estimating volatilities for the key-rates?

Step 1: Estimate the volatility for each key rate as well as the correlation for each pair of key

rates.

Step 2: Compute the key-rate 01s of the portfolio

Step 3: Compute the variance and volatility of the portfolio

For example, let’s make some assumptions:

There are two key-rates C1 and C2

The key rates of the portfolio are KR011 and KR012.

P gives the value of our portfolio

Then, by the definition of key rates,

ΔP = KR011 × ΔC 1 + KR01 2 × ΔC 2

258
© 2014-2020 AnalystPrep.
Furthermore, let σP2 , σ12 and σ22 denote the variances of the portfolio and of the key rates and let

ρ denote the correlation of the key rates. By applying the usual formula for finding the variance

of a portfolio, we can estimate portfolio variance:

σP2 = σ12 KR01 21 + σ22 KR01 22 + 2ρ1, 2 σ1 σ2 KR01 1 KR012

Note that this methodology can be applied equally well to partial ’01s or forward-bucket ’01s.

259
© 2014-2020 AnalystPrep.
Questions

Question 1

A risk manager at an Indian bank helps manage a portfolio of investment-grade

option-free bonds. After a lengthy market analysis, the manager strongly

recommends portfolio hedging using the key rates of 2-year, 5-year, 7-year, and 20-

year exposures. According to the manager, the 2-year rate has increased by 10 basis

points in the recent past.

How will the increase affect the 20-year rate?

A. It will increase by 10 basis points

B. It will decrease by 10 basis points

C. It will increase by 20 basis points

D. It will increase by zero basis points

The correct answer is D.

The key rate technique indicates that changes in each key rate will affect rates from

the term of the previous key rate to the term of the subsequent key rate. In this case,

the 2-year key rate will affect all rates from 0 to 5 years; the 5-year key rate affects

all rates from 2 to 7 years; the 7-year key rate affects all rates from 5 to 20 years; and

the 20-year key rate affects all rates from 7 years to the end of the curve.

Question 2

Suppose we have a 30-year option-free bond paying semiannual coupons of $4,000 in

a flat rate environment of 5% across all maturities. The following table provides the

initial price of the bond and its present value after application of a one basis point

260
© 2014-2020 AnalystPrep.
shift in four key rates:

Value Key Rate ‘01


Initial value 138, 200.55
2-year shift 138, 195.23 5.32
5-year shift 138, 187.33 13.22
7-year shift 138, 172.91 27.64
30-year shift 138, 180.25 20.30
Total 66.48

Suppose further that there are four other different bonds with the following key rate

exposures:

Security Exposure ( per 100 face value)


2-year key 5-year key 7-year 30-year
rate rate key rate key rate
2-year security 0.001
5-year security 0.0015 0.045
7-year security 0.002 0.001 0.1
30-year security 0.20

If we wish to fully hedge our initial position using these four securities, determine the

face amount of the 5-year security we need to short (assume that the 2-year bond and

the 30-year bond are trading at par):

A. 27,640

B. 10,150

C. 28,764

D. 30,000

The correct answer is C.

If we assume that the 2-year bond and the 30-year bond are trading at par, they are

only exposed to the key rate corresponding to their maturity dates (2 years and 30

years, respectively). The face amount we need for each security is given by Fi.

0.001 0.0015 0.002


261
© 2014-2020 AnalystPrep.
0.001 0.0015 0.002
2-year key rate exposure: × F2 + × F5 + × F7 = 5.32
100 100 100
0.045 0.001
5-year key rate exposure: × F5 + × F7 = 13.22
100 100
0.1
7 − year key rate exposure : × F7 = 27.64
100
0.2
30-year key rate exposure: × F30 = 20.30
100

⋯ ⋯ ⋯⋯ ⋯ ⋯⋯ ⋯ ⋯⋯

F30 = 10, 150


F7 = 27, 640
13.22 − 0.2764
F5 = = 28, 764
0.00045

262
© 2014-2020 AnalystPrep.
Reading 58: Binomial Trees

After completing this reading you should be able to:

Calculate the value of an American and a European call or put option using a one-step

and two-step binomial model.

Describe how volatility is captured in the binomial model.

Describe how the value calculated using a binomial model converges as time periods

are added.

Define and calculate the delta of a stock option.

Explain how the binomial model can be altered to price options on stocks with

dividends, stock indices, currencies, and futures.

Pricing Options Using the Binomial Model

The binomial option pricing model is a simple approximation of returns which, upon refining,

converges to the analytic pricing formula for vanilla options. The model is also useful for valuing

American options that can be exercised before expiry.

The model can be represented as:

P S0 u


S0

1− P S0 d

The notation used is as follows:

S0 =stock price today

P =probability of a price rise

263
© 2014-2020 AnalystPrep.
u=The factor by which the price rises

d=The factor by which the price falls

Over a small time interval Δt, the price today to one of only two potential future values: S0 u , and

S0 d.

The underlying price is assumed to follow a random walk.

Risk-neutral Valuation

The following formula are used to price options in the binomial model:

U =size of the up move factor=eσ√t , and

1 1
D=size of the down move factor= e−σ√ t = =
σ U
e √t

σ is the annual volatility of the underlying asset’s returns and t is the length of the step in the

binomial model.

ert −D
π u = probability of an up move= U−D

π d=probability of a down move=1 − π u

Example: Risk-neutral valuation

The price of an exchange-quoted zero-dividend share is $30. Over the past year, the stock has

exhibited a standard deviation of 17%. The continuously compounded risk-free rate is 5% per

annum. Compute the value of a 1-year European call option with a strike price of $30 using a

one-period binomial model:

The up- and down-move factors are:

U = e0.17 ×√1 = 1.1853


1
D= = 0.8437
1.1853

The risk-neutral probabilities of an up- and down-move are:

0.05 ×1

264
© 2014-2020 AnalystPrep.
(e0.05 ×1 ) − D
1.0513 − 0.8437
πu = = = 0.61
U−D 1.1853 − 0.8437
π d = 1 − 0.61 = 0.39

Exhibit 1: Binomial Tree – Stock

1.1853 × $30 = $35.60


$30

0.8437 × $30 = $25.30

Exhibit 2: Binomial Tree – Option

Max (0, $35.6 − $30) = $5.6


C0

Max (0, $25.3 − $30) = $0

The expected value of the option in one year is given by:

C U × πU + C D × πD = $5.6 × 0.61 + $0 × 0.39 = $3.42

The expected value of the option at present is given by:

C 0 = $3.42e(−0.05 ×1) = $3.25

Two-step Binomial model

In the two-period model, the tree is expanded to create room for a greater number of potential

outcomes. Exhibit 3 below presents the two-period stock price tree:

265
© 2014-2020 AnalystPrep.
SUU


SU

S0 ╱ SUD or SDU



SD

SDD

The two-step model uses the same formulae used in the one-step version to calculate the value of

an option.

Note: The value of a put can be calculated once the value of the call has been determined, using

the following formula:

P ut = Call − stock + Xe−r T

Where “stock” represents the stock price, X represents the strike price, r is the rate of return,

and t is the number of time periods.

Increasing the Number of Steps in the Binomial Model

Binomial models with one or two steps are unrealistically simple. Assuming only one or two steps

would yield a very rough approximation of the option price. In practice, the life of an option is

divided into 30 or more time steps. In each step, there is a binomial stock price movement.

As the number of time steps is increased, the binomial tree model makes the same assumptions

about stock price behavior as the Black– Scholes–Merton model. When the binomial tree is used

to price a European option, the price converges to the Black–Scholes–Merton price as the

number of time steps is increased.

266
© 2014-2020 AnalystPrep.
Delta

The delta, Δ, of a stock option, is the ratio of the change in the price of the stock option to

the change in the price of the underlying stock. It is the number of units of the stock an

investor/trader should hold for each option shorted in order to create a riskless portfolio. This

process is called delta-hedging.

The delta of a call option is always between 0 and 1 because as the underlying asset increases in

price, call options increase in price. The delta of a put option, on the other hand, is always

between -1 and 0 because as the underlying security increases, the value of put options

decrease.

For instance, suppose that when the price of a stock change from $20 to $22, the call option

price changes from $1 to $2. We can calculate the value of delta of the call as:

2 −1
= 0.5
22 − 20

This means that if the underlying stock increases in price by $1 per share, the option on it will

rise by $0.5 per share, all else being equal.

Suppose that an investor is long one call option on the stock above (with a delta of 0.5, or 50

since options have a multiplier of 100). The investor could delta hedge the call option by

shorting 50 shares of the underlying stock. Conversely, if the investor is long one put on the

stock (with a delta of -0.5, or -50), they would maintain a delta neutral position by purchasing

50 shares of the underlying stock.

How Volatility is Captured in the Binomial Model

As the standard deviation increases, so does the divide (dispersion) between stock prices in up

and down states ( SU and SD, respectively). Suppose there was no deviation at all. Would we have

a binomial tree in the first place? The answer is no.

With zero standard deviation, (SU would be equal to SD, and instead of a tree, we would have a

267
© 2014-2020 AnalystPrep.
straight line. But provided there’s some deviation, the gap between stock prices in the upstate

and stock prices in the downstate increasingly widens as the deviation increases.

To capture volatility, therefore, it would be paramount to evaluate stock prices at each time

period present in the tree.

How the Modified Binomial Model can be Altered to Price


Options on Non-zero Dividend Stocks, Stock Indices,
Currencies, and Futures

Given a stock that pays a continuous dividend yield q , the following formula can be used to price

the resulting option:

e(r−q)t − D
Probability of an up move = π u =
U−D
Probability of a down move = 1 − πu

U =size of the up move factor=eσ√t , and

1 1
D=size of the down move factor= e−σ√ t = =
eσ√t U

Note: The sizes of the up move factor and down move factor are the same as in the zero-dividend

model.

Sometimes it may also be necessary to price options constructed with a stock index as the

underlying, for instance, an option on the S&P 500 index. Such an option would be valued in a

manner similar to that of the dividend-paying stock. It’s assumed that the stocks forming part of

the index pay a dividend yield equal to q.

The binomial model can also be modified to incorporate the unique characteristics of options on

futures. Of note is the fact that futures contracts are largely considered cost-free to initiate, and

therefore in a risk-neutral environment, they are zero-growth instruments. The only formula that

changes is that of the probability of an up move, where:

1 −D
πu =
U−D

268
© 2014-2020 AnalystPrep.
When dealing with options on currencies, a plausible assumption is that the return earned on a

foreign currency asset is equal to the foreign risk-free rate of interest. As such, the probability of

an up move is given by:

e(r DC −r FC )t − D
πu =
U−D

American Options

To value an American option, we check for early exercise at each node. If the value of the option

is greater when exercised, we assign that value to the node. If that’s not the case, we assign the

value of the option unexercised. We then work backward through the tree as usual.

The Binomial Model When Time is Continuous

The binomial model is essentially a discrete-time model where we evaluate option values at

discrete times, say, intervals of one year, intervals of six months, intervals of three months, etc.

However, if we were to shrink the length of time intervals to arbitrarily small values, we’d end up

with a continuous-time model where the price can move at non-discrete times. The binomial

model converges to the continuous-time model when time periods are made arbitrarily small.

269
© 2014-2020 AnalystPrep.
Questions

Question 1

Suppose we have a 6-month European call option with K = $23. Suppose the stock

price is currently $22 and in two-time steps of three months, the stock can go up or

down by 10%. The up move factor is 1.1 while the down move factor is 0.9. The risk-

free rate of interest is 12%.

Compute the value of the call today.

A. $2

B. $1.54

C. $1.45

D. $0

The correct answer is C.

Suu = $26.62


Su = $24.2

S0 = $22
╱ Su d = $21.78

╲ Sd u = $21.78


Sd = $19.8

Sdd = $17.82

Su = 22 × 1.1 = 24.2,

Suu = 22 × 1.1 × 1.1

270
© 2014-2020 AnalystPrep.
Other values at other nodes are calculated using the relevant up/down factors.

ert − D e0.12 ×0.25 − 0.9


πu = = = 0.6523,
U −D 1.1 − 0.9
π d = 1 − 0.6523 = 0.3477

Let f represent the value of the call

Suu = 26.62 fuu = max ($26.62 − $23 , 0) fuu = $3.62


Su d = $21.78 fu d = max ($21.78 − $23 , 0) fu d = $0
Sd u = $21.78 fd u = max ($21.78 − $23 , 0) fd u = $0
Sdd = $17.82 fdd = max ($17.82 − $23 , 0) fdd = $0

The expected value of the call six months from now is given by:

0.6523 × 0.6523 × $3.62 + 0.6523 × 0.3477 × $0 + 0.3477 × 0.6523 × $0 + 0.3477 × 0.3477


= $1.54

$1.54
The value of the call today = = $1.45
e0.12×0.5

Question 2

A 1-year $50 strike European call option exists on ABC stock currently trading at $49.

ABC pays a continuous dividend of 3% and the current continuously compounded

risk-free rate is 4%. Assuming an annual standard deviation of 3%, compute the value

of the call today.

A. $0.31

B. $0.30

C. $0.47

D. $0

The correct answer is B.

u = eσ√t = e0.03 ×1 = 1.03


1
d= = 0.97
1.03

271
© 2014-2020 AnalystPrep.
Note that the stock is dividend-paying, and therefore the formula for the probability

of an up move is given by:

e( r−q) t−d e( 0.04−0.03) 1 −0.97


Probability of an up move=π u = = 1.03−0.97 = 0.67
u−d

Probability of a down move =1 − 0.67 = 0.33

Let S represent the price of the stock and f represent the value of the call. This is a

one-step binomial process.

Su = $49 × 1.03 = $50.47 f u = max ($50.47 − $50 , 0) = $0.47


Sd = $49 × 0.97 = $47.53 f d = max ($47.53 − $50 , 0) = $0

Value of the call option one year from today = ($0.47 × 0.67 + $0 × 0.33) = $0.31
$0.31
Value of the call today = = $0.30
e0.04

272
© 2014-2020 AnalystPrep.
Reading 59: The Black-Scholes-Merton Model

After completing this reading you should be able to:

Explain the lognormal property of stock prices, the distribution of rates of return, and

the calculation of expected return.

Compute the realized return and historical volatility of a stock.

Describe the assumptions underlying the Black-Scholes-Merton option pricing model.

Compute the value of a European option using the Black-Scholes-Merton model on a

non-dividend-paying stock.

Define implied volatilities and describe how to compute implied volatilities from market

prices of options using the Black-Scholes-Merton model.

Explain how dividends affect the decision to exercise early for American call and put

options.

Compute the value of a European option using the Black-Scholes-Merton model on a

dividend-paying stock.

Describe warrants, calculate the value of a warrant, and calculate the dilution cost of

the warrant to existing shareholders.

Suppose we have a random variable X. This variable will have a lognormal distribution if its

natural log (ln X) is normally distributed. In other words, when the natural logarithm of a random

variable is normally distributed, then the variable itself will have a lognormal distribution.

The two most essential characteristics of the lognormal distribution are as follows:

It has a lower bound of zero, i.e., a lognormal variable cannot take on negative values.

The distribution is skewed to the right, i.e., it has a long right tail.

These characteristics are in direct contrast to those of the normal distribution, which is

symmetrical (zero-skew) and can take on both negative and positive values. As a result, the

273
© 2014-2020 AnalystPrep.
normal distribution cannot be used to model stock prices because stock prices cannot fall below

zero. The lognormal distribution is also used to value options.

The Lognormal Property of Stock Prices

A crucial part of the BSM model is that it assumes stock prices are log-normally distributed.

Precisely,

σ2
lnST ∼ N (lnS0 + (μ − ) T, σ 2T)
2

Where:

ST =stock price at time T

S0 =stock price at time 0

μ=expected return on stock per year

σ=annual volatility of the stock price

Note: The above relationship holds because mathematically, if the natural logarithm of a random

variable ?,(?? ?) is normally distributed, then ? has a lognormal distribution. It's also imperative

to note that the BSM model assumes stock prices are lognormally distributed, with stock returns

being normally distributed. Specifically, continuously compounded annual returns are normally

274
© 2014-2020 AnalystPrep.
distributed with:

σ2 σ2
a mean of [μ − ] and a variance of
2 T

Example: Mean and standard deviation given a lognormal distribution

ABC stock has an initial price of $60, an expected annual return of 10%, and annual volatility of

15%. Calculate the mean and the standard deviation of the distribution of the stock price in six

months.

σ2
lnST ∼ N (lnS0 + (μ − ) T, σ√T )
2
0.152
=N [ln60 + (0.10 − ) 0.5, 0.152 × 0.5]
2
lnST ∼ N [4.139, 0.011]
(In this case,we have μ = √ 0.011 = 0.1049)

Sometimes, the Global Association of Risk Professionals (GARP) may want to test your

understanding of the lognormal concept by involving confidence intervals. Since lnST is log-

normally distributed, 95% of values will fall within 1.96 standard deviations of the mean.

Similarly, 99% of the values will fall within 2.58 standard deviations of the mean. For example, to

obtain the 99% confidence interval for stock prices using the above data, we will proceed as

follows:

lnST ∼ N [4.139 , 0.011]


lnST = μ ± z α × σ
(In this case,we have μ = √0.011 = 0.1049)
4.139 − zα × σ < lnST < 4.139 + zα × σ
e4.139−z α×σ < ST < e4.139+zα ×σ
47.86 < ST < 82.24
[e(lnx) = x]

Expected Stock Price

Using the properties of a lognormal distribution, we can show that the expected value of

275
© 2014-2020 AnalystPrep.
ST , E(ST ), is:

E(ST ) = S0 eμT

μ=expected rate of return

Example: Expected value of a stock

The current price of a stock is $40, with an expected annual return of 15%. What is the expected

value of the stock in six months?

Solution

Expected stock price = $40e0.15×0.5 = $43.11

Realized Portfolio Return

1
R pr = [r1 × r2 × r3 × ⋯ × rn ] n − 1

ri=portfolio return at the time i

The continuously compounded return realized over some time of length T is given by:

1 ST
ln ( )
T S0

Example: Realized return

The realized return of a stock initially priced at $50 growing, with volatility, to $87 over five

periods would simply be:

1 $87
Realized return = ln ( ) = 11.08%
5 $50

Estimating Historical Volatility

276
© 2014-2020 AnalystPrep.
We can calculate historical volatility from daily price data of stock. We simply need to calculate

continuously compounded returns per day and then determine the standard deviation.

The continuously compounded return for the day i is calculated as:

Si
ln ( )
Si−1

The volatility of short periods can be scaled to give the volatility of more extended periods.

For example,

Annual volatility = daily volatility × √( no.of trading days in a year)

Note that this formula is useful throughout the whole FRM part 1 and FRM part 2 exams in

estimating volatility.

Conversely,

annual volatility
daily volatility =
√ no.of trading days in a year

Black-Scholes-Merton Model

The Black-Scholes-Merton model is used to price European options and is undoubtedly the most

critical tool for the analysis of derivatives. It is a product of Fischer Black, Myron Scholes, and

Robert Merton.

The model takes into account the fact that the investor has the option of investing in an asset

earning the risk-free interest rate. The overriding argument is that the option price is purely a

function of the volatility of the stock's price (option premium increases as volatility increases).

Assumptions underlying the Black-Scholes-Merton Option Pricing Model

i. There is no arbitrage.

277
© 2014-2020 AnalystPrep.
ii. The price of the underlying asset follows a lognormal distribution.

iii. The continuous risk-free rate of interest is constant and known with certainty.

iv. The volatility of the underlying asset is constant and known.

v. The underlying asset has no cash flow, such as dividends or interest payments.

vi. Markets are frictionless – no transaction costs, taxes, or restrictions on short sales.

vii. Options can only be exercised at maturity, i.e., they are European-style. The model

cannot be used to value American options accurately.

Determining the Value of Zero-dividend European Options using the BSM


model

The value of a call option is given by:

C 0 = S0 × N (d1 ) − Ke−rT × N (d 2 )

The value of a put option is given by:

P 0 = Ke−rT × N (−d2 ) − S0 × N (−d 1 )

Where:

S0 σ2
ln ( ) + [r + ( )] T
K 2
d1 =
σ√ T
d 2 = d 1 − (σ√T)

T =time to maturity,assuming 365 days per year

S0 =asset price

K =exercise price

R cf=continuously compounded risk-free rate

σ=volatility of continuously compounded returns on the stock

N(di )=cumulative distribution function for a standardized normal distribution variable

278
© 2014-2020 AnalystPrep.
Example: Valuing a call option using the BSM model

Assume S0 = $100, K = $90, T = 6 months, r = 10%, and σ = 25%.

Calculate the value of a call option.

S0 σ2
ln ( ) + [r + ( )] T
K 2
d1 =
σ√ T
100 0.252
ln ( ) + [0.10 + ( )] 0.5
90 2 0.1053 + 0.0656
= = = 0.9672
0.25 √0.5 0.1768
d 2 = d 1 − (σ√ T) = 0.9672 − (0.25√0.5) = 0.7904

From a standard normal probability table, look up N(0.97) = 0.8333 and N(0.79) = 0.7852.

C 0 = S0 × N (d1 ) − Ke−rT × N (d2 )


= 100 × N (0.8340) − 90e−0.10 ×0.50 × N (0.7852) = $16.17

Note that the intrinsic value of the option is $10—our answer must be at least that amount.

Exam tips

Tip 1: Given one of either the put value or the call value, you can use the put-call parity to find

the other. Precisely,

279
© 2014-2020 AnalystPrep.
(−Rc T)
C 0 = P 0 + S0 − (Ke f )

(−Rc T)
P 0 = C 0 − S0 + (Ke f )

Tip 2: N(−d ?) = 1 − N(d i )

Tip 3: As S0 becomes very large, calls (puts) are extremely in-the-money (out-of-the-money)

Tip 4: As S0 becomes very small, calls (puts) are extremely out-of-the-money (in-the-money)

Tip 5: Although N(−d1 )and N(−d2 ) can easily be identified from statistical tables; sometimes

you'll be asked to compute d2 and d 2 without the formulas.

The Value of a European Option using the BSM Model on a Dividend-


paying Stock

Assume that we have a known dividend d distributed a time T1 , T1 < T where T is the maturity

date. To value calls and puts when there are such dividends, we modify the BSM model by

replacing S0 withS, where:

S = S0 − D

D is the sum of the PV(discounted at R cf ) of the dividend payments during the life of the option.

For example,with dividends D1 and D2 at times Δt1 nd Δt2 ,


Δt 1 Δt2
−(R c) −(Rc )
S = S 0 − D1 e f m − D2 e f m

Δti represents the amount of time until the ex-dividend date

m a division factor in bringing the Δt to a full year. e.g. Δt = 2Δt = 2 months, m=12 months, so
Δt1 2
= 0.1667 years .
m 12 =

After this, everything else in the computational formulas remains the same, i.e.,

The value of a call option is given by:

280
© 2014-2020 AnalystPrep.
∣ (−R c T) ∣
C 0 = [S0 × N (d1 )] − ∣K × e f × × N (d 2 )∣
∣ ∣

The value of a put option is given by:

(−Rc ×T)
P 0 = [K × e f × (1 − N (d 2 ))] − [S × (1 − N (d 1 ))]
S
ln ( ) + [R cf + (0.5 × σ2 )] T
K
d1 =
σ√ T
d2 = d1 − (σ√T )

S is simply S0 adjusted to include dividends payable.

The underlying argument here is that on the ex-dividend dates, the stock prices are expected to

reduce by the amounts of the dividend payments.

Exam tip: Sometimes, GARP will give you not a dollar amount "d" of the dividend, but a dividend

yield q. For example, you may be told that the dividend yield is 2%, continuously compounded. In

such a case, you're still expected to replace S0 with S where:

S = e−qT × S0

How Dividends affect the Early Exercise for American Calls and Puts

Call option holders have the right but not the obligation to buy shares as per the terms of the

contract, but they do not hold shares. As such, they cannot benefit from the rights of

shareholders, such as the right to receive dividends – as long as the call options have not been

exercised.

When the underlying stock pays dividends, a call option holder will not receive it unless they

exercise the contract before the dividend is paid. Whoever owns the stock as of the ex-dividend

date receives the cash dividend, so an investor who owns in-the-money call options may exercise

early to capture the dividend. In summary, a call option should only be exercised early to take

advantage of dividends if:

i. The option is in-the-money

281
© 2014-2020 AnalystPrep.
ii. The time value of the option needs to be less than the value of the dividend.

It wouldn't make sense to exercise an out-of-the-money call option and pay an above-market

price just to receive a dividend.

Suitable conditions for early exercise of a put option include:

i. The option must be deep in-the-money

ii. High-interest rates

iii. Sufficiently low volatility

Provided these conditions have been met, the holder of an American put option can exercise

early, but only after the dividend has been paid. It would make a whole lot more sense to exercise

the put option the day after the dividend is paid to collect the dividend, instead of exercising the

day before and missing out.

Black's Approximation in Calculating the Value of an American Call


Option on a Dividend-paying Stock

Black's approximation sets the value of an American call option as the maximum of two European

prices:

I. A European call with the same maturity as the American call being valued, but with the

stock price reduced by the present value of the dividend. This implies that S0 is reduced

by the present value of the dividends payable, but all other variables remain the same.

For example, if we anticipate two dividends, S = S0 − PV

Where

Δt 1 Δt2
−(r) −(r)
PV = D1e m + D2 e m

D1, 2 are the dividends on the ex-dividend dates

r is the risk-free rate

Δti represent the amount of time until the ex-dividend date

282
© 2014-2020 AnalystPrep.
m is a division factor in bringing the Δt to a full year. If Δt = 2 months, m=12 months, so
Δt 2
= = 0.1667 years
m 12

Note: All other variables (d 1 , d2 , C0 , K , etc.) remain the same.

II. A European option is maturing just before the final ex-dividend date of the American-

option. This implies that time to maturity is trimmed down to just before the final

dividend is paid. The PV of dividends other than the final one must be deducted from S0

The largest of the two values (I) and (II) above is the desired Black's approximation for the

American call.

Exam tips:

An American call on a non-dividend-paying stock should never be exercised early.

An American call on a dividend-paying stock should only be exercised immediately

before an ex-dividend date.

Options on Stock Indices, Currencies, and Futures

We can extend the BSM result to valuing other assets such as stock indices, currencies, and

futures. For a European option on a stock paying a continuous dividend yield at a rate of q, the

value of the call becomes:

C 0 = S0 e−qT × N (d 1 ) − Ke−rt × N (d 2 )

The value of a put option is given by:

P 0 = Ke−rt × N (-d2 ) − S0 e−qT × N (-d1)

Where:

σ2
283
© 2014-2020 AnalystPrep.
S0 σ2
ln ( ) + [r − q + ( )] T
K 2
d1 =
σ√T
d 2 = d 1 − (σ√T)

Note that we can also use these formulas to value a European option on a stock index paying

dividends at the rate of q when S0 is the value of the index.

When dealing with an option on foreign currency, we take note that it behaves like a stock paying

a dividend yield at the risk-free foreign rate (rf ). We, therefore, set q= (rf ), and we have the

following equations of valuation:

C 0 = S0 e−r fT × N (d1 ) − Ke−rT × N (d2 )


P 0 = Ke−rT × N (−d 2) − S0 e−rf T × N (−d 1 )
S0 σ2
ln ( ) + [r − rf + ( )] T
K 2
d1 =
σ√ T
d2 = d1 − (σ√T )

Example: Valuing an option on foreign currency

Suppose the current exchange rate for a currency is 1.100 and the volatility of the exchange rate

is rate is 20%. Calculate the value of a call option to buy 1000 units of the currency in 3 years at

an exchange rate of 2.200. The domestic and foreign risk-free interest rates are 2% and 3%,

respectively.

Solution

In this case S0 =1.100, K=1.200 , r=0.02, rf =0.03, σ=0.2 and T=3

S0 σ2
ln ( ) + [r − rf + ( )] T
K 2
d1 =
σ√ T
1.100 0.22
ln ( ) + [0.02 − 0.03 + ( )] 3
2.200 2
=
0.2√ 3
d2 = d1 − σ√T = −1.9143 − 0.2√3 = −2.2607

284
© 2014-2020 AnalystPrep.
From standard normal tables,

N (d1 ) = N (−1.91) = 1 − 0.9719 = 0.0281

N (d2 ) = N (−2.26) = 1 − 0.9881 = 0.0119

The value of the call is therefore given by:

C 0 = S0 e−rf T × N (d 1) − Ke−rT × N (d2 )


= 1.1e−0.03 ×3 × 0.0281– 2.2e−0.02×3 × 0.0119 = 0.0036

This is the value of the option to buy one unit of the currency. The value of an option to buy 1000

units is 0.0036×1000=$3.60

When we are considering an option on futures, we realize that the futures price F is typical to a

stock paying a dividend yield at the risk-free domestic rate (r). We, therefore, set q=r and

S0 = F0, so that we have the following valuation equations:

C 0 = F0 e−rT × N (d 1 ) − Ke−rT × N (d2 )


P 0 = Ke−rT × N (−d 2) − F0 e−rT × N (−d1 )
F0 σ 2T
ln ( ) +
K 2
d1 =
σ√T
d2 = d1 − σ√T

Complications Involving the Valuation of Warrants

Warrants are securities issued by a company on its own stock, which give their owners the right

to purchase shares in the company at a specific price at a future date. They are much like

options, the only difference being that while options are traded on an exchange, warrants are

issued by a company directly to investors in bonds, rights issues, preference shares, and other

securities. They are basically used as sweeteners to make offers more attractive.

When warrants are exercised, the company issues more shares, and the warrant holder buys the

shares from the company at the strike price. An option traded by an exchange does not change

the number of shares issued by the company. However, a warrant allows new shares to be

285
© 2014-2020 AnalystPrep.
purchased at a price lower than the current market price, which dilutes the value of the existing

shares. This is known as dilution.

In an efficient market, the share price reflects the potential dilution from outstanding warrants.

We are not necessarily required to consider these when valuing the outstanding warrants. This

implies that we can value warrants just like exchange-traded options.

For detachable warrants, their value can be estimated as the difference between the market

price of bonds with the warrants and the market price of the bonds without the warrants.

The Black-Scholes-Merton Model can also be used to value warrants using the BSM call/put

option formulas, i.e.

c
C 0 = [S0 × N (d 1 )] − ∣∣K × e−R f ×T × N (d2 )∣∣

However, the following adjustments must be made:

1. The stock price S0 s replaced by an "adjusted" stock price. Suppose a company has N

outstanding shares worth S0 . This means that the value of the company's equity is NS 0 .

Further, assume that the company has decided to issue M number of warrants with each

warrant giving the holder the right to buy one share for K. If the stock prices change to

ST at time T, the (adjusted) stock price which accounts for the dilution effect of the issued

warrants, is:

(NS0 + MK)
Sadjusted =
N+M

2. The volatility input is calculated on equity (volatility of the value of the shares plus the

warrants, not just the shares).


N
3. A multiplier (haircut) that captures dilution, given by .
N +M

Implied Volatility

The volatility of the stock price is the only unobservable parameter in the BSM pricing formula.

286
© 2014-2020 AnalystPrep.
The implied volatility of an option is the volatility for which the BSM option price equals the

market price.

Implied volatility represents the expected volatility of a stock over the life of the option. It is

influenced by market expectations of the share price as well as by supply and demand of the

underlying options. As expectations rise, and the demand for options increases, the implied

volatility increases. The opposite is true.

If we use the observable parameters in the BSM formula (S0 , K, r, and T ) and set the BSM

formula equal to the market price, then it's possible to solve for volatility that satisfies the

equation. However, there is no closed-form solution for the volatility, and the only way to find it is

through iteration.

287
© 2014-2020 AnalystPrep.
Practice Questions

Question 1

ABC stock is currently trading at $70 per share. A dividend of $1 is expected after

three months and another one of $1 after six months. An American option on ABC

stock has a strike price of $65 and 8 months to maturity. Given that the risk-free rate

is 10% and the volatility is 32%, compute the price of the option:

A. $9.85

B. $12.5

C. $10

D. $10.94

The correct answer is D.

The current price of the share must be adjusted to take into account the expected

dividends.

The present value of the dividends is

e−0.25 ×0.1 + e−0.50×0.1 = 1.9265


S0 = 70 − 1.9265 = 68.0735

Next, calculate the variables required,

S0 =68.0735

K=65

σ=0.32

r=0.1

T=0.6667

0.322
288
© 2014-2020 AnalystPrep.
68.0735 0.322
ln ( ) + [0.1 + ( )] 0.6667
65 2
d1 = = 0.5626
0.32 √0.6667
d2 = d1 − 0.32√0.6667 = 0.3013

N(d1 )=0.7131

N(d2 )=0.6184

The call price is

68.0735 × 0.7131 − 65e−0.1×0.6667 × 0.6184 = 10.94

Question 2

The stock price is currently $100. Assume that the expected return from the stock is

35% per annum, and its volatility is 20% per annum. Calculate the mean and

standard deviation of the distribution, and determine the 95% confidence interval for

the stock price

Mean Standard 95% CI


deviation
A 150 30 110 < ST < 330
B 201.38 58.12 112.30 < ST < 336.57
C 5.27 0.28 4.7212 < ST < 5.8188
D 0.35 0.2 112.30 < ST < 336.57

The correct answer is B.

In this case,

S0 =100

μ=0.35 and,

σ=0.20

The probability distribution of the stock price in two years is lognormal and is given

by:

289
© 2014-2020 AnalystPrep.
0.22
lnST ∼ {ln100 + (0.35 − ) 2, 0.22 × 2} ∼ (5.27 , 0.28 2 )
2
The mean stock price = S0 eμT = 100e(0.35 ×2) = 201.38

The standard deviation of the stock price

2×2
= 100e(0.35×2)√(e0.2 − 1) = 58.12

The 95% confidence interval for ln ST is given by

5.27 − 1.96 × 0.28 < lnST < 5.27 + 1.96 × 0.28

4.7212 < lnST < 5.8188

e4.7212 < ST < e5.8188

112.30 < ST < 336.57

290
© 2014-2020 AnalystPrep.
Reading 60: Option Sensitivity Measures: The “Greeks”

After completing this reading you should be able to:

Describe and assess the risks associated with naked and covered option positions.

Describe the use of a stop-loss hedging strategy, including its advantages and

disadvantages, and explain how this strategy can generate naked and covered option

positions.

Describe delta hedging for an option, forward, and futures contracts.

Compute the delta of an option.

Describe the dynamic aspects of delta hedging and distinguish between dynamic

hedging and hedge-and-forget strategy.

Define and calculate the delta of a portfolio.

Define and describe theta, gamma, vega, and rho for options positions and calculate

the gamma and vega for a portfolio.

Explain how to implement and maintain a delta-neutral and a gamma-neutral position.

Describe the relationship between delta, theta, gamma, and vega.

Describe how portfolio insurance can be created through option instruments and stock

index futures.

Risks Associated with Naked and Covered Option Positions

A naked option position occurs when a trader sells a call option without insurance in the form of

a holding of the underlying shares. If the option position is backed by ownership of the

underlying shares, the position is said to be covered (clothed).

Selling naked call options is laden with risks on the part of the trader. If the market price is

below the pre-agreed strike price on the expiration date, the seller makes a gain equal to the

291
© 2014-2020 AnalystPrep.
premium received. However, if the market price soars above the strike price as at expiration, the

buyer exercises the option. When that happens, the seller has to deliver the agreed number of

shares to the buyer, even if it means buying from the market. Depending on the extent of the

price increase, the entire loss arising from a naked option can be absorbed by the premium

received. In other cases, the price increase may be so high that the seller is left with a net loss.

On the contrary, the trader may insure themselves by selling covered options. This may be a

safer strategy but one that’s also laden with downside risk. If the stock falls, the seller will get to

keep the entire premium, but the shares under their ownership will now be worthless.

Sometimes the price fall may be too high such that the total value lost in the long position

exceeds the premium received.

Example:

Suppose a firm sells 10,000 naked call options on a stock on a stock currently going for $30 a

share. The strike price is $33 and the option premium is $4

292
© 2014-2020 AnalystPrep.
Scenario 1: Price at expiry = $29

The buyer will not exercise the premium. Total income generated = 10 , 000 × $4 = $40, 000

Scenario 2: Price at expiry = $37

The buyer will exercise the options, and the seller is obliged to honor the contract. As such, the

seller buys 10,000 shares from the market at a cost of $37 per share and hands them to the

buyer.

Net loss = premium received + contract proceeds − cost of shares delivered


= $4 × 10 , 000 + $33 × 10, 000 − $37 × 10 , 000 = $0

In this scenario, the loss is absorbed by the premium, resulting in a net loss of $0

Scenario 3: Price at expiry = $38

Again, the buyer exercises the options.

Net loss = premium received + contract proceeds − cost of shares delivered


= $4 × 10 , 000 + $33 × 10, 000 − $38 × 10 , 000 = $10 , 000

How Naked and Covered Option Positions Generate a Stop


Loss Trading Strategy

A stop-loss trading strategy is a strategy where the trader initially gets into a naked option

position but later on seeks cover when the option moves in-the-money. In other words, protection

is sought only when market conditions are such that the call writer stands to lose.

With a naked call position, this strategy requires the purchase of the underlying asset

immediately the market price rises above the option’s strike price. But as soon as the market

price returns to a position that’s below the strike price, the trader sells the underlying asset.

Although this sounds like a simple, executable plan on paper, it’s a lot more complicated in

practice thanks to transaction costs and price uncertainty. In practice, buy/sell costs increase as

fluctuations in the strike price increase. It also becomes even more difficult to predict whether

293
© 2014-2020 AnalystPrep.
the option will be in-the-money or out-of-the-money at expiration.

Delta of an Option

Delta is a measure of the degree to which an option is exposed to changes in the price of the

underlying asset. It’s the ratio of the change in the price of the call option to the change in the

price of the underlying.

change in the call option price


Delta = Δ =
change in the price of the underlying

For example, if we have a delta value of 0.5, it means that when the price of the underlying

moves by a point, the price of the corresponding call option will change by half a point. If delta =

0.5, a $1 increase in the underlying’s price triggers a $0.5 increase in the price of the call option.

How to Compute Delta of an Option

Call Option

Delta of a Call option is closely related to N (d1 ) in the Black-Scholes Pricing model. Precisely,

Δc = e−qT N (d1 )

Where

σ2
ln S +( r+ )T
K 2
d1 =
σ√T

q is the dividend yield(%)

S =price of the underlying

K =strike price of the option

r=risk-free interest rate

294
© 2014-2020 AnalystPrep.
σ=volatility of the underlying

T =time to option's expiry

Exam tips:

I. Delta of a call option is always positive (between 0 and 1). The delta of an at-the-money

call option is close to 0.5. Delta moves to 1 as the call goes deep in-the-money (ITM). It

moves to zero as the call goes deep out-of-the-money (OTM).

II. If the underlying does not pay dividends the delta of a call option simplifies to:

Δc = e−qT N (d1 ) = e0 N (d 1 ) = N (d1 )

(q is zero in this case, and any number raised to power zero is equal to 1)

Put option

The delta of a put option is

Δp = e−qT [N (d 1 ) − 1]

It behaves similar to the call delta, except for the sign (between 0 and -1). As with the call delta,

if there are no dividends,

Δ p = e0 [N (d1 ) − 1] = [N (d 1 ) − 1]

Exam tip: The delta of an at-the-money put option is close to -0.5. Delta moves to -1 as the put

goes deep in-the-money. It moves to zero as the put goes deep out-of-the-money.

Delta of a Forward

The delta of a forward contract is given by:

Δf = e−qT

295
© 2014-2020 AnalystPrep.
Where q is the dividend yield and T is time to expiry.

By definition, all forward positions have a delta of approximately 1. What does that imply?

It means the underlying asset and the corresponding forward contract have a one-to-one

relationship. As a result, a forward sale position can always be perfectly hedged by buying the

same number of securities at the spot price.

Delta of a Futures Contract

Unlike forward contracts, the delta value of a futures contract is not ordinarily equal to 1. This is

because futures and spot prices move in lockstep, but are not exactly identical.

For a futures position on a stock that does not pay dividends,

Δf utures = erT

Where r is the risk-free rate and T is the time to maturity.

For a futures position on a stock that pays a dividend,

Δf utures = e(r−q)T

Where q is the dividend yield.

The Dynamic Aspects of Delta Hedging

Delta hedging is an attempt to reduce (hedge) the risk associated with price movements in the

underlying, by offsetting long and short positions. For instance, a long call position could be

offset by shorting the underlying stock. Since delta is actually a function of the price of the

underlying asset, it continually changes as the underlying’s price changes.

When delta changes, the initially option-hedged position is, again, thrust into a state of

imbalance. In other words, the number of stocks is no longer matched with the right number of

options, exposing the trader to possible loss.

296
© 2014-2020 AnalystPrep.
The overall goal of delta-hedging (a delta-neutral position) is to combine a position in the

underlying with another position in an option such that the value of the portfolio remains fixed

even in the face of constant changes in the value of the underlying asset.

Delta Hedging with Stock

An options position can be hedged using shares of the underlying. A share of the underlying has

a delta equal to 1 because of the value changes by $1 for a $1 change in the stock. For instance,

suppose an investor is long one call option on a stock whose delta is 0.6. Because options are

usually held in multiples of 100, we could say that the delta is 60. In such a scenario, the investor

could delta hedge the call option by shorting 60 shares of the underlying. The converse is true: If

the investor is long one put option, he would delta hedge the position by going long 60 shares of

the underlying.

Delta Hedging with Options

Sometimes an options position can be delta hedged using another options position that has a

delta that’s opposite to that of the current position. This effectively results in a delta-neutral

position. For instance, suppose an investor holds a one call option position with a delta of 0.5. A

call with a delta of 0.5 means it is at-the-money. To maintain a delta neutral position, the trader

can purchase an at-the-money put option with a delta of -0.5, so that the two cancel out.

Delta of a Portfolio

Suppose we want to determine the delta of a portfolio of options, all on a single underlying. The

portfolio delta is equivalent to the weighted average of the deltas of individual options.

n
portfolio delta = Δ portfolio = ∑ w iΔ i
i =1

w i represents the weight of each option position while Δ i represents the delta of each option

position.

Portfolio delta gives the change in the overall option position caused by a change in the price of

297
© 2014-2020 AnalystPrep.
the underlying.

Theta, Gamma, Vega, and Rho for Option Positions

Theta

Theta, θ , tells us how sensitive an option is to a decrease in time to expiration. It gives us the

change in the price of an option for a one-day decrease in its time to expiration.

Options lose value as expiration approaches. Theta estimates the value lost per day if all other

factors are held constant. Time value erosion is nonlinear, and this has implications on theta. As

a matter of fact, the theta of in-the-money, at-the-money, and slightly out-of-the-money options

generally increases as expiration nears. On the other hand, the theta of far out-of-the-money

options generally decreases as expiration nears.

For a call option,

∂c
θ=
∂t

Where:

∂c=change in call price

∂ t=change in time

For European call options that have zero dividends, the Black-Scholes Merton model can be used

to calculate theta. Precisely,

S0 N ′ (d1 ) σ
θcall =− − rXe−rT N (d 2 )
2√T
S0 N ′ (d1 ) σ
θput = − + rXe−rT N (d −2 )
2√T

Where:

298
© 2014-2020 AnalystPrep.
y2
′ 1 −(
2
)
N (y) = e , y = d 1 , d2
√2π

In the above equations, the resulting value for theta is measured in years because T is also

measured in years. To covert theta into a daily value, divide by 252, assuming 252 trading days

in a year.

Gamma

Gamma, Γ, measures the rate of change in an option's Delta per $1 change in the price of the

underlying stock. It tells us how much the option’s delta should change as the price of the

underlying stock or index increases or decreases. Options with the highest gamma are the most

responsive to changes in the price of the underlying stock.

Mathematically,

∂ 2c
Γ=
∂ 2s

Where the numerator and denominator are the partial derivatives of the call and stock prices,

respectively.

For European calls and puts on stocks with zero dividends,

N ′ (d1 )
Γ=
S0 σ√T

While delta neutral positions hedge against small changes in stock price, gamma-neutral

positions guard against relatively large stock price moves. As such, a delta-neutral position is

important, but even more important is one that’s also gamma-neutral, because it will be

insulated from both small and large price moves.

The number of options that must be added to an existing portfolio to generate a gamma-neutral

position is given by:

Γ
299
© 2014-2020 AnalystPrep.
Γp
−( )
ΓT

Where:

Γp =gamma of the existing portfolio position

ΓT =gamma of a traded option that can be added

Example on Delta-Gamma Hedging

A trader has a short option position that’s delta-neutral but has a gamma of -800. In the market,

there’s a tradable option with a delta of 0.8 and a gamma of 2. To maintain the position gamma-

neutral and delta-neutral, what would be the trader’s strategy?

The number of options that must be added to an existing portfolio to generate a gamma-neutral

position is given by:

?? −800
− =− = 400
?? 2

Buying 400 calls, however, increases delta from zero to 320 (=400×0.8). Therefore, the trader

has to sell 320 shares to restore the delta to zero. Positions in shares always have zero gamma.

Relationship among Delta, Theta, and Gamma

The relationship between the three Greeks can best be expressed in the following equation:

rP = θ + rS Δ + 0.5σ2 S 2 Γ

Where:

r=risk neutral risk free rate

P =price of the option

θ=option theta

300
© 2014-2020 AnalystPrep.
S =price of the underlying stock

Δ=option delta

σ 2=variance of the underlying stock

Γ=option Gamma

If a position is delta-neutral, then Δ = 0, and the above equation narrows down to:

rP = θ + 0.5σ2 S 2 Γ

Vega

Vega measures the rate of change in an option's price per 1% change in the implied volatility of

the underlying stock. And while Vega is not a real Greek letter, it tells us how much an option’s

price moves in response to a change in volatility of the underlying.

As an example, a Vega of 6 indicates that for a 1% increase in volatility, the option’s price will

increase by 0.06. For a given exercise price, risk-free rate, and maturity, the Vega of a call equals

the Vega of a put.

Mathematically,

∂c
V ega =
∂σ

Where:

∂c=change in call price

∂σ=change in volatility

For European calls and puts on stocks with zero dividends,

V ega = S0 N ′ (d1 ) √T

A drop in Vega will typically cause both calls and puts to lose value. An increase in Vega will

301
© 2014-2020 AnalystPrep.
typically cause both calls and puts to gain value.

Vega decreases with maturity, unlike gamma which increases with maturity. Vega is highest for

at-the-money options.

Rho

Rho measures the expected change in an option's price per 1% change in interest rates. It tells

us how much the price of an option should fall or rise in response to an increase or decrease in

the risk-free rate of interest.

As interest rates increase, the value of call options will generally increase. On the other hand, as

interest rates increase, the value of put options will usually decrease. Although rho is not a

dominant factor in the price of an option, it takes center stage when interest rates are expected

to change significantly.

Long-term options are far more sensitive to changes in interest rates than are short-term

options. Furthermore, in-the-money calls and puts are more sensitive to interest rate changes

compared to out-of-the-money calls and puts.

Mathematically,

∂c
rho =
∂r

Where:

∂c=change in call price

∂r=change in interest rate

For European calls and puts on stocks that do not pay dividends,

rhoca ll = XT e−r T N (d2 )


rhop ut = −XT e−rT N (−d2 )

Hedging Activities In Practice, And How Scenario Analysis Can

302
© 2014-2020 AnalystPrep.
Be Used to Formulate Expected Gains and Losses With Option
Positions

On paper, attaining neutrality to all the Greeks might appear a straight forward task but in

practice, this is hardly the case. Although delta-neutral positions are easy to create and maintain,

it’s quite difficult to find securities at reasonable prices that can help tame the negative effects

of gamma and vega.

Traders usually concentrate on maintaining delta neutrality and then purpose to continuously

and closely monitor the other Greeks.

Sometimes traders may use different values of a portfolio value determinant in order to assess

how sensitive the portfolio is to that determinant. For example, the traders may work with

different values of volatility to estimate the impact on portfolio value. This is called scenario

analysis.

Under scenario analysis, a single parameter may be varied at a time, but two or more

parameters can also be varied simultaneously to estimate their overall effect on the portfolio.

Creating Portfolio Insurance through Option Instruments and


Stock Index Futures

Portfolio insurance is the combination of (1) an underlying instrument and (2) either cash or a

derivative that generates a minimum value for the portfolio in the event that markets crash and

values decline, while still allowing the trader to make a gain in the event that market values rise.

The degradation of portfolio value is protected.

The most common insurance strategy involves using put options to lock in the value of an asset.

This way, the trader is able to maintain a limit on the portfolio value – even if the underlying’s

price tumbles, the trader is insulated from prices below the put’s strike.

To hedge a portfolio with index options, the trader selects an index with a high correlation to

their portfolio. For instance, if the portfolio consists of mainly technology stocks, the Nasdaq

Composite Index might be a good fit. If the portfolio is made up of mainly blue-chip companies,

303
© 2014-2020 AnalystPrep.
then the Dow Jones Industrial Index could be used.

Alternatively, a trader can use stock index futures with a similar end goal. Traders who want to

hedge their portfolios need to calculate the amount of capital they want to hedge and find a

representative index. Assuming an investor wants to hedge a $500,000 stock portfolio, she would

sell $500,000 worth of a specific futures index, such as the S&P 500.

304
© 2014-2020 AnalystPrep.
Questions

Question 1

The current stock price of a company is USD 100. A risk manager is monitoring call

and put options on the stock with exercise prices of USD 70 and 6 days to maturity.

Which of these scenarios is most likely to occur if the stock price falls by USD 1?

Scenario Call V alue P ut V alue


A Decrease by $0.9 I ncrease by $0.05
B Decrease by $1 I ncrease by $1
C Decrease by $0 I ncrease by $1
D Decrease by $0 I ncrease by $0

The correct answer is A.

The call option is deep-in-the-money and therefore must have a delta close to one.

The put option is deep out-of-the-money and will, therefore, have a delta close to zero.

Therefore, the value of the in-the-money call will decrease by close to USD 1, and the

value of the out-of-the-money put will increase by a much smaller amount close to 0.

Among the four choices, it’s A that is closest to satisfying both conditions.

Question 2

XY Z Inc., a non-dividend-paying stock, has a current price of $200 per share. Eric

Rich, FRM, has just sold a six-month European call option contract on 200 shares of

this stock at a strike price of $202 per share. He wants to implement a dynamic delta

hedging scheme to hedge the risk of having sold the option. The option has a delta of

0.50. He believes that the delta would fall to 0.40 if the stock price falls to $195 per

share.

Identify what action he should take NOW (i.e., when he has just written the option

305
© 2014-2020 AnalystPrep.
contract) to make his position delta-neutral

After the option is written, if the stock price falls to $195 per share, identify the

action Mr. Rich should take at that time, i.e. LATER, to rebalance his delta-hedged

position

A. NOW: buy 200 shares of stock, LATER: buy 100 shares of stock

B. NOW: buy 100 shares of stock, LATER: sell 20 shares of stock

C. NOW: sell 100 shares of stock, LATER: buy 100 shares of stock

D. NOW: sell 100 shares of stock, LATER: buy 20 shares of stock

The correct answer is B.

NOW: Eric sold a call on 200 shares, that means he’s short delta of 0.50 × 200, which

is delta = -100. To be delta neutral, he must long (i.e. buy) 100 shares of stock.

LATER: As price falls to $195, the delta moves to −80 = −0.40 × 200 . To be delta

neutral, Eric’s portfolio needs to have 80 shares of stock. He purchased 100 shares at

time 0. To rebalance, he must sell 20 shares.

306
© 2014-2020 AnalystPrep.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy