FULLTEXT01
FULLTEXT01
1 Introduction .................................................................................................. 7
2 Preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.1 Basics of Itô calculus . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.2 Black-Scholes theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
3 Stochastic volatility models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.1 Classical Heston model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
3.2 Rough volatility models . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.2.1 Fractional Brownian motion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
3.2.2 Riemann-Liouville fractional Brownian motion . . . . . . . . . . . 20
3.2.3 Rough volatility . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
3.2.4 Zumbach effect and super-Heston rough models . . . . . . . . . . 21
3.3 Quadratic Rough Heston model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
4 Simulation of the Quadratic Rough Heston model . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.1 Notation and numerical preliminaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
4.2 Hybrid scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
4.2.1 Simulating Riemann-Liouville process . . . . . . . . . . . . . . . . . . . . . . . . . . 28
4.2.2 Simulating quadratic rough Heston . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
4.3 Multifactor scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
4.3.1 Kernel approximations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
4.4 Multifactor Hybrid Scheme . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
5 Calibration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1 Provided data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1.1 SPX vs SPXW . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
5.1.2 Recovering interest rates . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
5.2 Calibration process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 42
5.3 Objective function . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
5.3.1 Variance reduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 44
5.4 Calibrating b . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
5.5 Summary of the algorithm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
5.6 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
5.6.1 Dependence of the smile on the calibration parameters 51
6 Conclusion ................................................................................................. 53
References ........................................................................................................ 54
1. Introduction
7
2. Preliminaries
One of the key features of Brownian motion is its lack of memory or de-
pendence on past values. In other words, the future movements of the process
are independent of its past movements. This property is known as the Markov
property.
8
Itô integral and SDEs
For option pricing, we need to model the underlying asset price as a stochas-
tic process, and this is where Itô calculus comes into play. The Japanese math-
ematician Kiyoshi Itô introduced the concept of stochastic integration in the
context of Brownian motion and established the Itô integral, which is a central
tool in stochastic calculus.
The Itô stochastic integral can be viewed as a stochastic generalization of
the Riemann-Stieltjes integral in analysis. The integrands and the integrators
are now stochastic processes:
Z T n
0
St dWt = lim
Dt!0
 Sti 1 (Wti Wti 1 ), (2.1)
i=1
Itô’s lemma
Theorem 2.1.1 (Itô’s lemma). Assume that the process X satisfies the stochas-
tic differential equation
dXt = µt dt + st dWt , (2.2)
where µ and s are adapted processes. Let f (t, x) be a C1,2 -function. Define the
process Z by setting Z(t) = f (t, Xt ). Then Z satisfies the stochastic differential
equation
⇢
∂f ∂f
d f (t, Xt ) = (t, Xt ) + µt (t, Xt )
∂t ∂x
1 ∂2 f ∂f
+ st2 2 (t, Xt ) dt + s (t, Xt ) dWt . (2.3)
2 ∂x ∂x
9
Proof. We will not provide a complete rigorous proof here, but will instead
provide the heuristic proof presented in “Arbitrage Theory in Continuous Time”
by Thomas Björk [5] instead.
If we do a Taylor expansion till the second-order terms we obtain the fol-
lowing:
∂f ∂f 1 ∂2 f 2 1 ∂2 f 2 ∂2 f
df = dt + dXt + (dXt ) + (dt) + dtdXt . (2.4)
∂t ∂x 2 ∂ x2 2 ∂ x2 ∂t∂ x
By definition we have
dXt = µt dt + st dWt ,
so, at least formally, we obtain
Using the assumptions of Itô’s lemma and following the formal multiplica-
tion rules 8 2
< (dt) = 0,
>
dt · dWt = 0,
>
:
(dWt )2 = dt,
equation (2.4) takes the form:
∂f ∂f 1 ∂2 f
df = dt + dXt + (dXt )2 .
∂t ∂x 2 ∂ x2
With the Itô calculus, we can derive the stochastic differential equation
(SDE) that describes the evolution of asset prices over time. Solving these
SDEs gives us derivative pricing models, which are used to determine the fair
prices of financial derivatives.
10
Brownian motion by ensuring the non-negativity of stock prices and incorpo-
rating a drift term to account for trends.
It can be formally defined as the solution to the SDE
11
If the underlying asset’s price at expiration T is greater than the strike price
K, the option will be exercised, and the payoff will be the difference between
ST and K. If the underlying asset’s price is less than or equal to the strike
price, the option will not be exercised, and the payoff will be zero. The payoff
functions for European call and put options are represented in Figure 2.1.
Figure 2.1. Payoff functions for European call and put options.
Black-Scholes model
The big question of how to price options has been a challenge for traders
and researchers for a long time. During the early days, people relied on their
intuition, experience, and on trade-offs between supply and demand. However,
as financial markets grew more complex, it became clear that this approach
was insufficient.
In 1973 the economists Fischer Black and Myron Scholes wrote a paper
“The Pricing of Options and Corporate Liabilities” [6] where they introduced
a mathematical model for the dynamics of a financial market. Later in 1997,
Scholes was awarded the Nobel Prize for this work, and though Black couldn’t
receive the award due to his passing in 1995, the Swedish Academy acknowl-
edged his significant contribution.
The Black-Scholes model consists of a risky asset St (stock) and a risk-free
asset Bt (bank account or bond) with dynamics given by:
dBt = r Bt dt.
The Black-Scholes differential equation is derived based on the following
set of assumptions:
12
1. The stock price follows a geometric Brownian motion.
2. There are no transaction costs or taxes.
3. There are no restrictions on short selling.
4. The risk-free interest rate r and the standard deviation s are known and
constant.
5. There are no arbitrage opportunities.
6. There are no dividends paid out during the life of the option.
7. Security trading is continuous.
8. The underlying asset is traded in a frictionless market.
9. European options are considered.
Risk-neutral valuation
Let F denote the payoff function. Then it can be shown that the price of the
corresponding option can also be expressed as
rT
F =e E[F(XT )],
where Xt denotes the solution to
13
available for solving the differential equation, but most of them are inspired
by the finite-difference techniques for solving ordinary differential equations.
In many cases, multiple (tens of thousands) simulations may be needed to
make sure that the average provides a good estimate for the expected value.
Since simulations are often expensive, various variance reduction techniques
are used, to reduce the variance without increasing the number of simulations.
An example of a variance reduction technique is the so-called method of an-
tithetic variates, where each random variable in the simulation is paired with
its negation, effectively canceling out some of the variance. This technique
leverages the negative correlation between the original and antithetic variates
to produce a more stable average. Thus, it can potentially reduce the number
of simulations required for accurate results. Other common variance reduc-
tion techniques include control variates, importance sampling, and stratified
sampling, each providing its own unique approach to mitigating variance in
the context of Monte Carlo simulations. By employing these techniques, re-
searchers and practitioners can obtain more reliable results with fewer simu-
lations. Consequently, these techniques can reduce computational costs and
increasing the efficiency of the overall process.
Quasi-Monte Carlo methods are another alternative to traditional Monte
Carlo simulations that can provide more accurate results with fewer simula-
tions. These methods replace the random number generation process with
deterministic sequences of low-discrepancy numbers, such as Sobol or Halton
sequences, which are specifically designed to cover the sample space more
uniformly than purely random samples. Quasi-Monte Carlo methods exploit
the structure and regularity of these low-discrepancy sequences to achieve
faster convergence rates and, in turn, more accurate estimates of the expected
value with fewer samples.
For additional details, see [13], which is a standard reference for using
Monte Carlo methods for pricing derivatives.
Volatility
Of all the variables used in the Black-Scholes model, the only one that is not
known with certainty is volatility. Volatility is a measure of the asset’s price
fluctuations and it is often used as an indicator of the level of risk associated
with that instrument. High volatility implies uncertainty about the option price
and hence high risk, while low volatility means relatively stable price and low
risk.
There are two approaches to getting an estimate of volatility - to use historic
volatility or to use implied volatility.
Historic or realized volatility measures the standard deviation of historic
returns over a specific period in the past. The main drawback of using histor-
ical volatility is that it does not consider market sentiment about future price
movements.
14
In many cases, the second approach, the so-called implied volatility, is more
favorable. The implied volatility is the volatility level that is implied by the
current market price of an option, based on the Black-Scholes option pricing
model. It can be calculated by getting market price data for another “bench-
mark” option written on the same time and the same underlying stock as the
option that we want to value. Denoting the price of the benchmark option
by p, the strike price by K, today’s observed value of the underlying stock
by S, and writing the Black-Scholes pricing formula for European calls by
c(S, t, T, r, a, K), we then solve the following equation for a:
p = c(S, t, T, r, a, K). (2.8)
In other words, we try to find the value of a which the market has implicitly
used for valuing the benchmark option. This value of a is called the implied
volatility, and we then use the implied volatility for the benchmark in order to
price our original option. Thus, we price the original option in terms of the
benchmark.
A plot of observed implied volatility as a function of the exercise price for
a fixed expiration date forms a curve which is called the volatility curve. If the
Black-Scholes model completely accurately represented the situation on the
stock market, this curve would be a straight line (as the volatility is assumed
constant in the model).
In practice, however, we can observe different types of curvature. When
the market is relatively stable and risk is higher for out-of-the-money (for
call/put options - the price of an underlying asset is lower/higher than the strike
price) and in-the-money (for call/put options - the price of an underlying asset
is higher/lower than the strike price) options we observe a U-shaped pattern
called the volatility smile. Another possible shape is when the smile is skewed
to one side. This is called a volatility smirk and it can be observed when the
market is turbulent or for options on volatile stocks. For example, the implied
volatility for out-of-the-money put options (downside) is typically higher than
for out-of-the-money call options (upside) due to the market’s perception of a
higher risk for large price drops. See Figure 2.2 for an example.
The collection of all implied volatilities of a set of maturities forms the so-
called volatility surface (Figure 2.3). The shape and dynamics of volatility
surfaces are subject of huge interest to both academics and practitioners.
The main use of a volatility surface is pricing and valuing options. The
implied volatilities derived from the surface are crucial inputs in option pricing
models. By understanding how the implied volatilities change across different
strike prices and expiration dates, traders and investors can identify mispriced
options and can better estimate the fair value of options.
Volatility index
One notable application of implied volatility in the financial markets is the
Volatility Index or VIX.
15
Figure 2.2. Volatility smile and volatility smirk.
The volatility Index or VIX was introduced in 1993 by the Chicago Board
Options Exchange. It measures the level of fear or stress in the stock market
and is often called the "Fear Index". Investors are able to trade VIX futures
since 2004 and VIX options since 2006. By definition, it is a derivative of the
SPX index, which can be represented as:
q
V IXt = 2E Q [log (St+D /St ) |Ft ] ⇥ 100, (2.9)
where St is the value of the SPX, D = 30 days, (Ft )t 0 is the natural filtra-
tion of the market, and E Q is the risk-neutral expectation. According to this
formula, the VIX represents the annualized square root of the price of a con-
tract with payoff equal log(St+D /St ) and measures the market’s expectation
of 30-day S&P500 volatility implicit in the prices of near-term S&P options.
When the VIX is up, it means that there are significant and rapid price fluc-
tuations in the S&P500. The VIX typically has a negative correlation with
the S&P500, so in periods of market stress, the VIX increases. VIX options
are a very useful diversification tool since they allow us to consider only the
direction of the volatility movement without taking into account market risk.
16
Figure 2.3. Volatility surface.
17
3. Stochastic volatility models
Since 1973, the Black-Scholes model’s assumptions have been relaxed and
generalized in many directions, leading to a myriad of models. We are partic-
ularly interested in the assumption of constant volatility which is one of the
main limitations of the Black-Scholes model.
In 1990 Hull and White [15] generalized the model by allowing volatility to
be stochastic. Since then, many stochastic volatility models have been intro-
duced and their popularity continues to grow alongside complexity. Among
others, the Heston model, SABR, and Bergomi model are notable examples.
where µt is the drift of stock price returns and the volatility Vt is stochastic,
and in its turn satisfies the SDE
p
dVt = l (q Vt ) dt + s Vt dZ2 (t).
18
of volatility to return to an average level in the long run) and the presence
of volatility clustering (periods of high volatility followed by periods of low
volatility, and vice versa, instead of being randomly distributed over time).
This model has become popular because of many reasons, including that it
does not require that stock prices follow a log-normal probability distribution
and has a closed-form characteristic function.
19
The correlation between the increments is closely related to the regularity
of the sample paths of the fractional Brownian motion.
A well-known property of the standard Brownian motion states that its paths
are almost surely Hölder continuous with parameter g for any g < 1/2. In other
words, for any e > 0,
20
3.2.3 Rough volatility
The concept of rough volatility is based on the observation that the realized
volatility can be characterized by a fractional Brownian motion with Hurst
parameter 0 < H < 1/2. As we have seen in the previous section, when H <
1/2, the time series has a long memory and exhibits so-called anti-persistence,
meaning that large changes in price are likely to be followed by large changes
in the opposite direction.
This observation was first demonstrated in 2018 by Gatheral, Jaisson, and
Rosenbaum in [11], where the authors also showed that H can vary over time,
but typically lies between 0.06 and 0.2, with higher values occurring during
financial crises.
Since then multiple rough volatility models have been introduced. One of
them is the so-called Rough Heston Model, introduced in [10]:
p
dSt = St Vt dWt ,
Z t
1
Vt = V0 + (t s)a 1
l (q Vs ) ds
G(a) 0
Z t
1 p
+ (t s)a 1
l n Vs dBs . (3.1)
G(a) 0
The parameters l , q ,V0 and n are positive real numbers, playing here the
same role as in the classical Heston model. Bt and Wt are two Brownian mo-
tions with the quadratic covariation process given by
dWt · dBt = r dt.
Since we want the volatility process to be rough, the parameter a = H + 1/2
should lie in the interval (1/2, 1). Setting a = 1, we can reduce the rough
Heston model to the classical Heston model.
21
They show that the classical Heston model is sufficient if one only needs to
take into account facts 1-3. However, the classical Heston model fails to catch
fact 4, and this is where the rough Heston comes to help.
However, there is also a fifth stylized fact, not reflected in the rough Heston
model, namely, the so-called Zumbach effect.
In 2003, Zumbach [17] described the effect, which was later named after
him, by measuring the impact of past price trends on volatility. He showed
that price trends induce an increase in volatility; specifically, price trends tend
to increase volatility. There are two types of this effect:
• Weak Zumbach effect: the predictive power of past squared returns on
future volatility is stronger than that of past volatility on future squared
returns. To check this on data, one typically shows that the covariance
between past squared price returns and future realized volatility (over a
given duration) is larger than that between past realized volatility and
future squared price returns.
• Strong Zumbach effect: the conditional law of future volatility depends
not only on the past volatility trajectory but also on past returns.
To be able to successfully model the strong Zumbach effect, Dandapani et
al. [8] introduced a class of super-Heston rough volatility models, based on
quadratic versions of so-called Hawkes processes. The rough Heston model is
a special case of the super-Heston models.
Without giving a formal definition, we will say that in super-Heston models,
the square root in the integral
Z t
1 p
(t s)a 1
l n Vs dBs ,
G(a) 0
Z t Z t
a 1 l l p
Zt = (t s) (q0 (s) Zs ) ds + (t s)a
h Vs dWs , 1
0 G(a) 0 G(a)
(3.3)
where a 2 (1/2, 1), l > 0, h > 0, and q0 is a deterministic function.
22
Contrary to, for example, the classical Heston model, in this model the asset
price S and its volatility V are driven by the same Brownian motion W .
The interpretation of parameters a, b, and c is as follows:
• a > 0 is the sensitivity of volatility to past price returns. The lower a is,
the flatter the volatility smile is.
• b > 0 is responsible for the asymmetry of the feedback effect. Volatility
is higher when Z is negative and lower when Z is positive considering
that they have the same absolute value (shifts of the volatility smile to
the left or to the right);
• c > 0 is the minimal instantaneous variance (shifts the volatility smile
upwards or downwards).
Apart from the simplicity of this model (e.g., the fact that the model is
driven by a single Brownian motion), the quadratic rough Heston model is in-
teresting because it was the first model that made it possible to simultaneously
solve the problem of fitting both SPX and VIX volatility smiles.
In the remaining part of the thesis, we focus on the questions of implement-
ing the quadratic rough Heston model.
23
4. Simulation of the Quadratic Rough Heston
model
In the literature, three main methods of simulating the quadratic rough He-
ston model are used:
1. Hybrid scheme by Bennedsen et al., [3];
2. Multifactor scheme by Abi Jaber et al., [2];
3. Hybrid Multifactor scheme by Rømer [22], combining the other two
schemes.
Z t Z t
l l p
Zt = (t s)a 1
(q0 (s) Zs ) ds + (t s)a 1
h Vs dWs ,
0 G(a) 0 G(a)
where
l l
l̃ = , h̃ = h.
G(a) G(a)
We can drop the tildes, and redefining l = l̃ , h = h̃, we can write
24
Z t Z t
a 1
Zt = l (t s) (q0 (s) Zs ) ds + h (t s)a 1
ss dWs .
0 0
As is typically done in the literature (e.g., [12, 21]), we can narrow the
space of admissible functions q0 (s) in such a way that the expression may be
rewritten as
Z t Z t
Zt = z0 l (t s)a 1
Zs ds + h (t s)a 1
ss dWs ,
0 0
where z0 2 R is a new calibration constant that replaces q0 (s).
Moreover, it is more convenient to substitute b for a 1, so that we can
write Z Z
t t
b b
Zt = z0 l (t s) Zs ds + h (t s) ss dWs .
0 0
Thus, we have reformulated the quadratic rough Heston model in the form
q
dSt = St st dWt , st = a (Zt b)2 + c, (4.1)
Z t Z t
b b
Zt = z0 l (t s) Zs ds + h (t s) ss dWs , (4.2)
0 0
where a, b, c, l , h > 0, b 2 (0, 1/2), z0 2 R.
dSt = (r q) dt + St st dWt .
Then, the forward price process is defined as Ft = e(r q)t St . By applying the
Itô formula to Ft , we see that it satisfies the same equation as in (4.1).
Using forward price processes allows us to simplify the notation, without
losing any information; we can always return to the stock price process using
multiplication by e (r q)t .
25
1 1 1 1 1 2 2
d log St = dSt dSt · dSt = st dWt S s dt.
St 2 St2 2 St2 t t
Then, integrating, we see that
✓Z t Z t ◆
1
St = S0 exp ss dWs ss2 ds . (4.3)
0 2 0
dSt = (r q) St dt + St st dWt ,
then ✓Z Z t ◆
t 1
St = S0 exp ss dWs ss2 ds + (r q)t .
0 2 0
Discretization
Suppose we want to simulate Y on an interval [0, T ], for some T > 0. We
divide the interval [0, T ] into N subintervals, creating the grid
{0 = t0 , t1 , . . . , tN = T }, ti = i · Dt,
where Dt := NT .
Our goal is to simulate
Yti , i = 0, 1, . . . , N.
26
First, we can write
Z ti+1 i Z tk+1
Yti+1 =
0
(ti+1 s) b
ss dWs = Â (ti+1 s) b
ss dWs ,
k=0 tk
and, assuming that s does not vary too much, approximate
i Z tk+1
Yti+1 ⇡ Â stk tk
(ti+1 s) b
dWs .
k=0
b
For all terms but the last one, we approximate (ti+1 s) by a constant func-
tion within each integration interval, i.e., write
i 1
 (ti+1 bk ) b
stk DWtk .
k=0
Here, DWtk = Wtk+1 Wtk , and (bk )Nk=01 is a sequence of freely chosen evalua-
tion points within the discretization intervals, bk 2 [tk ,tk+1 ].
In the last term, the kernel contains a singularity. We keep the singular term
intact within the integral:
Z ti+1
b
sti (ti+1 s) dWs .
ti
In the original paper, there was a suggestion to preserve the singularity term
intact in several last terms, but the paper’s authors noted that the performance
is already good when using only one term. We prefer to keep the algorithm as
simple as possible in this case and will not try to consider additional terms.
In other words, our approximation in the Hybrid scheme is given by
i 1 Z ti+1
 (ti+1 b b
(N)
Yti+1 := bk ) stk DWtk + sti (ti+1 s) dWs .
k=0 ti
If we denote Z ti+1
b b
DWti := (ti+1 s) dWs , (4.4)
ti
this simplifies to
i 1
b
 (ti+1 b
(N)
Yti+1 := bk ) stk DWtk + sti DWti . (4.5)
k=0
27
Z tk+1
b b 1
cov (DWtk , DWtk ) = (tk+1 s) 2b
ds = (Dt)1 2b
, (4.7)
tk 1 2b
and
!
i 1
1 b
WtH
i+1
=
G(1 b) Â (ti+1 bk ) b
DWtk + DWti , i = 0, . . . , N 1.
k=0
(4.10)
To be concrete, we can put bk = tk , and then we get ti+1 bk = ti+1 tk =
ti k+1 ,
resulting in
!
i 1
1 b b
WtH
i+1
=
G(1 b) Â ti k+1 DWtk + DWti , i = 0, . . . , N 1. (4.11)
k=0
Simulation results for various values of the Hurst parameter are shown in
Figures 4.1 - 4.4.
28
Figure 4.1. Riemann-Liuville fBm with Hurst parameter H = 0.05.
29
Figure 4.3. Riemann-Liuville fBm with Hurst parameter H = 0.5.
Z ti+1 i Z tk+1
0
(ti+1 s) b
Zs ds = Â (ti+1 s) b
Zs ds
k=0 tk
i Z tk+1
⇡ Â Ztk tk
(ti+1 s) b
ds.
k=0
The integrals inside of the sum are easy to calculate:
Z tk+1 h i
b 1 b b
(ti+1 s) ds = (ti+1 tk )1 (ti+1 tk+1 )1
tk 1 b
1 h i
1 b 1 b
= ti+1 k ti k ,
1 b
The last equality follows because ti = i · Dt and we can, e.g., write
ti+1 tk = (i + 1) · Dt k · Dt = (i + 1 k) · Dt = ti+1 k ,
therefore,
Z ti+1 h
i i
1 1 b 1 b
0
(ti+1 s) b
Zs ds ⇡
1 b  k ti+1
Zt k ti k . (4.13)
k=0
Finally, combining (4.12) and (4.13), the discretized equation can be written
as
l i h i i 1
1 b 1 b b b
Zti+1 = z0
1 b  Ztk ti+1 k ti k +h  ti k+1 stk DWtk + h sti DWti ,
k=0 k=0
q
sti+1 = a (Zti+1 b)2 + c.
The above expressions thus represent the hybrid scheme for the quadratic
rough Heston model.
31
4.3 Multifactor scheme
4.3.1 Kernel approximations
The Multifactor scheme, introduced by Abi Jaber and El Euch in [2], relies
on approximating the kernel function by a sum of exponentials:
m 1
t b
⇡ Â ci e gi t
. (4.14)
i=0
This representation is in fact a Laplace transform of the weighted sum of Dirac
measures µ = Âi cni dgi :
Z •
b tx
t ⇡ e µ(dx).
0
Original approximation
When the multifactor scheme was first introduced for simulating rough
volatility models in [2], its authors used the following algorithm for computing
the coefficients ci , gi :
Let
1 p !2 r !2
5 5
m 5 10 (1 2H) 1 10 b
pm = = .
T 5 2H T m2 b
Then the weights and exponents are given by
b
pm ⇣ ⌘ b pm (i + 1)b +1 ib +1
ci = (i + 1)b ib , gi = , i = 0, . . . , m 1.
b G(b ) b + 1 (i + 1)b ib
In Figure 4.6 we can see that the accuracy of the approximation decreases as
t ! 0, and that we probably need several hundreds of terms to get satisfactory
results.
Geometric approximation
In the follow-up paper [1], the authors proposed a more efficient choice of
approximation.
For a given b , m and T , fix a parameter xm .
32
Figure 4.5. Original approximation of fractional kernel.
33
Figure 4.6. Geometric approximation of fractional kernel.
Beylkin-Monzón Algorithm
Assume that we want to approximate t b on the interval [tmin ,tmax ], with
accuracy e > 0.
1. By applying a linear transformation to its argument, convert t b into a
function f : [0, 1] ! R:
b
f (t) := (t · (tmax tmin ) + tmin ) .
2. Pick a sufficiently large integer N (on the order of 100-200) and calculate
samples ✓ ◆
k
hk = f , k = 0, . . . , 2N.
2N
3. Construct the so-called Hankel matrix H = {hi+ j 2 }N+1 i, j=1 and sort its
eigenvalues (which are real and non-negative since the matrix is sym-
metric): s0 s1 . . . sN 0.
4. Find the smallest eigenvalue sm such that sm khk2 · e, and denote the
corresponding eigenvector by u 2 RN+1 .
5. Using the coordinates of u, construct the polynomial Pu of order N.
6. Select m distinct real roots {r0 , . . . , rm 1 } of Pu that lie in the interval
(0, 1].
7. By solving a least-square problem, find {c̃0 , . . . , c̃m 1 } that minimize
!2
2N m 1
 hk  c̃i rik .
k=0 i=0
34
8. Denote
g̃i = 2N log(ri ), i = 0, . . . , m 1.
9. Return back to the original coordinates by setting
g̃i
gi = , ci = c̃i · egi ·tmin , i = 0, . . . , m 1.
tmax tmin
The final approximation is then given by
m 1
t b
⇡ Â ci e gi t
.
i=0
It also works fairly well for approximating the values near the singularity.
On the interval [0.001, 10], to achieve the accuracy of 0.01 it is sufficient to
take 5 terms (Figure 4.8).
35
Figure 4.8. Beylkin-Monzón approximation of fractional kernel on the interval
[0.001, 10].
q
st = a (Zt b)2 + c,
where b 2 (0, 1/2), l , h, a, b, c > 0, and Z0 2 R.
The fractional kernel depending on t makes the process non-Markovian and
non-semimartingale. By representing the kernel as a sum of exponentials
m 1
t b
⇡ Â ci e gi t
,
i=0
we regularize the problem and make it more tractable.
Multifactor approximation
Substituting the multifactor approximation of the kernel into the equation,
we get
Zt Z t
Zt = z0 Â ci l e gi (t s) Zs ds + h e gi (t s) ss dWs .
i 0 0
We denote
Z t Z t
gi (t s) gi (t s)
Zti := zi0 l e Zs ds + h e ss dWs ,
0 0
so that
36
Zt = Â ci Zti , z0 = Â ci zi0 .
i i
Zti
We can derive an SDE for as follows. Note that
Z t Z t
Zti = zi0 + e gi t l egi s Zs ds + h egi s ss dWs .
0 0
so Zti satisfy:
⇥ ⇤
dZti = gi Zti l Zt dt + h st dWt , Zt = Â ci Zti .
i
Note that in this scheme the calibration parameter Z0 (which was playing
the role of the function q (s)) is replaced by multiple calibration parameters
Z0i . This gives us additional degrees of freedom when doing calibration.
With this, we finally arrive at the following multifactor approximation of
the quadratic rough Heston model:
⇥ ⇤
dZti = gi Zti + l Zt dt + h st dWt , i = 0, . . . m 1,
q
st = a (Zt b)2 + c,
where
Z0i = zi0 , i = 0, . . . m 1, (4.16)
Zt = Â ci Zti .
i
Euler scheme
Following [21], we discretize the scheme using the explicit-implicit Euler
scheme. The time interval [0, T ] is split into N parts. Dt = T /N, tk = k · Dt,
0 = t0 < . . . < tN = N. We simulate Brownian increments as DWk ⇠ N (0, Dt).
Then, if we denote Zki = Ztik , Zk = Ztk , etc., the implicit-explicit Euler scheme
is given by
i 1
Zk+1 = (Z i l Zk Dt + h sk DWk ),
1 + gi Dt k
q
sk = a (Zk b)2 + c,
37
m 1
Zk = Â ci Zki .
i=0
Assume we want to solve the equation on the interval [0, T ]. We split the
interval into N parts, and denote Dt = T /N, tk = k · Dt, 0 = t0 < . . . < tN = N.
As we did in the multifactor scheme, we denote
Z t Z t
gi (t s) gi (t s)
Zti := zi0 l e Zs ds + h e ss dWs .
0 0
and
Z tk Z tk
b b b
(tk s) ss dWs ⇡ stk 1
(tk s) dWs = stk 1 DWtk 1 ,
tk 1 tk 1
where we use the same notation as we did in the hybrid scheme. Next, we
can look at the non-singular terms, and use the multifactor approximation for
them.
38
Z tk 1
Z tk 1
b b
l (tk s) Zs ds + h (tk s) ss , dWs
0 0
Z tk Z tk
1 1
⇡ Â ci l e gi (tk s)
Zs ds + h e gi (tk s)
ss dWs
i 0 0
Z tk Z tk
1 1
= Â ci e gi Dt
l e gi (tk 1 s)
Zs ds + h e gi (tk 1 s)
ss dWs
i 0 0
= Â ci e gi Dt i
Ztk 1 .
i
Dt 1 b b
Ztk ⇡ Â ci e gi Dt i
Ztk 1 + Zt + stk 1 DWtk
i 1 b k 1 1
Dt 1 b b
Zk+1 = Â ci e gi Dt i
Zk + Zk + sk DWk .
i 1 b
The factors Zki
can be found using the same implicit-explicit scheme as in
the simple multifactor scheme:
i 1
Zk+1 = (Z i l Zk Dt + h sk DWk ).
1 + gi Dt k
b
The pairs of random variables (DWtk , DWtk ), where
and Z tk+1
b b
DWtk := (tk+1 s) dWs ,
tk
can, exactly like in the hybrid scheme, be generated as jointly normal with the
covariance matrix given by
" 1 b
#
1
Dt 1 b (Dt)
1 1 b 1 1 2b .
1 b (Dt) 1 2b (Dt)
39
(a)
(b)
(c)
40
5. Calibration
Even though the quadratic rough Heston is notable for the fact that it allows
to simultaneously fit SPX and VIX volatility surfaces, we only calibrate using
SPX smiles in this thesis.
41
When choosing which of the two option types to use for calibration, we
need to consider:
1. Time horizon: SPXW is more suitable for shorter timeframes, and SPX
is more suitable for analyzing long-term market behavior.
2. Liquidity: SPX options have higher trading volume than SPXW, which
leads to more accurate pricing and narrower bid-ask spreads.
In this thesis, we choose to calibrate using SPX only.
42
7. For the given expiry date, for all available strikes, calculate the option
prices based on the simulated stock prices.
8. From the simulated option prices, calculate the implied volatility using
the py_vollib library (also known as "Let’s be rational").
9. Calculate the objective function for each expiry date, and take a sum of
the results.
10. Proceed to the next step of the optimization method.
However, in order to use the Nelder-Mead method, we first need to come up
with a reasonable initial guess for the parameters. It turns out that simply tak-
ing arbitrary points from the parameter space will most likely lead to numeri-
cal overflow errors, which will keep appearing even after multiple iterations of
the Nelder-Mead algorithm. To address this, we first use the Differential evo-
lution method [23] until we get a somewhat satisfactory approximation of the
volatility smile, and then we use the resulting parameter values as the initial
guess for the Nelder-Mead method. Note that an implementation of the Dif-
ferential evolution method is also available in the scipy.optimize library.
Here, n denotes our set of calibration parameters, O SPX - the given set of SPX
options, s o, mid the market mid implied volatility for the option o, and s o,n is
the implied volatility of the option o in the quadratic rough Heston model with
the given parameter n.
This choice has certain drawbacks. For example, options that are far in-
the-money or far out-of-the-money are often accompanied by a higher degree
of uncertainty and larger bid-ask spreads in implied volatilities. By using an
objective function that treats all options equally, we end up reducing the accu-
racy of the calibration for at-the-money options while trying to fit the options
which even the market cannot price with certainty.
To address this, we can use the bid-ask-spread as a proxy measure of how
important a certain strike price is:
1 (s o, mid s o, n )2
F(n) =
#O SPX Â e + |s o, bid s o, ask |
.
o2O SPX
Here, the smaller the spread is, the higher the contribution of each option to
the total error. e is a small constant (e.g. 0.001) added to avoid division by
zero.
43
As an alternative to using the bid-ask spread values, we can use the trading
volume as the weights.
Finally, we can choose to ignore s o, mid completely and only penalize the
cases when s o, n leaves the bid-ask corridor [s o, bid ; s o, ask ]:
1
F(n) =
#O SPX Â |s o, bid s o, n |2+ + |s o, n s o, ask |2+ ,
o2O SPX
44
without losing much accuracy (especially if we look at at-the-money strike
prices).
Figure 5.1. Monte Carlo vs quasi-Monte Carlo when the number of simulations is
low.
Figure 5.2. Lower variance of implied volatility with quasi-Monte Carlo simulations.
Figure 5.3. Monte Carlo vs quasi-Monte Carlo when the number of simulations is
high.
45
In contrast, Sobol sequences showed really good performance for both large
and small time steps, resulting in smooth, well-distributed plots. However, for
a smaller percentage of time steps we observe undesirable “Moire” patterns
(Figure 5.5) on the scatter plots. Nevertheless, since the instances of Moire
patterns are infrequent and the general quality of samples is good, we prefer
to use Sobol sequences.
Figure 5.4. An example of the Sobol sampling performing better than Halton’s.
5.4 Calibrating b
Unlike the rest of the calibration parameters, calibrating of b is compli-
cated. This is because:
1. The generated Gaussians depend on b via their covariance matrix (4.9).
2. The exponential kernel decomposition (4.14) depends on b .
3. As a consequence of 2, the role of the calibration parameters zi0 (see
(4.16)), also depend on b .
To address this, we perform calibration in two steps. First, we set zi0 = 0
for all i and let b vary. To avoid regenerating Gaussian arrays with a new
covariance matrix at each step, we do the following:
1. At the beginning of the calibration process we generate an array with
dimensions N ⇥ M ⇥ 2, consisting of standard independent Gaussians.
46
2. Whenever b and the covariance matrix change, we transform the ar-
ray by using the same procedure that we used for generating correlated
quasi-Gaussians (multiplying with the Cholesky decomposition of the
covariance matrix).
Once this step is finished, we fix b and let zi0 vary.
With the market data that we used, the value of b that we found was b ⇡
0.395, which corresponds to the Hurst parameter H = 0.105.
5.6 Results
We obtain a fairly good fit to the volatility surface, see, e.g., Figures 5.6-5.8.
47
Figure 5.7. Simulated volatility surface, 154 to 245 days.
We note that we are getting a better fit for short expiries. This is in line
with the fact that the quadratic rough Heston model is typically used to model
options with short expiries.
48
The fit becomes even better if we limit ourselves to a single maturity date
(Figure 5.9 - 5.10).
49
(a) Calibrated for a single maturity date.
two of these three methods (or our implementations of these methods) are not
particularly accurate. We also notice that the simulated prices are clearly not
Gaussian, see Figure 5.11.
50
Figure 5.11. Histogram of simulated stock prices after t days.
51
Figure 5.12. Variation of l and h.
52
6. Conclusion
The field of simulating rough volatility models is very active, with multiple
competing approaches and ideas, and a large number of recent papers.
We have studied the quadratic rough Heston model, which is relatively sim-
ple to implement (compared to other rough Heston models), and yet provides
a fairly good fit for the volatility surface.
We have implemented the model using three different schemes that are com-
monly used for simulating rough volatility models. In particular, we simplified
and implemented Rømer’s Hybrid multifactor scheme, as well as Beylkin-
Monzón algorithm, in Python.
We also found that quasi-Monte Carlo methods can be efficiently used to
speed up the calibration process also in rough volatility models. We have
discovered that different schemes may require different sets of calibration pa-
rameters (meaning that at least some of these schemes may not be accurate).
We have also found that completely different sets of calibration parameters
may yield equally good fits to the market data (meaning potentially that there
may be redundancy in the model parameters).
Unfortunately, we have not calibrated VIX options, even though this is
known to be one of the strengths of the quadratic rough Heston model. This
was mostly due to theoretical complications when simulating VIX option prices.
The complications were not unsurmountable, however, and given more time,
this presents a promising future direction for our work.
53
References
[1] Eduardo Abi Jaber. Lifting the Heston model. Quantitative Finance,
19(12):1995–2013.
[2] Eduardo Abi Jaber and Omar El Euch. Multifactor approximation of rough
volatility models. SIAM Journal on Financial Mathematics, 10(2):309–349,
2019.
[3] Mikkel Bennedsen, Asger Lunde, and Mikko S. Pakkanen. Hybrid scheme for
Brownian semistationary processes. Finance and Stochastics, 21(4):931–965,
2017.
[4] Gregory Beylkin and Lucas Monzón. On approximation of functions by
exponential sums. Applied and Computational Harmonic Analysis,
19(1):17–48.
[5] Tomas Björk. Arbitrage theory in continuous time. Oxford University Press, 3rd
edition, 2009.
[6] Fischer Black and Myron Scholes. The pricing of options and corporate
liabilities. Journal of political economy, 81(3):637–654, 1973.
[7] Laure Coutin and Philippe Carmona. Fractional Brownian motion and the
Markov property. Electronic Communications in Probability, 3:12, 1998.
[8] Aditi Dandapani, Paul Jusselin, and Mathieu Rosenbaum. From quadratic
Hawkes processes to super-Heston rough volatility models with Zumbach
effect. Quantitative Finance, 21(8):1235–1247, 2021.
[9] Omar El Euch, Masaaki Fukasawa, and Mathieu Rosenbaum. The
microstructural foundations of leverage effect and rough volatility. Finance and
Stochastics, 22:241–280, 2018.
[10] Omar El Euch and Mathieu Rosenbaum. The characteristic function of rough
Heston models. Mathematical Finance, 29(1):3–38, 2019.
[11] Jim Gatheral, Thibault Jaisson, and Mathieu Rosenbaum. Volatility is rough.
Quantitative Finance, 18(6):933–949, 2018.
[12] Jim Gatheral, Paul Jusselin, and Mathieu Rosenbaum. The quadratic rough
Heston model and the joint S&P 500/VIX smile calibration problem, 2020.
arXiv:2001.01789.
[13] P. Glasserman. Monte Carlo Methods in Financial Engineering. Applications
of mathematics: stochastic modelling and applied probability. Springer, 2004.
[14] Steven Heston. A closed-form solution for options with stochastic volatility
with applications to bond and currency options. Review of Financial Studies,
6:327–343, 1993.
[15] John Hull and Alan White. The Pricing of Options on Assets with Stochastic
Volatilities. The Journal of Finance, 42(2):281–300, June 1987.
[16] S.C. Lim and V.M. Sithi. Asymptotic properties of the fractional Brownian
motion of Riemann-Liouville type. Physics Letters A, 206(5):311–317.
[17] Paul E Lynch and Gilles O Zumbach. Market heterogeneities and the causal
structure of volatility. Quantitative Finance, 3(4):320, 2003.
54
[18] Benoit B Mandelbrot and John W Van Ness. Fractional Brownian motions,
fractional noises and applications. SIAM review, 10(4):422–437, 1968.
[19] D. Marinucci and P.M. Robinson. Alternative forms of fractional Brownian
motion. Journal of Statistical Planning and Inference, 80(1):111–122.
[20] J. A. Nelder and R. Mead. A Simplex Method for Function Minimization. The
Computer Journal, 7(4):308–313, 01 1965.
[21] Mathieu Rosenbaum and Jianfei Zhang. Deep calibration of the quadratic rough
heston model, 2022. arXiv:2107.01611.
[22] Sigurd Emil Rømer. Hybrid multifactor scheme for stochastic Volterra
equations with completely monotone kernels. SSRN Electronic Journal, May
2020.
[23] Rainer Storn and Kenneth Price. Differential Evolution – A Simple and
Efficient Heuristic for global Optimization over Continuous Spaces. Journal of
Global Optimization, 11(4):341–359, 1997.
55