Fitting Copulas To Data
Fitting Copulas To Data
5.5.1
5.5.2
5.5.3
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
We have data vectors X1 , . . . , Xn with identical distribution function F , describing nancial loss or nancial risk factor returns. We write Xt = (Xt ,1 , . . . , Xt ,d ) for an individual data vector, and X = (X1 , . . . , Xd ) for a generic random vector with df F . F has continuous margins F1 , . . . , Fd and by Sklars Theorem, a unique representation F (x ) = C (F1 (x1 ), . . . , Fd (xd )).
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
We use rank (Xt ,i ) to denote the rank of Xt ,i in X1,i , . . . , Xn,i , we can calculate the correlation coecient for the rank data {(rank(Xt ,i ), rankXt ,j )}, and this gives us the Spermans rank correlation coecient: 12 2 n(n 1) 1 1 (rank (Xt ,i ) (n + 1))(rank (Xt ,j ) (n + 1)) 2 2 t =1 (5.49)
n
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Recall Denition 5.27: For rvs X1 and X2 Kendalls tau is given by 1 )(X2 X 2 ))), where (X 1 , X 2 ) is (X1 , X2 ) = E (sign((X1 X independent copy of (X1 , X 2). The standard estimator of Kendalls tau (Xi , Xj ) is Kendalls rank correlation coecient : n 2
1
sign((Xt ,i Xs ,i )(Xt ,j Xs ,j ))
1t <s n
(5.50)
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
1 S (Xi , Xj ) = (6/ )arcsin ij ij , 2 where the nal approximation is very accurate. This suggests we estimate P by the matrix of pairwise Spearmans rank correlation coecient R S .
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
(Xi , Xj ) = (2/ )arcsinij so that a possible estimator of P is the matrix R with component 1 given by rij = sin( 2 rij ).
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Algorithm 5.55(eigenvalue method). Let R be a so-called pseudo -correlation matrix, i.e. a symmetric matrix if pairwise correlation estimate correlation estimates with unit diagonal entries and o diagonal entries in [-1,1] that is not positive semidenite. Calcukate the spectral decomposition R = GLG where L is the matrix of eigenvalues and G is an orthogonal matrix whose columns are eigenvectors of R . Replace all negative eigenvalues in L by small value > 0 to . obtain L
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Algorithm 5.55(eigenvalue method). , which will be symmetric and positive Calculate Q = G LG denite but not a correlation matrix, since its diagonal elements will not necessarily equal one. Return the correlation matrix R = (Q ), where denotes the correlation matrix operator.,where dened ( ) = ( ( ))1 ( ( ))1 , ( ( )) := diag( 11 , . . . , dd )
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
We now turn to the estimation of parameteric copulas by maximum likelihood (ML). In this section we describe brifely some general approaches to the rst step of estimating margins and constructing a pseudo sample of observations from the copula. In the following section we describe how the copula parameters are estimated by ML from pseudo-sample.
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
1 , . . . , F d denote estimates of the marginal dfs. The Let F pseudo-sample from the copula consists of the vectors 1 , . . . , U n , where U t = (Ut ,1 , . . . , Ut ,d ) = (F 1 (Xt ,1 ), . . . , F d (Xt ,d )) U
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
I { X t ,i x }
t =1
Extreme value theorey for the tails. Empirical distribution functions are known to be poor estimators of the underlying distribution in the tails, and the tail are model using a generalized Pareto distribution, the body of distribution may be modelled empirically.
Y-J, Chao 5.5 Fitting Copulas to Data
5.5
5.5.1
5.5.2
5.5.3
Example 5.57. Five years of daily log-return data (1996-2000). Intel, Microsoft and General Electric stocks. The marginal distributions are estimated empirically (method(2)). The pseudo-sample from copula is shown in Figure 5.14.
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Let C denote a parameter copula, where is the vector of parameters to be estimated. The MLE is obtained by maximizing
n
1 , . . . , U n) = ln L(; U
t =1
t) ln c (U
One could envisage Using the two-stage method to decide on the most appropriate cpula family and then estimating all parameters (marginal and copula) in a nal fully parametric round of estimation.
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Example 5.58 (tting the Gaussian copula). In the case of a Gaussina copula implies that the log-likelihood is 1 , . . . , U n) ln L(P ; U
n n d
=
t =1
t ,1 ), . . . , 1 (U t ,d )) ln fP ( (U
1 t =1 j =1
t ,j )) ln (1 (U
where f will be used to denote the joint density of a random vector with Nd (0, ) distribution.
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Yt Yt
Y-J, Chao
5.5
5.5.1
5.5.2
5.5.3
Example 5.58 (tting the Gaussian copula). For example 5.57 by full ML, the estimated correlation matrix has 0.58(INTC-MSFT), 0.34(INTC-GE) and 0.4(MSFT-GE); the log- likelihood at the maximum is 376.65, using the alternative method gives a log likelihood value of 376.62
Y-J, Chao