Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
arXiv:0912.2816v1 [math.PR] 15 Dec 2009 The Bivariate Normal Copula Christian Meyer∗† December 15, 2009 Abstract We collect well known and less known facts about the bivariate normal distribution and translate them into copula language. In addition, we prove a very general formula for the bivariate normal copula, we compute Gini’s gamma, and we provide improved bounds and approximations on the diagonal. 1 Introduction When it comes to modelling dependent random variables, in practice one often resorts to the multivariate normal distribution, mainly because it is easy to parameterize and to deal with. In recent years, in particular in the quantitative finance community, we have also witnessed a trend to separate marginal distributions and dependence structure, using the copula concept. The normal (or Gauss, or Gaussian) copula has even come to the attention of the general public due to its use in the valuation of structured products and the decline of these products during the financial crisis of 2007 and 2008. The multivariate normal distribution has been studied since the 19th century. Many important results have been published in the 1950s and 1960s. Nowadays, quite some effort is spent on rediscovering some of them. In this note we will proceed as follows. We will concentrate on the bivariate normal distribution because it is the most important case in practice (applications in quantitative finance include pricing of options and estimation of asset correlations; the impressive list by Balakrishnan & Lai (2009) mentions applications in agriculture, biology, engineering, economics and finance, the environment, genetics, medicine, psychology, quality control, reliability and survival analysis, sociology, physical sciences and technology). We will give an extensive view on its properties. Everything will be reformulated in terms of the associated copula. We will also provide new results (at least if the author has not been rediscovering himself...), including a very general formula implying other well known ∗ DZ BANK AG, Platz der Republik, D-60265 Frankfurt. The opinions or recommendations expressed in this article are those of the author and are not representative of DZ BANK AG. † E-Mail: [email protected] 1 formulas, a derivation of Gini’s gamma, and improved and very simple bounds for the diagonal. For collections of facts on the bivariate (or multivariate) distribution we refer to the books of Balakrishnan & Lai (2009), of Kotz, Balakrishnan & Johnson (2000), and of Patel & Read (1996), and to the survey article of Gupta (1963a) with its extensive bibliography (Gupta 1963b). For theory on copulas we refer to the book of Nelsen (2006). We will use the symbols P, E and V for probabilities, expectation values and variances. 2 Definition and basic properties Denote by 2 x 1 , ϕ(x) := √ exp − 2 2π Φ(h) := Z h ϕ(x) dx −∞ the density and distribution function of the standard normal distribution, and by 2 1 x − 2̺xy + y 2 p ϕ2 (x, y; ̺) := , exp − 2(1 − ̺2 ) 2π 1 − ̺2 Z h Z k ϕ2 (x, y; ̺) dy dx, Φ2 (h, k; ̺) := −∞ −∞ the density and distribution function of the bivariate standard normal distribution with correlation parameter ̺ ∈ (−1, 1). The bivariate normal (or Gauss, or Gaussian) copula with parameter ̺ is then defined by application of Sklar’s theorem, cf. Section 2.3 of (Nelsen 2006): (2.1) C(u, v; ̺) := Φ2 Φ−1 (u), Φ−1 (v); ̺ For ̺ ∈ {−1, 1}, the correlation matrix of the bivariate standard normal distribution becomes singular. Nevertheless, the distribution, and hence the normal copula, can be extended continuously. We may define C(u, v, −1) := lim C(u, v; ̺) = max(u + v − 1, 0), (2.2) C(u, v, +1) := lim C(u, v; ̺) = min(u, v). (2.3) ̺−→−1 ̺−→+1 Hence C(·, ·; ̺), for ̺ −→ −1, approaches the lower Fréchet bound, W (u, v) := max(u + v − 1, 0), and, for ̺ −→ 1, approaches the upper Fréchet bound, M (u, v) := min(u, v). 2 Furthermore, we have C(u, v; 0) = u · v =: Π(u, v), (2.4) the independence copula. What happens in between can be described using the following differential equation derived by Plackett (1954): d d2 Φ2 (x, y; ̺) = ϕ2 (x, y; ̺) = Φ2 (x, y; ̺) d̺ dx dy (2.5) We find C(u, v; ̺) − C(u, v; σ) = I(u, v; σ, ̺) := Z ̺ σ and in particular, ϕ2 Φ−1 (u), Φ−1 (v); r dr, C(u, v; ̺) = W (u, v) + I(u, v; −1, ̺), = Π(u, v) + I(u, v; 0, ̺), = M (u, v) − I(u, v; ̺, 1). (2.6) (2.7) (2.8) (2.9) In other words, the bivariate normal copula allows comprehensive total concordance ordering with respect to ̺, cf. Section 2.8 of (Nelsen 2006): W (·, ·) = C(·, ·; −1) ≺ C(·, ·; ̺) ≺ C(·, ·; σ) ≺ C(·, ·; 1) = M (·, ·) for −1 ≤ ̺ ≤ σ ≤ 1, i.e., for all u, v, W (u, v) = C(u, v; −1) ≤ C(u, v; ̺) ≤ C(u, v; σ) ≤ C(u, v, 1) = M (u, v). 1 2 into (2.8) we obtain Z ̺ 1 1 1 1 √ C 12 , 12 ; ̺ = + dr = + arcsin(̺), 2 4 4 2π 2π 1 − r 0 By plugging u = v = (2.10) a result already known to Stieltjes (1889). Mehler (1866) and Pearson (1901a), among other authors, obtained the socalled tetrachoric expansion in ̺: C(u, v; ̺) = uv +̺ Φ −1 ∞ X ̺k+1 −1 (u) ̺ Φ (v) Hek Φ−1 (u) Hek Φ−1 (v) (k + 1)! k=0 (2.11) where [k/2] Hek (x) = X i=0 are the Hermite polynomials. k! i!(k − 2i)! 3 i 1 − xk−2i 2 The bivariate normal copula inherits the symmetries of the bivariate normal distribution: C(u, v; ̺) = C(v, u; ̺) = u − C(u, 1 − v; −̺) = v − C(1 − u, v; −̺) = u + v − 1 + C(1 − u, 1 − v; ̺) (2.12) (2.13) (2.14) (2.15) Here, (2.12) is a consequence of exchangeability, and (2.15) is a consequence of radial symmetry, cf. Section 2.7 of (Nelsen 2006). In the following sections we will discuss formulas for the bivariate normal copula, its numerical evaluation, bounds and approximations, measures of concordance, and univariate distributions related to the bivariate normal copula. It will be convenient to assume ̺ 6∈ {−1, 0, 1} unless explicitly stated otherwise (cf. (2.2), (2.4), (2.3) for the simple formulation in the missing cases). 3 Formulas If (X, Y ) are bivariate standard normally distributed with correlation ̺ then Y conditional on X = x is normally distributed with expectation ̺x and variance 1 − ̺2 . This translates into the following formulas for the conditional distributions of the bivariate normal copula: ! ∂ Φ−1 (v) − ̺ · Φ−1 (u) p , (3.1) C(u, v; ̺) = Φ ∂u 1 − ̺2 ! ∂ Φ−1 (u) − ̺ · Φ−1 (v) p (3.2) C(u, v; ̺) = Φ ∂v 1 − ̺2 The copula density is given by: ϕ2 Φ−1 (u), Φ−1 (v); ̺ ∂2 c(u, v; ̺) := C(u, v; ̺) = ∂u ∂v ϕ (Φ−1 (u)) ϕ (Φ−1 (v)) 1 = p exp 1 − ̺2 2̺Φ−1 (u)Φ−1 (v) − ̺2 Φ−1 (u)2 + Φ−1 (v)2 2(1 − ̺2 ) (3.3) ! In terms of its copula density, the bivariate normal copula can be written as Z uZ v c(s, t; ̺) dt ds. (3.4) C(u, v; ̺) = 0 0 4 Now let α, β, γ ∈ (−1, 1) with αβγ = ̺. Then the following holds: C(u, v; ̺) Z ∞Z = −∞ ∞ Φ −∞ Φ−1 (u) − α · x √ 1 − α2 Φ Φ−1 (v) − β · y p 1 − β2 ! ϕ2 (x, y; γ) dy dx (3.5) ! Φ−1 (v) − β · Φ−1 (t) Φ−1 (u) − α · Φ−1 (s) √ p c(s, t; γ) dt ds Φ = Φ 1 − α2 1 − β2 0 0 Z 1Z 1 ∂ ∂ = C(u, s; α) C(t, v; β) c(s, t; γ) dt ds ∂s ∂t 0 0 Z 1 Z 1 The right hand side of (3.5) has occurred in credit risk modelling but without the link to the bivariate normal copula, cf. (Bluhm & Overbeck 2003). A proof in that context will be provided in Section A.1. Other formulas for C(u, v; ̺) are obtained by carefully studying the limit as some of the variables α, β, γ are approaching the value one. The interesting cases, not regarding symmetry, are listed in Table 1. α =α =1 =α =1 =1 β =β =β =β =̺ =1 γ =γ =γ =1 =1 =̺ reference (3.5) (3.6) (3.7) (3.8) (3.4) Table 1: Limiting cases of (3.5) By approaching α = 1 in (3.5) we obtain: ! Φ−1 (v) − β · y p C(u, v; ̺) = Φ ϕ2 (x, y; γ) dy dx 1 − β2 −∞ −∞ ! Z uZ 1 Φ−1 (v) − β · Φ−1 (t) p c(s, t; γ) dt ds = Φ 1 − β2 0 0 Z uZ 1 ∂ = C(u, t; β) c(s, t; γ) dt ds ∂t 0 0 Z Φ−1 (u) Z ∞ (3.6) However, Equation (3.6) may be considered rather unattractive. By approaching 5 γ = 1 in (3.5) instead we obtain a more interesting formula: ! Z ∞ −1 Φ−1 (v) − β · z Φ (u) − α · z √ p ϕ(z) dz (3.7) C(u, v; ̺) = Φ Φ 1 − α2 1 − β2 −∞ ! Z 1 −1 Φ (u) − α · Φ−1 (t) Φ−1 (v) − β · Φ−1 (t) √ p = Φ dt Φ 1 − α2 1 − β2 0 Z 1 ∂ ∂ = C(t, v; α) C(u, t; β) dt ∂t ∂t 0 Equation (3.7) seems to have been discovered repeatedly, sometimes in more general context (e.g., multivariate, cf. (Steck & Owen 1962)), sometimes for special cases. Gupta (1963a) gives credit to Dunnett & Sobel (1955). By approaching α = γ = 1 in (3.5) we obtain: ! Z u Z u ∂ Φ−1 (v) − ̺ · Φ−1 (t) p dt = C(u, v; ̺) = Φ C(t, v; ̺) dt (3.8) 2 ∂t 1−̺ 0 0 ! Z v Z v ∂ Φ−1 (u) − ̺ · Φ−1 (t) p dt = C(u, t; ̺) dt (3.9) = Φ 2 1−̺ 0 ∂t 0 These formulas can also be derived from (3.1) and (3.2). Finally, by approaching α = β = 1 in (3.5) we rediscover (3.4). In order to evaluate the bivariate normal distribution function numerically, Owen (1956) defined the following very useful function, to which we will refer as Owen’s T -function: Z a Z a exp − 12 h2 (1 + x2 ) 1 ϕ(hx) T (h, a) = dx = ϕ(h) dx (3.10) 2 2π 0 1 + x2 0 1+x He proved that u+v − T Φ−1 (u), αu − T Φ−1 (v), αv − δ(u, v) 2 ( 1 , if u < 12 , v ≥ 21 or u ≥ 12 , v < 21 δ(u, v) := 2 0, else C(u, v; ̺) = where (3.11) (3.12) and 1 αu = p 1 − ̺2 Φ−1 (v) −̺ , Φ−1 (u) 1 αv = p 1 − ̺2 In particular, on the lines defined by v = hold: 1 2; ̺ u = −T 2 1 2 Φ−1 (u) −̺ . Φ−1 (v) and by u = v, the following formulas −1 ̺ ! , (u), − p 1 − ̺2 r 1−̺ C(u, u; ̺) = u − 2 · T Φ−1 (u), 1+̺ C u, Φ 6 (3.13) (3.14) (3.15) From (3.11) and (3.14) we can derive the useful formula C(u, v; ̺) = C u, 12 ; ̺u + C v, 21 ; ̺v − δ(u, v), (3.16) where αu ̺u = − p = sin(arctan(−αu )), 1 + α2u αv = sin(arctan(−αv )). ̺v = − p 1 + α2v On the diagonal u = v, (3.16) reads: C(u, u; ̺) = 2 · C u, 1 2; − r 1−̺ 2 Inversion of (3.17) using (2.13) gives: ( 1 C(u, u; 1 − 2̺2 ), 1 C u, 2 ; ̺ = 2 1 u − 2 C(u, u; 1 − 2̺2 ), ! (3.17) ̺ < 0, ̺ > 0. (3.18) Applying (3.8) to (3.17) we obtain, cf. also (Steck & Owen 1962), Z u C(u, u; ̺) = 2 · g(t; ̺) dt (3.19) r (3.20) 0 with g(u; ̺) := Φ We find 1−̺ −1 · Φ (u) . 1+̺ d C(u, u; ̺) = 2 · g(u; ̺). (3.21) du The function g will become important in Section 5. Note that if U , V are uniformly distributed on [0, 1] with copula C(·, ·; ̺) then g(u; ̺) = P(V ≤ u | U = u). Below we list some properties of g: lim g(u; ̺) = 0, (3.22) lim g(u; ̺) = 1, (3.23) g( 12 ; ̺) = 12 , (3.24) u−→0+ u−→1− g(1 − u; ̺) = 1 − g(u; ̺), g(g(u; ̺); −̺) = u (3.25) (3.26) In particular, (3.22) and (3.23) show that the bivariate normal copula does not exhibit (lower or upper) tail dependence (cf. Section 5.2.3 of Embrechts, Frey & McNeil (2005)). Substitution of t = g(s; ̺) in (3.19) and application of (3.26) lead to the identity, cf. also (Steck & Owen 1962): C(u, u; ̺) = 2u · g(u; ̺) − C (g(u; ̺), g(u; ̺); −̺) 7 (3.27) 4 Numerical evaluation The bivariate normal copula has to be evaluated numerically. To my knowledge, in the literature there is no direct approach. Hence for the time being we have to rely on the definition of the bivariate normal copula (2.1) and numerical evaluation of Φ2 and Φ−1 . There are excellent algorithms available for evaluation of Φ−1 , cf. (Acklam 2004); the main problem is evaluation of the bivariate normal distribution function Φ2 . In the literature on evaluation of Φ2 there are basically two approaches: application of a multivariate method to the bivariate case, and explicit consideration of the bivariate case. For background on multivariate methods we refer to the recent book by Bretz & Genz (2009). In most cases, bivariate methods will be able to obtain the desired accuracy in less time. In the following we will provide an overview on the literature. We will concentrate on methods and omit references dealing with implementation only. Comparisons of different approaches in terms of accuracy and running time have been provided by numerous authors, e.g., Aǧca & Chance (2003), Terza & Welland (1991), and Wang & Kennedy (1990). Before the advent of widely available computer power, extensive tables of the bivariate normal distribution function had to be created. Using (3.16) or similar approaches, the three-dimensional problem (two variables and the correlation parameter) was reduced to a two-dimensional one. Pearson (1901a) used the tetrachoric expansion (2.11) for small |̺|, and quadrature for large |̺|. Nicholson (1943), building on ideas of Sheppard (1900), worked with a two-parameter function, denoted V -function. Owen (1956) introduced the T -function (3.10) which is closely related to Nicholson’s V -function. For many years, quadrature of the T -function was the method of choice for evaluation of the bivariate normal distribution. Numerous authors, e.g. Borth (1973), Daley (1974), Young & Minder (1974), and Patefield & Tandy (2000), have been working on improvements, e.g. by dividing the plane into many regions and choosing specific quadrature methods in each region. Sowden & Ashford (1969) applied Gauss-Hermite quadrature to (3.7) and Simpson’s rule to (3.8). Drezner (1978) used (3.16) and Gauss quadrature. Divgi (1979) relied on polar coordinates and an approximation to the univariate Mills’ ratio. Vasicek (1998) proposed an expansion which is more suitable for large |̺| than the tetrachoric expansion (2.11). Drezner & Wesolowsky (1990) applied Gauss-Legendre quadrature to (2.8) for |̺| ≤ 0.8, and to (2.9) for |̺| > 0.8. Improvements of their method in terms of accuracy and robustness have been provided by Genz (2004) and West (2005). Most implementations today will rely on variants of the approaches of Divgi (1979) or of Drezner & Wesolowsky (1990). The method of Drezner (1978), although less reliable, is also very common, mainly because it is proposed in (Hull 2008) and other prevalent books. 8 5 Bounds and approximations Nowadays, high-precision numerical evaluation of the bivariate normal copula is usually available and there is not much need for low-precision approximations anymore, in particular if the mathematics is hidden behind strange constants derived from some optimization procedure. On the other hand, if the mathematics is general and transparent then the resulting approximations are often rather weak. As an example, take approximations to the multivariate Mills’ ratios applied to the bivariate case, cf. (Lu & Li 2009) and the references therein. Bounds, if they are not too weak, are more interesting than approximations because they can be used, e.g., for checking numerical algorithms. Moreover, derivation of bounds often provides valuable insight into the mathematics behind the function to be bounded. In the following we concentrate on bounds and approximations explicitly derived for the bivariate case. Throughout this section we will only consider the case ̺ > 0, 0 < u = v < 1/2. Note that by successively applying, if required, (3.16), (3.18), (3.27) and (2.15), we can always reduce C(u, v; ̺) to a sum of two terms of that form. Any approximation or bound given for the special case can be translated to an approximation or bound for the general case, with at most twice the absolute error. Note also that for many existing approximations and bounds the diagonal u = v may be considered a worst case, cf. (Willink 2004). Mee & Owen (1983) elaborated on the so-called conditional approach proposed by Pearson (1901b). If (X, Y ) are bivariate standard normally distributed with correlation ̺ then we can write Φ2 (h, k; ̺) = Φ(h) · P(Y ≤ k | X ≤ h). (5.1) The distribution of Y conditional on X = h is normal but the distribution of Y conditional on X ≤ h is not. Nevertheless, it can be approximated by a normal distribution with the same mean and variance. In terms of the bivariate normal copula, the resulting approximation is ! u · Φ−1 (u) + ̺ · ϕ Φ−1 (u) C(u, u; ̺) ≈ u · Φ p . (5.2) u2 − ̺2 · ϕ (Φ−1 (u)) · (u · Φ−1 (u) + ϕ (Φ−1 (u))) The approximation works well for |̺| not too large. For |̺| large there are alternative approximations, e.g. (Albers & Kallenberg 1994). The simpler of the two approximations proposed by Cox & Wermuth (1991) replaces the second factor in (5.1) by the mean of the conditional distribution (the more complicated approximation adds a term of second order). In terms of the bivariate normal copula, the resulting approximation is ! r 1 − ̺ u · Φ−1 (u) + ̺ · ϕ Φ−1 (u) . · C(u, u; ̺) ≈ u · Φ 1+̺ (1 + ̺) · u 9 Mallows (1959) gave two approximations to Owen’s T -function (3.10). In terms of the bivariate normal copula, the simpler one reads r 1−̺ 3 u 1 C(u, u; ̺) ≈ 2u · Φ − Φ−1 . · Φ−1 + 1+̺ 2 4 4 Further approximations were derived by Cadwell (1951) and Pólya (1949). There are not too many bounds available in the literature. For ̺ > 0 there are, of course, the trivial bounds (2.2) and (2.3): u2 ≤ C(u, u; ̺) ≤ u The upper bound given by Pólya (1949) is just the one above. His lower bound is too weak (even negative) on the diagonal. A recent overview on known bounds, and derivation of some new ones, is provided by Willink (2004). We will present some of his bounds below in more general context. Theorem 5.1 Let ̺ ≥ 0 and 0 ≤ u ≤ 1/2. Then C(u, u; ̺) is bounded as follows, where g(u; ̺) is defined as in (3.20): C(u, u; ̺) ≥ u · g(u; ̺) (5.3) C(u, u; ̺) ≤ u · g(u; ̺) · 2 (5.4) The lower bound (5.3) is tight for ̺ = 0 or u = 0. The maximum error of (5.3) equals 1/4 and is obtained for u = 1/2, ̺ = 1. The upper bound (5.4) is tight for ̺ = 1 or u = 0. The maximum error of (5.4) equals 1/4 and is obtained for u = 1/2, ̺ = 0. A proof of Theorem 5.1 is provided in Section A.2, together with a proof of the following refinement: Theorem 5.2 Let ̺ > 0 and 0 ≤ u ≤ 1/2. Then C(u, u; ̺) is bounded as follows, where g(u; ̺) is defined as in (3.20): 2 (5.5) C(u, u; ̺) ≥ u · g(u; ̺) · 1 + arcsin(̺) π C(u, u; ̺) ≤ u · g(u; ̺) · (1 + ̺) (5.6) These bounds are the optimal ones of the form u · g(u) · a(̺). They are tight for ̺ =q 0, ̺ = 1, or u = 0. The maximum error of (5.6) is obtained for u = 1/2, ̺ = 1 − π42 ≈ 0.7712, the value being 1 4 r 4 2 1 − 2 − arcsin π π r 4 1− 2 π The lower bound (5.5) is tight for u = 1/2. 10 !! ≈ 0.05263. The bounds (5.3) and (5.6) have been discussed, without explicit computation of the maximum error, by Willink (2004). The maximum error of (5.5) is difficult to grasp analytically. Numerically, the error always stays below 0.006. An alternative upper bound is given by the following theorem, the proof of which is provided in Section A.3: Theorem 5.3 Let ̺ > 0 and 0 ≤ u ≤ 1/2. Then u C(u, u; ̺) ≤ 2u · g ;̺ . 2 The bound is tight for ̺ = 0, ̺ = 1, or u = 0. The maximum error is obtained for u = 1/2, r 1−̺ 1+̺ −1 1 − =ϕ · Φ (4) , 1+̺ 2π · Φ−1 ( 14 ) i.e., ̺ ≈ 0.5961, the value being approx. 0.015. It is also possible to derive good approximations to C(u, u; ̺) by considering the family C(u, u; ̺) ≈ u · g(u; ̺) · (a(̺) + b(̺)u). In particular, the choice 2 4 a(̺) := 1 + ̺, b(̺) := 2 · 1 + arcsin(̺) − (1 + ̺) = arcsin(̺) − 2̺ π π is attractive because the resulting approximation 4 arcsin(̺) − 2̺ u C(u, u; ̺) ≈ u · g(u; ̺) · 1 + ̺ + π (5.7) is tight for ̺ = 0, ̺ = 1, u = 0, or u = 1/2, and for u −→ 0+ it has the same asymptotic behaviour as (5.6). By visual inspection we may conjecture that it is even an upper bound, with an error almost cancelling the error of the lower bound (5.5) most of the time. Consequently, an even better approximation (but not a lower bound, for ̺ large) is given by ̺ 2 C(u, u; ̺) ≈ u · g(u; ̺) · 1 + + arcsin(̺) − ̺ u . (5.8) 2 π Again, (5.8) is tight for ̺ = 0, ̺ = 1, u = 0, or u = 1/2. Numerically, the absolute error always stays below 0.0006. Hence (5.8) is comparable in performance with (5.2), and much better for ̺ large. 6 Measures of concordance In the study of dependence between (two) random variables, properties and measures that are scale-invariant, i.e., invariant under strictly increasing transformations of the random variables, can be expressed in terms of the (bivariate) copula of the random variables. Among these are the so-called measures of 11 concordance, in particular Kendall’s tau, Spearman’s rho, Blomqvist’s beta and Gini’s gamma. For background and general definitions and properties we refer to Section 5 of (Nelsen 2006). In this section we will provide formulas for measures of concordance for the bivariate normal copula, depending on the correlation parameter ̺. Blomqvist’s beta follows immediately from (2.10): β(̺) := 4 · C 12 , 21 ; ̺ − 1 2 = arcsin(̺) (6.1) π For the bivariate normal copula, Kendall’s tau equals Blomqvist’s beta: Z 1Z 1 τ (̺) := 4 C(u, v; ̺) dC(u, v; ̺) − 1 0 0 Z 1Z 1 ∂ ∂ C(u, v; ̺) C(u, v; ̺) du dv =1−4 ∂v ∂u 0 0 2 = arcsin(̺) π (6.2) For a proof of (6.1) and (6.2) cf. Section 5.3.2 of (Embrechts et al. 2005). Both Blomqvist’s beta and Kendall’s tau can be generalized to (copulas of) elliptical distributions, cf. (Lindskog, McNeil & Schmock 2003). This is not the case for Spearman’s rho, cf. (Hult & Lindskog 2002), which is given by: Z 1Z 1 C(u, v; ̺) − uv du dv ̺S (̺) := 12 0 = 12 Z 0 = 0 1 Z 0 1 C(u, v; ̺) du dv − 3 ̺ 6 arcsin π 2 (6.3) For proofs of (6.3) cf. (Kruskal 1958) or Section 5.3.2 of (Embrechts et al. 2005). Gini’s gamma for the bivariate normal copula is given as follows: Z 1 Z 1 1 γ(̺) := 4 C(u, 1 − u; ̺) du − C(u, u; ̺) du + 2 0 0 Z 1 Z 1 1 =4 u − C(u, u; −̺) du − C(u, u; ̺) du + 2 0 0 2 1+̺ 1−̺ = arcsin − arcsin (6.4) π 2 2 √ √ 1+̺ 1−̺ 4 = arcsin − arcsin (6.5) π 2 2 p 1 p 4 = arcsin (1 + ̺)(3 + ̺) − (1 − ̺)(3 − ̺) (6.6) π 4 12 I have not been able to find proofs in the literature, hence they will be provided in Section A.4. Equation (6.6) can be inverted which may be useful for estimation of ̺ from an estimate for γ(̺): r π π ̺ = sin γ(̺) · 3 − tan γ(̺) · (6.7) 4 4 Finally, we propose a new measure of concordance, similar to Gini’s gamma but based on the lines u = 1/2, v = 1/2 instead of the diagonals. For a bivariate copula C it is defined by: Z 1 Z 1 v u (6.8) C 21 , v − dv γ̃(C(·, ·)) := 4 C u, 21 − du + 2 2 0 0 Z 1 Z 1 1 C 21 , v dv − =4 C u, 21 du + 2 0 0 For the bivariate normal copula we obtain γ̃(̺) := γ̃(C(·, ·; ̺)) = 4 arcsin π ̺ √ 2 . (6.9) A proof is given implicitly in Section A.4. 7 Univariate distributions In this section we will discuss two univariate distributions being closely related to the bivariate normal copula (or distribution). 7.1 The skew-normal distribution A random variable X on R is skew-normally distributed with skewness parameter λ ∈ R if it has a density function of the form fλ (x) = 2ϕ(x)Φ(λx). (7.1) The skew-normal distribution was introduced by O’Hagan & Leonard (1976) and studied and made popular by Azzalini (1985, 1986). Its cumulative distribution function is given by P(X ≤ x) = Z x 2ϕ(x)Φ(λx) = 2 Z 0 −∞ 13 Φ(x) Φ λΦ−1 (t) dt. (7.2) In the light of (3.19) and (3.17), cf. also (Azzalini & Capitanio 2003), we find λ (7.3) P(X ≤ x) = 2Φ2 x, 0; − √ 1 + λ2 1 − λ2 , λ ≥ 0, Φ2 x, x; 2 1+λ = (7.4) 1 − λ2 , λ ≤ 0. 1 − Φ2 −x, −x; 1 + λ2 In particular, the bounds given in Theorem 5.2 can be applied. 7.2 The Vasicek distribution A random variable P on the interval [0, 1] is Vasicek distributed with parameters p ∈ (0, 1) and ̺ ∈ (0, 1) if Φ−1 (P ) is normally distributed with mean Φ−1 (p) E(Φ−1 (P )) = √ 1−̺ (7.5) ̺ . 1−̺ (7.6) and variance V(Φ−1 (P )) = In Section A.1 implicitly it is proved that E(P ) = p, E(P 2 ) = C(p, p; ̺), so that V(P ) = E(P 2 ) − E(P )2 = C(p, p; ̺) − p2 . Furthermore, we have P(P ≤ q) = P(Φ −1 (P ) ≤ Φ −1 √ 1 − ̺ · Φ−1 (q) − Φ−1 (p) . (q)) = Φ √ ̺ The (one-sided) α-Quantile qα of P , with α ∈ (0, 1), is therefore given by √ ̺ · Φ−1 (α) + Φ−1 (p) √ qα = Φ . (7.7) 1−̺ In particular, the median of P is simply −1 Φ (p) = Φ E(Φ−1 (P )) . q0.5 = Φ √ 1−̺ The density of P is d P(P ≤ q) = dq r 1−̺ ·ϕ ̺ √ 1 − ̺ · Φ−1 (q) − Φ−1 (p) 1 · . √ ̺ ϕ (Φ−1 (q)) 14 (7.8) The distribution is unimodal with the mode at √ 1−̺ Φ · Φ−1 (p) 1 − 2̺ for ̺ < 0.5, monotone for ̺ = 0.5, and U-shaped for ̺ > 0.5. Let P̃ be Vasicek distributed with parameters p̃, ̺˜, and let corr Φ−1 (P ), Φ−1 (P̃ ) = γ. Then and s ̺˜ ̺ · , 1 − ̺ 1 − ̺˜ p E P · P̃ = C p, p̃; γ · ̺ · ̺˜ , cov Φ−1 (P ), Φ−1 (P̃ ) = γ · p cov P, P̃ = C p, p̃; γ · ̺ · ̺˜ − p · p̃. The Vasicek distribution does not offer immediate advantages over other two-parametric continuous distributions on (0, 1), such as the beta distribution. Its importance stems from its occurrence as mixing distribution in linear factor models set up as in Section A.1. It is a special case of a probit-normal distribution; it is named after Vasicek who introduced it into credit risk modeling. For (different) details on the material in this section we refer to Vasicek (1987, 2002) and Tasche (2008). Estimation of the parameters p and ̺ is also discussed by Meyer (2009). References Acklam, P. (2004), An algorithm for computing the inverse normal cumulative distribution function. http://home.online.no/∼pjacklam/notes/invnorm/index.html Albers, W., Kallenberg, W.C.M. (1994), A simple approximation to the bivariate normal distribution with large correlation coefficient, Journal of Multivariate Analysis 49(1), pp. 87–96. Arnold, B.C., Lin, G.D. (2004), Characterizations of the skew-normal and generalized chi distributions, Sankhyā : The Indian Journal of Statistics 66(4), pp. 593–606. Aǧca, Ş., Chance, D.M. (2003), Speed and accuracy comparison of bivariate normal distribution approximations for option pricing, Journal of Computational Finance 6(4), pp 61–96. Azzalini, A. (1985), A class of distributions which includes the normal ones, Scandinavian Journal of Statistics 12(2), pp. 171–178. 15 Azzalini, A. (1986), Further results on a class of distributions which includes the normal ones, Statistica 46(2), pp. 199–208. Azzalini, A., Capitanio, A. (2003), Distributions generated by perturbation of symmetry with emphasis on a multivariate skew t distribution, Journal if the Royal Statistical Society Series B 65, pp. 367–389. Balakrishnan, N., Lai, C.D. (2009), Continuous Bivariate Distributions, 2nd ed., Springer. Baricz, Á. (2008), Mills’ ratio: Monotonicity patterns and functional inequalities, Journal of Mathematical Analysis and Applications 340(2), pp. 1362– 1370. Bluhm, C., Overbeck, L. (2003), Estimating systematic risk in uniform credit portfolios, in: Credit Risk; Measurement, Evaluation and Management, ed. G. Bol et al., Contributions to Economics, Physica-Verlag Heidelberg, pp. 35–48. Borth, D.M. (1973), A modification of Owen’s method for computing the bivariate normal integral, Journal of the Royal Statistical Society, Series C (Applied Statistics) 22(1), pp. 82–85. Bretz, F., Genz, A. (2009), Computation of Multivariate Normal and t Probabilities, Lecture Notes in Statistics 195, Springer. Cadwell, J.H. (1951), The bivariate normal integral, Biometrika 38(3-4), pp. 475–479. Cox, D.R., Wermuth, N. (1991), A simple approximation for bivariate and trivariate normal integrals, International Statistical Review 59(2), pp. 263– 269. Daley, D.J. (1974), Computation of bi- and tri-variate normal integrals, Journal of the Royal Statistical Society, Series C (Applied Statistics) 23(3), pp. 435–438. Divgi, D.R. (1979), Calculation of univariate and bivariate normal probability functions, Annals of Statistics 7(4), pp. 903–910. Drezner, Z. (1978), Computation of the bivariate normal integral, Mathematics of Computation 32(141), pp. 277–279. Drezner, Z., Wesolowsky, G.O. (1990), On the computation of the bivariate normal integral, Journal of Statistical Computation and Simulation 35, pp. 101–107. Dunnett, C.W., Sobel, M. (1955), Approximations to the probability integral and certain percentage points of a multivariate analogue of Student’s tdistribution, Biometrika 42(1-2), pp. 258–260. 16 Embrechts, P., Frey, R., McNeil, A.J. (2005), Quantitative Risk Management: Concepts, Techniques, Tools, Princeton University Press. Genz, A. (2004), Numerical computation of rectangular bivariate and trivariate normal and t probabilities, Statistics and Computing 14(3), pp. 151–160. Gupta, S. (1963a), Probability integrals of multivariate normal and multivariate t, Annals of Mathematical Statistics 34(3), pp. 792–828. Gupta, S. (1963b), Bibliography on the multivariate normal integrals and related topics, Annals of Mathematical Statistics 34(3), pp. 829–838. Hull, J. (2008), Futures, Options, and Other Derivatives, 7th ed., Prentice Hall. Hult, H., Lindskog, F. (2002), Multivariate extremes, aggregation and dependence in elliptical distributions, Advances in Applied Probability 34(3), pp. 587–608. Kotz, S., Balakrishnan, N., Johnson, N.L. (2000), Continuous Multivariate Distributions, Volume 1: Models and Applications, 2nd ed., Wiley Series in Probability and Statistics. Kruskal, W.H. (1958), Ordinal measures of association, Journal of the American Statistical Association 53(284), pp. 814–861. Lindskog, F., McNeil, A.J., Schmock, U., Kendall’s tau for elliptical distributions, in: Credit Risk; Measurement, Evaluation and Management, ed. G. Bol et al., Contributions to Economics, Physica-Verlag Heidelberg, pp. 149–156. Lu, D., Li., W.V. (2009), A note on multivariate Gaussian estimates, Journal of Mathematical Analysis and Applications 354, pp. 704–707. Mallows, C.L. (1959), An approximate formula for bivariate normal probabilities, Technical Report No. 30, Statistical Techniques Research Group, Princeton University. Mee, R.W., Owen, D.B. (1983), A simple approximation for bivariate normal probability, Journal of Quality Technology 15, pp. 72–75. Mehler, G. (1866), Reihenentwicklungen nach Laplaceschen Functionen höherer Ordnung, Journal für die reine und angewandte Mathematik 66, pp. 161– 176. Meyer, C. (2009), Estimation of intra-sector asset correlations, Journal of Risk Model Validation 3(4), pp. 47–79. Nelsen, R.B. (2006), An Introduction to Copulas, 2nd ed., Springer. Nicholson, C. (1943), The probability integral for two variables, Biometrika 33(1), pp. 59–72. 17 O’Hagan, A., Leonard, T. (1976), Bayes estimation subject to uncertainty about parameter constraints, Biometrika 63(1), pp. 201–203. Owen, D.B. (1956), Tables for computing bivariate normal probability, Annals of Mathematical Statistics 27, pp. 1075–1090. Patefield, M., Tandy, D. (2000), Fast and accurate computation of Owen’s T function, Journal of Statistical Software 5(5). Patel, J.K, Read, C.B. (1996), Handbook of the Normal Distribution, Dekker. Pearson, K. (1901a), Mathematical contributions to the theory of evolution. VII. On the correlation of characters not quantitatively measurable, Philosophical Transactions of the Royal Society of London Series A 195, pp. 1–47. Pearson, K. (1901b), Mathematical contributions to the theory of evolution. XI. On the influence of natural selection on the variability and correlation of organs, Philosophical Transactions of the Royal Society of London Series A 200, pp. 1–66. Pinelis, I. (2002), Monotonicity properties of the relative error of a Padé approximation for Mills’ ratio, Journal of Inequalities in Pure and Applied Mathematics 3(2). Plackett, R.L. (1954), A reduction formula for normal multivariate integrals, Biometrika 41(3), pp. 351–360. Pólya, G. (1949), Remarks on computing the probability integral in one and two dimensions, Proceedings of the First Berkeley Symposium on Mathematical Statistics and Probability, Univ. of California Press, pp. 63–78. Sheppard, W.F. (1898), On the application of the theory of error to cases of normal distributions and normal correlation, Philosophical Transactions of the Royal Society of London Series A 192, pp. 101–167. Sheppard, W.F. (1900), On the calculation of the double integral expressing normal correlation, Transactions of the Cambridge Philosophical Society 19, pp. 23–69. Sowden, R.R., Ashford, J.R. (1969), Computation of the bivariate normal integral, Journal of the Royal Statistical Society, Series C (Applied Statistics) 18(2), pp. 169–180. Steck, G.P., Owen, D.B. (1962), A note on the equicorrelated multivariate normal distribution, Biometrika 49(1-2), pp. 269–271. Stieltjes, T.S. (1889), Extrait d’une lettre adressée à M. Hermite, Bulletin des Sciences Mathématiques Series 2 13, p. 170. Tasche, D. (2008), The Vasicek distribution. http://www-m4.ma.tum.de/pers/tasche/ 18 Terza, J.V., Welland, U. (1991), A comparison of bivariate normal algorithms, Journal of Statistical Computation and Simulation 39(1-2), pp. 115–127. Vasicek, O. (1987), Probability of loss on loan portfolio. http://www.moodyskmv.com/research/portfolioCreditRisk wp.html Vasicek, O. (1998), A series expansion for the bivariate normal integral, The Journal of Computational Finance 1(4), pp. 5–10. Vasicek, O. (2002), The distribution of loan portfolio value, RISK 15(12), pp. 160–162. Wang, M., Kennedy, W.J. (1990), Comparison of algorithms for bivariate normal probability over a rectangle based on self-validated results from interval analysis, Journal of Statistical Computation and Simulation 37(1-2), pp. 13–25. West, G. (2005), Better approximations to cumulative normal functions, Wilmott Magazine, May, pp. 70–76. Willink, R. (2004), Bounds on the bivariate normal distribution function, Communications in Statistics; Theory and Methods 33(10), pp. 2281–2297. Young, J.C., Minder, Ch.E. (1974), An integral useful in calculating non-central t and bivariate normal probabilities, Journal of the Royal Statistical Society, Series C (Applied Statistics) 23(3), pp. 455–457. A A.1 Proofs Proof of (3.5) Let p 1 − α2 · ǫ, p X̃ = β · Ỹ + 1 − β 2 · ǫ̃, X =α·Y + where α, β ∈ (−1, 1) \ {0} are parameters and where Y , Ỹ , ǫ, ǫ̃ are all standard normal and pairwise independent, except γ := corr Y, Ỹ = cov Y, Ỹ . By construction, X and X̃ are standard normal again with corr X, X̃ = cov X, X̃ = αβγ. We define indicator variables Z = Z(X) ∈ {0, 1}, Z̃ = Z̃(X̃) ∈ {0, 1} calibrated to expectation values u, v: Z=1 :⇐⇒ Z̃ = 1 :⇐⇒ 19 X ≤ Φ−1 (u), X̃ ≤ Φ−1 (v) Conditional on (Y = y, Ỹ = ỹ), Z and Z̃ are independent. We find P(Z = 1 | Y = y) = P X ≤ Φ−1 (u) | Y = y p = P α · y + 1 − α2 · ǫ ≤ Φ−1 (u) −1 Φ (u) − α · y Φ−1 (u) − α · y √ √ =Φ . =P ǫ≤ 1 − α2 1 − α2 Now we define the random variables P := P (Y ) := Φ Φ−1 (u) − α · Y √ 1 − α2 Φ−1 (v) − β · Y p 1 − β2 P̃ := P̃ (Ỹ ) := Φ We find ! , . u = E(Z) = E(E(Z | Y )) = E(P(Z = 1 | Y )) = E(P ), v = E(P̃ ) and P Z = 1, Z̃ = 1 = P X ≤ Φ−1 (u), X̃ ≤ Φ−1 (v) = Φ2 Φ−1 (u), Φ−1 (v), cov(X, X̃) = Φ2 Φ−1 (u), Φ−1 (v), αβγ . On the other hand, P Z = 1, Z̃ = 1 = P Z Z̃ = 1 = E Z Z̃ = E E(Z Z̃ | Y, Ỹ ) = E E(Z | Y, Ỹ ) · E(Z̃ | Y, Ỹ ) = E E(Z | Y ) · E(Z̃ | Ỹ ) = E P · P̃ Z ∞Z ∞ = P (x) · P̃ (y) · ϕ2 (x, y, γ) dx dy −∞ = Z ∞ −∞ A.2 −∞ Z ∞ −∞ Φ Φ−1 (u) − α · x √ 1 − α2 Φ Φ−1 (v) − β · y p 1 − β2 ! Proof of Theorems 5.1 and 5.2 We will assume ̺ as fixed and write C(u) := C(u, u; ̺), 20 g(u) := g(u; ̺). ϕ2 (x, y, γ) dx dy. The upper bound (5.4) follows from (3.27). Regarding the lower bound (5.3) we note that r 1−̺ ̺ ′ g (u) = · exp · Φ−1 (u)2 > 0, 1+̺ 1+̺ 2̺ Φ−1 (u) g ′′ (u) = g ′ (u) · · < 0. 1 + ̺ ϕ(Φ−1 (u)) Hence g is increasing and concave on (0, 1/2) and we conclude Z u 1 C(u) = 2 g(t) dt ≥ 2 · · u · g(u) = u · g(u). 2 0 Now we define Da (u) := u · g(u) · a with a ∈ [1, 2], u ∈ [0, 1/2]. We start by noting that C ′ (u) = 2g(u) > 0, C ′′ (u) = 2g ′ (u) > 0, hence C is increasing and convex on (0, 1/2). Furthermore, Da′ (u) = a · (g(u) + u · g ′ (u)) > 0, Da′′ (u) = a · (2 · g ′ (u) + u · g ′′ (u)) 1 u · Φ−1 (u) = 2a · g ′ (u) · 1 + · 1 + ̺ ϕ (Φ−1 (u)) 1 · H −Φ−1 (u) = 2a · g ′ (u) · 1 − 1+̺ with H(x) = x · R(x), where 1 − Φ(x) ϕ(x) R(x) = is Mills’ ratio. Pinelis (2002) has shown that H ′ (x) > 0 for x > 0, H(0) = 0, and limx−→∞ H(x) = 1. Hence Da (u) is increasing and convex on (0, 1/2) as well. For u ∈ (0, 1/2), C ′ (u) = Da′ (u) is equivalent with f (u) := 2−a u · g ′ (u) = , g(u) a or a= 2 . 1 + f (u) We will show that f is strictly increasing on (0, 1/2). We have Φ Φ−1 (u) · ϕ λ · Φ−1 (u) f (u) = λ · = λ · Fλ −Φ−1 (u) −1 −1 ϕ (Φ (u)) · Φ (λ · Φ (u)) q with λ = 1−̺ 1+̺ ∈ [0, 1] and Fλ (x) := R(x) , R(λ · x) 21 x ≥ 0. We find R′ (x) · R(λ · x) − R(x) · R′ (λ · x) R(λ · x)2 ′ R (x) R′ (λ · x) = Fλ (x) · <0 −λ· R(x) R(λ · x) Fλ′ (x) = for λ < 1. Here we have used that Fλ (x) > 0 and that the function y 7→ y · R′ (y) R(y) is strictly decreasing on (0, ∞), cf. (Baricz 2008). We conclude Fλ′ −Φ−1 (u) ′ f (u) = −λ · > 0. ϕ (−Φ−1 (u)) Furthermore, we find f and 1 2 R(0) = =λ· R(0) r 1−̺ 1+̺ g ′ (u) + u · g ′′ (u) g ′ (u) u−→0+ u · Φ−1 (u) 2̺ · = lim + 1 + 1 + ̺ ϕ(Φ−1 (u)) u−→0 2̺ · H −Φ−1 (u) = lim 1 − 1+̺ u−→0+ 2̺ 1−̺ =1− = . 1+̺ 1+̺ lim f (u) = lim u−→0+ We have C(0) = Da (0) = 0, C ′ (0) = Da′ (0) = 0, and lim + u−→0 D′ (u) a a Da (u) = lim + a′ = lim + (1 + f (u)) = . C(u) 1+̺ u−→0 C (u) u−→0 2 By standard calculus we conclude that • For a ≥ 1 + ̺ we have Da′ (u) ≥ C ′ (u), and hence Da (u) ≥ C(u), for all u ∈ [0, 1/2]; q −1 • For a ≤ 2 · 1 + 1−̺ we have Da′ (u) ≤ C ′ (u), and hence Da (u) ≤ 1+̺ C(u), for all u ∈ [0, 1/2]; • For a∈ ! r −1 1−̺ ,1 + ̺ 2· 1+ 1+̺ 22 there exists u0 ∈ (0, 1/2) with Da′ (u) < C ′ (u) for u ∈ (0, u0 ), and Da′ (u) > C ′ (u) for u ∈ (u0 , 1/2). Consequently, the best lower bound for C of the form Da is obtained if C(1/2) = Da (1/2), i.e., a = 1 + π2 arcsin(̺). Moreover, the upper bound Da with a = 1 + ̺ can not be improved. The maximum error of Da with a = 1 + ̺ is attained if d 1 1 = 0, [Da (1/2) − C(1/2)] = − p d̺ 4 2π 1 − ̺2 q which is equivalent with ̺ = 1 − π42 . A.3 Proof of Theorem 5.3 We will assume ̺ as fixed and write C(u) := C(u, u; ̺), We have C(u) = 2 · with Z 0 g(u) := g(u; ̺). u g(t) dt = 2u · g(v(u)) v(u) := v(u; ̺) ≤ u. Since g is concave and increasing, we even know that u v(u) ≤ 2 and hence u . C(u) = 2u · g(v(u)) ≤ 2u · g 2 Moreover, for the same reason we have u u u u d · g′ 2u · g − C(u) = 2 − g(u) − g ≥ 0, du 2 2 2 2 and hence, for ̺ fixed, the maximum error is obtained for u = 1/2, the value being r 1−̺ 1 1 1 − − · Φ−1 arcsin(̺). Φ 1+̺ 4 4 2π Derivation of the above expression with respect to ̺ gives the result. Note that by (3.26) we can write C(u) ; −̺ v(u; ̺) = g 2u Unfortunately, for large ̺, the function v is not convex, and the approximation u C( 1 ) 2u · g · 12 2 g( 4 ) is not an upper bound for C(u). 23 A.4 Proof of (6.4), (6.5), (6.6) In a first step, using (2.8) we find: Z 1 Z 1 Z ̺ Φ−1 (u)2 1 1 2 √ dr du exp − C(u, u; ̺) du = u + 2π 0 1+r 1 − r2 0 0 Z ̺ Z ∞ 1 1 v2 1 √ = + dv dr ϕ(v) exp − 3 2π 0 1+r 1 − r2 −∞ r Z ̺ Z ∞ 1 1 1 1+r √ ds dr = + ϕ(s) 3 2π 0 3+r 1 − r2 −∞ Z ̺ 1 1 1 p = + dr 3 2π 0 (1 − r)(3 + r) Z 1+̺ 2 1 1 1 √ dr = + 3 2π 12 1 − r2 1+̺ 1 1 1 arcsin − arcsin = + 3 2π 2 2 1 1 1+̺ = + . arcsin 4 2π 2 We conclude that Z 1 Z 1 1 γ(̺) = 4 u − C(u, u; −̺) du − C(u, u; ̺) du + 2 0 0 1 1 1 1 1 1+̺ 1−̺ 1 + − − − + arcsin arcsin =4 4 2π 2 2 4 2π 2 2 2 1+̺ 1−̺ = arcsin − arcsin . π 2 2 In a similar way, using again (2.8), we can compute Z 1 Z ̺ Z 1 Φ−1 (u)2 1 1 u √ exp − C u, 21 ; ̺ du = dr du + 2π 0 2(1 − r2 ) 1 − r2 0 0 2 1 ̺ 1 , arcsin √ = + 4 2π 2 which, using (3.17), leads to √ Z 1 Z 1 q 1 1 1−̺ C(u, u; ̺) du = 2 · C u, 21 ; − 1−̺ . − arcsin du = 2 2 π 2 0 0 We obtain alternative formulas for Gini’s gamma, the second one using the addition theorem for the arcsin function: √ √ 4 1+̺ 1−̺ γ(̺) = arcsin − arcsin π 2 2 p p 1 4 = arcsin (1 + ̺)(3 + ̺) − (1 − ̺)(3 − ̺) π 4 24