Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
The multivariate normal distribution Characterizing properties of the univariate normal distribution Different definitions of normal random vectors Conditional distributions Independence Cochran’s theorem Probability theory 2011 The univariate normal distribution - defining properties A distribution is normal if and only if it has the probability density 1 ( x )2 f X ( x) exp( ) 2 2 2 where R and > 0. A distribution is normal if and only if the sample mean 1 n X Xi n i 1 and the sample variance 1 n s ( X i X )2 n 1 i 1 are independent for all n. 2 Probability theory 2011 The univariate normal distribution - defining properties Suppose that X1 and X2 are independent of each other, and that the same is true for the pair Y1 a11 a12 X 1 Y a a 2 21 22 X 2 where no coefficient vanishes. Then all four variables are normal. Corollary: A two-dimensional random vector that preserves independence under rotation must be normal x2 x1 Probability theory 2011 The univariate normal distribution - defining properties Let F be a class of distributions such that X F a + bX F Can F be comprised of distributions other than the normal distributions? cf. Cauchy distributions Probability theory 2011 The univariate normal distribution - defining properties X C m, 1 f X x 2 ; x 2 x m X t E eitX eimt t Put Y a bX Y t E eitY E eit a bX eita E eitbX eita X bt eitae imbt bt e i ( a bm) t b t Y C a bm, b Probability theory 2011 The multivariate normal distribution - a first definition A random vector is normal if and only if every linear combination of its components is normal Immediate consequences: Every component is normal The sum of all components is normal Every marginal distribution is normal Vectors in which the components are independent normal random variables are normal Linear transformations of normal random vectors give rise to new normal vectors Probability theory 2011 The multivariate normal distribution - a first definition X i 0,,0, 1 ,0,,0 X ith element Every component is normal The sum of all components is normal Every marginal distribution is normal X 1,,1 X 1 n 1 ' n i a X i1 ,i2 ,,ik i i X 1' i1 ,i2 ,,ik X X i1 ,, X ik ' is normal Vectors in which the components are independent normal random n n n X 1 ,, X n i.i.d. N i , i2 1 ai X i N 1 ai i , 1 ai2 i2 variables are normal ' X 1 ,, X n is normal Linear transformations of normal random vectors give rise to new normal vectors Probability theory 2011 Illustrations of independent and dependent normal distributions Probability theory 2011 Illustrations of independent and dependent normal distributions Probability theory 2011 Parameterization of the multivariate normal distribution Is a multivariate normal distribution uniquely determined by the vector of expected values and the covariance matrix? Is there a multivariate normal distribution for any covariance matrix? Probability theory 2011 Fundamental properties of covariance matrices Let be a covariance matrix of a random vector X Then is symmetric Moreover, is nonnegative-definite, i.e. a1 . a1, ..., a p Λ . Var a j X j 0 j . a p Probability theory 2011 Factorization of covariance matrices Let be a covariance matrix. Because is symmetric there exists an orthogonal matrix C (C’C = C C’ = I) such that C’ C = D and = CD C’ where D is a diagonal matrix. Beacuse is also nonnegative-definite, the diagonal elements of D must be non-negative. Consequently, there exists a symmetric matrix B such that BB= B is often called the square root of Probability theory 2011 Construction of a random vector with a given covariance matrix Let be a covariance matrix. Derive a matrix B such that B B’ = If X has independent components with variance 1, then Y = BX has covariance matrix B B’ = Probability theory 2011 The multivariate normal distribution - a second definition A random vector is normal if and only if it has a characteristic function of the form X (t ) E (ei t' X ) exp( i t' μ t' Λ t ) 2 where is a nonnegative-definite, symmetric matrix and is a vector of constants Proof of the equivalence of definition I and II: Let XN( , ) according to definition I, and set Z = t’X. Then E(Z) = t’u and Var(Z) = t’ t, and Z(1) gives the desired expression. Let XN( , ) according to definition II. Then we can derive the characteristic function of any linear combination of its components and show that it is normally distributed. Probability theory 2011 The multivariate normal distribution - a third definition Let Y be normal with independent standard normal components and set X Λ1/2Y μ Y Λ-1/2 ( X μ) Then 1 f X ( x) 2 n/2 ( x μ)' Λ 1( x μ) 1 exp 2 det Λ provided that the determinant is non-zero. Probability theory 2011 The multivariate normal distribution - a fourth definition Let Y be normal with independent standard normal components and set X AY μ Then X is said to be a normal random vector. Probability theory 2011 The multivariate normal distribution - conditional distributions All conditional distributions in a multivariate normal vector are normal The conditional distribution of each component is equal to that of a linear combination of the other components plus a random error Probability theory 2011 The multivariate normal distribution - conditional distributions and optimal predictors For any random vector X it is known that E(Xn | X1, …, Xn-1) is an optimal predictor of Xn based on X1, …, Xn-1 and that Xn = E(Xn | X1, …, Xn-1) + where is uncorrelated to the conditional expectation. For normal random vectors X, the optimal predictor E(Xn | X1, …, Xn-1) is a linear expression in X1, …, Xn-1 Probability theory 2011 The multivariate normal distribution - calculation of conditional distributions Let XN (0 , ) where 1 2 1 2 6 0 1 0 4 Determine the conditional distribution of X3 given X1 and X2 Set Z = a X1 + bX2 + c Minimize the variance of the prediction error Z - X3 Probability theory 2011 The multivariate normal vector - uncorrelated and independent components The components of a normal random vector are independent if and only if they are uncorrelated Probability theory 2011 The multivariate normal distribution - orthogonal transformations Let X be a normal random vector with independent standard normal components, and let C be an orthogonal matrix. Then Y = CX has independent, standard normal components Probability theory 2011 Quadratic forms of the components of a multivariate normal distribution – one-way analysis of variance Let Xijij, i = 1, …, k, j = 1, …, ni , be k samples of observations. Then, the total variation in the X-values can be decomposed as follows: k k ni 2 X n X n ( X X ) ( X X ) i i. .. ij i. 2 ij i, j 2 .. 2 i 1 i 1 j 1 X ' I X X ' A1 X X ' A 2 X X ' A 3 X Probability theory 2011 1 / n . A1 . . 1 / n . . . . . 1/ n . . . . 1 / n 1 1 1 n1 n1 1 1 n 1 n 1 1 . 1 . n 1 A3 . . 1 n1 . 1 n1 1 1 n1 . 1 n1 . . . . . 1 nk 1 nk . 1 1 nk 1 1 nk . Probability theory 2011 1 nk 1 nk 1 nk 1 . 1 nk . . . . . 1 nk 1 1 nk Decomposition theorem for nonnegative-definite quadratic forms Let where n x i 1 2 i Q1 ... Q p Qi x' Ai x are non - negative quadratic forms and p p Rank ( A ) r n i 1 i i 1 i Then there exists an orthogonal matrix C such that with x = Cy (y = C’x) Q1 y12 ... yr21 . . Q p yn2 rp 1 ... yr2p Probability theory 2011 Decomposition theorem for nonnegative-definite quadratic forms (Cochran’s theorem) Let X1, …, Xn be independent and N(0; 2) and suppose that n X i 1 2 i Q1 ... Q p where Qi X' Ai X are nonnegative quadratic forms and p p Rank ( A ) r n i 1 i i 1 i Then there exists an orthogonal matrix C such that with X = CY (Y = C’X) Q1 Y12 ... Yr12 . . Q p Yn2 rp 1 ... Yr2p Furthermore, Q1, …, Qp are independent and 22-distrubuted with r1, …rp degrees of freedom Probability theory 2011 Quadratic forms of the components of a multivariate normal distribution – one-way analysis of variance Let Xijij, i = 1, …, k, j = 1, …, ni , be independent and N( ,2). Then, 2 the total sum of squares X ij can be decomposed into three i, j quadratic forms k k ni nX ni ( X i. X .. ) ( X ij X i. ) 2 Q1 Q2 Q3 2 .. 2 i 1 i 1 j 1 which are independent and 22-distrubuted with 1, k-1, and n-k degrees of freedom Probability theory 2011 Exercises: Chapter V 5.2, 5.3, 5.7, 5.16, 5.25, 5.30 Probability theory 2011