Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
TRANSFORMATION OF FUNCTION OF TWO OR MORE RANDOM VARIABLES BIVARIATE TRANSFORMATIONS DISCRETE CASE • Let X1 and X2 be a bivariate random vector with a known probability distribution function. Consider a new bivariate random vector (U, V) defined by U=g1(X1, X2) and V=g2(X1, X2) where g1(X1, X2) and g2(X1, X2) are some functions of X1 and X2 . 2 DISCRETE CASE • If B is any subset of 2, then (U,V)B iff (X1,X2)A where AU ,V x1 , x2 : g1x1 , x2 , g2 x1 , x2 B 2 • Then, Pr(U,V)B=Pr(X1,X2)A and probability distribution of (U,V) is completely determined by the probability distribution of (X1,X2). Then, the joint pmf of (U,V) is fU ,V u ,v Pr U u ,V v Pr X1 , X 2 AU ,V f X 1 , X 2 x1 , x2 x1 ,x2 AU ,V 3 EXAMPLE • Let X1 and X2 be independent Poisson distribution random variables with parameters 1 and 2. Find the distribution of U=X1+X2. 4 CONTINUOUS CASE • Let X=(X1, X2, …, Xn) have a continuous joint distribution for which its joint pdf is f, and consider the joint pdf of new random variables Y1, Y2,…, Yk defined as Y1 g1 X1 , X 2 , , X n Y2 g 2 X1 , X 2 , , X n * Yk g k X1 , X 2 , , X n T X Y ~ ~ 5 CONTINUOUS CASE • If the transformation T is one-to-one and onto, than there is no problem of determining the inverse transformation. An and Bk=n, then T:AB. T-1(B)=A. It follows that there is a oneto-one correspondence between the points (y1, y2,…,yk) in B and the points (x1, x2,…,xn) in A. Therefore, for (y1, y2,…,yk)B we can invert the equation in (*) and obtain new equation as follows: 6 CONTINUOUS CASE 1 x1 g1 y1 , y2 , , yk x2 g 21 y1 , y2 , , yk * * 1 xn g n y1 , y2 , , yk n 1 g i / y i • Assuming that the partial derivatives exist at every point (y1, y2,…,yk=n)B. Under these assumptions, we have the following determinant J 7 CONTINUOUS CASE g11 g11 yn y1 J det 1 g 1 g n n yn y1 called as the Jacobian of the transformation specified by (**). Then, the joint pdf of Y1, Y2,…,Yk can be obtained by using the change of variable technique of multiple variables. 8 CONTINUOUS CASE • As a result, the function g is defined as follows: 1 1 1 f g , g , , g X1,, X n 1 2 n | J |, for y1, y 2 ,, y n B g y , y , , y 1 2 n 0, otherwise 9 Example • Recall that I claimed: Let X1,X2,…,Xn be independent rvs with Xi~Gamma(i, ). Then, X ~ Gamma , n i 1 i n i 1 i • Prove this for n=2 (for simplicity). 10 M.G.F. Method • If X1,X2,…,Xn are independent random variables with MGFs Mxi (t), then the MGF n of Y Xi is MY (t) MX (t)...MX (t) i 1 1 n 11 Example • Recall that I claimed: ~ Bin n , p . Then, independent Let X i i X ~ Bin n n k i 1 i 1 2 n , p . k • Let’s prove this. 12 Example • Recall that I claimed: Let X1,X2,…,Xn be independent rvs with Xi~Gamma(i, ). Then, X ~ Gamma , n i 1 i n i 1 i • We proved this with transformation technique for n=2. • Now, prove this for general n. 13 More Examples on Transformations • Example 1: • Recall that I claimed: If X~N( , 2), then Z X ~ N (0,1) • Let’s prove this. 14 Example 2 • Recall that I claimed: Let X be an rv with X~N(0, 1). Then, X ~ 2 2 1 Let’s prove this. 15 Example 3 Recall that I claimed: • If X and Y have independent N(0,1) distribution, then Z=X/Y has a Cauchy distribution with =0 and σ=1. Recall the p.d.f. of Cauchy distribution: f (x) 1 1 ( 1 x )2 , 0 Let’s prove this claim. 16 Example 4 • See Examples 6.3.12 and 6.3.13 in Bain and Engelhardt (pages 207 & 208 in 2nd edition). This is an example of two different transformations: • In Example 6.3.12: In Example 6.3.13: X1 & X2 ~ Exp(1) Y1=X1 Y2=X1+X2 X1 & X2 ~ Exp(1) Y1=X1-X2 Y2=X1+X2 17 Example 5 • Let X1 and X2 are independent with N(μ1,σ²1) and N(μ2,σ²2), respectively. Find the p.d.f. of Y=X1-X2. 18