Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
5 Sampling Distributions and Approximations 5.1 Distributions of Sums of Random Variables Functions of random variables are usually of interest in statistical applications; and, of course, we have already considered some functions, called statistics, of the observations of a random sample. Two important statistics arc the sample mean X and the sample variance S2. Of course, in a particular sample, say xi, x ^ , . . . , x ^ , we observed definite values of these statistics, x and s 2 ; however, we should recognize that each value is only one observation of the respective random variables, X and S2. That is, each X or S2 (or, more generally, any statistic) is also a random variable with its own distribution. IQ this chapter we determine the distributions of some of the important statistics and, more generally, distributions of functions of random variables. The derivations of distributions of these functions fall in that area of statistics usually referred to as sampling distribution theory. We begin with an example that illustrates how we can find the distributions of two different functions of the random variables in a random sample. Example 5.1-1 Let us cast an unbiased four-sided die two independent times and observe the outcomes on these trials, say X, and X , . This if equivalent to saying that X ^ and X^ are the observations of a random sample of size n = 2 from a distribution with p.d.f. 284 5.1 Distributions of Sums of Random Variables /M=^, x=l,2,3,4. We shall find the distribution of the sum Y = A"i + JC; and of the range W = max (A'i, A',) - min (A-;, Ay. We first consider the distribution of Y = X^ + X^ and immediately recognize that the space of Y is R = {2, 3, 4, 5, 6, 7, 8}. To determine the p.d.f. of V, say g(y) = P(Y = y), y e R, we first look at some special cases. Let y = 3. We note that the event {Y = 3} can occur in two mutually exclusive ways: {A'i = 1, X^ = 2} and [X^ = 2, X^ = 1}. Thus recalling that X , and X^ are independent, we obtain g(3) = P(Y = 3) = P(Xt = 1, A-, = 2) + P(A-i = 2, X; = 1) Similarly, 9(5) = F(r = 5) = P(JCi = 1, A-, = 4) + P(Xi = 2, X., = 3) + P(A-i = 3, X; = 2) + P(Jfi = 4, A'.; = 1) In general, g(y) = PfV = y) = 'S /(t)/(y - k), where some of the summands could equal zero. We can give the p.d.f. either by Table 5.1-1 or by the formula 8(>')=4—1——^1-, 10 y = 2, 3, 4, 5, 6, 7, 8. We can compute the mean of Y from the p.d.f. of Y to obtain E(Y)= ^ra(y) -%HiiH^)<Mil)<) =5. 285 Sampling Distributions and Approximations Y SI.Y) 2 3 4 5 6 7 8 1/16 2/16 3/16 4/16 3/16 2/16 1/16 To find the distribution of the range, W, note that the event {^=0} occurs if [ X ^ = X ^ } , and thus fi(0)=.P(W=0)=^. The event {W = 3} occurs if {A"i = 1, X, = 4} or (A", = 4, A', = 1}. Thus fc(3)=P(W=3)=^. Similarly, A(l) = P(»V = 1) = 6/16, and A(2) = P(iy = 2) = 4/16. To find the mean of W, we have ^ © © © © i In Example 5.1-1, since Y = X i + X;, we anticipate that £(V) = 5, which is computed using the p.d.f. of Y, would be equal to E(Xi + X^), which is computed using the joint p.d.f. ofA'i and X ^ , namely Of course, this is true because E(X, + X,) = E(X,) + E(X^ - J + | = 5 . This special result with a linear function of two random variables extends to more general functions of several random variables. We accept the following without proof. 286 6.1 Distributions of Sums of Random Variables THEOREM 5.1-1 Let X ^ , X ^ , ..., X.be n random variables with joint p.d.f. f ( x ^ , x;,..., jc,,). Let the random variable Y = u(X^, X ^ , ..., XJ have the p.d.f. g(y}. Then, in the discrete case, E(Y) = £ provided that these summations exist. For random variables of the continuous type, integrals replace the summations. Recall (Section 4.2) that if X ^ and X; are independent random variables, then E^(X,)u,(X,f] = £[Ui(Xi)]£[:U2(A-2):l, provided that these expectations exist. This can be extended to products of functions of n mutually independent random variables. THEOREM 5.1-2 If X ^ , X ^ , .... X , are mutually independent random variables having p.d.f:'s /i(xi), fi(x^, ...,/.(x,) and £[u,(X,)], i = 1, 2, ..., n, exist, then E^u,(XMX,) • Proof; In the discrete case, we have that Etu,(XMX,) • = £ £ • • • £ "l(»l)"2^2) • = £ = £[ui(A'i)]£[»2(^2)] • REMARK Sometimes students recognize that X2 = X • £(X2) is equal to [£(X)][£(A")] = [£(X)]2 because the above theorem states that the expected value of the product is the product of the expected values. However, note the hypothesis of independence in the theorem, and certainly X is not independent of itself. Incidentally, if E(X2) did equal [£W]2, then the variance of X a2 = E(X2) - [£(X)]2 would always equal zero. This really happens only in the case of degenerate (one point) distributions. 287 Sampling Distributions and Approximations I Example 5.1-2 It is interesting to note that these two theorems allow i determine the mean, the variance, and the moment-generating function function such as V = A\ 4- X ^ , where X i and X^ have been define Example 5.1-1. We have already seen that /iy = 5. The variance of Y is a1, = £[(Y - ^)2] = EUX, + X , - /i, - ^)2], where fi, = E(X,) = 5/2, i = 1, 2. Thus a2 = £[(A-i - 11,} + (Xi - ^)]2 = £[(A-i - ^i)2 + 2(A-, - pi)(-y, - /^) + (A-; - ^)2]. But, from an earlier result about the expected value being a linear opei we have a2 = £[(Jf, - /^i)2] + 2£[(X, - ^)(A-, - ^.;)] + £[(^ - fi;)2]. However, since Xi and X^ are independent, then £[(^i - fi)(A'2 - vS\ = Wt - ^WX, - fi,)] = (/'I - f l ) ( f 2 - /'2) = 0. Thus "i = "i + "i. where (rf = £[(^, - ^,)2], i = 1, 2. In the case in which X^ and X^ are the outcomes on two independent of a four-sided die, we have af = 5/4, i = 1,2, and, hence - 5 5 5 "^^^lFinally, the moment-generating function is M,(t} = E(e") = .E|>w.+.'•«>] = E(e•x•e""). 288 5.1 Distributions of Sums of Random Variables The independence of X ^ and X, implies that M^t) = E(e"W). For our example in which X^ and X; have the same p.d.f. f(x) = ^ x = 1, 2, 3, 4, and thus the same moment-generating function ^O = 4 \ <•' +4^ e2' +4\e1' +4 \ '•"• we have that My(i) equals [M^ = ^^ ^ ^ ^ ^.+ ^ ^- Note that the coefficient of e" is equal to the probability P(Y = b); for illustration, 4/16 = P(Y = 5). This agrees with the result found in Example 5.1-1, and thus we see that we could find the distribution of Y by determining its moment-generating function. In this section we shall restrict attention to those functions that are linear combinations of random variables. We shall first prove an important theorem about the mean and the variance of such a linear combination. THEOREM 5.1-3 I f X ^ , X ^ , . . . , X , are n independent random variables with respective means f i ^ ^,..., ^ and variances f f \ , ffj, ..., (T^ , then the mean and the variance of Y = ^^ i a, X ^ , where a^, a;,.. -, a,, are real constants, are fiy=^a,ti, and o-y = ^ a f a f , respectively. Proof: We have that (^Em^S 0 ,^.^ ia.£W)= E a,H, \.-i / .=1 i=i 289 Sampling Distributions and Approximations Ch, B| I because the expected value of the sum is the sum of the expected values (i.e., £ is a linear operator). Also i o2 = £[( V - (^)2] = J( Y, a, X, - 1 a, ft.) ] | = E^i a,(X, - ^)Tl = J 1 f; a,fl/X. - AX^ - n,)} ; Again using the fact that £ ^ = 1=1 £ j-1 However, if i 96 y, then from the independence of X,and X j we have £[(X, - /<,X-y; - ?<,)] = E(X, - /i,)£(A-, - /!,) = (p, - ^,X(<, - ^) = 0. Thus the variance can be written as IT? = ^ af£[(A-, - /i,)2] = S afff,2. D We give two illustrations of the theorem. Example 5.1-3 Let the independent random variables X i and -Y; have respectively means fi, = —4 and /i; = 3 and variances u2 = 4 and irj = 9. The mean and the variance of Y = 3X^ - IX 3 are, respectively, ^ y = ( 3 X - 4 ) + ( - 2 X 3 ) = -18 and ff 2 =(3) 2 (4)+(-2) 2 (9)=72. Example 5.1-4 Let X ^ , X , , . . . . X. be a random sample of size n from a distribution with mean y. and variance cr2. First let Y = X ^ — X^, then /i,=/x-/i=0 and ff 2 = (I)2;!2 + (-l) 2 ^ = 2ff 2 . Now consider the sample mean X , + X, + • X='290 S.1 Distributions of Sums of Random Variables which is a linear function with each a, = l/n. Then "^.I.C;)'"" and '^^O'^- That is, the mean of X is that of the distribution from which the sample arose, but the variance of X is that of the underlying distribution divided by n. In some applications it is sufficient to know the mean and variance of a linear combination of random variables, say V. However, it is often helpful to know exactly how Y is distributed. The next theorem can frequently be used to find the distribution of a linear combination of independent random variables. THEOREM 5.1-4 I f X ^ X ^ , . . . , X^ are independent random variables with respective moment-generating functions My;((), i = 1, 2, 3, ..., n, then the moment-generating function ofY = ^?=i Oi-^i is M,(t)=f\M^t). Proof: The moment-generating function of Y is given by My{t) = £[e'1'] = E^e1^1^^ ••-+a-('•'] =E\ieal'xlea2txl•••ea'•tx^ = £[e•'l'yl]JE[e''2tx2] ••- EEe0"'^] using Theorem 5.1-2. However, since E^e'xi) = M^(t), then E{ea•txi}=M^a,t). Thus we have that My(t} = M^t)M^ t) • =f\M^a,t). 1=1 D A corollary follows immediately, and it will be used in some important examples. 291 Sampling Distributions and Approximations Ch. 5 COROLLARY 5.1-1 //-^n -^2' • • • ' ^n are observations of a random sample from a distribution with moment-generating function M(t), then (i) the moment-generating function of Y = Y.1= i ^ils l My(t)= n^(o=[^(o]"; 1=1 (ii) the moment-generating function ofX == ^^i (l/n)Xi is ^'-n^KC)]" I , ; : Proof: For (i), let a, = 1, i = 1, 2,..., n, in Theorem 5.1-4. For (ii), take a, = l/n,i= l,2,...,n. D The following examples and the exercises give some important applications of Theorem 5.1-4 and its corollary. Example 5.1-5 Let A",, X ^ , . . . , X , denote the outcomes on n Bernoulli trials. The moment-generating function of X , , i == 1, 2,..., n, is M(t) = q + fie'. Y = £^,, then A'r(t)= FT (9 +;"'')= te+Pe')". 1=1 Thus we again see that Y is 6{n, p). Example 5.1-6 Let X ^ , X ^ , X y be the observations of a random sample of size n = 3 from the exponential distribution having mean 8 and, of course, moment-generating function M(r) = 1/(1 — generating function of Y = X^ + X^ + -¥3 is My(() = [(1 - fl()-1]3 = (1 - fitF3, t < 1/9, which is that of a gamma distribution with parameters a = 3 and ft Thus Y 292 5.1 Distributions of Sums of Random Variables has this distribution. On the other hand, the moment-generating function of xis ^o-rn-r t<3/0; and hence the distribution of X is gamma with parameters a = 3 and 0/3, respectively. Exercises 5.1-1 Let X ] and X ^ be observations of a random sample of size n = 2 from a distribution with p.d.f./(x) = x/6, x = 1, 2, 3. Find the p.d.f of Y = X , + X ^ . Determine the mean and the variance of the sum in two ways. 5.1-2 Let X , and X; be a random sample of size n = 1 from a distribution with p.d.f. f { x ) = 6x(l - x), 0 < x < 1. Find the mean and the variance of Y = Xi + X;. 5.1-3 Let Xp X;, A" 3 be a random sample of size 3 from the distribution with p.d.f. f { x ) = 1/4, x = 1, 2, 3, 4. For example, observe three independent rolls of a fair four-sided die. (a» Find the p.d.f. of Y = X^ + X, + X ^ . (b) Sketch a bar graph of the p.d.f. of Y. 5.1-4 Let Xi and X^ be two independent random variables with respective means /^i and ^ and variances o\ and a\. Show that the mean and the variance of Y = X ^ X ^ are /i^ and o^o-j + ^fffj + ^ a\, respectively. HINT: Note that E(Y} = E(Xi)E[X^} and £(V2) = E(Xi)E(Xi). 5.1-5 Let X ^ and X ^ be two independent random variables with respective means 3 and 7 and variances 9 and 25. Compute the mean and the variance of Y = —2Xi +X,. 5.1-6 Let X, and X ^ have independent distributions &(«i, p) and fc(n;, p), respectively. Find the moment-generating function of Y = X i + X ^ . How is Y distributed? 5.1-7 Let A\, X;, X y be mutually independent random variables with Poisson distributions having means 2,1, 4, respectively. (a) Find the moment-generating function of the sum Y = X ^ + X ^ + X ^ . (b) How is V distributed? (c) Compute P(3 < Y < 9). 5.1-8 Generalize Exercise 5.1-7 by showing that the sum of n independent Poisson random variables with respective means fi,, ^,..., ^ is Poisson with mean ^l + ^ 2 + - - - + / V 5.1-9 Let A'i, X-i, X , , , X ^ , X., be a random sample of size 5 from a geometric distribution with p = 1/3. (a) Find the moment-generating function of Y = X ^ + X ^ + X ^ + X^ + X ^ (b) How is r distributed? 293 Sampling Distributions and Approximations Ch. 5 5.1-10 Let W = X\ + X ^ + • distributed exponential random variables with mean 6. Show that W has a gamma distribution with mean hQ. 5.1-11 Let X ^ X ^ , X y denote a random sample of size 3 from a gamma distribution with a = 7 and 0 = 5 . (a) Find the moment-generating function of Y = X ^ + X^ + X ^ (b) How is r distributed? 5.1-12 Let X and Y, with respective p-d.f.'s f [ x ) and g(y), be independent discrete random variables, each of whose support is a subset of the nonnegative integers 0,1, 2 , . . . . Show that the p.d.f. of W = X + Y is given by the convolution formula hiw) = S Mg(w - x), w = 0. 1, 2,. HINT: Argue that /i(w) = P(W = w} is the probability of the w + 1 mutually exclusive events (x, w — 5.1-13 Let X ^ , X ^ , X ^ , X^ be a random sample from a distribution having p.d.f. /(x)=(x+l»/6,x=0, 1,2. (a) Use Exercise 5.1-12 to find the p.d.f. of W, = X ^ + X ^ . (b) What is the p-d.f. of W^ - X,, + X^ (c) Now find the p.d.f. of W = W^ + W^ = X ^ + X ^ + X ^ + X ^ . 5.1-14 Roll a fair 4-sided die eight times and denote the outcomes by A], X ^ , . . . , X y . Use Exercise 5.1-12 to find the p.d.f.'s of (a) X , + X , , (b) S^,, (c) Y,X,. 5.2 Random Functions Associated with Normal Distributions In statistical applications, it is often assumed that the population from which a sample is taken is normally distributed, N(^i, f f 2 ) . There is then interest in estimating the parameters p. and a2 or in testing conjectures about these parameters. The usual statistics that are used in these activities are the sample mean X and the sample variance S 2 ; thus we need to know something about the distribution of these statistics or functions of these statistics. "THEOREM 5.2-1 If' X^, X - ^ , ..,, X^ are observations of a random sample of size n from the normal distribution N{p., o"2), then the distribution of the sample mean X = (1/n) ^?_i X, is N(fi, o-Vn). 294 5.2 Random Functions Associated with Normal Distributions Proof Since the moment-generating function of each X is / ff't^ M^t) = exp ( ^ ( + -^-1, the moment-generating function of X= 1 YX, n i= 1 is, from the corollary of Theorem 5.1-4, equal to ^44(^];" 2 2 r (ff /")' '! = exp f t + ——j— . However, the moment-generating function uniquely determines the distribution of the random variable. Since this one is that associated with the normal distribution N(fi, o^fn), the sample mean X is N(/f, ^/n). D Theorem 5.2-1 shows that if X^ X ^ , ..., X^ is a random sample from the normal distribution, N(^, o1}, then the probability distribution of X is also normal with the same mean ^ but a variance a ^ f n . This means that X has a greater probability of falling in an interval containing p, than does a single observation, say X i . For example, if fi = 50, o2 = 16, n = 64, then P(49 < X < 51) = 0.9544, whereas F(49 < Xi < 51) = 0.1974. This is illustrated again in the next example. Example 5.2-1 Let X ^ , X ^ , . . . , X , be a random sample from the JV(50, 16) distribution. We know that the distribution o f X i s N(50, 16/n). To illustrate the effect of n, the graph of the p.d.f. of X is given in Figure 5.2-1 for n = 1, 4, 16, and 64. When n = 64, compare the areas that represent P(49 < X < 51) and P(49 < X\ < 51). Before we can find the distribution of S2, or (n — liminary theorems that are important in their own right. The first is a direct result of the material in Section 5,1 and gives an additive property for indept chi-square random variables. (S^EOREM 5.2-2 Let the distributions o f X , , X ,,..., X^ be ^(i-i), ^(r;),.... ^(i-i,), respectively. I f X ^ X ^ , . . . , X,, are independent, then Y = X, + A"; + • + X, is ^(r, + r; + • 295 Sampling Distributions and Approximations Proof We give the proof for k = 2 but it is similar for a general k. moment-generating function of Y is given by Mi,(t) = E[_e'^ == E^1^2^ = E^^^. Because Xi and X^ are independent, this last expectation can be factorei that M^) = E^W^ = (1 - It)-^! - 20-^ t < -, since X^ and X^ have chi-square distributions. Thus M^)^!^)^'^/2, « ^ the moment-generating function for a chi-square distribution with r = r, degrees of freedom. The uniqueness of the moment-generating function im] that Y is ^(r^ + '•2). For the general k, Y is -^-(r^ + r^ + ••• + r^). The next theorem combines and extends the results of Theorems 3.5-1 5.2-2 and gives one interpretation of degrees of freedom. 296 6.2 Random Functions Associated with Normal Oixtributions THEOREM 5.2-3 Lei Z^, Z;, ..., Z, have standard normal distributions, N(0, 1). If these random variables are mutually independent, then W = 7.\ + 7.\ +• Proof: By Theorem 3.5-1, Z? is /'(I) for i = 1, 2,..., r. From Theorem 5.2-2, with k = r, V = W, and r, = 1, we see that W is /•'(r). 0 COROLLARY 5.2-1 If X ^ , X , , .... X. have mutually independent normal distributions N(u,, a1), i = 1, 2,..., r, respectively, then the distribution of w -^ is^(r). Proof We simply note that Z, = ( X , - u,)/a, is N(0, 1), i = 1, 2,..., r. Q Note that the number of terms in the summation and the number of degrees of freedom are equal in Theorem 5.2-3 and the corollary. The following theorem gives an important result that will be used in statistical applications. In connection with those, we will use the sample variance sl=-li(x,-x)l n ~ l 1=1 2 to estimatethe variance, f f , when sampling from the normal distribution, N{^, a2). More will be said about S2 at the time of its use. THES&REM 5.2-4 ^/-^i> ^2, •••, ^n are observations of a random sample of size nfrom the normal distribution N(^i, a2), ^4,t^ and s'-^IW-W ,„ ns2 ZW-^) 2 1 5 ( i) * " — — ' — — ^ — — — — — — i s ^ n l ) . 297 ] Sampling Distributions and Approximations Ch,^ Proof: We are not prepared to prove (i) in this book; so we accept it without proof here. To prove (ii), note that y (x' - t1}2 „ 1-1 \ " ) .=i L " J -IK-^}^ '-' because the cross-product term is equal to ^-^.-^2(^^_^ ,=i " " i-i But Y, = (X, — normal variables. Hence W = Y.^i y? i5 /("I by Corollary 5.2-1. Moreover, since X is A^/i, a2/"), then ^ _ / . g - ^ Y- _ ^ - ^ "L/^;^ ^ is y2;!) by Theorem 3.5-1. In this notation, equation (5.2-1) becomes ^(»_^z'. 2 a 2 However, from (i), X and S are independent; thus Z2 and S2 are also independent. In the moment-generating function of W, this independence permits us to write ^^e•l<"-l>sa;•2+^'l] = ^[(..(•-Ds'/.'e.z'] = E[.e•(•-w2"'~}E[,e'z"l. Since W and Z 2 have chi-square distributions, we can substitute their moment-generating functions to obtain (1 - 2t)-"'2 = El:^"-11""2^! - 2t)-"2. Equivalently, we have £[6""-""'°'] = (1 - 2t)-'"-'"2, ( < •1. This, of course, is the moment-generating function of a ^(n - 1) variable, and accordingly (n - IIS2^2 has this distribution. Q 298 6.2 Random Functions Associated with Normal Distributions Combining the results of Corollary 5.2-1 and Theorem 5.2-4, we see that when sampling from a normal distribution, ,^M is ^(n), and ^^_D! is ^(n - 1). That is, when the population mean, fi, in ^ (X, - ^)2 is replaced by the sample mean, X , one degree of freedom is lost. There are more general situations in which a degree of freedom is lost for each parameter estimated in certain chi-square random variables (e.g., see Section 8.5). Example 5.2-2 Let X^, X ^ , X ^ , X^ be a random sample of size 4 from the normal distribution, JV(76.4, 383). Then u=T. (X, - 76.4)2 383 is rt4), -(X^O. -i 383 is it^), and, for examples, P(0.711 < U < 7.779) = 0.90 - 0.05 = 0.85, P(0.352 :£ W & 6.251) = 0.90 - 0.05 = 0.85. In later sections we shall illustrate the importance of the chi-square distribution in applications. It is often true in statistical applications that there is interest in estimating the mean y. or testing hypotheses about the mean when the variance a1 is unknown. Theorem 4.7-1, in which a T random variable is defined, along with Theorems 5.2-1 and 5.2-4, provide a random variable that can be used in such situations. That is, if X^, X ^ , ..., X^ is a random sample from the normal distribution N(^, a1), then (X - tt)l(a!^n) is N(0, 1), (n - W/a^ is ^(n - 1), and the two are independent. Thus we have that ^._ (X - it)l(al^n) ^(n - W/a^n - 1) _X-ii S/^n 299 Sampling Distributions and Approximations Ch. 5 has a r distribution with r = n — able like this one that motivated Gosset's search for the distribution of T. This t statistic will play an important role in statistical applications. We now prove another theorem which deals with linear functions of independent normally distributed random variables. THEOREM 5.2-5 //^i» X ^ , .-., X^ are n mutually independent normal variables with means ^i, /;;, ..., u^ and variances a}, o-j, ..., o-^, respectively, then the linear function Y = ^c,X, has the normal distribution MS'-,/,,, ^c.^2 . \l-l r=l / Proof: By Theorem 5.1-4, we have M,(t) = flM^(c,t) = flexp (fi,c,t + aWl-t) because M , f t ) = exp (;;, t + aft 2 /!), i = 1, 2,..., n. Thus MM = exp [( E'-.f,)' + [l.^i)^')] This is the moment-generating function of a distribution which is N(ic,^,, icw}/ \i=i 1=1 and thus Y has this normal distribution. Q From Theorem 5.2-5 we can make the observation that the difference of two independent normally distributed random variables, say Y = X i — the normal distribution N(u^ — ahead to statistical applications in which we compare the means for two normal distributions. This, in turn, is used to construct the following t random variable, which is also important in statistical applications. Let -X\, X y _ , ..., Xn and Y^ Y^, ..., Y^ be random samples from the independent normal distributions, N{ux, ffjt) and N{uy, a^}, respectively. The distribution of X is N(ujc, a^/n) and the distribution of Y is N(^y, a^/m). Since X and V are independent, the distribution X — 300 5.2 Random Functions Associated with Normal Distributions and _ x-y-fe-^,) Z=- ^/ffi/n + s\lm is N(0, 1). The statistics ,„ 1,C2 SW--?)2 I" - I f i x i-l_________ and (.-1^ ^-y)2 have distributions that are '^(n — normal distributions are independent, these chi-square statistics are independent, and thus the distribution of ^ _ (n - l)Sj (m - 1)S; is ^(n + m — r =n +m — T=^U/(" + m - 2) [ X - Y - (^ - )ir)]/v/'d/" + "i/m ^/[(n - 1)S^ + (m - l)S?/<i?]/[n + m - 2] In the statistical applications we sometimes assume that the two variances are the same, say a\ = a\ =- a2, in which case X - Y - (fi, - 11,) 7-=^{[(n - 1)S^ + (m - l)Sa/(n + m - 2)}[(l/n) + (1/m)] and neither T nor its distribution depend on a2. Example 5.2-3 Let X i, X ^ , . . . , Xg and Yi, Y;, . - -, Vie be random samples of sizes « 301 Sampling Distributions and Approximations 2 2 Ch. 5 2 A/(^y, (7 ) and /V(/iy, o" ), respectively, where ff is unknown. Then p( -2.306 ^ x——^ < 2.306) = 0.95 V s,/^ ) (5.2.2) because (X — freedom. Also because ^_ X-Y~{^-^) ^[(SS2 + 15S2)/23][(1/9) + (1/16)] has a t distribution with r = 9 + 16—2= 23 degrees of freedom, P(- 1.714 < T < 1.714) = 0.90. (5.2-3) Note that in both equation (5.2-2) and equation (5.2-3), after the data have been observed, X, Y, S 2 -, and S2 can be calculated so that only p.^ and ^ will be unknown. In the next chapter we shall use this information to construct interval estimates of unknown means and differences of means. The F distribution also has many important applications in normal sampling theory, one of which uses the following random variable. Let X ^ , X ^ , ..., X^ and YI, y;, ..., y,n be random samples of sizes n and m from two independent normal distributions /V(/^y, o^) and .N(;Uy, o-2), respectively. We know that (n - l)S^/o-^ is ^(n - 1) and (m - l^y/o 2 is ^(m - 1). The independence of the distributions implies that S^ and Sy are independent so that (m - l^/^m - 1) _ SW (n - l)S^{n - 1) Si/o2 has an F distribution with r^ = m — This result is used in the next chapter. Exercises 5.2-1 Let Xi, X;, ..., X^ be a random sample from a normal distribution N{77, 25). Compute (a) P(77 < X < 79.5), (b) P(74.2 < X < 78.4). 5.2-2 Let X be N(50, 36). Using the same set of axes, sketch the graphs of the probability density functions of (a) X ; (b) X , the mean of a random sample of size 9 from this distribution; (e) X , the mean of a random sample of size 36 from this distribution. 5.2-3 Let X equal the widest diameter (in millimeters) of the fetal head measured between the 16th and 25th weeks of pregnancy. Assume that the distribution of X is 302 5.2 Random Functions Associated with Normal Distributions N(46.58, 40.96), Let X be the sample mean of a random sample of n = 16 observations of X . (a) Give the value of £(-?). (b) Give the value of Var(^). (c) Find P(44.42 ^ X ^ 48.98). 5.2-4 Let X equal the weight of the soap in a "6-pound" box. Assume that the distribution o f X is N(6.05, 0.0004). (a) Find P(X < 6.0171). (b) If nine boxes of soap are selected at random from the production line, find the probability that at most two boxes weigh less than 6.0171 each. (c) Let X be the sample mean of the nine boxes. Find P[X ^ 6.035). 5.2-5 Let Zp Z;, . . . , Z-, be a random sample of size n = 1 from the standard normal distribution N(0, 1). Let W = Z\ + 2| + • 5.2-6 If A], X ^ , ..., X,6 is a random sample of size n = 16 from the normal distribution N(50, 100), determine / 16 \ (a) P) 796.2 ^ S (X; - 50)2 ^ 2630 I, (b) P 726.1 < ^ (X, - X}2 ^ 2500 1, \ 1=1 / 5.2-7 Let X and S2 be the sample mean and sample variance associated with a random sample of size n = 16 from a normal distribution Nf/i, 225). (a) Find constants a and b so that P(a ^ S2 ^ b) - 0.95. (b) Find a constant c so that ^_^^^.0.95. 5.2-8 Let X and Y denote the wing lengths (in millimeters) of a male and a female gallinule, respectively. Assume that the respective distributions of X and Y are N(184.09, 39.37) and N(171.93, 50.88) and that X and Y are independent. Let X and Y equal the sample means of random samples of sizes n = m = 16 birds of each sex. (a) FindP(X- V^9.12). (b) Let S2. and S^ be the respective sample variances. Find a constant c so that ( X - Y- 12.16 \ P( -c ^ . ^ c - 0.95. \ V[(15S^-H5^)/30](2/16» / 5.2-9 Let X i , X ^ , ..., X q be a random sample of size 9 from a normal distribution N(54, 10) and let Vp Y ^ , Y ^ , Y^ be a random sample of size 4 from an independent 303 Sampling Distributions and Approximations Ch. 5 normal distribution N(54, 12). Compute s^-?) 2 / \ 1 4 l — — l ^ ^ 5.2-10 Compute P{X < Y) == P(0 < Y — variables with normal distributions /V(4, 9) and N(7, 16), respectively. 5.2-11 Let X denote the wing length in millimeters of a male gallinule and Y the wing length in millimeters of a female gallinule. Assume that X is N(184.09, 39.37) and V is N(171.93, 50.88), and that X and V are independent. If a male and a female gallinule are captured, what is the probability that X is greater than V? 5.2-12 Suppose that for a particular population of students SAT mathematics scores are N{529, 5732) and SAT verbal scores are N(474, 6368). Select two students at random, and let X equal the first student's math score and Y the second student's verbal score. Find P(X > Y). 5.2-13 Suppose that the length of life in hours, say X , of a light bulb manufactured by company A is N(800, 14,400) and the length of life in hours, say V, of a light bulb manufactured by company B is N(850, 2500). One bulb is selected from each company and is burned until "death." (a) Find the probability that the length of life of the bulb from company A exceeds the length of life of the bulb from company B by at least 15 hours. (b) Find the probability that at least one of the bulbs " lives " for at least 920 hours. 5.3 The Central Limit Theorem In Section 5.1 we found that the mean -Y of a random sample of size n from a distribution with mean f i and variance a1 > 0 is a random variable with the properties that E(X)= f i and Var(J?)=°-. n Thus, as n increases, the variance of X decreases. Consequently, the distribution of X clearly depends on n, and we see that we are dealing with sequences of distributions. In Theorem 5.2-1 we considered the p.d.f. of X when sampling from the normal distribution JV(/i, a1}. We showed that the distribution of X is N(fi, f f 2 / ! ! } , and in Figure 5.2-1, by graphing the p.d.f.'s for several values of n, we illustrated that as n increases, the probability becomes concentrated in a small interval centered at ^. That is, as n increases, X tends to converge to (i, or (X — 304 5.3 The Central Limit Theorem In general, if we let W=^(J?-,)=^ (7 ff/^/n where X is the mean of a random sample of size n from some distribution with mean p. and variance a2, then for each positive integer n, ^^1=^^=0 "f^P Lff/VnJ -ff/yij a/,/)) (7/,/n and VarW = E(^) = fe^l =E^-^ |_ (T-/n J a~/n ^ = 1. (r-/n Thus, while X — "spreads out" the probability enough to prevent this degeneration. What then is the distribution of W as n increases? One observation that might shed some light on the answer to this question can be made immediately. If the sample arises from a normal distribution then, from Theorem 5.2-1, we know that X is N{^i, a^/n) and hence W is N(0, 1) for each positive n. Thus, in the limit, the distribution of W must be N(0, 1). So if the solution of the question does not depend on the underlying distribution (i.e., it is unique), the answer must be /V(0, 1). As we will see, this is exactly the case, and this result is so important it is called the Central Limit Theorem, the proof of which is given in Section 5.5. THEOREM 5.3-1 (Central Limit Theorem) // X is the mean of a random sample Xp X^ , ..., X,, of size n from a distribution with a finite mean ^ and a finite positive variance o"2, then the distribution of .^^ ^/v" v" ° is N(0, 1) i« the limit as n -* oo. A practical use of the Central Limit Theorem is approximating, when n i "sufficiently large," the distribution function of W, namely ,r—^ P(W < w) » J-^J2n ' J^ 305 Sampling Distributions and Approximations Ch. We present some illustrations of this application, discuss " sufficiently large and try to give an intuitive feeling for the Central Limit Theorem. Example 5.3-1 Let X denote the mean of a random sample of size n = 1 from the distribution whose p.d.f. is/(x) = (3/2)x2, — l < x < l . Here ji = and a2 = 3/5. Thus P(0.03 < X < 0.15) = P(4^ < -£-*- < 4^) \7375/^/T5 ^3/5/^15 ^/WyW =P(0.15^ W<0.75) »<1)(0.75) - <I'(0.15) = 0.7734 - 0.5596 = 0.2138. Example 5.3-2 Let Xi, X ^ , . . . , X^o denote a random sample of size ^ from the uniform distribution U(0, 1). Here E(X,) = 1/2 and Var(^",) = 1/1 j = 1,2, . . . , 2 0 . I f r = ^ i + X ^ +••• +X2o,then -p p(y < 0,) = pf-^2) , 9___^ ^ ^ ^ _^ a<Ii(-0.697) = 0.2423. Also, 0 1 ^^y^ii.^p^ ^ -^^11-7-^ 8 -( ^5/3 ^Sft ^5/3 =P(-1.162< W < 1.317) »<t(1.317)-<!)(-1.162) =0.9061 -0.1226=0.7835. Example 5.3-3 Let X denote the mean of a random sample of size 25 froi the distribution whose p.d.f. is/(x) = x3/^, 0 < x < 1. It is easy to show thi ^1=8/5= 1.6 and cr2 = 8/75. Thus / 1.5 - 1.6 X - 1.6 1.65 - 1.6 \ P(1.5 $ X < 1.65 = P( ,—— ,— & .—— ^- < ,—— \V8/75/y25 ^S/TS/^IS V8/75/v'25/ =P(-1.531 < W ^ 0.765) »1>(0.765)-<1)(-1.531) =0.7779-0.0629=0.7150. 306 , 6.3 The Central Limit Theorem These examples have shown how the Central Limit Theorem can be used for approximating certain probabilities concerning the mean X or the sum Y = ^=1 X : of a random sample. That is, X is approximately N(^, o^/n), and V is approximately N(n^i, na2) when n is "sufficiently large," where /t and a2 are the mean and the variance of the underlying distribution from which the sample arose. Generally, if n is greater than 25 or 30, these approximations will be good. However, if the underlying distribution is symmetric, unimodal, and of the continuous type, a value of n as small as 4 or 5 can yield a very adequate approximation. Moreover, if the original distribution is approximately normal, X would have a distribution very close to normal when n equals 2 or 3. In fact, we know that if the sample is taken from N(^i, a2), X is exactly N(^, a2/^ for every n = 1, 2, 3 , . . . . The following examples will help to illustrate the previous remarks and will give the reader a better intuitive feeling about the Central Limit Theorem. In particular, we shall see how the size of n affects the distribution of X and V = ^ X , for samples from several underlying distributions. Example 5.3-4 Let X ^ , X ^ , X ^ , X^ be a random sample of size 4 from the uniform distribution (7(0, 1). Then ^ = 1/2 and a2 = 1/12. We shall compare the graph of the p.d.f. of Y= "LX, with the graph of the N[n(l/2), n(l/12)] p.d.f. for n = 2 and 4, respectively. To find the p.d.f. of Y = X ^ + X ^ , we first find the distribution function of V. The joint p.d.f. of-Yi and X^ is f { x ^ , x;) = 1, 0 < x, < 1, 0 <. X2 < 1. The distribution function of Y is G(y)=P(Y<y)=P(X,+X^<y). I f 0 < > - < 1, ^ G(y)= and if 1 < y < 2, v-^ -I:J:.,' 1 G(y) = 1 - ^ Idx^dx^^-, • r r' J,-i J>-»i (2 - v}2 1 dx, dx, = 1 - * — — " 2 307 Sampling Distributions and Approximations Ch. 5 Thus the p.d.f. of Y is (v ,(,)=GW=^ 0 < v < 1, ,^^ Adding the point g(l) = 1, we obtain the triangular p.d.f. that is graphed in Figure 5.3-l(a). In this figure the N[2(l/2), 2(1/12)] p.d.f. is also graphed. 308 I The Central Limit Theorem The p.d.f. of y = Xi + X^ + X , + X^ could be found in a similar, but somewhat more difficult, manner. It is y_ 0<y<1, 6' 3 2 -3y + 12y - lly + 4 9W = i 1 < V < 2, 2 3j,a - 24y + 60y - 44 3 2 < y < 3, 2 -y + 12y - 48y + 64 3 ^ y < 4. This p.d.f. is graphed'in Figure 5.3-l(b) along with the N[4(1/2), 4(1/12)] p.d.f. If we are interested in finding F(1.7 <; Y < 3.2), this could be done by evaluating 1 g(y) dy, r- Jl-7 which is tedious (see Exercise 5.3-10). It is much easier to use a normal approximation. In Example 5.3-4 we showed that even for a small value of n, like n = 4, the sum of the sample items has an approximate normal distribution. The following example illustrates that for some underlying distributions n must be quite large. In order to keep the scale on the horizontal axis the same for each value of n, we will use the following result. Let f { x ) and F(x) be the p.d.f. and distribution function of a random variable, X , of the continuous type having mean y. and variance <r2. Let W = (X — G(w) == P(W <w)= p(x—ajj- < w} \ / = P(X <vw+fi)= F(ffw + /i). Thus the p.d.f. of W is given by g(w) = F'(cw + f^)= of (aw + /i). 309 Sampling Distributions and Approximations Ch. 5 Example 53-5 Let X ^ , X ^ , ..., A\oo be a random sample of size 100 from a chi-square distribution with one degree of freedom. If Y= f^X,, then y is ^(n), and £(r) = n, Var(V) = In. Let ^y-'. V2n The p.d.f. of W is given by ^^^^^""'e-^^, -./^2,;<>V<«. r@2Note that w > - n/^/2n corresponds to y > 0. In Figure 5.3-2(a) and (b), the graph of W is given along with the N(0, 1) p.d.f. for n = 20 and 100, respectively. The next example will also help to give the reader a better intuitive feeling about the Central Limit Theorem. We shall again see how the size of n affects the distribution of W = (X - ^/(a/^/n). Example 5.3-6 It is often difficult to find the exact distribution of W = (X — distribution of W by simulating random samples on the computer. Let A";, A" 2, ..., X^ denote a random sample of size n from the distribution with p.d.f./(x), mean ^i, and variance (T2. We shall generate 1000 random samples of size n from this distribution and compute a value of W for each sample, thus obtaining 1000 observed values of W. A histogram of these 1000 values is constructed using 21 intervals of equal length. We depict the results of this experiment as follows: (1) In Figure 5.3-3, f ( x ) = (x + 1)/2, - 1 < x < 1, fi = 1/3, a2 = 2/9, for n = 2 and 12 (see Exercise 8.5-10). (2) In Figure 5.3-4,/(x) = (3/2)x2, - 1 < x < 1, ii = 0, <r2 = 3/5, for n = 2 and 12 (see Exercise 8.5-9). ; The N(0, 1) p.d.f. has been superimposed on each histogram. , Note very clearly that these examples have not proved anything. They are presented to give evidence of the truth of the Central Limit Theorem. So far all the illustrations have concerned distributions of the continuous type. However, the hypotheses for the Central Limit Theorem do not require the 310 I The Central Limit Theorem distribution to be continuous. We shall consider applications of the Central Limit Theorem for discrete-type distributions in the next section. Exercises 5.3-1 Let X be the mean of a random sample of size 12 from the uniform distribution on the interval (0, 1). Approximate P(l/2 < X < 2/3). 311 5.3-2 Let Y = A\ + X^ + - - • the distribution whose p.d.f. is/fx) = (3/2)x2, — P ( - 0 . 3 ^ Y ^ 1.5). 5.3-3 Let X be the mean of a random sample of size 36 from an exponential distribution with mean 3. Approximate P(2.5 < X < 4). 312 The Central Limit Theorem 0.4 f 7 \/-~ -3.0 -2.4 -1,8 1 \] 0.2 - / y ^< '\WO, l) ^ / ^ ^ -1,2 -0.6 n= 12 ( b) 0.6 1.2 >^ 2.4 1.8 3,0 F igur e 5 3 -4 53-4 Approximate F(39.75 $ X $ 41.25), where X is the mean of a random sample of size 32 from a distribution with mean f i = 40 and variance o-2 = 8. 5.3-5 Let Xj, X;, ..., X ^ y be a random sample of size 18 from a chi-square distribution with r = 1. Recall that f t = 1, a2 = 2. (a) How is V = ^ X , distributed? 313 Sampling Distributions and Approximations Ch. 6 (b) Using the result of part (a), we see from Table V in the Appendix that P{Y ^ 9.390) = 0.05 and P(Y < 34.80) = 0.99. Compare these two probabilities with the approximations found using the Central Limit Theorem. 53-6 A random sample of size n = 18 is taken from the distribution with p.d.f. f ( x ) = 1 - x/2,0 < x < 2. (a) Find ^ and ff 2 . (b» Find, approximately, P(2/3 £ 5.3-7 Let X equal the maximal oxygen intake of a human on a treadmill, where the measurements are in milliliters of oxygen per minute per kilogram of weight. Assume that for a particular population the mean of X is f i = 54.030 and the standard deviation is a = 5.8. Let X be the sample mean of a random sample of size n = 47. Find P(52.761 < X ^ 54.453), approximately. 53-8 Let Xt, X ^ , . . . , X^ be a random sample of size n = 25 from a population that has a mean of f i = 71.43 and a variance of a2 = 56.25. Let X be the sample mean. Find (a) E{X\_ (b) Var(X). (c) P(68.91 < X ^ 71.97), approximately. 53-9 Let X equal the birth weight in grams of a baby born in the Sudan. Assume that E(X) = 3320 and Var(Z) = 6602. Let X be the sample mean of a random sample of size n = 225. Find P(3233.76 s£ X ^ 3406.24), approximately. 5.3-10 In Example 5.3-4, compute P(1.7 ^ Y <. 3.2) and compare this answer with the normal approximation of this probability. 5.4 Approximations for Discrete Distributions In this section we illustrate how the normal distribution can be used to ; approximate probabilities for certain discrete-type distributions. One of the i most important discrete distributions is the binomial distribution. To see how | the Central Limit Theorem can be applied, recall that a binomial random | variable can be described as the sum of Bernoulli random variables. That is, | let -YI, X ^ , . . . , X^ be a random sample from a Bernoulli distribution with a | mean ju = p and a variance a2 = p(l — 314 Approximations for Discrete Distributions Figure 5.4-1 is b(n, p). The Central Limit Theorem states that the distribution of ,/np(l - p) ,/p(l - p)/n is N(0, 1) in the limit as n -»• co. Thus, if n is "sufficiently large," probabilities for the binomial distribution b{n, p) can be approximated using the normal distribution N[np, np(l — large " if np > 5 and n(l — Note that we shall be approximating probabilities for a discrete distribution with probabilities for a continuous distribution. Let us discuss a reasonable procedure in this situation. For example, if V is N{^1, a2), P(a < V < b) is equivalent to the area bounded by the p.d.f. of V, the v axis, v = a, and y = b. If V is b(n, p), recall that the probability histogram for Y was denned as follows. For each y such that k — ^-fe)-^-^' fe =o'l^•••'"• Then P(Y == k) can be represented by the area of the rectangle with a height of P(Y = I;) and a base of length 1 centered at k. Figure 5.4-1 shows the graph of the probability histogram for the binomial distribution b(3, 1/3). When using the normal distribution to approximate probabilities for the binomial distribution, areas under the p.d.f. for the normal distribution will be used to approximate areas of rectangles in the probability histogram for the binomial distribution. Example 5.4-1 Let Y be t>(10, 1/2). Then, by the Central Limit Theorem, P(a < Y < b) can be approximated using the normal distribution with mean 10(1/2) = 5 and variance 10(1/2)(1/2) = 5/2. Figure 5.4-2 shows the graph of the probability histogram for h(10, 1/2) and the graph of the p.d.f. of 315 Sampling Distributions and Approximations 0.15- / / 0.10 1 \ \ ^ 0.05- - W,}) / ^ 0.20- 0 ] 2 3 4 5 6 7 \ ^ 8 ^ , 9 10 11 Figure 5.4-2 N(5, 5/2). Note that the area of the rectangle whose base is H'*l) and the area under the normal curve between k - 1/2 and k + 1/2 are approximately equal for each integer k. Example 5.4-2 Let Y be fc(18, 1/6). Because up = 18(1/6) = 3 < 5, the normal approximation is not as good here. Figure 5.4-3 illustrates this by depicting the probability histogram for i)(18, 1/6) and the p.d.f. of N(3, 5/2). 0.05 ^ 4 5 6 Figure 5.4-3 316 7 8 9 10 5.4 Approximations for Discrete Distributions Example 5.4-3 Let Y have the binomial distribution of Example 5.4-1 and Figure 5.4-2, namely 6(10, 1/2). Then P(3 < Y < 6) = P(2.5 < Y s; 5.5) because P(Y = 6) is not in the desired answer. But the latter equals Pi2'5——5 < Y—^ <5'5—5) » \^10/4 yio/4 yi0/4/ = 0.6240 - 0.0570 = 0.5670. Using the binomial formula, we find that P(3 < Y < 6) = 0.5683. Example 5.4-4 Let Y be 6(36, 1/2). Then P(12 < V <; 18) = P(12.5 < V < 18.5) /12.5-18 /12.5 - 18 'r - 18 18.5 - 1S\ \ ^9 - ^9 ^ ) "i-^^ »1>(0.167)- W, -1.833) = 0.5329. Note that 12 was increased to 12.5 because P(Y = 12) is not included in the desired probability. Using the binomial formula, we find that P(12 < Y < 18) = P(13 < V < 18) = 0.5334. Also, P(Y = 20) = P(19.5 < V < 20.5) _/19.5 - 18 V - 18 20.5 - 18\ =P — — j = — < ——-/=- < ——T=— \ ^9 ^/9 ^/9 1 w 1)(0.833) - <D(0.5) = 0.1060. Using the binomial formula, we have P(Y = 20) = 0.1063. 317 Sampling Distributions and Approximations Ch. 5 Note that, in general, if Y is f)(n, p), (k + l / 2 - n p \ P(Y <. k) » ^-, and P(Y < k) - np\ ^k - 1/2 fnpq I We now show how the Poisson distribution with large enough mean can be approximated using a normal distribution. Example 5.4-5 A random variable having a Poisson distribution with mean 20 can be thought of as the sum Y of the items of a random sample of size 20 from a Poisson distribution with mean 1. Thus w^^ ,/20 has a distribution that is approximately N(0, 1), and the distribution of Y is approximately N(20, 20) (see Figure 5.4-4). So, for illustration, P(16 < V S 21) = P(16.5 <. Y < 21.5) _ A6.5 - 20 \ ^20 ss y-20 ^/20 - 21.5 - 20\ V'20 / »(1(0.335)-<I'(-0.783) = 0.4142. Note that 16 is increased to 16.5 because Y = 16 is not included in the event 16 < r < 21. The answer using the Poisson formula is 0.4226. In general, if Y has a Poisson distribution with mean ^, then the distribution of " ^ is approximately N(0, 1) when ). is sufficiently large. Exercises 5.4-1 Let the distribution of Y be b(25, 1/2). Find the following probabilities in two ways: exactly using Table II and approximately using the Central Limit Theorem. Compare the two results in each of the three cases. 318 Approximations for Discrete Distributions I7 \ \1 / \t / 12 14 ^ 6 8 10 / 7 \ \ \ Poisson, \ = 20 \A' (20,20) \ \ \ 16 18 20 22 24 26 ,'b-^- 28 30 32 34 Figure 5.4-4 (a) P(10 < V ^ 12), (b) P(12< r<15), (c) P(V = 12). 5.4-2 Let Y be &(18,1/3). Use the normal distribution to approximate (a) P(4 < V < 6), (b) P(5 < r < 8), (c) P(Y = 7). 5.4-3 Suppose that 20% of the American public believes that we could survive a nuclear war. Let Y equal the number of people in a random sample of size n = 25 who believe this way. Use the Central Limit Theorem to find approximately P(6 < V < 9) and compare this probability with that found using Table II. Since b(25, 0.2) is a skewed distribution, the approximation is not as good as for the symmetric binomial distribution (i.e., when p = 0.5), 5.4-4 A die is rolled independently 240 times. Approximate the probability that (a) more than 40 rolls are fives, (b) the number of twos and threes is from 75 to 83, inclusive. 5.4-5 Let X ^ , X ^ , .,., X^ be a random sample of size 48 from the distribution with p.d.f./(x) = 1/x2, 1 < x < oo. Approximate the probability that at most 10 of these random variables have values greater than 4. HINT : Let the dh trial be a success if X , > 4, i = 1,2,..., 48. 319 Sampling Distributions and Approximations Ch. 5 5.4-6 A candy maker produces mints that have a label weight of 20.4 grams. Assume that the distribution of the weights of these mints is N{21.37, 0.16). (a) Let X denote the weight of a single mint selected at random from the production line. Find P(X < 20.857). (b) During a particular shift 100 mints are selected at random and weighed. Let Y equal the number of these mints that weigh less than 20.857 grams. Find approximately P(Y < 5). (c) Let X equal the sample mean of the 100 mints selected and weighed on a particular shift. Find P(21.31 ^ X ^ 21.39). 5.4-7 Let X equal the number of alpha particles emitted by barium-133 per second and counted by a Geiger counter. Assume that X has a Poisson distribution with X = 49. Find approximately P(45 < X < 60). 5.4-8 Let X equal the number of alpha particles counted by a Geiger counter during 30 seconds. Assume that the distribution of X is Poisson with a mean of 4829. Find approximately P(4776 -£ X <. 4857). 5.4-9 Let A"], X-i, ..., X^o be a random sample of size 30 from a Poisson distribution with a mean of 2/3. Approximate (a) P ( l 5 < £-y,£22), (b) (c) < < 30 \ 21^ S^-<27|, 30 \ 15 £ i==1 / 5.4-10 In the casino game of roulette, the probability of winning with a bet on red is p = 18/38. Let V equal the number of winning bets out of 1000 independent bets that are placed. Find approximately P(Y > 500). 5.4-11 Let X denote the payoff for $1 bets in the game of chuck-a-luck. Then X can equal -1, 1, 2, or 3, such that E(X) = -0.0787 and Var(X) = 1.2392. Given 300 independent observations of X , say X , , X^ ..., X ^ , let Y = ^?^ X,. Then Y represents the amount won after 300 bets. Find approximately (a) P ( V ^ -21), (b) P(r>21). 5.4-12 Let Y equal the number of $499 prizes won by a gambler after placing n straight independent $1 bets in the Michigan daily lottery, in which a prize of $499 is won with probability 0.001. Thus V is b{n, 0.001). If n = 4200, a gambler is behind if V <. 8; use the Poisson distribution to approximate P(Y < 8). 5.4-13 If X is b{l00,0.1), find the approximate value of P(12 < X ^ 14) using (a) the normal approximation, (b) the Poisson approximation, (c) the binomial p.d.f. 320 5,5 Limiting Moment-Generating Functions 5.4-14 Let X ^ , X ^ , . . . , X^ be a random sample of size 36 from the geometric distribution with p.d.f./M = (1/4)*' '(3/4), x = 1, 2, 3, . . . . Approximate (a)P('46£ ^X,£49\ (b) P(1.25 £ HINT: Observe that the distribution of the sum is of the discrete type. S.4-15 A die is rolled 24 independent times. Let V be the sum of the 24 resulting values. Recalling that V is a random variable of the discrete type, approximate (a) P(Y 2 86), (b) P(V < 86), (c) P(70 < y < 86). *5.5 Limiting Moment-Generating Functions We would like to begin this section by showing that the binomial distribution can be approximated by the Poisson distribution when n is sufficiently large and p fairly small by taking the limit of a moment-generating function. Consider the moment-generating function of Y, which is b{n, p). We shall take the limit of this as n-> oo such that up =^isaconstant;thusp-»O.The momentgenerating function of Y is M ( f ) = (1 - p + pe')'. Because p = ^/n, we have that M^fl-^c'T L " " J =[i^]" Since limfl+''Y=^ we have lim M(() = e" 321 Sampling Distributions and Approximations Ch. 5 which exists for all real (. But this is the moment-generating function of a Poisson random variable with mean L Hence this Poisson distribution seems like a reasonable approximation to the binomial one when n is large and p is small. This approximation is usually found to be fairly successful if n > 20 and p < 0.05 and very successful if n > 100 and np < 10. Obviously, it could be used in other situations too; we only want to stress that the approximation becomes better with larger n and smaller p. Example 5.5-1 Let Y be fc(50, 1/25). Then P(^D=V ^oyy =0.400. /24\50 / 1 \/24\49 Since 2 = np = 2, the Poisson approximation is P(Y < 1)% 0.406, from Table III in the Appendix. The preceding result illustrates the theorem we now state: If a sequence of moment-generating functions approaches a certain one, say M(t), then the limit of the corresponding distributions must be the distribution corresponding to M(i). This statement certainly appeals to one's intuition! In a more advanced course, the proof of this theorem is given and there the existence of the moment-generating function is not even needed, for we would use the characteristic function ^i(t) = £(e'") instead. The preceding theorem is used to prove the Central Limit Theorem. To help in the understanding of this proof, let us first consider a different problem, that of the limiting distribution of the mean X of a random sample X^, X ^ , . . . , Xy from a distribution with mean a. If the distribution has moment-generating function M(r), the moment-generating function of X is [M(t/n)]". But, by Taylor's expansion, there exists a number ti between 0 and t/n such that M ( " ) = M(0) + M'(ti) L W " _ ^ at ^ [M'(ti) - M'(0)]t because M(0) = 1 and M'(0) = in. Since M'(t) is continuous at t = 0 and since ti -»0 as n -* oo, we know that lim [M'(ti) - M'(0)] = 0. 322 5.5 Limiting Moment-Generating Functions Thus, using a result from advanced calculus, we obtain ,-hm r./^M" r \ J, 1 +^— [M'(f,) - M'{0)-]t\" M = hm .-a, L WJ n^ [ n n J =e"', for all real t. But this limit is the moment-generating function of a degenerate distribution with all of the probability on ^.. Accordingly, X has this limiting distribution, indicating that X converges to ^ in a certain sense. This is one form of the law of large numbers. We have seen that, in some probability sense, X converges to ^ in the limit, or, equivalently, X — by some function of n so that the result will not converge to zero. In our search for such a function, it is natural to consider ^ _ X - fi _ ^/n{X -^ ^Y-n^i a/^/n o ^nrr ' where Y is the sum of the items of the random sample. The reason for this is that W is a standardized random variable and has mean 0 and variance 1 for each positive integer n. Proof of the Central Limit Theorem. We first consider W'1-'}}} / ...„-^...„l • [(W\-'[(7^1 -«.„ - ^ - . ...../• v^ [(W}}'W^^ £[exp (tW)-] == E^exp ( — — — ) ( ^ X, - nfi I L\^/na/\i= which follows from the mutual independence of X i , X ^ , . . . , X ^ . Then a .[expW] =[.(-)]", where m(t) = f{exp r/^^——^ll •[<' -„<-<„, -h < t < h, 323 Sampling Distributions and Approximations Ch. is the common moment-generating function of each ^^, ,=l,2,...,n. (T Since £(^) = 0 and £(Y?) = 1, it must be that m(0) = 1, m'(0) = E \ ' ~ 1 } = 0, \ " / m"(0) = E\ (x!—p-\ = 1. L\ " ! J Hence, using Taylor's formula with a remainder, we can find a number t between 0 and t such that .(0=.(0)+»w+'"^=l+''f(il)t2. By adding and subtracting t'A we have that ^^^m^y. Using this expression of m(t) in £[exp (tW)], we can represent the momeni generating function of W by £[exp (tW)-] = {l + ' (——) + \ [m"(ti) - l]f——Y}" I '• \^/n/ -' \Vn/ J f , t 2 [""(ti) - lit 2 ')" = ^ 1 + — + — — — — — — > , ^ In 2n J r, ^, -^nh<t<^nh, where now t^ is between 0 and i/^/n. Since ffi"(t) is continuous at t = 0 an r; -* 0 as n -> oo, we have that lim [m"((i) - 1] = 1 - 1 = 0. Thus, using a result from advanced calculus, we have that limECexp^fl^im^l^'"""-'-1^ n--co n-oi {, in In J ^Ji^l"^, .-„ l " J 324 5.5 Limiting Moment-Generating Functions for all real (. This means that the limiting distribution of ^ ^ Y n i^.-^ x - ^ _i=l_______ a/^/n ^/no- is N(0, 1). This completes the proof of the Central Limit Theorem. Q Examples of the use of the Central Limit Theorem as an approximating distribution have been given in Sections 4.5 and 4.6. Exercises 5.5-1 Let V be the number of defectives in a box of 50 articles taken from the output of a machine. Each article is defective with probability 0.01. What is the probability that V = 0,1, 2, or 3 (a) using the binomial distribution? (b) using the Poisson approximation? 5.5-2 The probability that a certain type of inoculation takes effect is 0.995. Use the Poisson distribution to approximate the probability that at most 2 out of 400 people given the inoculation find that it has not taken effect. HINT: Let p = 1 - 0.995 = 0.005. 5.5-3 Let S2 be the sample variance of a random sample of size n from N(^i, a2). Show that the limit, as n -* GO, of the moment-generating function of S2 is e"2'; thus, in the limit, the distribution of S1 is degenerate with probability 1 at o-2. 5.5-4 Let V be ^(n). Use the Central Limit Theorem to demonstrate that W = {Y - n)A/2n has a limiting distribution that is JV(0, • HINT: Think of Y as being the sum of a random sample from a certain distribution, 5.5-5 Let Y have a Poisson distribution with mean 3n. Use the Central Limit Theorem to show that W = (Y — 325