Download Central Limit Theorem (as an “experimental fact”)

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

History of statistics wikipedia , lookup

Bootstrapping (statistics) wikipedia , lookup

Taylor's law wikipedia , lookup

Central limit theorem wikipedia , lookup

Student's t-test wikipedia , lookup

Transcript
11.4 CENTRAL LIMIT THEOREM
A theorem is a statement of some mathematical property or principle that
can be proven by logical arguments. The central limit theorem is a very
important theorem in statistics. We will not prove it in this book, but we will
look at examples that demonstrate what it tells us and its great usefulness.
Central Limit Theorem (as an “experimental fact”)
Suppose we take a very large number of random samples, each of size n, from a population with mean ␮ and standard deviation ␴ .
Then if the sample size n is relatively large
(n ⱖ 20), the shape of the relative frequency
histogram of the sample means X will ap-
proximate the shape of a normal density
whose mean is ␮ and whose standard deviation is ␴ /冪n. Equivalently, the relative
frequency histogram of (X ⫺ ␮ )/(␴ / 冪n) will
approximate the standard normal density
tabulated in Appendix E.
The fact that the relative frequency histogram of a very large number
of Xs must be close to an appropriately scaled normal density provided n,
the sample size used to form each X, is relatively large (n ⱖ 20) amounts to
asserting that we can use the normal distribution to calculate probabilities
involving X as if X were distributed exactly normally with mean ␮ and
standard deviation ␴ / 冪n. Thus the experimental fact claimed in the above
statement of the central limit theorem translates into the practical result
that we can compute probabilities involving an X based on a sample of at
least 20 using the normal distribution regardless of the shape of the original
population the Xs were sampled from. Thus to compute p(X ⱕ x) we let
z ⳱ (x ⫺ ␮ )/(␴ / 冪n) and enter Table E using this value of z. Here ␮ and
␴ are the population mean and standard deviation and n is the sample size
used to obtain X.
Example 11.8
Using a box model of size 200, we constructed a nonnormal population having
a population mean ␮ ⳱ 63.5 and a population standard deviation ␴ ⳱ 12.0. We
then randomly drew 100 samples of size 4, 16, and 36 with replacement from this
population, and we calculated the sample mean for each of the 100 samples at each
sample size. The resulting sample means are presented in the stem-and-leaf plots
in Tables 11.1, 11.2, and 11.3. From these stem-and-leaf plots we constructed the
relative frequency histograms of Figures 11.2, 11.3, and 11.4.
Let’s make some observations about the results of our sampling as we look at
these stem-and-leaf plots and the associated relative frequency histograms.
1. The “grand” mean of the 100 sample means is in all three cases close to the
population mean of 63.5. For example, the mean of the 100 sample means of
size 4 is 63.3, as can be computed from Table 11.1. This is what the central limit
theorem tells us to expect because it says that the relative frequency histogram
of 100 Xs will be close to a normal density with mean ␮ .
2. As the sample size increases from n ⳱ 4, a number much too small for the
central limit theorem to apply, to n ⳱ 16, a number for which the central limit
theorem should begin to hold, the relative frequency histogram goes from a
very non-normal shape in Figure 11.2 to a fairly normal shape in Figure 11.3.
Similarly, Figure 11.3 appears somewhat more bell-shaped. This is what the
central limit theorem tells us to expect.
3. As the sample size increases, the variation among the sample means decreases.
More specifically, as the sample size increases, the standard error (the standard
deviation of the sample means) decreases from 5.7 to 2.9 to 2.0 as n increases
from 4 to 16 to 36. This is precisely the kind of behavior that ␴ /冪n in the central
limit theorem predicts. Indeed, the calculated value of the standard error found
by computing the standard deviation of the 100 sample means for each sample
size is close to the value of ␴ /冪n asserted for the normal density in the central
limit theorem. For example, the calculated standard error from Table 11.1 is 5.7,
compared with ␴ / 冪n ⳱ 6.0, while the calculated standard error from Table
11.3 is 2.0, compared with ␴ /冪n ⳱ 2.0—an exact (and lucky) match.
The central limit theorem is a very powerful tool in statistical analysis.
In the present context we will use the central limit theorem to calculate
confidence intervals for population means, for proportions, and for the
difference between two means.
Sample Means of
Table 11.1
100 Samples of Size 4 (Population
Values: ␮ ⳱ 63.5, ␴ ⳱ 12.0)
Stem
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
Leaf
8
7
1,5,6
2,3,3,5,6,7
4,6,6,6,9
1,1,1,3,8,9
1,3,5,8,8,9
1,1,3,4,5,5,6,8
3
3
2,2,5,5,8
2,2,4,5,5,5,7,9
0,0,0,2,4,7,9,9
1,1,3,3,4,7
1,3,5,5,6,6,9,9
0,2,7,9
0,2,3,5
3,5,5,6,6,8
8
1,2,4,5,6,6,8,9
2,9
1
3
Frequency
0
0
1
1
3
0
6
5
6
6
8
1
1
5
8
8
6
8
4
4
6
1
8
0
2
1
1
0
0
Total 100
Sample values:
n⳱4
mean ⳱ 63.3
standard error ⳱ 5.7
Table 11.2
Sample Means of 100 Samples of Size 16
(Population Values: ␮ ⳱ 63.5, ␴ ⳱ 12.0)
Stem
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
Leaf
Frequency
0
0
1
1
0
8
6
7
11
25
10
8
6
6
8
1
1
0
1
0
0
9
7
0,2,2,5,6,6,7,9
0,2,5,7,7,7
1,2,5,8,9,9,9
0,1,3,3,4,5,5,6,7,8,8
2,2,2,2,3,3,3,3,3,4,4,4,4,4,5,6,6,7,7,8,8,9,9,9,9
2,2,3,4,4,5,6,6,7,8
0,0,0,3,3,3,5,7
0,1,2,4,4,5
0,4,4,5,6,7
2,3,3,4,5,6,6,9
5
2
1
Sample values:
n ⳱ 16
mean ⳱ 63.9
standard error ⳱ 2.9
Total 100
0.08
0.07
0.06
0.05
0.04
0.03
0.02
0.01
50
55
60
65
70
75
Figure 11.2
Relative frequency histogram of sample means
of 100 samples of size 4.
80
Table 11.3
Sample Means of 100 Samples of Size 36
(Population Values: ␮ ⳱ 63.5, ␴ ⳱ 12.0)
Stem
56
57
58
59
60
61
62
63
64
65
66
67
68
69
Leaf
Frequency
1,6,9
7,8,9
0,0,3,6,7,7,7,8,8,9
0,1,2,4,5,8,9
0,0,2,2,2,2,4,5,6,6,7,7,7,8,8,8
1,1,1,1,2,3,3,4,4,5,5,7,8,9
0,0,1,1,2,2,2,3,3,4,4,4,4,4,4,5,5,6,6,6,7,8,8
0,0,0,2,3,3,4,4,4,4,5,5,8,9
0,3,4,5,5,8,8
1,2,4
0
0
3
3
10
7
16
14
23
14
7
3
0
0
Total 100
Sample values:
n ⳱ 36
mean ⳱ 63.4
standard error ⳱ 2.0
0.19
0.18
0.16
0.14
0.12
0.10
0.08
0.06
0.04
0.02
50
60
70
Figure 11.3
Relative frequency histogram of sample means
of 100 samples of size 16.
80
0.25
0.20
0.15
0.10
0.05
50
60
70
80
Figure 11.4
Relative frequency histogram of sample means of
100 samples of size 36.
SECTION 11.4 EXERCISES
1. Suppose the mean weight of all male Americans is 175 pounds with a standard deviation
of 15 pounds. What are the estimated theoretical mean and standard deviation of a sample
mean from this population based on 50 observations? How is the sample mean distributed?
2. Suppose the sample mean in Exercise 1 is
based on 10 observations instead of 50. Can
you still apply the central limit theorem to
answer the questions in Exercise 1? Explain
your answer.
3. Which relative frequency histogram will look
more like a normal distribution—the histogram of sample means based on 20 observations per mean or the histogram based on
40 observations per mean? (Assume that the
samples are taken from the same population).
4. What are the mean and standard deviation
of a sample mean of 40 observations based
on a population with mean 0 and standard
deviation 1? What is the distribution of the
sample mean approximately like?