Download 11.2 OINT ESTIMATES AND CONFIDENCE INTERVALS Point

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Taylor's law wikipedia , lookup

Bootstrapping (statistics) wikipedia , lookup

Resampling (statistics) wikipedia , lookup

Misuse of statistics wikipedia , lookup

German tank problem wikipedia , lookup

Transcript
11.2 POINT ESTIMATES AND CONFIDENCE INTERVALS
Point Estimates
Suppose we want to estimate the proportion of Americans who approve of
the president. In the previous section we took a random sample of size 1200
from the population and used the proportion of the people in the sample
who approved of the president to estimate the proportion of the people in
the entire country who approve of the president. The proportion of people
in the sample who approve of the president is an example of what is called
a point estimate. It is called a point estimate because it is a single number
that estimates a population parameter, here the proportion of the people in
the country who approve of the president.
Example 11.3
Suppose that out of 1200 people in the sample, 615 approve of the president. The
point estimate of the proportion of the people in the country who approve of the
president is 615/1200 ⳱ 0.5125.
Example 11.4
Suppose that in the example of the 10 containers of sampled water, the sample
average of the 10 observed E. coli bacteria densities is 1500/cm2 . The point estimate
of the density of E. coli bacteria in the swimming pool is 1500/cm2 .
Example 11.5
It was desired to find the population mean and standard deviation of the ages
of the students enrolled at a particular community college. It was impossible to
survey all the students, so a random sample of 100 students was taken and the
mean and the standard deviation of the ages of the students in the sample were
calculated. The sample mean was 20.25 and the sample standard deviation was
2.25. The sample mean age, X ⳱ 20.25, is a point estimate of the population mean
age, ␮ . The sample standard deviation of the ages, S ⳱ 2.25, is a point estimate of
the population standard deviation, ␴ .
In summary, a point estimate for a population parameter of interest
is a statistic computed from the sample. It is believed to be an effective
estimate of the unknown population parameter. Often, what formula to
use for the statistic is obvious. For example, for the population parameter
␮ , which is the mean over a (presumably large) population, it seems that
the sample mean will be a good estimate. However, in more complex
situations the choice of an estimate of a population parameter is not always
clear. Statisticians may find that either they have no idea how to use the
sample data to estimate the population parameter of interest, or they may
in fact have several equally plausible competing estimates to select from.
A relatively simple example of the latter case is the estimation of ␴ 2 . Let n
denote the sample size. Many statisticians would use S2 ⳱ sum(Xi ⫺ X)2 /n,
while many others would use sum(Xi ⫺ X)2 /(n ⫺ 1). Here Xi denotes the
ith observation of the random sample. A solid argument can be made for
either estimate. In fact, a significant amount of statistical theory is devoted
to finding the “best” estimate of a population parameter.
Let’s consider an example in which the formula for a good estimate is
not obvious.
Example 11.6
Suppose the waiting time for a train that goes to the parking area from a particular
terminal at Chicago’s O’Hare International Airport obeys a continuous uniform
distribution on the time interval [0, T], where T is an unknown population parameter. Thus a train always arrives within T minutes, but it is equally likely to arrive
at any moment during this waiting time. Suppose we interview five randomly
selected passengers and find that their waiting times were (in minutes) 1.4, 4.7, 4.2,
5.1, and 2.1. We want to estimate the parameter T. Clearly the sample mean is a poor
choice for an estimate of T. What formula should we use? A widely used technique
for producing an estimate when we have none in mind, called maximum likelihood
estimation, leads us to an estimate that is equal to the maximum of the five observed
times, which is 5.1 minutes. This choice of the maximum of the observed values as
our point estimate is not obvious. Even though it might not be the “best” choice
for an estimate of T, it seems clearly better than X. We will not consider maximum
likelihood estimation in this book.
A point estimate reports one single value that we estimate to be the
true value of the population parameter. However, point estimates have the
important limitation of not informing us how much the estimate is likely
to be in error. Whenever we estimate a population parameter, we lack
total accuracy. Thus our estimate will almost always be different from the
actual population parameter. So the estimate will almost always have some
amount of error associated with it. By simply reporting the point estimate
of a parameter, we have essentially ignored the important issue of the likely
error size associated with the estimate.
Example 11.7
Reconsider Example 11.5. A second random sample, this time of 200 students, was
taken. Suppose, to keep our explanation simple, the point estimate of the population
mean is the same in both cases, namely 20.25. The standard deviation of the second
sample mean would be much lower than the standard deviation of the first mean,
because the second sample is larger. So even though the two point estimates of the
population mean are the same, the second one is surely more accurate, because it is
based on more information—namely, twice as many observations.
Confidence Intervals
To improve on point estimates, statisticians usually report an interval of
values that they believe the parameter is highly likely to lie in. Usually the
point estimate is the middle point of the interval and the endpoints of the
interval communicate the size of the error associated with the estimate (recall
that point estimates ignore this error) and how “confident” we are that the
population parameter is in the interval. The intervals are called confidence
intervals. Typical confidence levels used in practice for confidence intervals
are 90%, 95%, or 99%, with 95% occurring most frequently in applications.
In Examples 11.1 and 11.3, if we are given a 95% confidence interval for the
proportion of the population that approves of the president, which can be
shown to be the interval (0.48, 0.54), we say we are 95% confident that the
population proportion is contained in the confidence interval. In Examples
11.2 and 11.4, given a 90% confidence interval for the density of bacteria
in the swimming pool, (1490, 1510) say, we say we are 90% confident that
the density of bacteria in the swimming pool is contained in the confidence
interval. (We will learn how to calculate confidence intervals for different
population parameters later in the chapter.)
What does it really mean to state a 95% confidence interval for the
unknown population proportion approving of the president? Although
95% sounds impressive, we cannot be satisfied unless we understand what
it means. Theoretically, it means that the probability is 0.95 that such a
confidence interval, which will be random because the sample it is formed
from is random, will contain (surround) the unknown proportion in the
population approving of the president.
Our experimental view of probability based on the five-step method
will help us more clearly and deeply understand what this probability of
0.95 means practically. Just as we do simulations over and over in the
five-step method, imagine that a statistician does the sampling experiment
of Example 11.1 over and over—1000 times, say—and each time computes
a 95% confidence interval from the 1200 sampled people. Now we can
find the experimental probability given by the proportion of the 1000
confidence intervals that actually covers the true fraction of the population
favoring the president. Since 0.95 is the theoretical probability of the interval
containing the population proportion, this experimental probability of the
1000 confidence intervals including the true value will also be close to 0.95.
(Below, we will simulate 100 such 95% confidence intervals and calculate
the experimental confidence interval probability.)
Of course, in a real application there will only be one random sample
and hence only one such confidence interval, such as the (0.48, 0.54)
interval of Examples 11.1 and 11.3. But the statistician obtaining this one
sample knows, because of the experimental probability viewpoint, that this
confidence interval is very likely to be correct in the sense that it contains
the true value (since about 95% of such confidence intervals would cover
the true population proportion). In the case of the (0.48, 0.54) interval of
Examples 11.1 and 11.3, we know it is very likely that the true proportion of
people favoring the president lies between 0.48 and 0.54.
Now in light of this insight into how to interpret the confidence interval
percentage, let’s return to the Key Problem. The St. Louis Post Dispatch
explained the concept of its reported confidence interval this way: “[A 95%
confidence interval] means if the survey were taken 100 times, the results
for the [random] group of respondents would each vary no more than 5.7
percent in either direction from the true population percentage opposing
the stadium [about] 95% of these times.” The Post Dispatch quote is a bit
roundabout and hence forces us to go through a slightly tricky piece of
logic (draw yourself a picture if needed). If the interval, which extends 5.7%
in either direction from its midpoint, indeed varies no more than 5.7% in
either direction from the true population mean about 95% of the time, then
about 95% of these intervals must contain the true population percentage as
desired. The Post Dispatch could have more simply and more directly told its
readers that such an interval can be expected to contain the true population
parameter about 95% of the time.
In other words, just as explained above in the presidential popularity
example, the Post Dispatch is pointing out that if you take 100 random
samples from the same population and calculate the confidence interval for
the population proportion for each sample, about 95% of the confidence
intervals will include the true population proportion. That is, you will be
correct in your claim that the unknown population proportion is in the
interval computed using the sample for about 95 of the 100 samples.
Using our five-step method, we now simulate 100 confidence intervals and determine how many of them contain the true population parameter. Suppose that for the Key Problem the true proportion of people in
the county who are opposed to the stadium is 50% (remember that the parameter value is never known to the statistician). In that case the probability
that a person in the sample will be opposed to the stadium is the same as
the probability of heads seen in flipping a fair coin: 0.5. Our goal is to obtain
100 samples of 301 people and calculate the confidence interval for each of
these 100 samples (we will learn to compute such confidence intervals in
Section 11.6 below). Each sampling is the same as flipping a coin 301 times,
recording the number of heads, and calculating the confidence interval for
the proportion of heads in the sample. We then repeat the process 100 times.
The 100 confidence intervals obtained from this process are represented in
Figure 11.1.
0.65
0.60
0.55
0.50
0.45
0.40
0.35
Figure 11.1
One hundred simulated 95% confidence intervals for the Key Problem
assuming a 50/50 population split.
The line across the middle of the graph represents the true population
proportion, 0.50. The 100 confidence intervals are the vertical lines on
the graph. If the confidence interval covers the 0.50 line, then we say that
the true population proportion is contained in the interval. Likewise, if the
confidence interval does not cover the 0.50 line, then we say that the true
population proportion is not contained in the interval. For the graph of
Figure 11.1, we see that 96 out of the 100 confidence intervals cover the
0.50 line. So for the 100 confidence intervals, 96% of them (about 95%, as
expected) contain the true population parameter, 0.50.
Thus, this example of the five-step method clearly illustrates how we are
to correctly interpret a 95% confidence interval. As already discussed above,
it is a sort of statisticians’ success rate or batting average. If a statistician
constructs 100 95% confidence intervals during a year’s work, then, as
our five-step simulation confirms, we can expect about 95% of them to be
“hits”: cases in which the population parameter is contained in the interval.
Similarly we would expect about 5% of them to be “outs,” or misses: cases in
which the population parameter is not contained in the interval. Compared
with baseball, in which a batting average of 0.300 is considered great, a 0.950
confidence interval coverage rate for statisticians is what is usually required.
Batting 0.300 is never guaranteed in baseball, but in fact a statistician can
guarantee a 95% confidence interval, as we shall see.
Confidence intervals have two basic characteristics that we need to
understand. First, given the same set of data, a 95% confidence interval
is wider than a 90% confidence interval, and a 99% confidence interval is
wider than a 95% confidence interval. Thus, the higher the confidence level
we require, the wider the interval we are forced to accept! Of course, a very
wide interval is of little use to the scientist who has sought statistical advice.
Thus there is no “free lunch” in specifying a 99% confidence instead of a
95% confidence, because the price paid is a wider interval. Here are the
90%, 95%, and 99% confidence intervals for the proportion of people in the
country who approve of the president in Example 11.1:
90%: (0.49, 0.53)
95%: (0.48, 0.54)
99%: (0.47, 0.55)
As you can see from this example, the higher the level of confidence, the
wider the confidence interval needs to be in order to contain the population
proportion with the specified confidence.
The second characteristic of confidence intervals is that, given the
same confidence level, a shorter and hence more informative confidence
interval is associated with more data points. Suppose in the situation of
Example 11.1 a sample of 2400 people was taken from the population and
the number of people in the sample that approved of the president was
1230. The point estimate of the population proportion would be the same:
1230/2400 ⳱ 0.5125. However, a 95% confidence interval for the population
proportion based on this sample of 2400 can be shown to be (0.4925, 0.5325).
This is shorter than the 95% confidence interval for the original sample
of 1200, which can be shown to be (0.4842, 0.5412). Indeed, the result of
increasing the sample size is a shorter interval (of length 0.04 compared with
0.057) for the same confidence level of 95%.
In summary, point estimates provide only a single number to estimate
the value of a population parameter. By contrast, confidence intervals give
a range of values that we reasonably expect will contain the population
parameter. Again, a 95% confidence interval means that if we were to take
a large number of samples (like 100 or 1000) of equal size from the same
population and calculate a confidence interval for the population parameter,
about 95% of the confidence intervals would contain the true value of the
population parameter.
SECTION 11.2 EXERCISES
1. Suppose we want to estimate the proportion
of a city’s residents who drive to work. What
is a good choice for the point estimate of this
proportion?
2. Suppose, instead, we want to estimate the “average” number of miles people living in a city
drive to work. What are two possible choices
for the point estimate of this “average”?
3. Explain the meaning of a 99% confidence
level.
4. Suppose you want a confidence interval for
a population proportion. You want to be as
accurate as possible, so you select a 100% confidence level. What would your confidence
interval have to be?
5. Which confidence interval, when based on
the same data, is wider: an 80% or an 85%
confidence interval?
6. True or false: If you flip a fair coin 100 times,
then calculate a 95% confidence interval, there
is an approximate 95% chance 1/2 will be in
the interval.
7. A Gallup poll of 1013 adults found 61% of
the people in the sample drink alcoholic beverages, which yields a confidence interval of
(58%, 64%). True or false: There is an approximate 95% chance that the percentage of
adults in the population who drink alcoholic
beverages is between 58% and 64%.