Download STAT111 Introductory Statistics

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Central limit theorem wikipedia , lookup

Transcript
STAT 111 Introductory Statistics
Lecture 8: More on the Binomial Distribution
and Sampling Distributions
June 1, 2004
Today’s Topics
• More on the binomial distribution
– Mean and variance
•
•
•
•
•
Sample proportion
Normal approximation of the binomial
Continuity correction
Sampling distribution of sample means
Central Limit Theorem
Recall: The Binomial Setting
• There are a fixed number n of trials.
• The n trials are all independent.
• Each trial has one of two possible outcomes,
labeled “success” and “failure.”
• The probability of success, p, remains the same
for each trial.
Recall: The Binomial Distribution
• The distribution of the count X of successes in
the binomial setting is called the binomial
distribution with parameter n and p, where
– n is the number of trials
– p is the probability of a success on any trial
• The count X is a discrete random variable,
typically abbreviated as X ~ B(n, p).
• Possible values of X are the whole numbers from
0 to n.
The Binomial Distribution
• If X ~ B(n,p), then
n x
n!
n x
x
n x


P( X  x)    p (1  p) 
p (1  p)
x!(n  x)!
 x
• Examples: Let n = 3.
 n   3
3!
3  2 1
x  0 :      

1
 x   0  0!3! (1)(3  2 1)
 n   3
3!
3  2 1
x  1 :      

3
 x   1  1!2! (1)( 2 1)
Developing Binomial Probabilities for
n=3
S1 p
p
1-p
p
1-p
F1
1-p
S2
p
S3
F2 1-p
p
S2 1-p
p
1-p
p
F3
S3
F3
S3
F3
S3
F2 1-p
F3
P(SSS) = p3
P(SSF) = p2(1 – p)
P(SFS) = p2(1 – p)
P(SFF) = p(1 – p)2
P(FSS) = p2(1 – p)
P(FSF) = p(1 – p)2
P(FFS) = p(1 – p)2
P(FFF) = (1 – p)3
Binomial Probabilities for n = 3
• Let X be the number of successes in three trials.
P(X = 0) = (1 – p)3
P(X = 1) = 3p(1 – p) 2
P(X = 2) = 3p2(1 – p)
P(X = 3) = p3
P(FFF) = (1 – p)3
X=0
P(SSF) = p2(1 – p)
P(SFS) = p2(1 – p)
X=1
P(SFF) = p(1 – p)2
P(FSS) = p2(1 – p)
X=2
P(FSF) = p(1 – p)2
P(FFS) = p(1 – p)2
X=3
P(SSS) = p3
Example: Rolling a Die
• Roll a die 4 times, let X be the number of times the
number 5 appears.
• “Success” = get a roll of 5, so P(Success) = 1/6.
X=0
X=1
X=2
X=3
X=4
4!
P( X  0) 
(1 / 6)0 (1  1 / 6) 40  0.4823
0!(4  0)!
4!
P( X  1) 
(1 / 6)1 (1  1 / 6) 41  0.3858
1!(4  1)!
4!
(1 / 6) 2 (1  1 / 6) 42  0.1157
2!(4  2)!
4!
P( X  3) 
(1 / 6)3 (1  1 / 6) 43  0.0154
3!(4  3)!
P( X  2) 
P( X  4) 
4!
(1 / 6) 4 (1  1 / 6) 44  0.0008
4!(4  4)!
Example: Rolling a Die
• Find the probability that we get at least 2 rolls of
5.
Expected Value and Variance of a
Binomial Random Variable
• If X~B(n,p),then
E ( X )   X  np
Var ( X )   X2  np(1  p )
 X  np(1  p)
Set-up for Derivation
• Let Xi indicate whether the i th trial is a success or
failure by,
Xi =1, if ith trial is a success
i = 1,2,….,n.
Xi =0, if ith trial is a failure.
• X1, …, Xn are independent and identically
distributed with probability distribution
Outcome:
1
Probability: p
0
1-p
Binomial Example: Checkout Lanes
• A grocery store has 10 checkout lanes. During a
busy hour the probability that any given lane is
occupied (has at least one customer) is 0.75.
Assume that the lanes are occupied or not
occupied independently of each other.
– What is the probability that a customer will find at
least one lane unoccupied?
– What is the expected number of occupied lanes?
– What is the standard deviation of the number of
occupied lanes?
Sample Proportions
• In statistical sampling we often want to estimate
the proportion p of “successes” in a population.
• The sample proportion is defined as
X count of successes in sample
pˆ 

n
size of sample
• If the count X is B(n, p), then the mean and
standard deviation of the sample proportion are
E ( pˆ )   pˆ  p
 pˆ  p (1  p ) n
Sample Proportions
• Our sample proportion is an unbiased estimator
of the population proportion p.
• The variability of our estimator decreases as
sample size increases.
• In particular, we must multiply the sample size by
4 if we want the cut the standard deviation in
half.
Sample Proportions
• The histogram of the distribution of the sample
proportion when n = 1000, p = 0.6
0.30
0.25
0.15
0.10
0.05
p_hat
1
0.9
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.00
0
P(X)
0.20
Normal Approximation for Counts,
Proportions
• Let X be the number of successes in a SRS of size
n from a large population having proportion p of
successes, and let the sample proportion of
successes be denoted by
• Then for large n,
X
pˆ 
n
– X is approximately normal with mean np and variance
np(1 – p).
– p̂ is approximately normal with mean p and variance
p(1 – p) / n.
Normal Approximation: Rule of
Thumb
• The accuracy of the approximation generally
improves as the sample size n increases.
• For any fixed sample size, the approximation is
most accurate when p is close to 0.5, and least
accurate when p is near 0 or 1.
• As a general rule of thumb, then, we use the
normal approximation for values of n and p such
that np ≥ 10 and n(1 – p) ≥ 10.
Example
• The Laurier Company’s brand has a market share
of 30%. Suppose that in a survey, 1,000
consumers of the product are asked which brand
they prefer. What is the probability that more than
32% of the respondents will say they prefer the
Laurier brand?
Another Example
• A quality engineer selects an SRS of size 100
switches from a large shipment for detailed
inspection. Unknown to the engineer, 10% of the
switches in the shipment fail to meet the
specifications. The actual binomial probability
that no more than 9 of the switches in the sample
fail inspection is P(X ≤ 9) = .4513.
• How accurate is the normal approximation for
this probability?
Another Example (cont.)
• Let X be the number of bad switches; then
X ~ B(100, 0.1).
 X  np

9

100
(
0
.
1
)

P( X  9)  P

 np (1  p)

100
(
0
.
1
)(
1

0
.
1
)


 P( Z  0.33)  0.3707
• It’s not that accurate. Note that np = 10, so n and
p are on the border of values for which we are
willing to use the approximation.
Continuity Correction
• While the binomial distribution places probability
exactly on X = 9 and X = 10, the normal
distribution spreads probability continuously in
that interval.
• The bar for X = 9 in a probability histogram goes
from 8.5 to 9.5, but calculating P(X ≤ 9) using the
normal approximation only includes the area to
the left of the center of this bar.
• To improve the accuracy of our approximation,
we should let X = 9 extend from 8.5 to 9.5, etc.
Continuity Correction
• Use continuity correction to approximate the
binomial probability P(X=10) when n=100,
p=0.1
• Using the normal approximation to the binomial
distribution, X is approximately distributed as
N(10, 3).
Continuity Correction
The exact binomial probability is P(X=10)=0.13187
P(9.5<Xnormal<10.5)=0.13237
9.5 10 10.5
P(Xbinomial=10)=0.13187
Continuity Correction
8
8.5
P( X binomial  8)  P( X normal  8.5)
Q: what about continuity correction for P(X<8)?
Continuity Correction
P( X binomial  14)  P( X normal  13.5)
13.514
Q: what about continuity correction for P(X>14)?
Example Re-visited
• Using the continuity correction, the probability
that no more than 9 of the switches in the sample
fail inspection is
P( X binomial  9)  P( X normal  9.5)


9
.
5

100
(
0
.
1
)

 P Z 


100
(
0
.
1
)(
1

0
.
1
)


 P( Z  .1667)  0.4338
Example: Inspection of Switches
• Find the probability that at least 5 but at most 15
switches fail the inspection.
P(5  X binom  15)  P( X binom  15)  P( X binom  4)
Sampling Distributions
• Counts and proportions are discrete random
variables; used to describe categorical data.
• Statistics used to describe quantitative data are
most often continuous random variables.
• Examples: sample mean, percentiles, standard
deviation
• Sample means are among the most common
statistics.
Sampling Distributions
• Regarding sample means,
– They tend to be less variable than individual
observations.
– Their distribution tends to be more normal than that
of individual observations.
• We’ll see why later.
Sampling Distributions of Sample
Means
• Let x be the mean of an SRS of size n from a
population having mean µ and standard deviation
σ.
• The mean and standard deviation of x are
x  
x 
• Why?

n
Sampling Distributions of Sample
Means
• The shape of the distribution of the sample mean
depends on the shape of the population
distribution itself.
• One special case: normal population distribution
If a population has the N (  ,  ) distributi on, then
the sample mean x of n independen t observatio ns
has the N (  , / n ) distributi on.
• Because: any linear combination of independent
normal random variables is normal distributed.
Example
• The foreman of a bottling plant has observed that
the amount of soda pop in each “32-ounce”
bottle is actually a normally distributed random
variable, with a mean of 32.2 ounces and a
standard deviation of .3 ounce.
– If a customer buys one bottle, what is the probability
that that bottle contains more than 32 ounces?
– If that same customer instead buys a carton of 4
bottles, what is the probability that the mean of those
4 bottles is greater than 32 ounces?
Example
• The starting salaries of M.B.A.s at Wilfrid Laurier
Univ.(WLU) are normally distributed with a mean
of $62,000 and a standard deviation of $14,500.
The starting salaries of M.B.A.s at the University
of Western Ontario (UWO) are normally
distributed with a mean of $60,000 and a standard
deviation of $18,300.
– A random sample of 50 WLU M.B.A.s and a random
sample of 60 UWO M.B.A.s are selected
– What is the probability that the sample mean of WLU
graduates will exceed that of the UWO graduates?
Central Limit Theorem
• When the population distribution is normal, so is
the sampling distribution of x
• What about when the population distribution is
non-normal?
• For large sample sizes, it turns out that the
distribution of x gets closer to a normal
distribution.
• As long as the population has finite standard
deviation, this will be true regardless of the actual
shape of the population distribution
Central Limit Theorem
• Formally, draw an SRS of size n from any
population with mean µ and finite standard
deviation σ.
• As n approaches infinity (gets very large)
  
x is approximat ely N   ,


n
• This can hold even if the observations are not
independent or identically distributed.
• This is why normal distributions are common
models for observed data.