Download 7 Point Estimation of Parameters and Sampling Distributions

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

History of statistics wikipedia , lookup

Statistics wikipedia , lookup

Transcript
3/1/2016
Applied Statistics and
Probability for Engineers
Sixth Edition
Douglas C. Montgomery
George C. Runger
Chapter 7
Point Estimation of Parameters and Sampling Distributions
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
7
CHAPTER OUTLINE
Point Estimation of
Parameters and
Sampling
Distributions
7-1 Point Estimation
7-3.4 Mean Squared Error of an
7-2 Sampling Distributions and
Estimator
the Central Limit Theorem
7-4 Methods of Point Estimation
7-3 General Concepts of Point
7-4.1 Method of Moments
Estimation
7-4.2 Method of Maximum
7-3.1 Unbiased Estimators
Likelihood
7-3.2 Variance of a Point
7-4.3 Bayesian Estimation of
Estimator
Parameters
7-3.3 Standard Error: Reporting
a Point Estimate
2
Chapter 7 Title and Outline
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
1
3/1/2016
Learning Objectives for Chapter 7
After careful study of this chapter, you should be able to do
the following:
1. General concepts of estimating the parameters of a population or a
probability distribution.
2. Important role of the normal distribution as a sampling distribution.
3. The central limit theorem.
4. Important properties of point estimators, including bias, variances,
and mean square error.
5. Constructing point estimators using the method of moments, and
the method of maximum likelihood.
6. Compute and explain the precision with which a parameter is
estimated.
7. Constructing a point estimator using the Bayesian approach.
Chapter 7 Learning Objectives
3
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Point Estimation
• A point estimate is a reasonable value of a
population parameter.
• X1, X2,…, Xn are random variables.
• Functions of these random variables, x-bar
and s2, are also random variables called
statistics.
• Statistics have their unique distributions
which are called sampling distributions.
Sec 7-1 Point Estimation
4
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
2
3/1/2016
Point Estimator
As an example,suppose the random variable X is normally distributed with
an unknown mean μ. The sample mean is a point estimator of the unknown
population mean μ. That is, μ  X . After the sample has been selected,
the numerical value x is the point estimate of μ.
Thus if x1  25, x2  30, x3  29, and x4  31, the point estimate of μ is
x
25  30  29  31
 28.75
4
Sec 7-1 Point Estimation
5
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Some Parameters & Their Statistics
Parameter
μ
σ2
σ
p
μ1 - μ2
p1 - p2
•
•
Measure
Mean of a single population
Variance of a single population
Standard deviation of a single population
Proportion of a single population
Difference in means of two populations
Difference in proportions of two populations
Statistic
x-bar
s2
s
p -hat
x bar1 - x bar2
p hat1 - p hat2
There could be choices for the point estimator of a parameter.
To estimate the mean of a population, we could choose the:
– Sample mean.
– Sample median.
– Average of the largest & smallest observations in the sample.
Sec 7-1 Point Estimation
6
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
3
3/1/2016
Some Definitions
• The random variables X1, X2,…,Xn are a
random sample of size n if:
a) The Xi ‘s are independent random variables.
b) Every Xi has the same probability
distribution.
• A statistic is any function of the
observations in a random sample.
• The probability distribution of a statistic is
called a sampling distribution.
Sec 7-2 Sampling Distributions and the Central Limit Theorem
7
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Central Limit Theorem
Sec 7-2 Sampling Distributions and the Central Limit Theorem
8
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
4
3/1/2016
Example 7-2: Central Limit Theorem
Suppose that a random variable X has a continuous
uniform distribution:
1 2, 4  x  6
f  x  
0,
otherwise
Find the distribution of the sample mean of a random
sample of size n = 40.
By the CLT the distribution X is normal .

ba 64

5
2
2
b  a 
 
2
12
2
 6  4

12
2

13
1
 X2 


n
40 120
2
13
Figure 7-5 The distribution
of X and X for Example 7-2.
Sec 7-2 Sampling Distributions and the Central Limit Theorem
9
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Sampling Distribution of a Difference in Sample Means
• If we have two independent populations with means μ1
and μ2, and variances σ12 and σ22, and
• If X-bar1 and X-bar2 are the sample means of two
independent random samples of sizes n1 and n2 from
these populations:
• Then the sampling distribution of:
is approximately standard normal, if the conditions of the
central limit theorem apply.
• If the two populations are normal, then the sampling
distribution of Z is exactly standard normal.
Sec 7-2 Sampling Distributions and the Central Limit Theorem
10
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
5
3/1/2016
Example 7-3: Aircraft Engine Life
The effective life of a component used
in jet-turbine aircraft engine is a
random variable with mean 5000 and
SD 40 hours and is close to a normal
distribution. The engine manufacturer
introduces an improvement into the
Manufacturing process for this
component that changes the
parameters to 5050 and 30. Random
samples of size 16 and 25 are
selected.
What is the probability that the
difference in the two sample means is
at least 25 hours?
Figure 7-6 The sampling distribution
of X2 − X1 in Example 7-3.
Process
Old (1) New (2) Diff (2-1)
x -bar = 5,000 5,050
50
s=
40
30
n=
16
25
Calculations
s / √n =
10
6
11.7
z=
-2.14
P(xbar2-xbar1 > 25) = P(Z > z) = 0.9840
= 1 - NORMSDIST(z)
Sec 7-2 Sampling Distributions and the Central Limit Theorem
11
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Unbiased Estimators Defined
Sec 7-3.1 Unbiased Estimators
12
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
6
3/1/2016
Example 7-4: Sample Mean & Variance Are Unbiased-1
• X is a random variable with mean μ and variance σ2. Let
X1, X2,…,Xn be a random sample of size n.
• Show that the sample mean (X-bar) is an unbiased
estimator of μ.
Sec 7-3.1 Unbiased Estimators
13
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-4: Sample Mean & Variance Are Unbiased-2
Show that the sample variance (S2) is a unbiased estimator of σ2.
Sec 7-3.1 Unbiased Estimators
14
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
7
3/1/2016
Minimum Variance Unbiased Estimators
• If we consider all unbiased estimators of θ,
the one with the smallest variance is called
the minimum variance unbiased estimator
(MVUE).
• If X1, X2,…, Xn is a random sample of size
n from a normal distribution with mean μ
and variance σ2, then the sample X-bar is
the MVUE for μ.
Sec 7-3.2 Variance of a Point Estimate
15
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Standard Error of an Estimator
Sec 7-3.3 Standard Error Reporting a Point Estimate
16
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
8
3/1/2016
Example 7-5: Thermal Conductivity
• These observations are 10
measurements of thermal
conductivity of Armco iron.
• Since σ is not known, we use s to
calculate the standard error.
• Since the standard error is 0.2%
of the mean, the mean estimate is
fairly precise. We can be very
confident that the true population
mean is 41.924 ± 2(0.0898) or
between 41.744 and 42.104.
xi
41.60
41.48
42.34
41.95
41.86
42.18
41.72
42.26
41.81
42.04
41.924 = Mean
0.284 = Std dev (s )
0.0898 = Std error
Sec 7-3.3 Standard Error Reporting a Point Estimate
17
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Mean Squared Error
Conclusion: The mean squared error (MSE) of
the estimator is equal to the variance of the
estimator plus the bias squared.
Sec 7-3.4 Mean Squared Error of an Estimator
18
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
9
3/1/2016
Relative Efficiency
• The MSE is an important criterion for
comparing two estimators.
• If the relative efficiency is less than 1, we
conclude that the 1st estimator is superior
than the 2nd estimator.
Sec 7-3.4 Mean Squared Error of an Estimator
19
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Optimal Estimator
• A biased estimator can
be preferred than an
unbiased estimator if it
has a smaller MSE.
• Biased estimators are
occasionally used in
linear regression.
• An estimator whose MSE
is smaller than that of
any other estimator is
called an optimal
estimator.
Figure 7-8 A biased estimator that
has a smaller variance than the
unbiased estimator .
Sec 7-3.4 Mean Squared Error of an Estimator
20
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
10
3/1/2016
Moments Defined
• Let X1, X2,…,Xn be a random sample from the
probability distribution f(x), where f(x) can be either
a:
– Discrete probability mass function, or
– Continuous probability density function
• The kth population moment (or distribution
moment) is E(Xk), k = 1, 2, ….
n
The k th sample moment is 1 / n   X i k , k  1, 2,...
i 1
• If k = 1 (called the first moment), then:
– Population moment is μ.
– Sample moment is x-bar.
• The sample mean is the moment estimator of the
population mean.
Sec 7-4.1 Method of Moments
21
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Moment Estimators
Sec 7-4.1 Method of Moments
22
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
11
3/1/2016
Example 7-8: Normal Distribution Moment Estimators
Suppose that X1, X2, …, Xn is a random sample
from a normal distribution with parameter μ and
σ2 where E(X) = μ and E(X2) = μ2 + σ2.
X 
1 n
 Xi
n i 1
2  2 
and
1 n 2
 Xi
n i 1
1 n

X

n
Xi 



n
2
1
 n i 1 
   X i2  X 2  i 1
n i 1
n
n
2
2
i
2

 n
 

  Xi  
1 n
   X i2   i 1   

n  i 1
n




n
 X
i 1
i
X
n
2
(biased)
Sec 7-4.1 Method of Moments
23
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-9: Gamma Distribution Moment Estimators-1
Suppose that X1, X2, …, Xn is a random sample from a gamma distribution with
parameter r and λ where E(X) = r/ λ and E(X2) = r(r+1)/ λ2 .
r
 E  X   X is the mean

r
2
 E  X 2   E  X  is the variance or
2
r  r  1
 E  X 2  and now solving for r and  :
2
X2
r
n
1/ n   X i2  X 2
i 1

X
n
1/ n   X i2  X 2
i 1
Sec 7-4.1 Method of Moments
24
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
12
3/1/2016
Example 7-9: Gamma Distribution Moment Estimators-2
Using the time to failure data in
the table. We can estimate the
parameters of the gamma
distribution.
x-bar =
xi
21.646
2
ΣX = 6645.4247
r
X2
n
1/ n   X i2  X 2
11.96
5.03
67.40
16.07
31.50
7.73
11.10
22.38
2
xi
143.0416
25.3009
4542.7600
258.2449
992.2500
59.7529
123.2100
500.8644

21.6462
 1.29
1 8 6645.4247  21.6462

21.646
 0.0598
1
8
6645.4247
 21.6462
 
i 1

X
n
1/ n   X i2  X 2
i 1
Sec 7-4.1 Method of Moments
25
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Maximum Likelihood Estimators
• Suppose that X is a random variable with
probability distribution f(x;θ), where θ is a single
unknown parameter. Let x1, x2, …, xn be the
observed values in a random sample of size n.
Then the likelihood function of the sample is:
L(θ) = f(x1;θ) ∙ f(x2; θ) ∙…∙ f(xn; θ)
• Note that the likelihood function is now a
function of only the unknown parameter θ. The
maximum likelihood estimator (MLE) of θ is the
value of θ that maximizes the likelihood function
L(θ).
Sec 7-4.2 Method of Maximum Likelihood
26
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
13
3/1/2016
Example 7-10: Bernoulli Distribution MLE
Let X be a Bernoulli random variable. The probability mass function is
f(x;p) = px(1-p)1-x, x = 0, 1 where P is the parameter to be estimated.
The likelihood function of a random sample of size n is:
L  p   p x1 1  p 
1 x1
 p x2 1  p 
1 x2
 ...  p xn 1  p 
1 xn
n
n
  p xi 1  p 
1 xi
n
 xi
n  xi
 p i1 1  p  
i 1
i 1
n
 n 


ln L  p     xi  ln p   n   xi  ln 1  p 
i 1
 i 1 


n


 n   xi 
i 1

i 1

p
1  p 
n
d ln L  p 

dp
x
i
Sec 7-4.2 Method of Maximum Likelihood
27
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-11: Normal Distribution MLE for μ
Let X be a normal random variable with unknown mean μ and
known variance σ2. The likelihood function of a random sample of
n
size n is:
1  x     2 
L  
i 1

ln L    
 2
1
 2 
2
2
e
i
1
e 2
n2
2
n
2
  xi   
2
i 1
n
1 n
2
ln  2 2   2   xi   
2
2 i 1
d ln L    1 n
 2   xi   
d
 i 1
Sec 7-4.2 Method of Maximum Likelihood
28
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
14
3/1/2016
Example 7-12: Exponential Distribution MLE
Let X be a exponential random variable with parameter λ.
The likelihood function of a random sample of size n is:
n
L      e
  xi
 e
n

n
 xi
i 1
i 1
n
ln L     n ln       xi
i 1
d ln L    n
   xi
d
 i 1
n
Equating the above to zero we get
n
n
x
i 1
i
 1 X (same as moment estimator)
Sec 7-4.2 Method of Maximum Likelihood
29
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-13: Normal Distribution MLEs for μ & σ2
Let X be a normal random variable with both unknown mean μ
and variance σ2. The likelihood function of a random sample of
n
 x     2 
1
size n is:
L   , 2  
e
2

i
2
i 1

ln L   ,  2  
 ln L   ,  2 

 ln L   ,  2 
 
2



1
 2 
2
1
e
n2
2 2
n
  xi   

2
2
i 1
n
1
ln  2 2   2
2
2
1
2
n
x  
i 1
n
 x    0
i 1
i
n
1
 4
2
2
2
n
 x  
i 1
i
n
2
  X and  
Sec 7-4.2 Method of Maximum Likelihood
2
i
2
0
x  X 
i 1
2
i
n
30
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
15
3/1/2016
Properties of an MLE
Notes:
• Mathematical statisticians will often prefer MLEs because of
these properties. Properties (1) and (2) state that MLEs are
MVUEs.
• To use MLEs, the distribution of the population must be known
or assumed.
Sec 7-4.2 Method of Maximum Likelihood
31
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Invariance Property
This property is illustrated in Example 7-13.
Sec 7-4.2 Method of Maximum Likelihood
32
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
16
3/1/2016
Example 7-14: Invariance
For the normal distribution, the MLEs were:
Sec 7-4.2 Method of Maximum Likelihood
33
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Complications of the MLE Method
The method of maximum likelihood is an
excellent technique, however there are two
complications:
1. It may not be easy to maximize the likelihood
function because the derivative function set
to zero may be difficult to solve algebraically.
2. It may not always be possible to use
calculus methods directly to determine the
maximum of L(ѳ).
The following example illustrate this.
Sec 7-4.2 Method of Maximum Likelihood
34
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
17
3/1/2016
Example 7-16: Gamma Distribution MLE-1
Let X1, X2, …, Xn be a random sample from a gamma
distribution. The log of the likelihood function is:
 n  r xir 1e   x 
ln L  r ,    ln  

i

i 1
 r 

n
n
i 1
i 1
 nr ln      r  1  ln  xi   n ln   r      xi
 ln L  r ,  
 'r 
 n ln      ln  xi   n
r
 r 
i 1
n
 ln L  r ,   nr n
   xi

 i 1
Equating the above derivative to zero we get
Sec 7-4.2 Method of Maximum Likelihood
35
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-16: Gamma Distribution MLE-2
Figure 7-11 Log likelihood for the gamma distribution using the
failure time data. (a) Log likelihood surface. (b) Contour plot.
Sec 7-4.2 Method of Maximum Likelihood
36
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
18
3/1/2016
Bayesian Estimation of Parameters-1
• The moment and likelihood methods interpret
probabilities as relative frequencies and are called
objective frequencies.
• The random variable X has a probability
distribution of parameter θ called f(x|θ).
• Additional information about θ is that it can be
summarized as f(θ), the prior distribution, with
mean μ0 and variance σ02. Probabilities associated
with f(θ) are subjective probabilities.
• The joint distribution is f(x1, x2, …, xn|θ).
• The posterior distribution is f(θ|x1, x2, …, xn) is our
degree of belief regarding θ after observing the
sample data.
7-4.3 Bayesian Estimation of Parameters
37
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Bayesian Estimation of Parameters-2
• Now the joint probability distribution of the sample is
f(x1, x2, …, xn, θ) = f(x1, x2, …, xn |θ) ∙ f(θ)
• The marginal distribution is:
  f  x1 , x2 ,..., xn ,θ  , for θ discrete 
 θ
f  x1 , x2 ,..., xn    
  f  x1 , x2 ,..., xn ,θ  dθ, for θ continuous


• The desired posterior distribution is:
f  θ | x1 , x2 ,..., xn  
f  x1 , x2 ,..., xn ,θ 
f  x1 , x2 ,..., xn 
The Bayesian estimator of θ is θ, the mean of the posterior distribution.
7-4.3 Bayesian Estimation of Parameters
38
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
19
3/1/2016
Example 7-16: Bayes Estimator for the mean of a Normal Distribution -1
Let X1, X2, …, Xn be a random sample from a normal distribution unknown mean μ and
known variance σ2. Assume that the prior distribution for μ is:
f μ  
2
1
  
e  0
2 0
2 02

1
2 02
e

  2  2 0  02

2 02
The joint probability distribution of the sample is:
f  x1 , x2 ,..., xn |   

1
 2 
2 n2
1
 2 
2 n2
e
e

 1 2 2
n
 ( x   )
i
2
i 1
 n
 1 2 2  xi2  2 

 i 1


n

i 1

 xi  n 2 
7-4.3 Bayesian Estimation of Parameters
39
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-17: Bayes Estimator for the mean of a Normal Distribution-2
Now the joint probability distribution of the sample and μ is:
f  x1 , x2 ,..., xn ,    f  x1 , x2 ,..., xn |    f  μ 

 2 
e

1
2 n2
2 0
 
1
1
1/2   2  2  2
   0  n
e

  1

xi 
xi2  2 
n 
1/2   2  2  2   2   02  2   2  02 
0   
    
 0 


  0


  2   0  x

 2  2

 0
n



  h ( x , x ,..., x , 2 ,  , 2 )
n
0 0
 1 1 2


Upon completing the square in the exponent,
2
f
 x1 , x2 ,..., xn ,    e
 1

x 02
1   2  ( 2 / n )  0
1/ 2 2  2
 2
     2
  h2 ( x1 , x2 ,..., xn , 2 , 0 , 02 )
2
2

 0  / n  
  0  / n  0  / n  
where hi ( x1 , x2 ,..., xn ,  2 , 0 ,  02 )is a function of the observed
values and the parameters  2 , 0 and  02 .
Since f
 x1 , x2 ,..., xn ,  
f   | x1 , x2 ,..., xn   e
does not depend on 
 1
1   2  ( 2 / n ) 0  x 02  
1/ 2 2  2
  
h ( x , x ,..., xn , 2 , 0 , 02 )

 
  3 1 2
 02  2 / n
 0  / n  


7-4.3 Bayesian Estimation of Parameters
40
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
20
3/1/2016
Example 7-17: Bayes Estimator for the mean of a Normal Distribution-3
which is recognized as a normal probability density
function with posterior mean
and posterior variance
 0  n 
 1
1 
V     2  2   2
0  2 n
 0  n 
1
2
2
7-4.3 Bayesian Estimation of Parameters
41
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
Example 7-17: Bayes Estimator for the mean of a Normal Distribution-4
To illustrate:
– The parameters are: μ0 = 0, σ02= 1
– Sample: n = 10, x-bar = 0.75, σ2 = 4
 n  

2
0
  02 x
 02   2 n

 4 10  0  1 0.75  0.536
1   4 10 
7-4.3 Bayesian Estimation of Parameters
42
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
21
3/1/2016
Important Terms & Concepts of Chapter 7
Bayes estimator
Bias in parameter estimation
Central limit theorem
Estimator vs. estimate
Likelihood function
Maximum likelihood estimator
Mean square error of an
estimator
Minimum variance unbiased
estimator
Moment estimator
Normal distribution as the
sampling distribution of the:
– sample mean
– difference in two sample
means
Parameter estimation
Point estimator
Population or distribution
moments
Posterior distribution
Prior distribution
Sample moments
Sampling distribution
An estimator has a:
– Standard error
– Estimated standard error
Statistic
Statistical inference
Unbiased estimator
Chapter 7 Summary
43
Copyright © 2014 John Wiley & Sons, Inc. All rights reserved.
22