* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Topic 2. Distributions, hypothesis testing, and sample size determination
Sufficient statistic wikipedia , lookup
Degrees of freedom (statistics) wikipedia , lookup
Foundations of statistics wikipedia , lookup
History of statistics wikipedia , lookup
Bootstrapping (statistics) wikipedia , lookup
Taylor's law wikipedia , lookup
German tank problem wikipedia , lookup
Misuse of statistics wikipedia , lookup
Topic 2. Distributions, hypothesis testing, and sample size
determination
The Student - t distribution (ST&D pg 56 and 77)
Consider a repeated drawing of samples of size n = 5 from a normal distribution.
For each sample compute Y , s, sY , and another statistic, t:
t (n-1)= ( Y - )/ s
Y
(Remember Z = ( Y - )/ Y )
The t statistics is the number of standard error that separate Y and its hypothesized
mean µ.
df=n-1=4
Critical values for |t|>|Z| -> less
sensitivity. This is the price we pay for
being uncertain about the population
i
Fig. 1. Distribution of t (df=4) compared to Z. The t distribution is symmetric, but wider
and flatter than the Z distribution, lying under it at the center and above it in the tails.
1
When N increases the t distribution tend towards the N distribution
2
2. 2. Confidence limits based on sample statistics (ST&D p.77)
The general formula for any parameter is:
Estimated Critical value * Standard error of the estimated
So, for a population mean estimated via a sample mean:
Y t
2
, n 1
sY
The statistic Y is distributed about according to the t distribution so it satisfies
P{ Y - t /2, n-1
sY
Y + t /2, n-1 s }= 1-
Y
For a confidence interval of size 1-, use a t value corresponding to /2.
Therefore the confidence interval is
Y-
t /2, n-1
sY
Y + t /2, n-1
sY
These two terms represent the lower and upper 1- confidence limits of the mean.
The interval between these terms is called confidence interval (CI).
Example: Data Set 1 of Hordeum 14 malt extraction values
Y = 75.94 sY = 1.23 /
14 = 0.3279 A table gives the t0.025,13 value of 2.16
95% CI for = 75.94 ± 2.160 * 0.3279 [75.23- 76.65]
If we repeatedly obtained samples of size 14 from the population and constructed
these limits for each, we expect 95% of the intervals to contain the true mean.
True mean
Fig. 2 Twenty 95% confidence intervals. One out of 20 intervals does not include
the true mean.
3
2. 3. Hypothesis testing and power of the test (ST&D 94).
Example Barley data. Y = 75.94, sY = s 2 / n = 0.3279, t0.025,13 = 2.160, CI: [75.23- 76.65]
1) Choose a null hypothesis: Test Ho = 78 against the H1 78.
2) Choose a significance level: Assign = 0.05
3) Calculate the test statistic t:
Y 75.94 78.00
t
sY
0.3279
6.28
(interpretation: the sample mean is 6.3 SE from the hypothetical mean of 78. Too far!).
4) Compare the absolute value of the test statistic to the critical statistic:
| - 6.28 | > 2.16
5) Since the absolute value of the test statistic is larger, we reject H0.
This is equivalent to calculate a 95% confidence interval for the mean.
Since o (78) is not within the CI [75.23- 76.65] we reject Ho.
is called the significance level of the test (<0.05): probability of incorrectly
rejecting a true Ho, a Type I error.
is the Type II error: to incorrectly accept Ho when it is false
Accepted
Null hypothesis
Null hypothesis
Rejected
True Correct decision
Type I error=
False Type II error=
Correct decision= Power= 1-
Power of the test: 1- is the power of the test, and represents the probability of
correctly rejecting a false null hypothesis. It is a measure of the ability of the test to
detect an alternative mean or a significant difference when it is real
Note that for a given Y and s, if 2 of the 3 quantities , , and n are specified then
the third one can be determined.
Choose the right number of replications to keep Type I error and Type II error
under the desired limits (e.g. <0.05 & <0.20).
4
Power of a test (ST&D pg 118-119)
Power 1 P( Z Z / 2
1 0
Y
) or
P(t t / 2
1 0
sY
)
between 2
means in SE
units
What is the power of a test for Ho: = 74.88 in the barley data set against H1: =
75.94. Since = 0.05, n = 14 (t 0.025,13 = 2.160), and
Power 1 P(t 2.160
75.94 74.88
0.32795
sY = 0.32795.
) P(t 1.072) 0.85
The probability of the Type II error we are looking for is the shaded area to the
left of the lower curve.
/2
Ho is true
Fig. 3. Type I and Type II
errors in the Barley
data set.
Ho is false
Acceptance
1-
Rejection
The area 1- in the
rejection region = the
probability that > 75.588 under H1 = power =
P(t>(75.588-75.94)/0.32795)= P(t>-1.072)=0.85 same as above!
The magnitude of depends on
1. The Type I error rate
2. The distance between the two means under consideration
3. The number of observations (n) sY
s
n
5
When the distance between the two means is reduced, increases
Variation of power as a function of the distance between the
alternative hypotheses (Biometry Sokal and Rohlf)
n=35
SE=0.7
n=5
SE=1.7
6
2.3.2. Power of the test for the difference between the means of two samples
Two types of alternative hypothesis
H0: 1 - 2=0 versus H1: 1 - 2 0 (two tail test) -> (t value: top t Table)
H1: 1 - 2 >0 (one tail test) -> (t value: bottom t Table)
The general power formula for both equal and unequal sample sizes reads as:
Power P(t t
2
| 1 2 |
| 2 |
) P(t t 1
)
2
,
sY 1Y 2
s
2
pooled
N
where s
2
pooled
and N
is a weighted variance given by: s
2
pooled
(n1 1) s12 (n2 1) s 22
(n1 1) (n2 1)
n1 n2
.
n1 n 2
When n1 = n2 = n (equal sample sizes) that the formulas reduce to:
s
2
pooled
(n1 1) s12 (n2 1) s 22 (n 1)( s12 s 22 ) s12 s 22
(n1 1) (n2 1)
2(n 1)
2
n1 n2
n2 n
N
n1 n2 2n 2
Power P (t t
2
| 1 2 |
| 2 |
)
) P (t t 1
2
sY 1Y 2
2 s pooled
2
n
The variance of the difference between two random variables is the sum of the
variances (error are always added) (ST&D 113-115).
The degrees of freedom for the critical t /2 are
General case: (n1-1) + (n2-1)
Equal sample size: 2*(n-1)
7
2. 4. 2 Sample size for estimating µ, when is known.
Using the z statistic
If the population variance is known the Z statistic may be used.
Z
Y
so
CI =
Y Z / 2 Y
or
Y Z / 2
n
The formula for d= half-length of the confidence interval for the mean is
[ d
Y d
]
d Z Y Z
2
2
n
This can be rearranged to estimate the confidence interval in terms of the
population variance. For =0.05:
n = z 2/2 2 / d2 = z 2/2 (/d)2 = (1.96)2(/d)2= 3.8(/d)2
So if d= n 4
d= 0.5 n 16
d= 0.25 n 64
2
The equation may be expressed in terms
of the coefficient of variation
2
n Z
Z 2
2
2 d
2
CV 2
d
2
CV= s / Y (as a proportion not as a %)
d/ is the confidence interval as a fraction of the population mean.
For example d/ < 0.1 means that the length of the confidence interval should not
be larger than two tenth of the population mean.
d/ < 0.1 and so 2d/ < 0.2
Example: The CVs of yield trials in our experimental station are never higher than
15%. How many replications are necessary to have a 95% CI for the true mean of
less than 1/10 of the average yield?
2d= 0.1
so
d= 0.1/2= 0.05
n= 1.962 0.152/0.052 = 34.6 35
8
2. 4. 3 Sample size for the estimation of the mean
Unknown 2. Stein's Two-Stage Sample
Consider a (1 - )% confidence interval about some mean µ:
Y-
t
2
s Y + t sY
, n 1
, n 1 Y
2
The half-length (d) of this confidence interval is therefore:
d t
2
,n 1
sY t
2
,n 1
s
n
This formula can be rearranged to estimate necessary sample size n
2
s2
2
Z 2
n t
2
,n1 d
d
2
2
2
Stein's Two-Stage procedure involves using a pilot study to estimate s2.
Note that n is now present at both sides of the equation: iterative approach
Example: An experimenter wants to estimate the mean height of certain plants. From a
pilot study of 5 plants, he finds that s = 10 cm. What is the required sample size, if he
wants to have the total length of a 95% CI about the mean be no longer than 5 cm?
Using n = t2 /2,n-1 s2 / d2, n is estimated iteratively,
initial-n
5
123
62
64
t5%, df
2.776
1.96
2.00
2.00
n
(2.776)2 (10)2 /2.52 = 123
(1.96)2 (10)2 / 2.52 = 62
64
64
Thus with 64 observations, he could estimate the true mean with a CI no
longer than 5 cm at =0.05.
To accelerate the iteration you can start with Z:
n = z2 s2 / d2 = (1.96)2 (10)2 / 2.52 = 62
9
2. 4. 4 Sample size estimation for a comparison of two means
When testing the hypothesis Ho: o, we can take into account the possibility of
a Type I and Type II error simultaneously.
To calculate n we need to known the alternative 1 or at least the minimum
difference we wish to detect between the means = |o - 1
The formula for computing n, the number of observations on each treatment, is:
n = 2 ( / (Z/2 + Z)
2
For = 0.05 and = 0.20, z= 0.8416 and z /2 = 1.96, (Z/2 + Z)2=7.85 8
We can define in terms of
If δ = 2σ, n ≈ 4
If δ = 1σ, n ≈ 16
If δ = 0.5σ, n ≈ 64
We rarely know 2 and must estimate it via sample variances:
s
n 2 pooled
2
t
t
,n1n 22 ,n1n 22
2
2
, where
s pooled
s12 s 22
2
n is estimated iteratively. If no estimate of s is available, the equation may be
expressed in terms of the CV, and as a proportion of the mean:
n 2 [(/) / ((Z/2 + Z)2 2(CV/%)2(Z/2 + Z)2
2
Example: Two varieties are compared for yield, with a previously estimated s =
2.25 (s=1.5). How many replications are needed to detect a difference of 1.5
tons/acre with a = 5%, and = 20%?
Approximate: n 2 (/(Z/2 + Z) = 2 (1.5/1.5)2(1.96+0.8416)= 15.7
2
Then use n = 2 (s / (t/2 + t) to estimate the sample size iteratively.
2
guesstimate n df = 2(n - 1)
16
30
17
32
t0.025
2.0423
2.0369
t0.20
0.8538
0.8530
estimated n
16.8
16.7
The answer is that there should be 17 replications of each variety.
10
2. 4. 5. Sample size to estimate population standard deviation
The chi-squared (2) distribution is used to establish confidence intervals around the
sample variance as a way of estimating the true, unknown population variance.
2. 4. 5. 1. The Chi- square distribution (ST&D p. 55)
0.5
2 df
4 df
6 df
0.4
0.3
0.2
0.1
0.0
1
2
3
4
5
6
C hi- sq uar e
Distribution of 2 , for 2, 4, and 6 degrees of freedom.
Relation between the normal and chi-square distributions.
The 2 distribution with df = n is defined as the sum of squares of n independent,
normally distributed variables with zero means and unit variances.
2α, df=1= Z2(0,1) α/2 2α/2, df=
2
1, 0.05 =
3.84 and Z2(0,1), 0.025 = t2, 0.025 = 1.962 = 3.84
Note: Z values from both tails go into the upper tail of the χ2 because of the disappearance of the minus sign in the
squaring. For this reason we use for the 2 and /2 for Z and t.
n
Z
2
i
i 1
(Yi ) 2
2
1
2
(Y )
2
i
If we estimate the parametric mean with a sample mean, we obtain:
1
2
(Yi Y )
2
(n 1) s 2
2
n
…due to: s 2
i 1
(Yi Y ) 2
n 1
n
(Y Y )
i 1
i
2
(n 1) s 2
This expression, which has a 2n-1 distribution, provides a relationship between the
sample variance and the parametric variance.
11
2. 4. 5. 2. Confidence interval for 2
We can make the following statement about the ratio (n-1) s2/2 that has2n-1
distribution,
P { 21-/2, n-1 (n-1) s2/2 2/2, n-1} = 1 -
Simple algebraic manipulation of the quantities in the inequality yields
P { 21-/2, n-1 /(n-1) s2/2 2/2, n-1/(n-1)} = 1 -
which is useful when the precision of s2 can be expressed in terms of the % of 2.
Or inverting the ratio and moving (n-1):
P {(n-1) s2/ 2/2, n-1 2 (n-1) s2/ 21-/2, n-1} = 1 -
2
which is useful to construct 95% confidence intervals for .
Example: What sample size is required to obtain an estimate of that
deviates no more than 20% from the true value of with 90% confidence?
2
2
Pr {0.8 < s/ < 1.2} = 0.90 = Pr {0.64 < s / < 1.44} = 0.90
thus
2
(1 - /2, n-1) / (n-1)=
2
0.64 and
(/2, n-1) / (n-1)=
1.44
2
Since is not symmetrical, the above two solutions may not identical for
small n. The computation involves an arbitrary initial n and an iterative
process:
n
21
31
41
36
35
df
(n-1)
20
30
40
35
35
1 - /2 = 95%
2
(n-1)
(n-1) /(n-1)
10.90
0.545
18.50
0.616
26.50
0.662
22.46
0.642
21.66
0.637
2
/2 = 5%
2
(n-1)
(n-1) /(n-1)
31.4
1.57
43.8
1.46
55.8
1.40
49.8
1.42
48.6
1.43
2
Thus a rough estimate of the required sample size is ~ 36.
12