Download HERE

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Central limit theorem wikipedia , lookup

Transcript
Practice & Communication of Science
From Distributions to Confidence
@UWE_KAR
The Normal Distribution

Represents the idealised distribution of a large
number of things we measure in biology


many parameters approximate to the ND
Is defined by just two things…

population mean



µ (mu)
the centre of the distribution (mean=median=mode)
population standard deviation (SD)




σ (sigma)
the distribution ‘width’ (mean  point of inflexion)
encompasses 68% of the area under the curve
95% of area found within 1.96 σ either side of mean
The Normal Distribution

Is symmetrical

mean=median=mode
The Normal Distribution

One SD either side of mean includes 68% of
represented population

SD boundary is inflexion point



2* SD covers 95%


curvature changes direction
the ‘s’ bit
(*actually 1.96)
3 SD covers 99.7%
The Normal Distribution

All Normal Distributions are similar

differ in terms of…



mean
SD (governs how ‘spikey’ curve is)
Fig below…

4 different SDs, 2 different means
Standardising Normal Distributions

Regardless of what they measure, all Normal
Distributions can be made identical by…

subtracting the mean from every reading


dividing each reading by the SD


the mean then becomes zero
a reading one SD bigger  +1
Called Standard Scores or z-scores

amazing! Different measurements  same ‘view’
Standard (z) Scores

A ‘pure’ way to represent data distribution







the actual measurements (mg, m, sec) disappear!
replaced by number of SDs from the mean (zero)
For any reading, z = (x - µ) / σ
A survey of daily travel time had these results (in minutes):
26,33,65,28,34,55,25,44,50,36,26,37,43,62,35,38,45,32,28,34
The Mean is 38.8 min, and the SD is 11.4 min
To convert the values to z-scores…




eg to convert 26
first subtract the mean: 26 - 38.8 = -12.8,
then divide by the Standard Deviation: -12.8/11.4 = -1.12
So 26 is -1.12 Standard Deviations from the Mean
Familiarity with the Normal Distribution

95% of the class are between 1.1 and 1.7m tall


what is the mean and SD?
Assuming normal distribution…


the distribution is symmetrical, so mean height is
(1.7 - 1.1) / 2 = 1.4m
the range 1.1  1.7m covers 95% of the class,
which equals ± 2 SDs



one SD = (1.7 – 1.1) / 4
= 0.6 / 4
= 0.15m
Familiarity with the Normal Distribution

One of that class is 1.85m tall


what is the z-score of that measurement?
Assuming normal distribution…




z-score = (x - µ) / σ
z = (1.85m - 1.4m) / 0.15m
= 0.45m / 0.15m
=3


note there are no units
3 SDs cover 99.7% of the population

only 1.5 in 1000 of the class will be as tall/taller

a big class, with fractional students! 
Familiarity with the Normal Distribution

36 students took a test; you were 0.5 SD above
the average; how many students did better?




from the curve, 50% sit above zero
from the curve, 19.1% sit between 0 and 0.5 SD
so 30.9% sit above you
30.9% of 36 is about 11
Familiarity with the Normal Distribution

Need to have a ‘feel’ for this…
Populations and Samples – a Diversion?

A couple of seemingly pedantic but important
points about distributions…

population

the potentially infinite group on which measurements
might be made


don’t often measure the whole population
sample

a sub-set of the population on which measurements
are actually made


most studies will sample the population
n is the number studied


n-1 called the ‘degrees of freedom’
often extrapolate sample results to the population
Populations and Samples – so what?

They are described/calculated differently…



μ is the population mean, x is the sample mean
σ or σn is population SD, s or σn-1 is sample SD
Calculating the SD is different for each


most calculators do it for you…
as long as you choose the right type (pop vs samp)
Populations and Samples – choosing

Analysing the results of a class test…


Analysing the results of a drug trial…


sample, since you expect the conclusions to apply
to the larger population
A national census collects information about
age


population, since you don’t intend extrapolating the
results to all students everywhere
population, since by definition the census is about
the population taking part in the survey
If in doubt, use the sample SD

and as n increases, the difference decreases
Populations and Samples – implications

If the sample observed is the population, then
the mean and SD of that sample are the
population mean and the population SD


ie you calculate σ (or σn)
If the sample is part of the bigger population,
then the sample mean and SD are estimates
of the population mean and the population SD


ie you calculate σn-1 (or s)
(the presumption here is you chose your sample to
reflect the population!)
Implications of Estimating Pop Mean

For a sample, the ‘quality’ of its estimate of the
population mean and SD depends on the
number of observations made




the mean from a sample of 1 member of the
population is unlikely to be close to the pop mean
if you sampled the whole population, your sample
estimate is the population mean
in between, adding extra readings to a sample will
improve your estimate of the population mean
also, repeated sampling  a collection of means



eg you sampled 200 FVCs and did this 100 times
the 100 means are ND and will have its own SD (!)
called the Standard Error of the Mean (SEM)
The Standard Error of the Mean

Immediate recap…



repeated samples of a distribution will produce
different estimates of the population mean
the SD of those estimates of pop mean called the
SEM
Surprisingly easy to calculate




we don’t need to keep repeating our sampling
can derive it from SD of data in one sample
SEM = sample standard dev / square root of
number of samples
SEM = s / √ N

eg if N=16, then SEM is 4x smaller than SD
Remember, Samples  Estimates

When we sample a population, we end up
with a sample mean, x

our ‘best guess’ estimate of the real pop mean, µ


Our sample also has a measure of the
variability of the data that comprises it

the sample Standard Deviation, s


the ‘real’ mean of the population is ‘hidden’
which is also an estimate of the population SD, σ
s can be also be used to indicate the
variability of the population mean itself

SEM = s / √ N

can then use SEM to determine confidence limits
Confidence Limits and the SEM

The SEM reflects the ‘fit’ of a sample mean, x,
to the underlying population mean, µ

if we calculate two sample means and they are
similar, but for one the SEM is high,


we are less ‘confident’ about how well that sample
mean estimates the population mean
Just like the ‘raw’ data used to calculate a
sample mean follows a distribution, so do
repeat estimates of the population mean itself


this is the t Distribution
similar to the Normal Distribution
The t Distribution

Yet another distribution! 

but distributions are important because they define
how we expect our data to behave


if we know that, then we gain insight into our expts!
Generally ‘flatter’ than the Normal Distribution

any particular area is more ‘spread out’ (less clear)

the more ‘pointed’ a curve, the clearer the peak
t Distribution Pointedness Varies!

Logic…


the number of samples influences the ‘accuracy’ of
our estimate of the population mean from the
sample mean
as N increases, the ‘peak’ becomes sharper


a given area of the curve is less ‘spread out’
At large N, t Distribution = Normal Distribution

and for both, 95% of curve contained in ± 1.96 SD
Using the t Distribution

When we calculate a sample mean and call it
our estimate of the population mean…


it’s nice to know how ‘confident’ we are in that
estimate
One measure of confidence is the 95%
Confidence Interval (95%CI)

the range over which we are 95% confident the
‘true’ population mean lies



derived from our sample mean (we calculate)
and our SEM (we calculate)
and the N (though it’s the ‘degrees of freedom’, N-1)

we use this to look up a ‘critical t value’
From t Distribution to 95%CI


The t Distribution is centred around our mean
and its shape is influenced by N-1
95%CI involves chopping off the two 2.5% tails



Need a t table to look up how many SEMs (SDs)
along the x-axis this point will be
Value varies with N-1
And level of
confidence sought
Step 1: The t Table

t value varies with…

Row…


DoF is N-1
Column…


level of ‘confidence’
95%CI involves
chopping off the
two 2.5% tails


α = 0.05 (5%)
For N = 10, α = 0.05


t(N-1),0.05 = 2.262
when N large, t=1.96
Step 2: Using the t value

The t value is the number of SEMs along the xaxis (in each direction) that encompasses that
% of the t distribution centred on our mean


2.262 in the case of t(N-1),0.05
Eg we measure the FVC (litres) of 10 people…

mean = 3.83, SD = 1.05, N = 10



SEM = 1.05/√10 = 0.332 litres
t(N-1),0.05 = 2.262 standard errors to cover 95% curve
So, litres either side of the mean = 2.262 * 0.332


= 0.751 litres either side of mean covers 95% of dist
So, 95%CI is 3.83 ± 0.751 = 3.079  4.581 litres

(3.079, 4.581)
Effect of Bigger N

A larger sample size gives us greater
confidence in any population mean we estimate


In previous example…



so 95%CI should be smaller
mean = 3.83, SD = 1.05, N =10, SEM = 0.332
95%CI is (3.079, 4.581)
But say we measured 90 more people…

mean = 3.55, SD = 0.915, N = 100




mean and SD similar to before, but
SEM now a lot smaller, at 0.915/√100 = 0.0915
so too is t(N-1),0.05 = 1.96 (rather than 2.262)
95%CI = 3.55 ± (1.96 * 0.0915) = 3.371  3.729
Effect of Bigger N

A bigger N ‘sharpens’ the t distribution so that
the 95% boundaries are less far apart


ie our confidence interval will become smaller
95%CI also shrinks because SEM = SD/√N
Summary

Normal Dist fully defined just by mean and SD



SD calculation differs for sample vs population
Estimation of population mean from a sample is
always prone to uncertainty


Standard Error of Mean (s/√N) reflects uncertainty
Estimates of means follow the t Distribution


Transformation to z-scores makes all NDs identical
t Distribution becomes ‘sharper’ with higher N
‘Width’ of t dist covering 95% is called 95%CI



range in which 95/100 mean estimates would fall
95%CI = mean ± (t(N-1),0.05 * SEM)
t is the number of SEMs along dist covering that %