Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Practice & Communication of Science From Distributions to Confidence @UWE_KAR The Normal Distribution Represents the idealised distribution of a large number of things we measure in biology many parameters approximate to the ND Is defined by just two things… population mean µ (mu) the centre of the distribution (mean=median=mode) population standard deviation (SD) σ (sigma) the distribution ‘width’ (mean point of inflexion) encompasses 68% of the area under the curve 95% of area found within 1.96 σ either side of mean The Normal Distribution Is symmetrical mean=median=mode The Normal Distribution One SD either side of mean includes 68% of represented population SD boundary is inflexion point 2* SD covers 95% curvature changes direction the ‘s’ bit (*actually 1.96) 3 SD covers 99.7% The Normal Distribution All Normal Distributions are similar differ in terms of… mean SD (governs how ‘spikey’ curve is) Fig below… 4 different SDs, 2 different means Standardising Normal Distributions Regardless of what they measure, all Normal Distributions can be made identical by… subtracting the mean from every reading dividing each reading by the SD the mean then becomes zero a reading one SD bigger +1 Called Standard Scores or z-scores amazing! Different measurements same ‘view’ Standard (z) Scores A ‘pure’ way to represent data distribution the actual measurements (mg, m, sec) disappear! replaced by number of SDs from the mean (zero) For any reading, z = (x - µ) / σ A survey of daily travel time had these results (in minutes): 26,33,65,28,34,55,25,44,50,36,26,37,43,62,35,38,45,32,28,34 The Mean is 38.8 min, and the SD is 11.4 min To convert the values to z-scores… eg to convert 26 first subtract the mean: 26 - 38.8 = -12.8, then divide by the Standard Deviation: -12.8/11.4 = -1.12 So 26 is -1.12 Standard Deviations from the Mean Familiarity with the Normal Distribution 95% of the class are between 1.1 and 1.7m tall what is the mean and SD? Assuming normal distribution… the distribution is symmetrical, so mean height is (1.7 - 1.1) / 2 = 1.4m the range 1.1 1.7m covers 95% of the class, which equals ± 2 SDs one SD = (1.7 – 1.1) / 4 = 0.6 / 4 = 0.15m Familiarity with the Normal Distribution One of that class is 1.85m tall what is the z-score of that measurement? Assuming normal distribution… z-score = (x - µ) / σ z = (1.85m - 1.4m) / 0.15m = 0.45m / 0.15m =3 note there are no units 3 SDs cover 99.7% of the population only 1.5 in 1000 of the class will be as tall/taller a big class, with fractional students! Familiarity with the Normal Distribution 36 students took a test; you were 0.5 SD above the average; how many students did better? from the curve, 50% sit above zero from the curve, 19.1% sit between 0 and 0.5 SD so 30.9% sit above you 30.9% of 36 is about 11 Familiarity with the Normal Distribution Need to have a ‘feel’ for this… Populations and Samples – a Diversion? A couple of seemingly pedantic but important points about distributions… population the potentially infinite group on which measurements might be made don’t often measure the whole population sample a sub-set of the population on which measurements are actually made most studies will sample the population n is the number studied n-1 called the ‘degrees of freedom’ often extrapolate sample results to the population Populations and Samples – so what? They are described/calculated differently… μ is the population mean, x is the sample mean σ or σn is population SD, s or σn-1 is sample SD Calculating the SD is different for each most calculators do it for you… as long as you choose the right type (pop vs samp) Populations and Samples – choosing Analysing the results of a class test… Analysing the results of a drug trial… sample, since you expect the conclusions to apply to the larger population A national census collects information about age population, since you don’t intend extrapolating the results to all students everywhere population, since by definition the census is about the population taking part in the survey If in doubt, use the sample SD and as n increases, the difference decreases Populations and Samples – implications If the sample observed is the population, then the mean and SD of that sample are the population mean and the population SD ie you calculate σ (or σn) If the sample is part of the bigger population, then the sample mean and SD are estimates of the population mean and the population SD ie you calculate σn-1 (or s) (the presumption here is you chose your sample to reflect the population!) Implications of Estimating Pop Mean For a sample, the ‘quality’ of its estimate of the population mean and SD depends on the number of observations made the mean from a sample of 1 member of the population is unlikely to be close to the pop mean if you sampled the whole population, your sample estimate is the population mean in between, adding extra readings to a sample will improve your estimate of the population mean also, repeated sampling a collection of means eg you sampled 200 FVCs and did this 100 times the 100 means are ND and will have its own SD (!) called the Standard Error of the Mean (SEM) The Standard Error of the Mean Immediate recap… repeated samples of a distribution will produce different estimates of the population mean the SD of those estimates of pop mean called the SEM Surprisingly easy to calculate we don’t need to keep repeating our sampling can derive it from SD of data in one sample SEM = sample standard dev / square root of number of samples SEM = s / √ N eg if N=16, then SEM is 4x smaller than SD Remember, Samples Estimates When we sample a population, we end up with a sample mean, x our ‘best guess’ estimate of the real pop mean, µ Our sample also has a measure of the variability of the data that comprises it the sample Standard Deviation, s the ‘real’ mean of the population is ‘hidden’ which is also an estimate of the population SD, σ s can be also be used to indicate the variability of the population mean itself SEM = s / √ N can then use SEM to determine confidence limits Confidence Limits and the SEM The SEM reflects the ‘fit’ of a sample mean, x, to the underlying population mean, µ if we calculate two sample means and they are similar, but for one the SEM is high, we are less ‘confident’ about how well that sample mean estimates the population mean Just like the ‘raw’ data used to calculate a sample mean follows a distribution, so do repeat estimates of the population mean itself this is the t Distribution similar to the Normal Distribution The t Distribution Yet another distribution! but distributions are important because they define how we expect our data to behave if we know that, then we gain insight into our expts! Generally ‘flatter’ than the Normal Distribution any particular area is more ‘spread out’ (less clear) the more ‘pointed’ a curve, the clearer the peak t Distribution Pointedness Varies! Logic… the number of samples influences the ‘accuracy’ of our estimate of the population mean from the sample mean as N increases, the ‘peak’ becomes sharper a given area of the curve is less ‘spread out’ At large N, t Distribution = Normal Distribution and for both, 95% of curve contained in ± 1.96 SD Using the t Distribution When we calculate a sample mean and call it our estimate of the population mean… it’s nice to know how ‘confident’ we are in that estimate One measure of confidence is the 95% Confidence Interval (95%CI) the range over which we are 95% confident the ‘true’ population mean lies derived from our sample mean (we calculate) and our SEM (we calculate) and the N (though it’s the ‘degrees of freedom’, N-1) we use this to look up a ‘critical t value’ From t Distribution to 95%CI The t Distribution is centred around our mean and its shape is influenced by N-1 95%CI involves chopping off the two 2.5% tails Need a t table to look up how many SEMs (SDs) along the x-axis this point will be Value varies with N-1 And level of confidence sought Step 1: The t Table t value varies with… Row… DoF is N-1 Column… level of ‘confidence’ 95%CI involves chopping off the two 2.5% tails α = 0.05 (5%) For N = 10, α = 0.05 t(N-1),0.05 = 2.262 when N large, t=1.96 Step 2: Using the t value The t value is the number of SEMs along the xaxis (in each direction) that encompasses that % of the t distribution centred on our mean 2.262 in the case of t(N-1),0.05 Eg we measure the FVC (litres) of 10 people… mean = 3.83, SD = 1.05, N = 10 SEM = 1.05/√10 = 0.332 litres t(N-1),0.05 = 2.262 standard errors to cover 95% curve So, litres either side of the mean = 2.262 * 0.332 = 0.751 litres either side of mean covers 95% of dist So, 95%CI is 3.83 ± 0.751 = 3.079 4.581 litres (3.079, 4.581) Effect of Bigger N A larger sample size gives us greater confidence in any population mean we estimate In previous example… so 95%CI should be smaller mean = 3.83, SD = 1.05, N =10, SEM = 0.332 95%CI is (3.079, 4.581) But say we measured 90 more people… mean = 3.55, SD = 0.915, N = 100 mean and SD similar to before, but SEM now a lot smaller, at 0.915/√100 = 0.0915 so too is t(N-1),0.05 = 1.96 (rather than 2.262) 95%CI = 3.55 ± (1.96 * 0.0915) = 3.371 3.729 Effect of Bigger N A bigger N ‘sharpens’ the t distribution so that the 95% boundaries are less far apart ie our confidence interval will become smaller 95%CI also shrinks because SEM = SD/√N Summary Normal Dist fully defined just by mean and SD SD calculation differs for sample vs population Estimation of population mean from a sample is always prone to uncertainty Standard Error of Mean (s/√N) reflects uncertainty Estimates of means follow the t Distribution Transformation to z-scores makes all NDs identical t Distribution becomes ‘sharper’ with higher N ‘Width’ of t dist covering 95% is called 95%CI range in which 95/100 mean estimates would fall 95%CI = mean ± (t(N-1),0.05 * SEM) t is the number of SEMs along dist covering that %