Download (Feb 4: t-distributions) I`m confused about the use of the t

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Foundations of statistics wikipedia , lookup

Psychometrics wikipedia , lookup

Taylor's law wikipedia , lookup

Bootstrapping (statistics) wikipedia , lookup

History of statistics wikipedia , lookup

Statistical inference wikipedia , lookup

Degrees of freedom (statistics) wikipedia , lookup

Resampling (statistics) wikipedia , lookup

Misuse of statistics wikipedia , lookup

Student's t-test wikipedia , lookup

Transcript
(Feb 7: Non-parametric statistics) It seems like ranking observations reveals relatively little
about the spread of the data. Even so, in what types of circumstances would a non-parametric
statistic test be better than a parametric one? In research, how common is it to find data that
doesn’t seem to follow any sort of distribution?
Very common, to answer your last question first. Most distributions are summarized by a few
parameters (for the normal it’s just 2) and it’s often unrealistic to think that an entire rich dataset
can be ``explained’’ by just a few summary statistics or parameters. If a dataset has many outliers,
heavy skewing, more than one mode, or just shows some unusual pattern, a parametric procedure
may be unreliable and give poor results. In this setting non-parametric procedures are preferred.
They are also often easier to compute and their underlying theory more accessible.
Non-parametric procedures can also give conservative “bounds” on results. For instance, if
both your parametric and non-parametric test shows that an effect is significant then this gives
added confidence to the result. You are correct that ranking observations reveals little about the
spread, but that is precisely the point. It means that the procedure is insensitive to the actual
values of the data. It’s analogous to why the median is useful as a measure of center. The median
reveals little about the extreme values of the data, and therefore is not sensitive to outliers. If the
data isn’t symmetric, the median is often preferred over the mean as a measure of center. If the
data isn’t easily summarized by a few parameters, non-parametric procedures are preferred.
One other advantage to the rank-based methods presented in Chapter 11 is that they are
sometimes the only procedures that can be used with ordinal-type data. Ordinal data refers to
measurements where only the comparisons “greater”, “less”, or “equal” being measurements are
relative. For instance, when a person is asked to assign a number between 1 and 10 to their
political position, where 1=left-wing and 10=right-wing, one is using an ordinal scale of
measurement. Instead of 1 to 10 we could have used 10 to 100 or any increasing list of 10
numbers. For such data, what’s important isn’t the value of the numbers themselves but the
comparisons between numbers. Rank-based methods are often the most appropriate procedures to
use with ordinal data. Note, however, that the applicability of non-parametric statistics is much
wider than this.
(Feb 4: t-distributions) I’m confused about the use of the t-distribution. The t-distribution is like
a normal distribution, but why do we use the t? What makes this distribution so useful?
If you have a sample from a normal population, the mean of that sample will also have a
normal distribution. And if you normalize the mean, that is subtract off the population mean
(mu) and divide by the population standard deviation over root n (sigma/sqrt(n)) then that
random quantity will have a standard normal distribution ( N(0,1) )
However, if you replace the population standard deviation with the sample standard
deviation (which is itself random, being based on the sample), that quantity
(x-bar - mu) / (s / sqrt(n) )
does not have a normal distribution. Theory tells us it has a t-distribution. That t-distribution
also depends on n through some parameter which we call the
"degrees of freedom". We say it has a t-distribution with n-1 degrees of freedom.
The lab a few classes ago was meant to look at the differences between the normal
distribution and the t-distribution. When n is large they're both fairly close. But when n is small
the t-distribution has a lot more spread to it. So for small sample sizes, when the population is
fairly normal, to get accurate confidence intervals and hypothesis test results when we use the
sample mean, we need to use t-related procedures, as opposed to z-related (normal) procedures.
. . . Related to this question is . . .
(Feb 4: Degrees of freedom) When samples of two populations are used in the t test, they are
assumed to have the same variance. When this isn’t the case, the degrees of freedom for a t
distribution must be calculated from the estimated variances of the populations. Conceptually,
what does variance have to do with degrees of freedom?
Also, what exactly are degrees of freedom and why do we consider them in t-distributions?
There are two answers to this: (a) One simple explanation is “it’s just a word!’’ That is, the tdistribution is a distribution that depends on n through some parameter (we could have called it
theta). We just happen to give that parameter the ungodly name of “degrees of freedom.” The
reason the t-distribution in statistical inference depends on the sample size n (unlike the normal
distribution), is because we are using an estimator for the population variance (that is, using s for
sigma), which depends on n. (The larger n is the better the variance estimate becomes.)
(b) To give more explanation behind the terminology, the concept of degrees of freedom is
widespread in statistics. Roughly it is used for the number of “independent” or “free”
measurements needed to estimate the variability. With a sample of size n we start off with n
independent measurements. However, the estimator for the sample variance relies on x-bar,
which, in a sense, uses up one piece of information. That is, if I have n measurements, and the
average of those measurements, I have a redundant piece of information. From n-1 measurements
and x-bar I can calculate the nth measurement. In a two-sample t-test with two samples of sizes n
and m, respectively, we have n+m total measurements. But we use x-bar and y-bar, the means of
the two groups, giving up 2 “degrees of freedom.” That’s why the test statistic for that test has
n+m-2 degrees of freedom.
(Feb 4, what if the experiment isn’t truly random) If after an experiment was done we realize
that the controls were not picked at random, what can we do to correct for this error? Do we
have to start over again?
A good statistical design must put its top priority on getting good data. This means getting the
sample as representative of the population as possible with as little bias as possible. The most
basic assumptions underlying all the procedures we do is that we have a
genuine random sample. Without that, all of the inference fails. So to answer your question we
might very well have to start back over at square one.