Download Important Formulas for Statistics

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Bootstrapping (statistics) wikipedia , lookup

Foundations of statistics wikipedia , lookup

History of statistics wikipedia , lookup

Sufficient statistic wikipedia , lookup

Central limit theorem wikipedia , lookup

Resampling (statistics) wikipedia , lookup

Law of large numbers wikipedia , lookup

Transcript
Important Formulas for Statistics
Basic stuff
π‘‰π‘Žπ‘Ÿ(𝑋) = πΆπ‘œπ‘£(𝑋, 𝑋) = 𝐸[(𝑋 βˆ’ πœ‡)(𝑋 βˆ’ πœ‡)] = 𝐸[(𝑋 βˆ’ πœ‡)2 ] = 𝐸[𝑋 2 ] βˆ’ (𝐸[𝑋])2
Variance
Properties
Expected
value
Properties
Is a linear operator, E[XY]=E[X]E[Y] if independant.
Limit in
probability
A sequence {Xn} of random variables converges in probability towards X if for all Ξ΅ > 0
Properties of
prob.
convergence
Gaussian
Integral
Convergence in probability implies convergence in distribution
, x can be a linear function of the integrand variable
Joint,
Marginal and
Conditional
probability
Expectation
and var
conditional
Conditional
Probaibility
=
𝑓(π‘Ž,𝑏)
𝑓(𝑏)
Sum of a RVits mean
βˆ‘(π‘₯𝑖 βˆ’ π‘₯Μ… ) = 0
Sample
Variance
Prop of log
Liebniz rule
on the
differential of
an integral
Binomial
Coefficient
(helpful in convolution of poisson)
Expectation
(useful for the proof of Rao Blackwell thm)
in
conditional
Transformations
One Variable
(discreet)
One Variable
(continuous)
Two Variables
Jacobian
Order Statistics
𝑛[𝐹(π‘₯)]π‘›βˆ’1 𝑓(π‘₯)
Maximum
In general
Same distribu.
Min
Between two
(diff
distribution)
=
Between min
and max
Idea behind
multiplication
of supports
𝑛(𝑛 βˆ’ 1)[𝐹(𝑦) βˆ’ 𝐹(π‘₯)]π‘›βˆ’2 𝑓(π‘₯)𝑓(𝑦)
If 1(xi<ΞΈ) then multiplication can be summarized in 1(x(n)<ΞΈ). Conversely if
1(xi<1/ΞΈ)οƒ  1(ΞΈ <1/x(n)).
Convolutions
Always check that the supports are compatible
to establish the limits of the
integration/summation
Remember that in a convolution the support can be seen as a straight line crossing the plane in (z,0)
and (0,z). (e.g. when adding two uniforms the limits of integration vary when z>1)
Sufficiency
Sufficiency (I) Conditional
distribution approach (By
definition)
π‘ƒπœƒ (𝑋 = π‘₯|𝑇 = 𝑑) = 𝑃(𝑋 = π‘₯|𝑇 = 𝑑). That is this pmf/pdf is free of ΞΈ.
𝑃(𝑋 = π‘₯) βˆ— 𝑃(𝑇 = 𝑑|𝑋 = π‘₯) 𝑃(𝑋 = π‘₯)
π‘ƒπœƒ (𝑋 = π‘₯|𝑇 = 𝑑) =
=
βˆ— 1(𝑇(π‘₯) = 𝑑)
𝑃(𝑇 = 𝑑)
𝑃(𝑇 = 𝑑)
β€²
𝑓𝑋|𝑇=𝑑 (π‘₯) π‘‘π‘œπ‘’π‘ π‘› π‘‘π‘–π‘›π‘£π‘œπ‘™π‘£π‘’ πœƒ
Sufficiency (II) Neyman
πΉπ‘Žπ‘π‘‘π‘œπ‘Ÿπ‘–π‘§π‘Žπ‘‘π‘–π‘œπ‘›: 𝐿(πœƒ, π‘₯) = β„Ž(π‘₯)𝑔(𝑇(π‘₯), πœƒ)οƒ  T(x) is suff.
Factorization theorem
Hint for proof: Dicreete case only: X is contained in T(x). Then
make g(T(x),ΞΈ)=PΞΈ(T(X)=T(x)).
οƒŸ: Use the distribution of T:Pπœƒ (T(X) = t). = βˆ‘
β†’:𝑇(β†’)=𝑑
𝑦
𝐿(πœƒ) to divide in
𝑦
𝐿(πœƒ, π‘₯) , to ultimately find that PΞΈ(X=x|T(X)=t) doesn’t depend on ΞΈ.
Minimal sufficiency:
Lehmann Scheffé (also
β„Ž(π‘₯, π‘₯𝑑 ; πœƒ) =
𝐿(πœƒ,π‘₯)
𝐿(πœƒ,π‘₯𝑑 )
↔ 𝑇(π‘₯) = 𝑇(π‘₯𝑑 ) 𝑠. 𝑑. β„Ž 𝑖𝑠 π‘“π‘Ÿπ‘’π‘’ π‘“π‘Ÿπ‘œπ‘š πœƒ β†’
𝑇 𝑖𝑠 π‘šπ‘–π‘›π‘ π‘’π‘“π‘“ π‘“π‘œπ‘Ÿ πœƒ.
works with non identically
If we T(x)=T(y) only happens for x=y then we can look at the order
distr)
statistics.
Prof: The sufficiency is with a partition and the β†’, the minimal is
with ← and using the fact that T is a function of another sufficient.
Information
Helpful in: CRLB, Asymptotic MLE
Arcillarity
Good trick: If the distribution of Y=ΞΈX does not involve ΞΈ, then T=Xi/Xj, doesn’t either.
Families
i)Location,
ii)Scale
iii)Locationscale.
Ancillary in each family is easy to achieve by subtraction and/or
division by another element of the sample (or an order stat)
Completeness
A stat both Complete and
A function 1-1 of a minsuff is a minsuff
sufficient οƒ  minsuff
Basus theorem:
w is complete and sufficient for ΞΈ, U is ancillary, οƒ  U and W are
independent
Hint proof: WTS: Conditional distribution of U on W is the distribution
of U. Then use completeness using these distributions as functions.
UMVUE and Estimators
UMVUE: Smallest variance in the family of unbiased estimator.
How to find estimators?
Method of moments
Equate the theoretical moments to the sample ones
Maximum likelihood
1. Find the L 2. Ln(L) 3.Maximize (and check is maximum) using FOC
and SOC
Invariance of MLE
You can apply a function to the MLE st you get the MLE of the function
How to compare them?
Mean Squared error
Bias
Μ‚) βˆ’ 𝜽
B(𝜽)=𝑬(𝜽
BLUE
Sumatory (same weight) of the unbiased estimators
How to improve them?
Rao Blackwell (improve)
T is an unbiased est. of Ο„(ΞΈ), U is (jointly) sufficient for ΞΈ. π‘Š =
𝑔(𝑒) = 𝐸𝑒 [𝑇|π‘ˆ = 𝑒]. W is unbiased for Ο„, and has lower variance
than T.
Proof: Use Confitional expectation to check that W is unbiased,
2
and use the Var(T)=E[(T- Ο„(ΞΈ)) ] with W on the middle.
CRLB (compare)
Hint proof: Use chain rule of τ’(ΞΈ) so τ’(ΞΈ)=Cov(TY)<V(T)V(Y) then
equate with the V(Y)=nE(…)
Lehman Scheffé (find)
T is an unbiased est. of Ο„(ΞΈ), U is (jointly) complete and sufficient
for ΞΈ. π‘Š = 𝑔(𝑒) = 𝐸𝑒 [𝑇|π‘ˆ = 𝑒]. W is UMVUE.
Hint proof: Suppose W* another UMVUE , use completeness to
prove W*=W
β€˜How to apply?
1) Find a complete and suff stat for ΞΈ
2) Find a function of this compsuff stat that is unbiased for
the thing you are trying to estimate. Sometimes the
function could be a simple multiplication or sum of 1)
THE EXPONENTIAL FAMILY (and reasons to love it)
𝑓(π‘₯, πœƒ) = π‘Ž(πœƒ) βˆ— 𝑔(π‘₯) βˆ— 𝑒 𝑏(πœƒ)βˆ—π‘…(π‘₯)
1. We can Easily find the Expected value πΈπœƒ (𝑅(π‘₯)) =
βˆ’π‘Žβ€²(πœƒ)
,
π‘Ž(πœƒ)𝑏′(πœƒ)
2. T(x)=Ξ£R(xi) is minsuff and complete.
3. It follows the MLR property
Tests
Errors
Simple vs simple:
Neymann Pearson Lemma
1)Define the test
STEPS
2)Find the Likelihood ratio
3) Choose the statistics of interest and define Rejection region: An MP
a level test always depend on the sufficient statistic.
4)Operate to a form that is implementable either by knowing the
distribution of the statistic (take it to chi or normal) or by using
πœƒ
π‘ƒπœƒ0 (𝑅) = 𝛼 (e.g. βˆ«π‘˜ 0 𝑓(π‘₯)𝑑π‘₯
Remember that if the statistic of interest is decreasing because of the
other constants then k must behave properly. (eg. In a normal when
µ0< µ1, then the rejection is (…)>z, but if µ0> µ1 then the rejection is
(…)<-z.
One side composite
UMP via Neymann Person
Fix a value ΞΈ1>ΞΈ0. Then we can find the simple vs simple, and say that if
it doesn’t matter which value of theta we choose it also applies here.
MLR property
One side composite with
or
MLR property (Karlin
Rubin)
If MLR Non decreasing:
IF MLR non increasing:
With T(x)
being a suff stat.
GLRT
1)Define the Rejection region
2)Calculate
3) The sup(…) is the MLE, so in the case in which the null hypothesis is
simple then is simply replacing, while for the alternative we shall use
the MLE.
4)Operate to a form that is implementable by operating the K with the
KNOWN variables, that is express the test in terms of the MLE and the
parameter of the null. Check if the exponents are making a two sided
inequality.
πœƒ
5)Apply π‘ƒπœƒ0 (𝑅) = 𝛼 (e.g. βˆ«π‘˜ 0 𝑓(π‘₯)𝑑π‘₯ )
Asymptotic
FOR MLE
Delta Method
You can then operate the left inside affecting the right as regular normal. E.g. multipliying by a
constant will multiply the µ but will be squared with the Οƒ
You can also apply the delta method in this case, in which case you arrive to MLE being Efficient (by
CRLB).
Confidence intervals
Assymptotic: