Download Convergence in Probability and Convergence in Distribution

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Divergent series wikipedia , lookup

Series (mathematics) wikipedia , lookup

Transcript
STAT 516: Multivariate Distributions
Lecture 7: Convergence in Probability and Convergence in
Distribution
Prof. Michael Levine
November 12, 2015
Levine
STAT 516: Multivariate Distributions
Convergence in Probability
I
p
A sequence Xn → X (converges in probability to X ) if, for
any ε > 0
lim P(|Xn − X | ≥ ε) = 0
n→∞
I
In this context, X may be a constant a - a degenerate random
variable
I
Chebyshev’s inequality is a common way of showing
convergence in probability
Levine
STAT 516: Multivariate Distributions
Examples
1. Let Xn = X +
1
n
where X ∼ N(0, 1)
p
2. Easy to verify that (by Chebyshev’s inequality) Xn → X
1. For {Xn } s.t. the mean µ and variance σ 2 are finite
p
X̄n → µ
2. The weak law of large numbers - second moment must exist;
strong law does not require that - will not be proved in this
course
Levine
STAT 516: Multivariate Distributions
Convergence in probability is closed under linearity
p
p
p
I
Xn → X and Yn → Y implies Xn + Yn → X + Y
I
If a is a constant and Xn → X aXn → aX
I
Conclusion: convergence in probability is closed under linearity
p
Levine
p
STAT 516: Multivariate Distributions
Continuous Mapping Theorem for Convergence in
Probability
p
p
I
If g is a continuous function, Xn → X then g (Xn ) → g (X )
I
We only prove a more limited version: if, for some constant a,
p
g (x) is continuous at a, g (Xn ) → g (a)
I
Can be viewed as one of the statements of Slutsky theorem the full theorem to be stated later
Levine
STAT 516: Multivariate Distributions
Another useful property
p
p
p
I
If Xn → X and Yn → Y , Xn Yn → XY
I
Only prove in this form but can be generalized to
p
g (Xn , Yn ) → g (X , Y )
Levine
STAT 516: Multivariate Distributions
Consistency and convergence in probability
I
For X ∼ F (x; θ), θ ∈ Ω a statistic Tn is a consistent
estimator of θ if
p
Tn → θ
I
I
Weak Law of Large Numbers → X̄ is a consistent estimator of
µ
1 Pn
2
2 p
A sample variance S 2 = n−1
i=1 (Xi − X̄ ) → σ
I
By continuous mapping theorem S → σ
p
Levine
STAT 516: Multivariate Distributions
Example
I
I
I
I
Let X1 , . . . , Xn ∼ Unif [0, 1] and Yn = max{X1 , . . . , Xn }
n
The cdf of Yn is FYn (t) = θt for 0 < t ≤ θ
n
θ - Yn is a biased estimator
Check that EYn = n+1
Direct computation implies that
Yn is consistent...and so is
the unbiased estimator n+1
Yn
n
Levine
STAT 516: Multivariate Distributions
Convergence in Distribution
I
If for a sequence {Xn } with cdf FXn (x) , and a random
variable X ∼ FX (x)
lim FXn (x) = FX (x)
n→∞
D
for all points of continuity of FX (x), Xn → X - Xn converges
in distribution or in law to X
I
We say that FX (x) is the asymptotic distribution or the
limiting distribution of Xn
I
Occasional abuse of notation: Xn → N(0, 1)
Levine
STAT 516: Multivariate Distributions
Convergence in distribution and convergence in probability
I
Convergence in distribution is only concerned with
distributions and not at all with random variables
I
For a symmetric fX (x), X and −X have the same distribution
I
Let
Xn =
X
−X
if n is odd
if n is even
D
I
Clearly, Xn → X but there is no convergence in probability!
Levine
STAT 516: Multivariate Distributions
Example
I
Let X̄ ∼ N(0, σ 2 /n)
I
Check that
I
Conclude that Fn (x̄) converges to the point mass at zero

 0 x̄ < 0
1
lim Fn (x̄) =
x̄ = 0
n→∞
 2
1 x̄ > 0
Levine
STAT 516: Multivariate Distributions
Example
I
Convergence of pdfs/pmfs does NOT mean convergence in
distribution!
I
Define the pmf
pn (x) =
1
0
x = 2 + n1
elsewhere
I
limn→∞ pn (x) = 0 for any x
I
However, the limiting function of cdf’s is F (x) = 0 if x < 2
and F (x) = 1 if x ≥ 2 which is a cdf!
I
Convergence in distribution does take place!
Levine
STAT 516: Multivariate Distributions
Example
I
However...for Tn ∼ tn we have
Z t
Γ[(n + 12 )]
1
dy
Fn (t) =
√
n
2 /n)(n+1)/2
(1
+
y
πnΓ
−∞
2
I
Stirling’s formula:
Γ(k + 1) ≈
√
2πk k+1/2 exp (−k)
I
The limit under the sign of integral is the normal pdf...so
Z t
1
√ exp (−y 2 /2) dy
lim Fn (t) =
n→∞
2π
−∞
I
The limiting distribution of tn is N(0, 1)
Levine
STAT 516: Multivariate Distributions
Example
I
Recall that for X1 , . . . , Xn ∼ Unif [0, θ] Yn = max1≤i≤n is the
consistent estimator of θ
I
Now we can say more...let Zn = n(θ − Yn )
I
For any t ∈ (0, nθ)
t/θ n
P(Zn ≤ t) = P(Yn ≥ θ − (t/θ)) = 1 − 1 −
n
I
Since limn→∞ P(Zn ≤ t) = 1 − exp (−t/θ) for some
D
Z ∼ exp(θ) Zn → Z
Levine
STAT 516: Multivariate Distributions
Relationship between convergence in probability and
convergence in distribution
p
D
I
If Xn → X , Xn → X
I
The converse is not true in general - see an earlier example!
I
However, if for a constant b Xn → b it also true that Xn → b
D
Levine
STAT 516: Multivariate Distributions
p
Basic properties of convergence in distribution
D
p
D
I
If Xn → X and Yn → 0, Xn + Yn → X
I
This is the magic wand if it is hard to show that Xn → X but
p
D
easy to show that some other Yn → X and Xn − Yn → 0
I
If Xn → X and g (x) is a continuous function on the support
of X ,
D
g (Xn ) → g (X )
I
If a and b are constants, Xn → X , An → a, and Bn → b,
D
D
p
D
p
D
An + Bn Xn → a + bX
Levine
STAT 516: Multivariate Distributions
Boundedness in probability
I
For any X ∼ FX (x), we can always find η1 and η2 s.t.
FX (x) < ε/2 for x ≤ η1 and FX (x) > 1 − ε/2 for x ≥ η2
I
Thus, for η = max{|η1 |, |η2 |}
P[|X | ≤ η] ≥ 1 − ε
I
Formal definition: {Xn } is bounded in probability if for any
ε > 0 there exist Bε > 0 and an integer Nε s.t. if n ≥ Nε
P[Xn ≤ Bε ] ≥ 1 − ε
I
Can show immediately that if Xn → X then {Xn } is bounded
in probability...the converse is not always true
D
Levine
STAT 516: Multivariate Distributions
Why a sequence that is bounded in probability may not
converge
1
2m
and X2m−1 = 1 +
1
2m
w.p.1
I
Define X2m = 2 +
I
All of the mass of this sequence is concentrated in [1, 2.5] and
so it is bounded in probability
I
Xn consists of two subsequences that converge to degenerate
RV’s Y = 2 and W = 1 in distribution
Levine
STAT 516: Multivariate Distributions
A useful property
I
Xn a sequence of random variables bounded in prob. and Yn a
sequence that converges to zero in probability
I
Then, Xn Yn → 0
I
Analog from the world of calculus: if A is a constant and
ξn = n1 then limn→∞ n1 = 0 and limn→∞ An = 0
p
Levine
STAT 516: Multivariate Distributions
MGF technique
I
If Xn has mgf MXn (t) for |t| ≤ h, X has mgf MX (t) for
|t| ≤ h1 < h, and limn→∞ MXn (t) = M(t) for |t| ≤ h1 , then
D
Xn → X
I
I
Take Yn ∼ b(n, p) with fixed µ = np for every n
h
in
t
Check that MYn (t) = [(1 − p) + pe t ]n = 1 + µ(e n−1)
Levine
STAT 516: Multivariate Distributions
Poisson approximation of the binomial: an example
I
1
Y ∼ b 50, 25
;
I
P(Y ≤ 1) =
I
Since µ = np = 2, we have the Poisson approximation
24 50
25
+ 50
1
25
24 49
25
= 0.4000
e −2 + 2e −2 = 0.406
Levine
STAT 516: Multivariate Distributions
Central Limit Theorem (CLT)
I
I
2
If X1 , . . . , Xn ∼ N(µ, σ 2 ) we know that X̄ ∼ N µ, σn
CLT: if X1 , . . . , Xn are independent, bE Xi = µ and
Var Xi = σ 2 , we have
√
n(X̄ − µ)
∼ N(0, 1)
σ
Levine
STAT 516: Multivariate Distributions