Download Solutions to Problems 1-9 1. Write down the Law of Iterated

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Central limit theorem wikipedia , lookup

Transcript
The University of Hong Kong – Department of Statistics and Actuarial Science – STAT2802 Statistical Models – Tutorial Solutions
The best way to solve a problem at hand is to look at a harder problem and an easier problem.
Solutions to Problems 1-9
1.
Write down the Law of Iterated Expectations and prove it. Then write down the Law of Total Probability and prove it.
Law of Iterated Expectations:
. Proof. The inner expectation is expanded as
transformation on the random variable . Mount the outer expectation and then expand: (
. Note that this is a
)
.
where { }
∑
Law of Total Probability (of events):
Then
entails the Law of Total Probability.
2.
. Therefore (
Chebyshev’s Inequality: {
.
)
}
(
, for any
!):
,
(
)
)
. Proof. Lemma: For any positive r.v.
,
. Thus
.
. Proof. {
}
 {
}
.
Write down the Weak Law of Large Numbers and prove it.
Weak Law of Large Numbers: Let { }
→
and
Write down Markov’s Inequality and prove it. Then write down Chebyshev’s Inequality and prove it.
Markov’s Inequality (Only for positive random variable:
3.
is a partition of the sample space . Proof. Let
be
. Proof. From Chebyshev’s Inequality,
i.i.d. random variables with finite variance. Let
{|
|
}
(
)
be their common mean. Then
→
,
{|
|
}
.
1
4.
What is the Moment Generating Function of a distribution? What is the Characteristic Function of a distribution? What
are the Moment Generating Function and the Characteristic Function of a univariate normal distribution?
Moment Generating Function: The moment generating function of the distribution of the r.v.
. Its r-th derivative with respect to at
,
, equals the r-th moment
Characteristic Function: The characteristic function of the distribution of the r.v.
the complex unit.
.
)
, where is
; its Characteristic function is
.
Write down the density of a univariate normal distribution and show that it integrates to 1.
∫
and
6.
is a function of some number : (
, its Moment Generating Function is
Density of the univariate normal distribution:
√
.
.
For the univariate normal distribution
5.
is a function of some number :
√
∫
√
∫
. Next to show it integrates to 1: ∫
, for
√
√
∫
√
∫
√
√
∫
√
( )
√
, where
.
Write down the Gamma function
Gamma function:
. “Base case”:
∫
[
]
∫
and show that
for any positive integer .
. Next to show the increment relationship “
.”:
. Then apply induction principle on the combination of the
base case and the increment relationship.
2
7.
Write down the Central Limit Theorem and prove it.
Central Limit Theorem: Let { }
(
)→
√
be
i.i.d. random variables with finite variance. Let
. Lemma.
be their common mean,
be their common variance. Then
. Pf of Lemma.
. Corollary.
CLT:
√
√
→
8.
√
,
√
√
√
which is the MGF of N(0,1). In the proof we have also used classical results of the Euler’s number: (
) →
.
Write down the density of the d-dimensional normal distribution. What are the scalar parameters of a general bivariate
normal density? Show that the components of a bivariate normal r.v. are independent of each other iff their correlation
coefficient is 0.
Density of the d-dimensional normal distribution:
For d=2: the vector-matrix parameters are
Next we show that “Let [
[
(only if)
→
√
√
. Proof of
][
, but
]
, then
]
[
√
],
[
if and only if
.
√
], the scalars involved are
.” (if) If
and then
(√
then
) √
[
.
] and then

√
, therefore
√
.
.
3
9.
Show that the correlation coefficient is absolutely bounded by 1.
Denote the two r.v.s by
and . Consider the non-negative expression
.
: For any :

4