Download PHY 4105: Quantum Information Theory Lecture 2

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Mathematics of radio engineering wikipedia , lookup

Proofs of Fermat's little theorem wikipedia , lookup

Infinite monkey theorem wikipedia , lookup

Karhunen–Loève theorem wikipedia , lookup

Tweedie distribution wikipedia , lookup

Expected value wikipedia , lookup

Law of large numbers wikipedia , lookup

Transcript
PHY 4105: Quantum Information Theory
Lecture 2
Anil Shaji
School of Physics, IISER Thiruvananthapuram
(Dated: August 6, 2013)
A.
Laws of large numbers
The ability to estimate the probabilities p(x) for a random variable taken on each of its allowed
values depend on the various forms of the laws of large numbers. It tells you how the probability
vector - whether it is an artifact of your mind or otherwise - can take shape based on a finite
number of observational data on events, atomic or otherwise. Take care not to confuse laws of
large numbers with the central limit theorem, which in many cases end up saying the same thing.
In particular the weak law says that the mean taken over several instances of a random variable
will converge to the expected value. It guarantees stable averages even for random events.
Chebyshev inequality: For any non-negative random variable X with finite mean,
P (x > a) ≤
hxi
.
a
Proof
∞
Z
Z
dx p(x) ≤
P (x > a) =
a
a
∞
x
dx p(x) ≤
a
Z
∞
0
x
hxi
dx p(x) =
.
a
a
For a random variable X with a finite mean and variance
P (|x − hxi| > a) = P ((x − hxi)2 > a2 ) ≤
h(∆x)2 i
,
a2
where the variance of X is
h(∆x)2 i = h(x − hxi)2 i = hx2 i − hxi2 .
Let us consider N tosses of a six sided die. Each toss may be considered as a random variable,
Xj which takes values 1 through 6. Alternatively one can think of N identical dies being tossed
and the results of each being taken as random variable. What matters is that in both cases, the
probability distribution for the outcomes is identical, either from trial to trial or from die to die as
the case may be. Implicit in this statement is the fact that the next toss of the die does not depend
on the previous one or the toss of one die does not depend on the other ones as the case may be.
In other words X1 , X2 , . . . , XN are Independent, Identically Distributed (iid) random variables.
Furthermore let these random variables have a finite mean hxi and finite variance h(∆x)2 i.
We can define the a new variable; the sample mean,
S=
N
1 X
Xj .
N
j=1
2
The sample mean has mean,
hsi =
N
1 X
hXl i = hxi,
N
j=1
and second moment
hs2 i =
1 X
1
hXj Xk i = hxi2 + h(∆x)2 i,
2
N
N
j,k
using
X
hXj Xk i = h(hxi + ∆Xj )(hxi + ∆Xk )i = hxi2 + h∆Xj ∆Xk i = hxi2 + δjk h(∆x)2 i.
jk
So the variance of the sample mean is
h(∆s)2 i = hs2 i − hsi2 =
1
h(∆x)2 i.
N
This is the first form of the weak law of large numbers that we are going to learn.
A different form for the weak law is
lim P (|SN − hxi| ≤ ) = 1
for any
N →∞
> 0.
The central limit theorem tells you a bit more than what the weak law of large numbers does.
It tells you not only the mean and the variance of the sample mean but it tells that the sample
mean is a random variable distributed according to the Normal distribution. Let us push things a
bit forward and ask how the measured frequencies fi = ni /N obtained from the number of times
ni that a random variable X took on a value xi out of N trials are connected to the probabilities Pi
associated with that outcome. Let the possible values taken by the random variables X be labeled
x1 , x2 , . . . , xD . In N trials we get a sequence X1 , . . . XN of the D possible outcomes in which the
value xj occurs nj times, xk occurs nk times and so on. We have
N
X
nj = N.
j=1
Since we are worried about the frequencies fj = nj /N of the occurrence of each of the D values,
let us look at the probability that the sequence we obtain has occurrence numbers n1 , . . . , nD :
p(n1 , . . . , nD ) =
N!
pn1 · · · pnDD .
n1 ! · · · nD ! 1
Normalization of the probability distribution means that
X
p(n1 , . . . , nD ) = (p1 + · · · + pD )N = 1.
n1 ,...nD
The mean values of the occurrence number can be computed as
X
∂ X
hnj i =
nj p(n1 , . . . , nD ) = pj
p(n1 , . . . , nD ) = N pj .
∂pj n ,...n
n ,...n
1
D
1
D
3
Similarly for the second moment matrix,
X
hnj nk i =
nj nk p(n1 , . . . , nD )
n1 ,...nD
∂
∂ X
= pj
pk
p(n1 , . . . , nD )
∂pj
∂pk n ,...n
1
D
∂ = pj
N pk (p1 + . . . + pD )N −1
∂pj
= N pj δjk (p1 + . . . + pD )N −1 + N (N − 1)pj pk (p1 + . . . + pD )N −2
= N 2 pj pk + N (pj δjk − pj pk ).
The correlation matrix:
h∆nj ∆nk i = h(nj − hnj i)(nk − hnk i)i
= hnj nk i − hnj ihnk i
= N (pj δjk − pj pk ).
So the variances are
h(∆nj )2 i = N pj (1 − pj ).
Now looking at the frequency of occurrence of each alternative, fj = nj /N ,
hfj i =
h∆fj ∆fk i =
hnj i
= pj ,
N
1
(pj δjk − pj pk ),
N
h(∆fj )2 i =
1
pj (1 − pj ).
N
So we have a third form of the weak law of large numbers,
P (|fj − pj | ≤ , j = 1, . . . , D) ≥ 1 −
1−
PD
2
j=1 pj
.
N 2
The probability that all the frequencies lie within of the corresponding probability is now bounded
by the expression on the right side of the above inequality. This form of the weak law says that
lim P (|fj − pj | ≤ , j = 1, . . . , D) = 1,
N →∞
for any > 0.
Proof
X
P (|fj − pj | ≤ , j = 1, . . . , D) ≥ P ( (fj − pj )2 ≤ 2 )
j
X
= 1 − P ( (fj − pj )2 > 2 )
j
P
≥ 1−
j h(∆fj )
2
2i
=
1−
P
j
N 2
p2j
.
4
The last inequality comes form Chebyshev inequality.
The strong law of large numbers deals directly with limits of a sequence of frequencies and is
stated in the form
(N )
P ( lim fj
N →∞
= pj ) = 1,
while the weak laws refer to a limit of a sequence of probabilities (see the third form above).
Typically we are interested in limits of probabilities and so we are interested in the weak laws and
not on the strong laws.