Download THE CIRCULAR LAW PROOF OF THE REPLACEMENT PRINCIPLE by ZHIWEI TANG

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Indeterminism wikipedia , lookup

Birthday problem wikipedia , lookup

Inductive probability wikipedia , lookup

Random variable wikipedia , lookup

Ars Conjectandi wikipedia , lookup

Infinite monkey theorem wikipedia , lookup

Randomness wikipedia , lookup

Probability interpretations wikipedia , lookup

Central limit theorem wikipedia , lookup

Conditioning (probability) wikipedia , lookup

Law of large numbers wikipedia , lookup

Transcript
THE CIRCULAR LAW
PROOF OF THE REPLACEMENT PRINCIPLE
by
ZHIWEI TANG
A thesis submitted to the Department of Mathematics and Statistics
in conformity with the requirements for
the degree of Master of Science
Queen’s University
Kingston, Ontario, Canada
July, 2009
c Zhiwei Tang, 2009
Copyright °
Abstract
It was conjectured in the early 1950s that the empirical spectral distribution
(ESD) of an n × n matrix whose entries are independent and identically
distributed with mean zero and variance one, normalized by a factor of
√1 ,
n
converges to the uniform distribution over the unit disk on the complex
plane, which is called the circular law. The goal of this thesis is to prove
the so called Replacement Principle introduced by Tao and Vu which is a
crucial step in their recent proof of the circular law in full generality. It gives
a general criterion for the difference of the ESDs of two normalised random
matrices
√1 An , √1 Bn
n
n
to converge to 0.
i
Acknowledgements
I would like to sincerely thank my supervisors Professor James Mingo and
Professor Roland Speicher for their enormous help and valuable suggestions
in the construction of this thesis.
ii
Contents
Abstract
i
Acknowledgements
ii
1 Introduction
1
2 Notations and Definitions
5
3 The replacement principle
9
4 Proof of the replacement principle
13
5 Conclusion
34
Bibliography
35
iii
Chapter 1
Introduction
In 1955, Wigner first observed that the empirical spectral distribution of an
n × n Hermitian matrix with upper diagonal entries being independent and
identically distributed Gaussian random variables converges to the semicircle law (see Figure 1). Wigner’s result was later extended to many other
forms according to the distributions of the entries, the most general one only
requires mean zero and variance one.
Theorem 1.1 (Wigner’s semicircle law)
Let An be the n × n Her-
mitian random matrix whose upper diagonal entries are independent and
identically distributed complex random variables with mean zero and variance one. Then the empirical spectral distribution of
semi-circle distribution.
1
√1 An
n
converges to the
0.3
0.2
0.1
-1
0
1
2
Figure 1.1: The distribution of eigenvalues of a randomly generated 2000 by
2000 Hermitian matrix with entries chosen from a standard normal distribution is illustrated above.
Circular Law Conjecture: Suppose we have an n × n matrix An , where
each entry aij of An is an independently identically distributed copy of
a single random variable with mean 0 and variance 1. Numerical evidence
reveals that the eigenvalues of such a matrix appear to distribute themselves
uniformly on the unit disk in the limit n → ∞. (see Figure 2)
Conjecture 1.2 (Circular Law) Let An be the n×n random matrix whose
entries are independently identically distributed complex random variables
with mean zero and variance one. Then the empirical spectral distribution
of
√1 An
n
converges in probability to the uniform distribution on the unit disk.
There are significant difficulties in giving a rigorous proof to this conjecture since the main techniques used to handle Hermitian matrices cannot
be applied to the non-Hermitian model. However the conjecture has been
worked on intensively for many decades. The circular law was verified for
the complex Gaussian distribution in [8] and the real Gaussian distribution
2
in [4] by using Ginibre’s formula for the joint density function of the eigenvalues of An . In 1984, Girko [5] suggested a strategy to prove the general
case. In 1997 Bai succeeded in making this rigorous for continuous distributions with bounded sixth moment in [1] and this hypothesis was lowered
to (2 + η)-th moment for any η > 0 in [2]. The removal of the hypothesis of continuous distribution required some new ideas. Important partial
results were obtained by Götze-Tikhomirov ([6], [7]), Pan-Zhou [9] and TaoVu [11]. All these results are established under additional assumptions on
the distribution of the entries. Tao-Vu finally proved the conjecture in full
generality in [12]. The most important step in their proof is the so called
replacement principle. Namely it shows that the difference of the empirical
spectral distribution of two normalised random matrices
√1 An , √1 Bn
n
n
verges to 0 if they satisfy certain conditions. It is proven in [12] that if
is a Gaussian random matrix and
√1 Bn
n
con-
√1 An
n
satisfies the assumptions of conjec-
ture 1.2, then these conditions always hold, therefore we know the empirical
spectral distribution for
for the
√1 An
n
√1 Bn
n
converges to the same limiting measure as
which is the uniform distribution on the unit disk. The proof
of the replacement principle in [12] is quite sketchy, with many details left
to the reader. In this thesis we will give a complete proof of the replacement
principle with all the necessary steps and calculations.
3
-1.0
1.0
1.0
0.5
0.5
0.5
-0.5
1.0
-1.0
1.0
0.5
0.5
-0.5
1.0 -1.0
0.5
-0.5
-0.5
-0.5
-0.5
-1.0
-1.0
-1.0
Figure 1.2: Eigenvalue plots of three randomly generated 1000 by 1000
matrices. Each entry of the left one is an iid Gaussian normal random
2
variable, with probability density function √12π e−x /2 . Each entry of the
middle one is an iid uniform random variable. Each entry of the right one
1
is an iid Cauchy random variable, with probability density function π(1+x
2) .
The third example shows that one can not expect the circular law to hold
in case where the second moment does not exist.
4
1.0
Chapter 2
Notations and Definitions
Empirical spectral distributions
This thesis is focused on the convergence
in probability of empirical spectral distributions of random matrices. Let us
first fix our definitions of convergence in probability and weak boundedness
[3].
Definition 2.1 (Convergence in probability) The sequence {Xn } of random
variables is said to converge in probability to X if for every ² > 0 we have
lim P (|Xn − X| > ²) = 0.
n→∞
Definition 2.2 (Weak boundedness). Let {Xn } be a sequence of random
variables on probability space (Ω, P ), we say {Xn } is weakly bounded if we
5
have
lim lim inf P (|Xn | ≤ C) = 1
C→∞ n→∞
Remark 2.3. In order to make Definition 2.2 more convenient to use, one
can rewrite Definition 2.2 as the following equivalent statement:
Proposition 2.4 (Weak boundedness). Let {Xn } be a sequence of random
variables on probability space (Ω, P ). {Xn } is weakly bounded if and only
if given ² > 0, ∃n0 , C0 , such that P (|Xn | ≤ C0 ) > 1 − ² for n ≥ n0 .
Proof: (=⇒) Let ² > 0 be given, ∃C0 , such that for C > C0 ,
lim inf P (|Xn | ≤ C) > 1 − ²
n→∞
thus ∃n0 , such that
P (|Xn | ≤ C) > 1 − ² (for n ≥ n0 )
by the definition of lim inf.
(⇐=) Let ² > 0 be given. By assumption ∃n0 , C0 , such that for n ≥ n0 , we
have P (|Xn | ≤ C0 ) > 1 − ². Therefore
lim inf P (|Xn | ≤ C0 ) ≥ 1 − ²
n→∞
Clearly P (|Xn | ≤ C) ≥ P (|Xn | ≤ C0 ) for all C ≥ C0 , hence lim inf n→∞ P (|Xn | ≤
C) ≥ 1 − ², i.e.
lim lim inf P (|Xn | ≤ C) = 1
C→∞ n→∞
6
Notation 2.5 Let Mn (C) denote the set of n × n complex matrices. For
A ∈ Mn (C) and s, t ∈ R, we let
µA (s, t) :=
1
|{1 ≤ i ≤ n, Reλi ≤ s, Imλi ≤ t}|
n
be the empirical spectral distribution of its eigenvalues λi ∈ C, i = 1, · · · , n,
counted by algebraic multiplicity (we use |B| to denote the cardinality of a
finite set B). Thus for any complex valued function f
Z
n
1X
f dµA =
f (λi ).
n
i=1
Clearly, µA is a discrete probability measure on C.
Notation 2.6 (Convergence in probability for µA (s, t)) Let us give the space
of probability measures on C the usual vague topology. Hence a sequence of
R
R
measures µn converges to µ vaguely if C f dµn converges to C f dµ for every
continuous and compactly supported function f : C → R.
Remark 2.7 According to Definition 2.1, we see that the random probability
measures µ √1
n
An
converge to µ∞ in probability if for every continuous and
compactly supported function f : C → R, the sequence of random variables
Z
Z
C
f (z)dµ √1
n
An
7
−
C
f (z)dµ∞
converges to zero in probability, i. e.
Z
lim P (|
n→∞
C
Z
f (z)dµ √1
n
An
−
C
f (z)dµ∞ | > ²) = 0
for every ² > 0.
We will use the following lemma, whose proof is given in the appendix of
[12].
Lemma 2.8 Let A = (aij )1≤i,j≤n ∈ Mn (C) have eigenvalues λ1 , · · · , λn ∈ C
repeated according to multiplicity. Then
n
X
|λj |2 ≤ ||A||22 =
j=1
n X
n
X
i=1 j=1
8
|aij |2 .
Chapter 3
The replacement principle
Now we are going to give the formal statement of the replacement principle.
Theorem 3.1 (Replacement principle). Suppose for each n that An , Bn
are sequences of n × n random matrices. Assume that
(i) the sequence of random variables
n1
o
1
2
2
||A
||
+
||B
||
n 2
n 2
n2
n2
is weakly bounded; and
(ii) for almost all complex numbers z, the sequence of random variables
n1
o
1
1
1
log | det( √ An − zI)| −
log | det( √ Bn − zI)|
n
n
n
n
converges to zero in probability.
Then the sequence of random measures {µ √1
n
9
An
− µ √1
n
Bn }
converges to zero
in probability, i.e. given ² > 0, we have for every continuous function f with
compact support
Z
lim P (|
n→∞
C
Z
f (z)dµ √1
n
An (z)
−
C
f (z)dµ √1
n
Bn (z)|
> ²) = 0.
Note that we do not need to require independence among the entries of An
and Bn . The proof of this theorem relies on the Stieltjes transform technique
that analyses the random probability measure µ √1
n
determinants
1
n
An
in terms of the log-
log | det( √1n An −zI)|, combined with tools from real analysis
such as the dominated convergence theorem. The details are given in the
next chapter.
From the weak boundedness of Theorem 3.1 (i) we see that for every ² > 0
there exists C² > 0, such that for n large enough,
P(
1
||An ||22 ≤ C² ) > 1 − ².
n2
Thus by Lemma 2.8, with probability greater than 1 − ², we have for the
eigenvalues λ1 , · · · , λn of An the bound
n
X
1
|λj |2 ≤ C²
n2
(3.1)
j=1
or, equivalently, that
Z
C
|z|2 dµ √1
n
An (z)
10
≤ C² .
Similarly, we have
Z
C
|z|2 dµ √1
n
Bn (z)
≤ C²
with probability greater than 1 − ². For any R > 0, we have
Z
2
R µ √1 An {z ∈ C : |z| ≥ R} ≤
n
Z
≤
|z|2 dµ √1 An (z) ≤ C² .
|z|≥R
|z|2 dµ √1
n
An (z)
n
C
Thus we have the bounds
µ √1
An {z
∈ C : |z| ≥ R} ≤ C² /R2
µ √1
Bn {z
∈ C : |z| ≥ R} ≤ C² /R2
n
and
n
for all R > 0 with probability greater than 1 − ².
Let Ω be the underlying probability space. We have that, as an immediate
consequence, for fixed ² ≥ 0 one can find an R ≥ 1 large enough such that
P (Ω1 ) > 1 − ², where Ω1 ⊆ Ω,
Ω1 = {ω|µ √1
n
An (ω) (z
∈ C : |z| ≥ R)+µ √1
n
Bn (ω) (z
∈ C : |z| ≥ R) ≤ ²}. (3.2)
In the following chapter we will establish the proof of the replacement principle. We will pass from the empirical spectral distribution to the characteristic function and show the equivalence between them in the sense of
11
convergence in probability. Then we can significantly simplify our problem
by using Girko’s identity which gives the relationship between the characteristic function and the Stieltjes-like transform.
12
Chapter 4
Proof of the replacement
principle
We begin with a probabilistic version of the dominated convergence theorem.
Lemma 4.1 (Dominated convergence). Let (X, ν) be a finite measure space.
For each integer n ≥ 1, let fn (x, ω) : X × Ω → R be a sequence of random
functions which are jointly measurable with respect to X and Ω. Assume
that
(i) (Uniform integrability) There exists δ > 0 such that the sequence of
R
random variables { X |fn (x)|1+δ dν(x)} is weakly bounded.
(ii) (Pointwise convergence in probability) For ν-almost every x ∈ X, the
sequence of random variables fn (x) converges to zero in probability.
R
Then the sequence of random variables { X fn (x)dν(x)} converges to zero
in probability.
13
Proof: We can normalise ν to be a probability measure. It suffices to show
that given ² > 0, δ > 0, ∃N , such that:
¯
³¯ Z
´
¯
¯
P ¯
fn (x)dν(x)¯ < ² > 1 − δ (for n ≥ N)
X
Fix ² > 0. By our first assumption, ∃n0 , C² , such that for n ≥ n0 , we have
³Z
P
X
´
|fn (x)|1+δ dν(x) ≤ C² > 1 − ²
For every M > 0 and ω ∈ Ω, we have:
Z
M
1+δ
ν({x | |fn (x, ω)| ≥ M }) ≤
Z
≤
X
{x| |fn (x,ω)|≥M }
|fn (x, ω)|1+δ dν(x)
|fn (x, ω)|1+δ dν(x)
i.e.
Z
ν{x| |fn (x, ω)| ≥ M } ≤
{x| |fn (x,ω)|≥M }
|fn (x, ω)|1+δ dν(x)/(M 1+δ )
Now Hölder’s inequality tells us:
¯Z
¯ Z
¯
¯
fn (x, ω)I(|fn (x, ω)| ≥ M )dν(x)¯ ≤
|fn (x, ω)I(|fn (x, ω)| ≥ M )|dν(x)
¯
X
X
´1/(1+δ) ³ Z
³Z
´ δ
1+δ
1+δ
1+δ
|I(|fn (x, ω)| ≥ M )| δ dν(x)
|fn (x, ω)| dν(x)
≤
X
X
´1/(1+δ)
³Z
δ
(ν{x| |fn (x, ω)| ≥ M }) 1+δ
|fn (x, ω)|1+δ dν(x)
≤
X
14
1
≤ δ
M
Z
X
|fn (x, ω)|1+δ dν(x)
where I(E) denotes the indicator of an event E.
By our second hypothesis, we have limn→∞ P (|fn (x, ω)| ≥ ²) = 0 except for
a subset of X of ν-measure 0. The dominated convergence theorem implies
that
Z
0=
lim P (|fn (x, ω)| ≥ ²)dν(x)
X n→∞
Z
= lim
n→∞ X
Z
= lim
n→∞ X
P (|fn (x, ω)| ≥ ²)dν(x)
Z
[ I{ω̃| |fn (x, ω̃)| ≥ ²}dP (ω̃)]dν(x)
(4.1)
Ω
By Fubini’s theorem and Markov’s inequality, we conclude that
Z Z
[ I{x| |fn (x, ω̃)| ≥ ²} dν(x)] dP (ω)
n→∞ Ω X
³Z
´
= lim E
I{x| |fn (x, ω̃)| ≥ ²}dν(x)
n→∞
XZ
² ´
² ³
P
I{x| |fn (x, ω̃)| ≥ ²}dν(x) ≥
≥ lim
n→∞ M
M
X
(4.1) = lim
which tells us that, for every M , limn→∞ P (ν({x | |fn (x)| ≥ ²}) ≥
²
M)
= 0.
Choose M > (C² /²)1/δ . There exists n1 , such that P (ν({x | |fn (x)| ≥ ²}) ≥
²
M)
< ² for n ≥ n1 . Now fix n ≥ max{n0 , n1 }.
Now we let Ω1 = {ω |
R
X
|fn (x, ω)|1+δ dν(x) ≤ C² }. Then P (Ω1 ) > 1 − ².
15
For ω ∈ Ω1 we have
¯Z
¯
¯
¯
fn (x, ω)I(|fn (x, ω)| ≥ M )dν(x)¯ ≤ C² /M δ < ²
¯
X
Next let Ω2 = {ω | ν({x | |fn (x, ω)| ≥ ²}) <
²
M )}.
We have P (Ω2 ) > 1 − ².
Then for ω ∈ Ω2 , we have
¯Z
¯
¯
{x| |fn (x,ω)|≤M }
¯
¯
fn (x, ω)dν(x)¯
¯Z
¯
≤¯
{x| |fn (x,ω)|<²}
¯ ¯Z
¯ ¯
fn (x, ω)dν(x)¯ + ¯
{x| ²≤|fn (x,ω)|≤M }
Z
≤²+M
X
¯
¯
fn (x, ω)dν(x)¯
I{x| |fn (x, ω)| ≥ ²}dν(x)
≤ 2²
Now let Ω3 = Ω1 ∩ Ω2 , clearly P (Ω3 ) > 1 − 2².
¯R
¯
¯
¯
For the ω ∈ Ω3 and n as above, we have ¯ X fn (x, ω) dν(x)¯ < 3². Thus for
every n ≥ max{n0 , n1 } we have found an Ω3 with P (Ω3 ) > 1 − 2² such that
¯R
¯
¯
¯
for ω ∈ Ω3 we have ¯ X fn (x, ω) dν(x)¯ < 3². Hence for n ≥ max{n0 , n1 }
¯
³¯ R
´
¯
¯
we have P ¯ X fn (x) dν(x)¯ < 3² > 1 − 2². Therefore we have completed
the proof of Lemma 3.1.
We now take the standard step of passing from the ESDs µ √1
n
the characteristic functions m √1
n
An ,
m √1
n
Bn
Z
n
An (u, v)
=
C
eiuRe(z)+ivIm(z) dµ √1
n
16
µ √1
n
Bn
to
: R2 → C, which are defined by
the formulae:
m √1
An ,
An (z)
Z
m √1
n
Bn (u, v)
thus the functions m √1
n
An ,
=
C
m √1
n
eiuRe(z)+ivIm(z) dµ √1
n
Bn
Bn (z)
are continuous and are bounded uni-
formly in magnitude by 1.
Lemma 4.2 The following statements are equivalent:
(i) The sequence of random probability measures µ √1
n
An
− µ √1
n
Bn
converges
to zero in probability.
(ii) For almost every u, v, the sequence of random characteristic functions
m √1
n
An (u, v)
− m √1
n
Bn (u, v)
converges to zero in probability.
P roof. We first show that (i) implies (ii). Fix u, v, and let ² > 0 be arbitrary.
We can find an R depending on C² and ², n0 such that
µ √1
n
An ({z
∈ C : |z| ≥ R}) + µ √1
n
Bn ({z
∈ C : |z| ≥ R}) ≤ ² (n ≥ n0 )
In particular, we have
Z
eiuRe(z)+ivIm(z) [dµ √1 Bn − dµ √1 An ]
m √1 Bn (u, v) − m √1 An (u, v) =
n
n
n
n
C
Z
=
eiuRe(z)+ivIm(z) [dµ √1 Bn − dµ √1 An ]
n
n
|z|<R
Z
+
eiuRe(z)+ivIm(z) [dµ √1 Bn − dµ √1 An ]
n
n
|z|≥R
Z
= ψ(z/R)eiuRe(z)+ivIm(z) [dµ √1 Bn − dµ √1 An ] + ²
n
17
n
where ψ is any smooth compactly supported function that equals one on
the unit ball. But since by hypothesis µ √1
n
An
− µ √1
n
Bn
converges to zero
in probability, so the above integral converges to zero in probability. The
claim then follows.
Now we prove (ii) implies (i). Since continuous compactly supported functions are the uniform limit of smooth compactly supported functions, it
R
R
suffices to show that C f dµ √1 An − C f dµ √1 Bn converges to zero in proban
n
bility for every smooth compactly supported function f : C → C.
Now fix a smooth compactly supported function f : C → C. Let fˆ denote
the Fourier transform of f . Then using Parseval’s identity [10] , we can
write
Z
Z
C
f dµ √1
n
An −
C
Z Z
f dµ √1
n
Bn
=
R
R
fˆ(u, v)(m √1
n
An (u, v)−m √1n Bn (u, v))dudv
Note that the Fourier transform of a smooth compactly supported function is
smooth and rapidly decreasing. In particular, the measure dν = fˆ(u, v)dudv
is finite. By the hypothesis and the fact that the function m √1
n
m √1
n
Bn (u, v)
An (u, v)
−
is bounded and so clearly obeys the moment condition required
in Lemma 4.1, the claim now follows.
Definition 4.3 Now we are going to use the Stieltjes-like transform g √1
n
An (w)
C → R, defined almost everywhere by the formula that
Z
g √1
n
An (w) (z) := 2Re
C
1
n
z−w
2 X Re(z − √n λj )
.
dµ
=
1
|z − w|2 √n An (w) n
|z − √1n λj |2
j=1
18
(4.2)
:
Lemma 4.4 (Girko’s identity). For every non-zero u, v we have
u2 + v 2
m √1 An (u, v) =
n
4πiu
Z Z
( g √1
R
n
R
An (s
+ it)eius+ivt dt)ds,
where the inner integral is absolutely integrable for almost every s, and the
outer integral is absolutely convergent.
P roof. Since
n
1 X i(uRe( √1n λj )+vIm( √1n λj ))
e
m √1 An (u, v) =
n
n
j=1
it suffices from (4.2) to show that
i(uRe(w)+vIm(w))
e
u2 + v 2
=
2πiu
Z Z
Re(s + it − w) ius+ivt
(
e
dt)ds
2
R R |s + it − w|
for each complex number w. Note that
R
|s|
itv
R s2 +t2 e dt
= πe−|sv| , so this gives
us:
Z
R
Re(s + it − w) ius+ivt
e
dt = πsgn(s − Re(w))e−|v||s−Re(w)| eius eivIm(w)
|s + it − w|2
for every s 6= Re(w), and the claim follows by the Laplace transform
Z
∞
sin(su)e−|v|s ds =
0
u2
u
+ v2
.
Similarly we can define g √1 Bn in the same way. To conclude the proof of
n
19
Theorem 3.1, it thus suffices to show
Z Z
( (g √1
R
n
R
An (s
+ it) − g √1
n
Bn (s
+ it))eius+ivt dt)ds
converges to zero in probability under the assumptions of Theorem 3.1.
Now let ψ : R → R+ be a smooth cutoff function which equals 1 on [−1, 1],
|ψ(x)| ≤ 1 for all x and is supported on [−2, 2].
Lemma 4.5 (Truncation in s,t). Fix u 6= 0, v 6= 0. Let w ∈ C.
(i) The integral
Z
Z
|
R
R
Re(s + it − w) ius+ivt
e
dt|(1 − ψ(s/R2 ))ds
|s + it − w|2
(4.3)
is bounded as a function of w, and for |w| ≤ R, is less than any given ² > 0,
if R is large enough.
(ii)
Z
Z
|
R
R
Re(s + it − w) ius+ivt
e
(1 − ψ(t/R2 ))dt|ψ(s/R2 )ds
|s + it − w|2
(4.4)
is bounded, and for |w| ≤ R, is less than any given ² > 0, if R is large
enough.
20
Proof: (i) Since
Z
R
Re(s + it − w) ius+ivt
e
dt = πsgn(s − Re(w))e−|v||s−Re(w)| eius eivIm(w)
|s + it − w|2
for every s 6= Re(w), then we have
Z
Re(s + it − w) ius+ivt
e
dt| = πe−|v||s−Re(w)| .
|s + it − w|2
|
R
Now it suffices to show
Z
πe−|v||s−Re(w)| (1 − ψ(s/R2 ))ds
(4.5)
R
is bounded. When |Re(w)| < R2
Z
Z
(4.5) =
πe
−|v||s−Re(w)|
R
³Z
R
∞
−|v|(s−Re(w))
≤π
e
Z
πe−|v||s−Re(w)| ψ(s/R2 )ds
−
−
Re(w)
ds +
Re(w)
R2
Z
e−|v|(Re(w)−s) ds
´
−∞
πe−|v||s−Re(w)| ds
−R2
³Z
R2
= 2π/|v| − π
Z
−|v|(s−Re(w))
e
Rew)
ds +
e−|v|(Re(w)−s) ds
−R2
Re(w)
2π −|v|(R2 −Re(w))
2
(e
+ e−|v|(R +Re(w)) )
|v|
4π
≤
.
|v|
=
When Re(w) ≥ R2
Z
−R2
(4.5) =
πe−|v||s−Re(w)| (1 − ψ(s/R2 ))ds
−∞
21
´
Z
+
∞
πe−|v||s−Re(w)| (1 − ψ(s/R2 ))ds
R2
Z −R2
Z
−|v|(Re(w)−s)
≤ π(
e
ds +
−∞
Z ∞
+
e−|v|(s−Re(w)) )ds
Re(w)
e−|v|(Re(w)−s) ds
R2
Re(w)
4π
≤
.
|v|
Similarly we have (4.5) ≤
4π
|v|
for Re(w) ≤ −R2 . Therefore we get an explicit
upper bound for (4.3), namely
4π
|v| .
If |w| ≤ R, then |Re(w)| ≤ R, and thus one can easily see that for fixed w,
(4.3) goes to 0 if R is large enough.
(ii)First we are trying to show that there are positive constants A,B,C such
that:
Z
|
R
is bounded by
A
R4
Re(w − (s + it)) ius+ivt
e
(1 − ψ(t/R2 ))dt|
|w − (s + it)|2
+
B
R2 |x−s|
+
C
,
|x−s|2
(4.6)
where w = x + yi. Since the eius term
has magnitude 1 and can be discarded, to prove (4.6) is bounded, it suffices
to show
Z
I=|
R
is bounded by
A
R4
x−s
eivt (1 − ψ(t/R2 ))dt|
(x − s)2 + (y − t)2
B
C
+ R2 |x−s|
+ |x−s|
2 . Using integration by parts, integrating
the eivt term and differentiating the others, we have:
¯1
x−s
¯
I = ¯ eivt
(1 − ψ(t/R2 )) |∞
−∞
iv
(x − s)2 + (y − t)2
22
Z
(s − x)(2t − 2y)
1 ivt
e (
(1 − ψ(t/R2 ))
((x − s)2 + (y − t)2 )2
R iv
¯
x−s
1
¯
0
2
+
(−
)ψ
(t/R
))dt
¯
2
2
2
(x − s) + (y − t)
R
−
Note that the first term appears in I vanishes, so
Z
1 ivt (s − x)(2t − 2y)
e
(1 − ψ(t/R2 ))dt|
((x − s)2 + (y − t)2 )2
R iv
Z
1 ivt
x−s
1
+|
e
(− 2 )ψ 0 (t/R2 )dt|
2
2
(x − s) + (y − t)
R
R iv
I≤|
Now we integrate the integrals above by parts again separately:
¯Z
¯
I1 = ¯
¯
1 ivt (s − x)(2t − 2y)
¯
2
e
(1
−
ψ(t/R
))dt
¯
2
2
2
((x − s) + (y − t) )
R iv
¯
1
(s − x)(2t − 2y)
¯
= ¯ − 2 eivt
(1 − ψ(t/R2 )) |∞
−∞
v
((x − s)2 + (y − t)2 )2
Z
1 ivt
(s − x)2 − 3(t − y)2
+
e
2(s
−
x)
(1 − ψ(t/R2 ))dt
2
((x − s)2 + (y − t)2 )3
R v
Z
¯
1 ivt (s − x)(2t − 2y)
1
¯
0
2
+
e
(−
)ψ
(t/R
))dt
¯
2
2 + (y − t)2 )2
2
v
((x
−
s)
R
R
Again, the first term in the RHS disappears.
Z
I1 ≤
Since
1
(s − x)2 − 3(t − y)2
(2(s
−
x)
(1 − ψ(t/R2 ))|dt
2
2 + (y − t)2 )3
v
((x
−
s)
R
Z
1
(s − x)(2t − 2y)
1
+ | 2
( 2 )ψ 0 (t/R2 ))|dt.
2
2
2
R v ((x − s) + (y − t) ) R
|
√
(s − x)2 − 3(t − y)2
3 3
|2(s − x)
|dt =
((x − s)2 + (y − t)2 )3
4|s − x|2
R
Z
23
and by the fact that
Z
|
0
∞
(a2
2ax
1
dx| =
2
2
+x )
|a|
we obtain:
√
1
(s − x)2 − 3(t − y)2
3 3
2
| 2 (2(s − x)
(1 − ψ(t/R ))|dt ≤ 2
((x − s)2 + (y − t)2 )3
4v |s − x|2
R v
Z
Z
|
R
1
(s − x)(2t − 2y)
2
1
1
1
( 2 )ψ 0 (t/R2 ))|dt ≤ 2
A1
2
2
2
2
v ((x − s) + (y − t) ) R
v |x − s| R2
Let A1 > 0 be such that |ψ 0 (t/R2 )| ≤ A1 . Similarly, we have
Z
1 ivt
x−s
1
e
(− 2 )ψ 0 (t/R2 )dt|
2
2
(x − s) + (y − t)
R
R iv
1
x−s
1
= | 2 eivt
( )ψ 0 (t/R2 ) |∞
−∞
v
(x − s)2 + (y − t)2 R2
Z
1 ivt
(x − s)(2t − 2y)
1
+
e (
(− 2 )ψ 0 (t/R2 )
2
2
2
2
((x − s) + (y − t) )
R
R v
x−s
1
+
(− )ψ 00 (t/R2 ))dt|
(x − s)2 + (y − t)2 R4
1
2
1
π 1
≤ 2
A1 + 2 4 A2
2
v |x − s| R
v R
I2 = |
Again let A2 > 0 be such that |ψ 00 (t/R2 )| ≤ A2 . Hence we finally get:
√
4A1
1
πA2 1
3 3
+ 2 2
+ 2 4
I≤ 2
4v |s − x|2
v R |x − s|
v R
which is exactly what we want.
Note that one can get an explicit upper bound for (4.6) which stems from
24
the classical Poisson integral identity
R∞
a
−∞ a2 +(x−x0 )2 dx
= πsgn(a). The
eius+ivt (1 − ψ(t/R2 )) term has magnitude at most 1 and thus can be discarded.
Now given M > 0. In order to get an upper bound for (4.4), we integrate
(4.4) as follows:
Z
Z
(4.4) =
|
Z
|s−Re(λj )|>M
Z
R
Re(λj − (s + it)) ius+ivt
e
(1 − ψ(t/R2 ))dt|ψ(s/R2 )ds
|λj − (s + it)|2
Re(λj − (s + it)) ius+ivt
e
(1 − ψ(t/R2 ))dt|ψ(s/R2 )ds
2
|s−Re(λj )|≤M
R |λj − (s + it)|
Z
1
1
≤K
( 2+
)2 ψ(s/R2 )ds + 2M π
|s − Re(λj )|
|s−Re(λj )|>M R
+
|
(for some positive constant K)
Z
1
1
≤ 2K
+
ds + 2M π
T
4
(s − Re(λj ))2
[−2R2 ,2R2 ] {|s−Re(λj )|≤M } R
1
2M
1
2
4
− 4 −
+
) + 2M π
= 2K( 2 −
2
2
R
2R + Re(λj )
R
2R − Re(λj ) M
As one can see (4.4) is bounded as R → ∞.
Finally, if |w| ≤ R, then |x| ≤ R and |y| ≤ R; also note that the integrand
of (4.4) vanishes unless |s| ≤ 2R2 and |t| ≥ R2 . So |s − x| ≤ 2R2 and
|t − y| ≥ R2 − R > R2 /2 for R sufficiently large. Consequently we have
1
(x−s)2 +(y−t)2
≤
4
.
R4
Now given M1 > 0, M2 > 0, we integrate I1 and I2 by
parts separately exactly as above to obtain
Z
I1 ≤
1
(s − x)2 − 3(t − y)2
2(s
−
x)
(1 − ψ(t/R2 ))|dt
2
2 + (y − t)2 )3
v
((x
−
s)
R
Z
1
(s − x)(2t − 2y)
1
+ | 2
( 2 )ψ 0 (t/R2 )|dt
2
2
2
R v ((x − s) + (y − t) ) R
|
25
Z
1
(s − x)2 − 3(t − y)2
1 − ψ(t/R2 )
| 2 2(s − x)
|dt
((x − s)2 + (y − t)2 )2 (s − x)2 + (t − y)2
R v
Z
1
1
(s − x)(2t − 2y)
ψ 0 (t/R2 )
(
|dt
+ | 2
)
2
2 3/2 R2 ((x − s)2 + (t − y)2 )1/2
R v ((x − s) + (y − t) )
π 4
3π 4
4 A1 1
≤ 2 4+ 2 4+ 2 2 2
v R
v R
v R R
M1
≤ 4
R
=
Similarly, we have
Z
(x − s)(2t − 2y)
1 0
1
|
ψ (t/R2 )|dt
I2 ≤ 2
v R ((x − s)2 + (t − y)2 )2 R2
Z
1
x−s
1
+ 2
|
( 4 )ψ 00 (t/R2 )|dt
2
2
v R (x − s) + (t − y) R
π A2
4 A1
≤ 2 4+ 2 4
v R
v R
M2
≤ 4
R
We conclude that I ≤ I1 + I2 ≤
M
R4
for some M > 0. Clearly (4.4) ≤
4R2 M
R4
since |s| ≤ 2R2 , the second part of claim (ii) now follows and this completes
the proof of Lemma 4.5.
Lemma 4.6 For some R > 1 large enough, we have
Z Z
( g √1
R
and
R
Z Z
( g √1
R
R
n
n
+ it)eius+ivt dt)(1 − ψ(s/R2 ))ds
(4.7)
+ it)eius+ivt (1 − ψ(t/R2 ))dt)ψ(s/R2 )ds
(4.8)
An (s
An (s
both converge to zero in probability.
26
Proof: Given ²1 > 0, ²2 > 0. From (3.2), (4.2), Lemma 4.4 and the triangle
inequality we obtain
Z Z
| ( g √1
R
n
R
+ it)eius+ivt dt)(1 − ψ(s/R2 ))ds|
An (s
Z Z
n
2X
Re(w − (s + it)) ius+ivt
=
| (
e
dt)(1 − ψ(s/R2 ))ds|
2
n
|w
−
(s
+
it)|
R R
j=1
Z
Z
n
Re(w − (s + it)) ius+ivt
2X
|
e
dt|(1 − ψ(s/R2 ))ds
≤
2
n
|w
−
(s
+
it)|
R
j=1 R
Z Z
n
X
Re(w − (s + it)) ius+ivt
2
|
=
e
dt|(1 − ψ(s/R2 ))ds
2
n
|w
−
(s
+
it)|
R
R
2
j=1,|w|≤R
n
X
2
+
n
Z
|
j=1,|w|>R2
≤ 2²1 +
Z
R
R
Re(w − (s + it)) ius+ivt
e
dt|(1 − ψ(s/R2 ))ds
|w − (s + it)|2
8π
²2
|v|
Similarly we have:
|
Z Z
( g √1
R
R
n
An (s
+ it)eius+ivt (1 − ψ(t/R2 ))dtψ(s/R2 )ds|
Z Z
n
Re(w − (s + it)) ius+ivt
2X
=
| (
e
(1 − ψ(t/R2 ))dt)ψ(s/R2 )ds|
2
n
|w
−
(s
+
it)|
R R
j=1
Z
Z
n
2X
Re(w − (s + it)) ius+ivt
≤
|
e
(1 − ψ(t/R2 ))dt|ψ(s/R2 )ds
n
|w − (s + it)|2
R
R
j=1
Z Z
n
2 X
Re(w − (s + it)) ius+ivt
=
|
e
(1 − ψ(t/R2 ))dt|ψ(s/R2 )ds
2
n
|w
−
(s
+
it)|
R
R
j=1,|w|≤R
n
X
2
+
n
Z
j=1,|w|>R R
Z
|
R
Re(w − (s + it)) ius+ivt
e
(1 − ψ(t/R2 ))dt|ψ(s/R2 )ds
|w − (s + it)|2
≤ 2²1 + K²2
27
for some K > 0. Now we have finished the proof of Lemma 4.6.
Now from Lemma 4.6 (and their counterparts for g √1
n
Bn )
and the triangle
inequality, we thus see that to prove that
Z Z
( (g √1
R
n
R
An (s
+ it) − g √1
n
Bn (s
+ it))eius+ivt dt)ds
(4.9)
converges to zero in probability, it suffices to show that
Z Z
R
R
(g √1
n
An (s
+ it) − g √1
n
Bn (s
+ it))eius+ivt ψ(t/R2 )ψ(s/R2 )dtds (4.10)
converges to zero in probability for every fixed R ≥ 1 large enough. Indeed,
assume (4.10) converges to zero in probability. Observe that
Z Z
|
(g √1
RZ RZ
=|
R
R
n
An (s
(g √1
n
+ it) − g √1
An (s
n
Bn (s
+ it) − g √1
n
+ it))eius+ivt |
Bn (s
+ it))eius+ivt (1 − ψ(s/R2 ) + ψ(s/R2 )
−ψ(t/R2 )ψ(s/R2 ) + ψ(t/R2 )ψ(s/R2 ))dtds|
Z Z
≤ | ( g √1 An (s + it) − g √1 Bn (s + it))eius+ivt dt)(1 − ψ(s/R2 ))ds|
n
RZ RZ n
+| ( g √1 An (s + it) − g √1 Bn (s + it))eius+ivt (1 − ψ(t/R2 ))dt)ψ(s/R2 )ds|
n
n
ZR Z R
+|
(g √1 An (s + it) − g √1 Bn (s + it))eius+ivt ψ(t/R2 )ψ(s/R2 )dtds|
R
R
n
n
Note that the integrands here are now jointly absolutely integrable in t, s,
and so we may now freely interchange the order of integration. Now from
the assumption and Lemma 4.6, one can see (4.9) converges to zero in prob-
28
ability.
Observe that g √1
An (w) (z)
n
A (w) (z)
n n
g √1
is a locally integrable function on C, and that
∂
2
1
log | det( √ An (w) − zI)|
∂Re(z) n
n
=
(4.11)
Fix R. Using (4.11) and integration by parts in the s variable, we have
Z Z
(4.10) =
R
2
=
n
R
(g √1
Z Z
n
An (s
+ it) − g √1
n
Bn (s
+ it))eius+ivt ψ(t/R2 )ψ(s/R2 )dtds
eius+ivt ψ(t/R2 )ψ(s/R2 )
R
R
1
1
∂
× (log | det( √ An − zI)| − log | det( √ Bn − zI)|)dt
∂s
n
n
Z
2
=
eius+ivt ψ(t/R2 )ψ(s/R2 )
n R
1
1
×(log | det( √ An − zI)| − log | det( √ Bn − zI)|) |∞
−∞ dt
n
n
Z Z
2
1
1
−
(log | det( √ An − zI)| − log | det( √ Bn − zI)|)
n R R
n
n
∂ ius+ivt
× e
ψ(t/R2 )ψ(s/R2 )dsdt
∂s
Therefore we can rewrite (4.10) in the form of
Z Z
R
R
fn (s, t)φu,v,R (s, t)dsdt,
29
where
fn (s, t) :=
1
1
2
2
log | det( √ An − zI)| −
log | det( √ Bn − zI)|
n
n
n
n
and
φu,v,R (s, t) := −
∂ ius+ivt
(e
ψ(t/R2 )ψ(s/R2 )).
∂s
Note: For the values of t and s where | det( √1n An − zI)| or | det( √1n Bn − zI)|
may equal zero the integration by parts is not justified, but this is not going
to be a problem since there are only finitely many values of such t and s,
with measure 0.
Thus to complete the proof of theorem 3.1 it will suffice to show that
Z Z
R
R
|fn (s, t)||φu,v,R (s, t)|dsdt
converges to zero in probability. We shall apply Lemma 4.1 to prove this.
Note that φu,v,R (s, t)dsdt = dν is a finite measure. It remains to check the
uniform integrability of fn , we will take δ = 1 in this case.
Lemma 4.7 Given any positive numbers c1 , c2 , there exists some M > 0
such that
Z
(log |z − (s + it)|)2 dsdt ≤ M (1 + |z|2 )
|s|≤c1 ,|t|≤c2
for all z ∈ C.
30
(4.12)
Proof: Fix z ∈ C, say z = x + yi . Put D = {(s, t)||z − (s + it)| ≤ 1},
R = [−c1 , c1 ] × [−c2 , c2 ]. We can rewrite the integral in (4.12) by setting
s0 = s − x, t0 = t − y. Thus it suffices to show that for all c1 , c2 > 0, there
exists M > 0 such that
Z
(log
p
s02 + t02 )2 ds0 dt0 ≤ M (1 + |z|2 )
R0
where R0 = [−c1 −x, c1 −x]×[−c2 −y, c2 −y]. We also put D0 = {(s0 , t0 )|s02 +
T
t02 ≤ 1}. Then for the region D0 R0 , where the function has a singularity,
we can estimate the integral by the local integrability of log2 in the plane,
i.e.
Z
Z
D
T
Z
(log |z − (s + it)|)2 dsdt =
R
Z
(log r)2 rdrdθ ≤
≤
D0
0
2π
Z
D
1
T
0
(log
p
s02 + t02 )2 ds0 dt0
R0
(log r)2 rdrdθ
0
= π/2
For the area R0 \D0 , we can estimate the integral by estimating log
√
s02 + t02
over R0 \D0 . For |z| ≤ max{2||(c1 , c2 )||, 1 + ||(c1 , c2 )||}, then the function
(log |z − (s + it)|)2 is a continuous function on the bounded region R0 \D0 .
R
So we can find some M 0 > 0, such that: R0 \D0 (log |z − (s + it)|)2 dsdt ≤ M 0 .
R
Thus in this case we can bound |s|≤c1 ,|t|≤c2 (log |z−(s+it)|)2 dsdt by M 0 +π/2.
For |z| > max{2||(c1 , c2 )||, 1 + ||(c1 , c2 )||}, we have
1 ≤ |z| − ||(c1 , c2 )|| ≤ ||(x, y) − (c1 , c2 )|| ≤
31
p
s02 + t02
and
p
s02 + t02 ≤ |z| + ||(s, t)|| ≤ |z| + ||(c1 , c2 )|| ≤ 2|z|
√
This implies that log s02 + t02 ≤ log(2|z|) ≤ 2|z|. Thus we obtain a very
√
R
crude estimation of R0 (log s02 + t02 )2 ds0 dt0 by 16c1 c2 |z|2 . Now we find
M = max{M 0 + π/2, 16c1 c2 } such that
Z
(log |z − (s + it)|)2 dsdt ≤ M (1 + |z|2 )
|s|≤c1 ,|t|≤c2
This completes the proof of Lemma 4.7.
From linear algebra, we have the identity
n
1
1
1
1X
log | det( √ An − zI)| =
log | √ λj − (s + it)|
n
n
n
n
(4.13)
j=1
and similarly for Bn . From the boundedness and compact support of φu,v,R
and Lemma 4.7, we observe that
Z Z
R
1
1
(log | √ λ − (s + it)|)2 |φu,v,R (s, t)|dsdt ≤ M (1 + |λ|2 )
n
n
R
for all λ ∈ C and some M > 0.
Now given ² > 0, there exist C1 , C2 and M such that
Z Z
R
R
|fn (s, t)|2 |φu,v,R (s, t)|dsdt
(triangle inequality)
32
(4.14)
Z Z
≤
4
1
(| log | det( √ An − zI)|
2
n
R R n
1
+| log | det( √ Bn − zI)|)2 |φu,v,R (s, t)|dsdt
n
(by (4.13))
Z Z
n
4 X
1 (A )
≤
(
| log | √ λj n − (s + it)||
2
n
R R n
j=1
+
n
X
1 (B )
| log | √ λj n − (s + it)||)2 |φu,v,R (s, t)|dsdt
n
j=1
(Cauchy − Schwartz inequality)
Z Z
n
1 (A )
8X
≤
((log | √ λj n − (s + it)|)2
n
R R n
j=1
1 (B )
+(| log | √ λj n − (s + it)|)2 )|φu,v,R (s, t)|dsdt
n
(by (4.14))
n
8M X
1 (A )
1 (B )
≤
(1 + |λj n |2 + 1 + |λj n |2 )
n
n
n
j=1
= 16M +
n
X
8M
j=1
n2
(An ) 2
|λj
| +
n
X
8M
j=1
n2
(Bn ) 2
|λj
|
By (3.1) we can estimate the last term by 8M (C1 + C2 + 2) with probability
greater than 1 − ², i.e. fn has uniform integrability with δ = 1. By the
hypothesis of Theorem 3.1 fn (s, t) converges to zero in probability for almost
every s, t. The claim now follows directly from Lemma 4.1. The proof of
Theorem 3.1 is now complete.
33
Chapter 5
Conclusion
Now we have shown that the difference of the empirical spectral distribution
of two normalised random matrices
√1 An , √1 Bn
n
n
converges to 0 if they
satisfy conditions (i) and (ii) in Theorem 3.1. The remainder of the proof of
Conjecture 1.2 is given in [12]. The main burden is to show that condition
(ii) of Theorem 3.1 is satisfied. We shall not discuss this part of the proof
in this thesis.
34
Bibliography
[1] Z. D. Bai, Circular law, Ann. Probab. 25, 494-529, 1997
[2] Z. D. Bai and J. Silverstein, Spectral analysis of large dimensional random matrices, Mathematics Monograph Series 2, Science Press, Beijing,
2006.
[3] K. L. Chung, A Course in Probability Theory (Third Edition), 68-90,
2001.
[4] A. Edelman, Probability that a random real Gaussian matrix has k real
eigenvalues, related distributions, and the Circular Law, Journal of Multivariate Analysis, 203-232, 1997.
[5] V. L. Girko, Circular law, Theory Probab. Appl. , 694-706, 1984.
[6] F. Götze and A.N. Tikhomirov, On the circular law, preprint, 2007
[7] F. Götze and A.N. Tikhomirov, The Circular Law for random matrices,
preprint, 2007
[8] M.L. Mehta, Random Matrices and the Statistical Theory of Energy
Levels, Academic Press, New York, NY, 1967.
35
[9] G. Pan and W. Zhou, Circular Law, extreme singular values and potential theory, preprint, 2007
[10] W. Rudin, Functional Analysis (Second Edition), 186-189, 1991.
[11] T. Tao and V. Vu, Random matrices: The Circular Law, Communication in Contemporary Mathematics 10, 261-307, 2008.
[12] T. Tao, V. Vu, M. Krishnapur, Random matrices: Universality of ESDs
and the circular law, arXiv:0807.4898v5, 2009.
36