Download The Probabilistic Method

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
The Probabilistic Method
prepared and Instructed by
Shmuel Wimer
Eng. Faculty, Bar-Ilan University
1
The Probabilistic Method
2015June
The probabilistic method comprises two ideas:
β€’ Any random variable assumes at least one value not
smaller than its expectation.
β€’ If an object chosen randomly from the universe
satisfies a property with positive probability, there
must be an object of the universe satisfying that
property.
Theorem. For any undirected graph 𝐺 𝑉, 𝐸 with 𝑛
vertices and π‘š edges, there is a partition of 𝑉 into 𝐴
and 𝐡 such that the edge cut-set has π‘š 2 edges at
least, namely 𝑒, 𝑣 ∈ 𝐸 𝑒 ∈ 𝐴 and 𝑣 ∈ 𝐡 β‰₯ π‘š 2.
2
The Probabilistic Method
2015June
Proof. Consider the following experiment. Each vertex is
independently and equiprobaly assigned to 𝐴 or 𝐡.
The probability that the end points of and edge 𝑒, 𝑣
are in different sets is ½.
By the linearity of expectation the expected number of
edges in the cut is π‘š 2.
It follows that there must a partition satisfying the
theorem.∎
Consider the satisfiability problem. A set of π‘š clauses is
given in conjunctive (sum) normal form over 𝑛 variables.
3
The Probabilistic Method
2015June
We have to decide whether there is a truth assignment
of the 𝑛 variables satisfying all the clauses (POS).
There is an optimization version called MAX-SAT where
we seek for a truth assignment maximizing the number
of satisfied clauses. This problem is NP-hard.
We subsequently show that there is always a truth
assignment satisfying at least π‘š 2 clauses. This is the
best possible universal guarantee (consider π‘₯ and π‘₯).
Theorem: For any set of π‘š clauses, there is a truth
assignment satisfying at least π‘š 2 clauses.
4
The Probabilistic Method
2015June
Proof: Suppose that every variable is set to TRUE or
FALSE independently and equiprobaly.
For 1 ≀ 𝑖 ≀ π‘š, let 𝑍𝑖 = 1 if the clause is satisfied, and
𝑍𝑖 = 0 otherwise.
Due to the conjunctive form, the probability that a
clause containing π‘˜ literals is not satisfied is 2βˆ’π‘˜ ≀ 1 2 ,
or 1 βˆ’ 2βˆ’π‘˜ β‰₯ 1 2 that it is satisfied, implying 𝐄 𝑍𝑖
β‰₯ 1 2.
The expected number of satisfied clauses is therefore
π‘š
𝑖=1 𝐄 𝑍𝑖 β‰₯ π‘š 2 , implying that there must be an
assignment for which π‘š
𝑖=1 𝑍𝑖 β‰₯ π‘š 2. ∎
5
The Probabilistic Method
2015June
Expanding Graphs
𝐺 𝑉, 𝐸 is called an expanding graph if there is a 𝑐 > 0
such that for any 𝑆 βŠ† 𝑉 there is Ξ“ 𝑆 β‰₯ 𝑐 𝑆 , where
Ξ“ 𝑆 is the set of 𝑆’s neighbors.
.
A particular type of expanding graph is a bipartite multi
graph 𝐺 𝐿, 𝑅, 𝐸 called an OR-concentrator.
It is defined by a quadruple 𝑛, 𝑑, 𝛼, 𝑐 , where 𝐿 = 𝑅
= 𝑛, such that
1. deg 𝑣 ≀ 𝑑 βˆ€π‘£ ∈ 𝐿, and
2. βˆ€π‘† βŠ† 𝑉 such that 𝑆 ≀ 𝛼𝑛 there is Ξ“ 𝑆 β‰₯ 𝑐 𝑆 .
In most applications it is desired to have 𝑑 as small as
possible and 𝑐 as large as possible.
6
The Probabilistic Method
2015June
Of particular interest are those graph where 𝛼, 𝑐 and 𝑑
are constants independent of 𝑛 and 𝑐 > 1.
Those are strict requirements and it is not trivial task to
construct such graphs. We rather show that such graphs
exist.
We show that a random graph chosen from a suitable
probability space has a positive probability of being an
𝑛, 18, β…“, 2 OR-concentrator. (The constants are
arbitrary, other combinations are possible.)
Theorem: There is an integer 𝑛0 such that for all 𝑛 > 𝑛0
there is an 𝑛, 18, β…“, 2 OR-concentrator.
7
The Probabilistic Method
2015June
Proof: The proof is carried out in terms of 𝑑, 𝑐, and 𝛼,
while the constants are pinned at the end of the proof.
Consider a random 𝐺 𝐿, 𝑅, 𝐸 , where 𝑣 ∈ 𝐿 choses its 𝑑
neighbors Ξ“ 𝑣 βŠ† 𝑅 independently and uniformly with
replacements, and avoid multi edges.
Let πœ€π‘  be the event that for 𝑆 βŠ† 𝐿, 𝑆 = 𝑠 there is Ξ“ 𝑆
< 𝑐𝑠, namely, an OR-connector does not exist.
Consider 𝑆 βŠ† 𝐿, 𝑆 = 𝑠 and any 𝑇 βŠ† 𝑅, 𝑇 = 𝑐𝑠. There
𝑛
𝑛
are
ways to choose 𝑆 and
ways to choose 𝑇.
𝑠
𝑐𝑠
Since 𝑑𝑠 edges emanate from 𝑆, the probability that 𝑇
contains Ξ“ 𝑆 is 𝑐𝑠 𝑛 𝑑𝑠 .
8
The Probabilistic Method
2015June
The probability of the event that all the 𝑑𝑠 edges
emanating from 𝑠 vertices of 𝐿 fall within any 𝑐𝑠
vertices of 𝑅 is bounded by
Pr πœ€π‘ 
𝑛
≀
𝑠
𝑛
𝑐𝑠
𝑐𝑠 𝑑𝑠
𝑛
𝑛
Substituition of the inequality
≀
π‘˜
Pr πœ€π‘ 
𝑛𝑒
≀
𝑠
𝑠
𝑛𝑒
𝑐𝑠
𝑐𝑠
𝑐𝑠
𝑛
𝑑𝑠
=
𝑠
𝑛
𝑛𝑒 π‘˜
π‘˜
obtains
π‘‘βˆ’π‘βˆ’1
𝑠
𝑒 1+𝑐 𝑐 π‘‘βˆ’π‘
Using 𝛼 = 1 3 and s ≀ 𝛼𝑛, there is
9
The Probabilistic Method
2015June
1
3
Pr πœ€π‘  ≀
𝑠
π‘‘βˆ’π‘βˆ’1
𝑒 1+𝑐 𝑐 π‘‘βˆ’π‘
𝑐
3
≀
𝑠
𝑑
3𝑒
𝑐+1
Using 𝑐 = 2 and 𝑑 = 18, there is
Pr πœ€π‘  ≀
Let π‘Ÿ = 2 3
18
2
3
𝑠
18
3𝑒
3
=
π‘Ÿ
1βˆ’π‘Ÿ
3𝑒 3 , there is
𝑠β‰₯1 Pr
πœ€π‘  ≀
𝑠β‰₯1
π‘Ÿπ‘ 
< 1,
showing that the desired OR-connector exists. ∎
10
The Probabilistic Method
2015June
Crossing Number
The crossing number π‘π‘Ÿ 𝐺 of a graph 𝐺 is the smallest
number of edge crossings in a planar embedding of 𝐺.
In VLSI it is the number of jumpers (via) required to
layout a circuit.
For a planar graph 𝐺 𝑉, 𝐸 , 𝑉 = 𝑛, 𝐸 = π‘š there is
π‘π‘Ÿ 𝐺 = 0.
Euler formula for planar graph states 𝑛 βˆ’ π‘š + 𝑓 = 2.
Since a face comprises a least 3 edges, and each edge is
shared by two faces, there is
0 = 𝑛 βˆ’ π‘š + 𝑓 βˆ’ 2 ≀ 𝑛 βˆ’ π‘š 3 βˆ’ 2.
11
The Probabilistic Method
2015June
Since π‘π‘Ÿ 𝐺 = 0 for a planar 𝐺, for any 𝐺 there is
π‘π‘Ÿ 𝐺 β‰₯ π‘š βˆ’ 3𝑛 + 6 for 𝑛 β‰₯ 3. (Homework)
Stronger lower bound can be derived with the aid of
expectation.
Lemma: (The Crossing Lemma, proof by N. Alon). Let 𝐺
be a simple graph with π‘š β‰₯ 4𝑛. Then
π‘π‘Ÿ 𝐺 β‰₯
1 π‘š3
.
2
64 𝑛
Proof: Let 𝐺 be a planar embedding of 𝐺 yielding π‘π‘Ÿ 𝐺 .
Let 𝑆 βŠ† 𝑉 be obtained by choosing 𝑣 ∈ 𝑉 randomly with
probability 𝑝 ≔ 4𝑛 π‘š. Let 𝐻 ≔ 𝐺 𝑆 and 𝐻 ≔ 𝐺 𝑆 .
12
The Probabilistic Method
2015June
𝐻 is a planar embedding of 𝐻 imposed by 𝐺.
Let 𝑋, π‘Œ and 𝑍 be the random variables of the number of
vertices, number of edges and the number of crossings
in 𝐻, respectively.
It follows from the trivial lower bound that 𝑍 β‰₯ π‘π‘Ÿ 𝐻
β‰₯ π‘Œ βˆ’ 3𝑋 + 6. By linearity of expectation there is 𝐸 𝑍
β‰₯ 𝐸 π‘Œ βˆ’ 3𝐸 𝑋 .
There is 𝐸 𝑋 = 𝑝𝑛 and 𝐸 π‘Œ = 𝑝2 π‘š (an edge is defined
by its two end vertices).
Since a crossing is defined by four vertices, there is 𝐸 𝑍
= 𝑝4 π‘π‘Ÿ 𝐺 .
13
The Probabilistic Method
2015June
All in all there is
𝑝4 π‘π‘Ÿ 𝐺 β‰₯ 𝑝2 π‘š βˆ’ 3𝑝𝑛.
Dividing by 𝑝4 yields
π‘π‘Ÿ 𝐺 β‰₯
π‘π‘šβˆ’3𝑛
𝑝3
=
𝑛
4𝑛 π‘š 3
=
1 π‘š3
.
64 𝑛2
∎
The Crossing Lemma is useful in combinatorial
geometry. Consider 𝑛 points in the plane and lines
passing through each pair of points.
𝑛
Some of those
at most distinct lines might pass
2
through more than two points.
14
The Probabilistic Method
2015June
Given π‘˜ β‰₯ 2, how many lines can pass through at least π‘˜
points?
If 𝑛 is a perfect square and the point are on a 𝑛 × π‘›
grid, there are 2 𝑛 + 2 lines passing through 𝑛 points.
Is there a configuration of 𝑛 points in the plane yielding
more lines passing through at least 𝑛 points?
Theorem: (Szemerédi and Trotter 1983). Let 𝑃 be a set
of 𝑛 points in the plane, and let 𝑙 be the number of lines
passing through at least π‘˜ + 1 points of 𝑃 , 1 ≀ π‘˜
≀ 2 𝑛. Then
𝑙 < 32𝑛2 π‘˜ 3 .
15
The Probabilistic Method
2015June
Proof: Form a graph 𝐺 with vertex set 𝑃.
𝐺’s edges are the segments between consecutive points
of the 𝑙 lines. 𝐺 has therefore at least π‘˜π‘™ edges and its
𝑙
crossing number is at most
.
2
If it happens that π‘˜π‘™ < 4𝑛 then 𝑙 < 4𝑛 π‘˜ ≀ 32𝑛2 π‘˜ 3 (1
≀ π‘˜ ≀ 2 𝑛).
Otherwise π‘˜π‘™ β‰₯ 4𝑛, and the Crossing Lemma applies.
𝑙
It follows from the lemma that 𝑙 2 >
β‰₯ π‘π‘Ÿ 𝐺
2
β‰₯ π‘˜π‘™ 3 64𝑛2 , yielding again 𝑙 ≀ 32𝑛2 π‘˜ 3 . ∎
2
16
The Probabilistic Method
2015June
Markov’s Inequality and Random Graphs
Let 𝛺𝑛 , 𝑃𝑛 , 𝑛 β‰₯ 1 be a probability space, 𝛺𝑛 is a
sample space and 𝑃𝑛 ∢ 𝛺𝑛 β†’ 0,1
a probability
function satisfying πœ”βˆˆπ›Ίπ‘› 𝑃𝑛 πœ” = 1.
We subsequently explore the existence of few
properties in random large graphs.
Large means 𝑉 𝐺 = 𝑛 β†’ ∞, whereas the probability
𝑝 of an edge depends on 𝑛 and satisfies 𝑝 𝑛 β†’ 0.
𝐆𝑛,𝑝 denotes the probability space of such graphs.
17
The Probabilistic Method
2015June
Let 𝑋 be a nonnegative random variable and 𝑑 ∈ ℝ, 𝑑
> 0. Then
𝐸 𝑋
𝑃 𝑋β‰₯𝑑 ≀
𝑑
Markov’s inequality is applied to show that a 𝐺 ∈ 𝐆𝑛,𝑝
almost surly has a particular property for a certain 𝑝.
It is obtained by setting 𝑋 = 𝑋𝑛 and 𝑑 = 1.
Corollary: Let 𝑋𝑛 ∈ β„• be a nonnegative random variable
in a probability space 𝛺𝑛 , 𝑃𝑛 , 𝑛 β‰₯ 1. If 𝐸 𝑋𝑛 β†’ 0 as
𝑛 β†’ ∞, then 𝑃 𝑋𝑛 = 0 β†’ 1 as 𝑛 β†’ ∞.
18
The Probabilistic Method
2015June
Asymptotic Behavior of Graphs
We are interested in the number 𝑋 of triangles in 𝐺
∈ 𝐆𝑛,𝑝 .
𝑋 can be expressed as the sum
𝑋=
𝑋𝑆 ∢ 𝑆 βŠ† 𝑉, 𝑆 = 3 ,
where 𝑋𝑆 is the indicator random variable for the event
𝐴𝑆 that 𝐺 𝑆 is a triangle.
𝑋𝑆 = 1 if 𝑆 imposes a triangle and 𝑋𝑆 = 0 otherwise. By
the expectation definition there is
𝐸 𝑋𝑆 = 𝑃 𝑋𝑆 = 1 .
There is 𝑃 𝐴𝑆 = 𝑝3 .
19
The Probabilistic Method
2015June
By linearity of expectation, there is
𝑛 3
𝐸 𝑋 = 𝐸 𝑋𝑆 ∢ 𝑆 βŠ† 𝑉, 𝑆 = 3 =
𝑝 < 𝑝𝑛 3 .
3
Thus if 𝑝𝑛 β†’ 0 as 𝑛 β†’ ∞, then𝐸 𝑋 β†’ 0, and 𝑃 𝑋 = 0
β†’ 1.
That means that if 𝑝𝑛 β†’ 0 as 𝑛 β†’ ∞, 𝐺 will almost surly
be triangle-free.
We subsequently are interested in the probability of
having the independent sets in a graph of 𝑛 vertices and
edge probability 𝑝, not exceeding a certain size.
20
The Probabilistic Method
2015June
Theorem: (ErdΕ‘s 1961). A random graph 𝐺 ∈ 𝐆𝑛,𝑝
almost surely has stability number (size of maximal
independent set) no larger than 2π‘βˆ’1 log 𝑛 .
Proof: Let 𝑆 βŠ‚ 𝑉 𝐺 , 𝑆 = π‘˜ + 1, π‘˜ ∈ β„•. π‘˜ is pinned
down later.
The probability that 𝑆 is an independent set is the
probability that none of the vertex pairs has a
connecting edge, namely, 1 βˆ’ 𝑝
π‘˜+1
2
.
Let 𝐴𝑆 be the event that 𝑆 is an independent set and let
𝑋𝑆 be the corresponding indicator random variable.
21
The Probabilistic Method
2015June
There is
𝐸 𝑋𝑆 = 𝑃 𝑋𝑆 = 1 = 𝑃 𝐴𝑆 = 1 βˆ’ 𝑝
π‘˜+1
2
.
Let 𝑋 be the number of independent sets of size π‘˜ + 1.
Then
𝑋 = 𝑋𝑆 ∢ 𝑆 βŠ‚ 𝑉, 𝑆 = π‘˜ + 1 .
By linearity of expectation there is
𝐸 𝑋 =
𝐸 𝑋𝑆 ∢ 𝑆 βŠ‚ 𝑉, 𝑆 = π‘˜ + 1 =
𝑛
π‘˜+1
22
1βˆ’π‘
π‘˜+1
2
The Probabilistic Method
.
2015June
𝑛
There is
≀
π‘˜+1
π‘›π‘˜+1
π‘˜+1 !
and 1 βˆ’ 𝑝 < 𝑒 βˆ’π‘ .
Substitution in 𝐸 𝑋 yields
𝐸 𝑋 ≀
𝑛
π‘˜+1 βˆ’π‘
π‘˜+1
2
𝑒
π‘˜+1 !
=
π‘˜+1
βˆ’π‘π‘˜
2
𝑛𝑒
π‘˜+1 !
Let us now pin down π‘˜, supposing π‘˜ = 2π‘βˆ’1 log 𝑛 .
Then π‘˜ β‰₯ 2π‘βˆ’1 log 𝑛, and by exponentiation there is
𝑛𝑒 βˆ’π‘π‘˜ 2 ≀ 1, hence
1
𝐸 𝑋 ≀
π‘˜+1 !
23
The Probabilistic Method
2015June
Since π‘˜ β‰₯ 2π‘βˆ’1 log 𝑛, π‘˜ grows at least as fast as log 𝑛,
hence 𝐸 𝑋 β†’ 0 as 𝑛 β†’ ∞.
Recall the corollary stating that if 𝐸 𝑋 β†’ 0 as 𝑛 β†’ ∞,
then 𝑃 𝑋 = 0 β†’ 1 as 𝑛 β†’ ∞.
That means that with probability approaching 1, 𝐆𝑛,𝑝
has no independent set larger than π‘˜, or in other words,
its stability number is at most π‘˜. ∎
24
The Probabilistic Method
2015June
Related documents