Download Random Variable A

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
Probability Theory Overview and
Analysis of Randomized Algorithms
Analysis of Algorithms
Prepared by
John Reif, Ph.D.
Probability Theory Topics
a) Random Variables: Binomial and
Geometric
b) Useful Probabilistic Bounds and
Inequalities
Readings
• Main Reading Selections:
– CLR, Chapter 5 and Appendix C
Probability Measures
• A probability measure (Prob) is a
mapping from a set of events to the
reals such that:
1) For any event A
0  Prob(A)  1
2) Prob (all possible events) = 1
3) If A, B are mutually exclusive
events, then
Prob(A  B) = Prob (A) + Prob (B)
Conditional Probability
• Define
Prob(A  B)
Prob(A | B) 
Prob(B)
for Prob(B) > 0
Bayes’ Theorem
• If A1, …, An are mutually exclusive and
contain all events then
Prob(A i | B) 
Pi
n
P
j
j=1
where Pj = Prob(B | A j )  Prob(A j )
Random Variable A
(Over Real Numbers)
• Density Function
f A (x) = Prob(A=x)
Random Variable A (cont’d)
• Prob Distribution Function
x
FA (x) = Prob(A  x) =  f A (x) dx
-
Random Variable A (cont’d)
• If for Random Variables A,B
x FA (x)  FB (x)
• Then “A upper bounds B” and
“B lower bounds A”
FA (x)  Prob (A  x)
FB (x)  Prob (B  x)
Expectation of Random Variable A

E(A) = A =  x f A (x) dx
-
• Ā is also called “average of A” and
“mean of A” = A
Variance of Random Variable A
σ 2A  (A  A) 2  A 2  (A) 2

where 2nd momoment A 
2
x
-
2
f A (x) dx
Variance of Random Variable A
(cont’d)
Discrete Random Variable A
Discrete Random Variable A
(cont’d)
Discrete Random Variable A
Over Nonnegative Numbers
• Expectation

E(A) = A =  x f A (x)
x=0
Pair-Wise Independent Random
Variables
• A,B independent if
Prob(A  B) = Prob(A) X Prob(B)
• Equivalent definition of independence
f A  B (x) = f A (x)  f B (x)
M A  B (s) = M A (s)  M B (s)
G A  B (z) = G A (z)  G B (z)
Bounding Numbers of Permutations
• n! = n x (n-1) x 2 x 1
= number of permutations of n objects
• Stirling’s formula
n! = f(n) x (1+o(1)), where
n -n
f(n) = n e
2πn
Bounding Numbers of Permutations
(cont’d)
• Note
– Tighter bound
f(n) e
n!
(n-k)!
1
(12n+1)
< n! < f(n) e
1
12n
= number of permutations of n
objects taken k at a time
Bounding Numbers of Combinations
n
n!
 
 k  k! (n-k)!
= number of (unordered)
combinations of n objects
taken k at a time
• Bounds (due to Erdos & Spencer, p. 18)
k 2 k3


k
2n 6n 2
n n e
 ~
k!
k
(1  o(1))
 
for k = o  n 
 
3
4
Bernoulli Variable
• Ai is 1 with prob P and 0 with prob 1-P
• Binomial Variable
• B is sum of n independent Bernoulli
variables Ai each with some probability p
procedure
begin
end
BINOMIAL with parameters n,p
B0
for i=1 to n do
with probability P do B  B+1
output
B is Binomial Variable with
Parameters n,p
mean   n  p
variance  2  np (1-p)
B is Binomial Variable with
Parameters n,p (cont’d)
n x
density fn  Prob(B=x) =   p (1-p) n-x
x
n k
n-k
distribution fn  Prob(B  x) =   p (1-p)
k=0  k 
x
Poisson Trial
• Ai is 1 with prob Pi
and 0 with prob 1-Pi
• Suppose B’ is the
sum of n independent Poisson trials
Ai with probability Pi for i > 1, …, n
Hoeffding’s Theorem
• B’ is upper bounded by a Binomial
Variable B
n
P
i
• Parameters n,p where
p=
i=1
n
Geometric Variable V
• parameter p
x  0
Prob(V=x) = p(1-p)x
procedure
GEOMETRIC parameter p
begin
V 0
loop with probability 1-p
goto exit
Probabilistic Inequalities
• For Random Variable A
mean   A
variance   A  (A)
2
2
2
Markov and Chebychev
Probabilistic Inequalities
• Markov Inequality
Prob (A  x) 
(uses only mean)

x
• Chebychev Inequality
and variance)
Prob ( A    ) 
(uses mean
2
2
Example of use of Markov and
Chebychev Probabilistic Inequalities
• If B is a Binomial with parameters n,p
np
Then Prob (B  x) 
x
np (1-p)
Prob ( B  np  ) 
2

Gaussian Density Function
 (x) =
1
e
2π
x2
2
Normal Distribution
x
 (X) =
 (Y) dY
-
• Bounds x > 0 (Feller, p. 175)
Ψ(x)
1 1 
 (x)   3   1  (x) 
x
x x 
x  [0,1]
x
1
 x (1)  (x) -   (0) =
2
2πe
x
2π
Sums of Independently Distributed
Variables
• Let Sn be the sum of n independently
distributed variables A1, …, An
2

• Each with mean
and variance
n
n
• So Sn has mean  and variance 2
Strong Law of Large Numbers:
Limiting to Normal Distribution
• The probability density function of
Tn =
(Sn - )

limits as n  
to normal distribution (x)
• Hence
Prob (
Sn -   x)  (x) as n  
Strong Law of Large Numbers
(cont’d)
• So
Prob (
Sn -   x)  2(1  (x))
 2(x)/x
(since 1- (x)  (x)/x)
Advanced Material
Moment Generating Functions and
Chernoff Bounds
Moments of Random Variable A
(cont’d)
• n’th Moments of Random Variable A

A 
n
x
n
f A (x) dx
-
• Moment generating function

M A (s)   esx f A (x) dx
-
= E(esA )
Moments of Random Variable A
(cont’d)
• Note
S is a formal parameter
 d M A (s) 


n
ds


n
A
n
s0
Moments of Discrete Random
Variable A
• n’th moment

A   x f A (x)
n
n
x=0
Probability Generating Function of
Discrete Random Variable A

G A (z) =  z f A (x)  E(z )
x
A
x=0
1st derivative G 'A (1) = A
2nd derivative G"A (1) = A2  A
variance σ  G (1)  G (1)  (G (1))
2
A
"
A
'
A
'
A
2
Moments of AND of Independent
Random Variables
• If A1, …, An independent with same
distribution
f A1 (x) = f Ai (x) for i = 1, ... n
Then if B = A1  A 2  ...  A n

f B (x) = f A1 (x)


n

n

M B (s) = M A1 (s) , G B (z) = G A1 (z)

n
Generating Function of Binomial
Variable B with Parameters n,p
n k
n-k
G(z) = (1-p+pz) = z   p (1-p)
k=0
k
x
n
k
• Interesting fact
 1 
Prob(B= ) = 

 n
Generating Function of Geometric
Variable with parameter p

p
G(Z) =  Z (p(1-p) ) =
1-(1-p)Z
k=0
k
k
Chernoff Bound of Random
Variable A
• Uses all moments
• Uses moment generating function
Prob (A  x)  e
-sx
=e
M A (s) for s  0
γ(s) - sx
 e
where γ(s) = ln (M A (s))
γ(s) - s γ'(s)
By setting x = ’ (s)
1st derivative minimizes bounds
Chernoff Bound of Discrete Random
Variable A
Prob (A  x)  z G A (z) for z  1
-x
• Choose z = z0 to minimize above
bound
• Need Probability Generating function
G A (z) =  z f A (x) = E(z )
x
x 0
A
Chernoff Bounds for Binomial B
with parameters n,p
• Above mean x  
Prob (B  x)
 n- 


 n-x 
n-x

 
x
x

 1
1
 e   since 1    e
x
 x
-x - μ
2
 e
for x  μe
x
x-μ
x
Chernoff Bounds for Binomial B
with parameters n,p (cont’d)
• Below mean x  
Prob (B  x)
 n- 


 n-x 
n-x

 
x
x
Anguin-Valiant’s Bounds for
Binomial B with Parameters n,p
• Just above mean
 = np for 0 <  < 1
Prob (B  (1+ ) )  e
-ε 2
μ
2
• Just below mean
 for 0 <  < 1
Prob (B   (1- ) )  e
-ε 2
μ
3
Anguin-Valiant’s Bounds for
Binomial B with Parameters n,p
(cont’d)
 Tails are bounded by Normal
distributions
Binomial Variable B with Parameters
p,N and Expectation  = pN
• By Chernoff
Bound for p < ½
N
N
Prob (B  )  2 p
2
n
2
• Raghavan-Spencer bound
for any  > 0


e
Prob (B  (1+ ) )  
(1+ ) 
 (1+ )


μ
Probability Theory
Analysis of Algorithms
Prepared by
John Reif, Ph.D.
Related documents