Download Lecture 3 Gaussian Probability Distribution Introduction

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

History of network traffic models wikipedia , lookup

Inductive probability wikipedia , lookup

Birthday problem wikipedia , lookup

History of statistics wikipedia , lookup

Infinite monkey theorem wikipedia , lookup

Student's t-distribution wikipedia , lookup

Exponential distribution wikipedia , lookup

Law of large numbers wikipedia , lookup

Negative binomial distribution wikipedia , lookup

Transcript
Lecture 3
Gaussian Probability Distribution
Introduction
l
l
Gaussian probability distribution is perhaps the most used distribution in all of science.
u also called “bell shaped curve” or normal distribution
Unlike the binomial and Poisson distribution, the Gaussian is a continuous distribution:
P(y) =
†
l
1
e
s 2p
-
( y-m) 2
2s 2
m = mean of distribution (also at the same place as mode and median)
s2 = variance of distribution
y is a continuous variable (-∞ £ y £ ∞)
Probability (P) of y being in the range [a, b] is given by an integral:
P(a < y < b) =
u
b -
1
Úe
s 2p a
( y-m) 2
2s 2 dy
Karl Friedrich Gauss 1777-1855
The integral for arbitrary a and b cannot be evaluated analytically
+ The value of the integral has to be looked up in a table (e.g. Appendixes A and B of Taylor).
†
1
p(x) =
e
s 2p
P(x)
(x - m )2
2
2s
gaussian
Plot of Gaussian pdf
x
K.K. Gan
L3: Gaussian Probability Distribution
1
l
The total area under the curve is normalized to one.
+ the probability integral:
†
( y-m) 2
2s 2 dy = 1
1
Úe
s 2p -•
We often talk about a measurement being a certain number of standard deviations (s) away
from the mean (m) of the Gaussian.
+ We can associate a probability for a measurement to be |m - ns|
from the mean just by calculating the area outside of this region.
ns Prob. of exceeding ±ns
0.67
0.5
It is very unlikely (< 0.3%) that a
1
0.32
measurement taken at random from a
2
0.05
Gaussian pdf will be more than ± 3s
3
0.003
from the true mean of the distribution.
4
0.00006
P(-• < y < •) =
l
• -
Relationship between Gaussian and Binomial distribution
l
l
The Gaussian distribution can be derived from the binomial (or Poisson) assuming:
u p is finite
u N is very large
u we have a continuous variable rather than a discrete variable
An example illustrating the small difference between the two distributions under the above conditions:
u Consider tossing a coin 10,000 time.
p(heads) = 0.5
N = 10,000
K.K. Gan
L3: Gaussian Probability Distribution
2
For a binomial distribution:
mean number of heads = m = Np = 5000
standard deviation s = [Np(1 - p)]1/2 = 50
+ The probability to be within ±1s for this binomial distribution is:
5000+50
10 4 !
m
10 4 -m
P=
0.5
0.5
= 0.69
Â
4
m=5000-50 (10 - m)!m!
n For a Gaussian distribution:
( y-m) 2
m
+
s
2
1
P(m - s < y < m + s ) =
Ú e 2s dy ª 0.68
s 2p m-s
†+ Both distributions give about the same
probability!
n
Central Limit Theorem
l
l
l
Gaussian
distribution is important because of the Central Limit Theorem
†
A crude statement of the Central Limit Theorem:
u Things that are the result of the addition of lots of small effects tend to become Gaussian.
A more exact statement:
Actually, the Y’s can
u Let Y1, Y2,...Yn be an infinite sequence of independent random variables
be from different pdf’s!
each with the same probability distribution.
u Suppose that the mean (m) and variance (s2) of this distribution are both finite.
+ For any numbers a and b:
È Y +Y + ...Yn - nm
˘
1 b - 12 y 2
lim PÍa < 1 2
< b˙ =
dy
Úe
˚
nÆ• Î
s n
2p a
+ C.L.T. tells us that under a wide range of circumstances the probability distribution
that describes the sum of random variables tends towards a Gaussian distribution
as the number of terms in the sum Æ∞.
† K.K. Gan
L3: Gaussian Probability Distribution
3
Alternatively:
È Y -m
˘
È
˘
Y -m
1 b - 12 y 2
lim PÍa <
< b˙ = lim PÍa <
< b˙ =
dy
Úe
nÆ• Î
˚ nÆ• Î
s/ n
sm
2p a
˚
n sm is sometimes called “the error in the mean” (more on that later).
l For CLT to be valid:
u m and s of pdf must be finite.
†u No one term in sum should dominate the sum.
l A random variable is not the same as a random number.
u Devore: Probability and Statistics for Engineering and the Sciences:
+ A random variable is any rule that associates a number with each outcome in S
n S is the set of possible outcomes.
l Recall if y is described by a Gaussian pdf with m = 0 and s = 1 then
the probability that a < y < b is given by:
+
1 b - 12 y 2
P ( a < y < b) =
dy
Úe
2p a
l
The CLT is true even if the Y’s are from different pdf’s as long as
the means and variances are defined for each pdf!
u See Appendix of Barlow for a proof of the Central Limit Theorem.
K.K. Gan
L3: Gaussian Probability Distribution
4
l
Example: A watch makes an error of at most ±1/2 minute per day.
After one year, what’s the probability that the watch is accurate to within ±25 minutes?
u Assume that the daily errors are uniform in [-1/2, 1/2].
n For each day, the average error is zero and the standard deviation 1/√12 minutes.
n The error over the course of a year is just the addition of the daily error.
n Since the daily errors come from a uniform distribution with a well defined mean and variance
+ Central Limit Theorem is applicable:
È Y1 +Y2 + ...Yn - nm
˘
1 b - 12 y 2
lim PÍa <
< b˙ =
dy
Úe
˚
nÆ• Î
s n
2p a
+ The upper limit corresponds to +25 minutes:
Y +Y + ...Yn - nm 25 - 365 ¥ 0
b= 1 2
=
= 4.5
1 365
s n
† + The lower limit corresponds to12-25 minutes:
Y +Y + ...Yn - nm -25 - 365 ¥ 0
a= 1 2
=
= -4.5
1 365
s n
12
† + The probability to be within ± 25 minutes:
1 4.5 - 12 y 2
P=
dy = 0.999997 = 1- 3¥10 -6
Ú e
2p
† + less than three-4.5
in a million chance that the watch will be off by more than 25 minutes in a year!
†
K.K. Gan
L3: Gaussian Probability Distribution
5
l
Example: Generate a Gaussian distribution using random numbers.
u Random number generator gives numbers distributed uniformly in the interval [0,1]
n m = 1/2 and s2 = 1/12
u Procedure:
n Take 12 numbers (ri) from your computer’s random number generator
n Add them together
n Subtract 6
+ Get a number that looks as if it is from a Gaussian pdf!
È Y +Y2 + ...Yn - nm
˘
PÍa <
< b˙
Î
˚
s n
A) 5000 random numbers
B) 5000 pairs (r1 + r2)
of random numbers
12
È
˘
1
 ri -12 ⋅ 2
Í
˙
i=1
= PÍa <
< b˙
1
⋅ 12
Í
˙
12
ÍÎ
˙˚
12
È
˘
= PÍ-6 < Â ri - 6 < 6˙
Î
˚
C) 5000 triplets (r1 + r2 + r3) D) 5000 12-plets (r1 + r2 +…r12)
i=1
of random numbers
1 6 - 12 y 2
=
dy
Úe
2p -6
E) 5000 12-plets
E
Thus the sum of 12 uniform random
numbers minus 6 is distributed as if it came
from
† a Gaussian pdf with m = 0 and s = 1.
K.K. Gan
of random numbers.
L3: Gaussian Probability Distribution
(r1 + r2 +…r12 - 6) of
random numbers.
Gaussian
m = 0 and s = 1
-6
0
+6
6
l
Example: The daily income of a "card shark" has a uniform distribution in the interval [-$40,$50].
What is the probability that s/he wins more than $500 in 60 days?
u Lets use the CLT to estimate this probability:
È Y +Y + ...Yn - nm
˘
1 b - 12 y 2
lim PÍa < 1 2
< b˙ =
dy
Úe
˚
nÆ• Î
s n
2p a
u The probability distribution of daily income is uniform, p(y) = 1.
+ need to be normalized in computing the average daily winning (m) and its standard deviation (s).
50
Ú yp(y)dy
†
m = -40
=
50
Ú p(y)dy
1 [50 2
2
- (-40)2 ]
50 - (-40)
=5
-40
50
2
2
Ú y p(y)dy
2
s = -4050
-m =
Ú p(y)dy
1 [50 3 - (-40) 3 ]
3
- 25 = 675
50 - (-40)
-40
u
†
u
†
+
The lower limit of the winning is $500:
Y +Y + ...Yn - nm 500 - 60 ¥ 5 200
a= 1 2
=
=
=1
s n
675 60
201
The upper limit is the maximum that the shark could win (50$/day for 60 days):
Y +Y + ...Yn - nm 3000 - 60 ¥ 5 2700
b= 1 2
=
=
= 13.4
s n
675 60
201
1 13.4 - 12 y 2
1 • - 12 y 2
P=
e
dy
ª
dy = 0.16
Ú
Úe
2p 1
2p 1
16% chance to win > $500 in 60 days
K.K. Gan
†
L3: Gaussian Probability Distribution
7