Download Exponential distribution

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Statistics wikipedia , lookup

History of statistics wikipedia , lookup

Probability wikipedia , lookup

Transcript
UNIT-1
Random variables:
Random variables:
Discrete
Continuous
Probability
Distribution
Discrete
Binomial D.
Continuous
Poisson D.
Related Problems, mean,
Normal D.
Variance, standard deviation
In a random experiment, the outcomes are governed by
chance mechanism and the sample space’s consists of all
outcomes of the experiment when the elements of the
sample space are non-numeric, they can be quantified by
assigning a real no to every event of the sample space. This
assignment rule is known as random variable.
Random Variable:
A random variable X on a sample space S is a function
X:S→R from S to the set of real no.s which assigns a real
number X(x) to each sample point of S.
s
r
x
x(x)
[i.e., the pre image of every element of R is an event of S]
Range:
The range space Rx is the set of all possible values of
X:Rx≤R
Note: Although X is called a random variable, it is intact a
single valued fn.
X denotes random variable and x denotes one of its values.
Discrete Random Variable:
A random variable X is said to be discrete random variable
if its set of all possible outcomes, the sample space S is
countable (termite or countable infinite).
Counting Problems give rise to discrete random variables.
Continuous Random variable:
A random variable X is said to be continuous random
variable if S contains infinite no.s equal to the no. of points
on a live segment.
i.e., it takes all the possible values in an interval.
(An internal contains uncountable no. of possible values).
Ex: Consider the experiment of tossing a coin twice.
Sample space S={HH,HT,TH,TT}. Define X:S→R by
X(s)=no. of heads in S
. X (s )  2 X (s )  1X (s )  1X (s )  0
Range of X   X (8) :8  s  0,1, 2
1
2
3
4
2. Consider the experiment of throwing a pair of dice and
noting sum
S={(1,1)(1,2)_ _ _ _(6,6)}
Then the random variable for this experiment is defined as
X : S  RbyX (i, j )  i  j(i, j )  S .
{1.If X and Y are two random variables defined on S and a
and b are two real no.s, then.
(i)ax+by is also a random variable . In particular x-y is also
a random variable.
(ii) xy is also a random variable
(iii) IfX (s)  0s  sthen 1/x is also a random a
variable.}
.
It an a random experiment, the event corresponding to
a number a occurs, then the corresponding random variable
X assumes a and the probability of that event is denoted by
p(x=a) similarly, the probability of the event x assuming
any value in the interval a<x<b).
The probability of the event x≤c is written as p(x.≤c).
Note that more than one random variable can be
defined in a sample space
Discrete random variable:
It we can count the possibilities, of x1 then x is discrete.
Ex: The random variable x=the sum of the dots on two dice
is discrete.
X can assume the value 2,3,4,5………12.
Continuous:
In an interval of real no.s, there are an infinite no. of
possible values.
Ex: The random variable x= the time that an athlete crosses
the winning live.
[Probability density function:
The p.d.f. of a random variable x denoted by of (x) has the
favoring properties.
f ( x) 
1.
 f ( x)dx  1
2.


3.
P( E )   f ( x)dx
Where E is any event.
E
Note: P(F) = 0 does not imply that E is nill event or
impossible event.]
Probability distribution or distribution:
The probability distribution or distribution f(x) of a random
variable X is a description of the set of possible values of
(range of x) along with the probability associated with each
of x
Ex: Let x=the no.of heads in tossing two coins
X=x
0
1
2
F(x) = P(X=x)
¼
2/4=1/2
1/4
Cumulative distribution function:
The cumulative distribution function for a random variable
x is defined by F(x) = P(x≤x) where x is any real no. (…….)
Properties:
1. If a<b then p(a<x≤b) = F(b)-F(a).
2. P(a≤x≤b) = p(x=a)+f(b)-f(a)
3. P(a<x<b) = F(b)-F(a)-p(x=b)
4. P(a≤x<b)=F(b)-F(a)-P(x)
According to the type of random variables, we have two
types of probability distributions.
1 Discrete probability distribution
2. Continuous probability distribution.
Discrete Probability distribution:
Let x be a discrete random variable. The discrete
probability distribution function f(x) for x is given by
satisfying the properties
f ( x)  p( x  x)xorf ( xi )  p( x  xi )i  1, 2...
1. P(xi) ≥0 V i
2. ∑P(xi) = 1
1. f(x)>,0 V x
2. ∑f(x) = 1 x←x
The discrete probability function can also be called as
probability mass function.
Any function satisfying the above 2 properties, will be a
discrete probability fn. or probability mass function.
X=x x1 x2 x3
P(x=x) P1 P2 P3
Ex: x: the sum of no.s which turn on tossing a pair of dice.
X=xi
2
3
4
5
6
7
8
9
10 11
12
P(x=xi)1/36 2/36 3/36 4/36 5/36 6/36 5/36 4/36 3/36 2/36
1/36
1 .P(xi)≥0
2. ∑P(xi)=1
X x1 x2 ---------------Xn
P(x) p(x1)
p(x2)
--------------P(Xn)
1. P (X<xi) = P (x1) + P (X2) + ----------+P(Xi-i)
2. P (X≤Xi) = P (x1) + P (X2)+ ---------+P (Xi-1) +
P(Xi)
3. P (X>Xi) = 1 – P (x≤xi)
Check whether the following can act as discrete probability
functions.
1. F(x) = X - 2/2 for X = 1,2,3,4
2. F(x) =XF2/25 x = 0,1,2,3,4
(1)
cannot f(x)<0 for x=1
(2)
cannot ∑f(x)≠1
General properties:
Expectation, mean, variance & Standard deviation :
Expectation:
Let a random variable x assumes the values x1,x2--------xn
with respective probabilities f1,f2,--------fn. Then the
expectation of X E(x) is defined as the sum of different
values of x and the corresponding probabilities.
E( X ) 
n

i 1
pi xi
Results:
1. It x is a random variable and k is a constant then
a. F(x+k) = F(x)+k
b. E(x)
2 If x and y are two discrete random variables, then E(x+y)
= E(x)+E(y)
Note:
1. E(x+y+z) =E(x) + E(y) + E(z)
2. E(ax+by) = a E(x) + bE(y)
3. E(xyz) = E(x) E(y) E(z)
Mean:
The mean value u of a discrete distribution
function is given by
n
px
i 1
i
i
 E( X )
Variance: The variance of a discrete distribution function is
given by
n
   pi x 2i   2
2
i 1
Standard deviation:
It is nothing but the positive square root of the variance.
  p x 
n
2
2
i 1
i
2
i
Continuous Probability distribution:
When a random variable x takes every value in an interval,
it is called an continuous random variable.
Ex: Temperature, heights and weights.
The continuous random variable will form a curve that can
be used to calculate the probabilities via areas.
Def: Let x be a continuous random variable. A function f(x)
is said to be a continuous probability function of x if for
any [a,b] a,b←R such that
1. f(x) ≥0
2.  f ( x)dx  1

The total area bounded by

graph f(x) and horizontal axis is
3. p(a≤x≤b) =  f (t )dt
a and b any two values of x
b
a
satisfying a<b.
Continuous probability function f(x) is also called as
probability density function.
Def: the cumulative distribution function of a continuous
random variable x is defined by f(x) = p (x≤xx)
x
F(x) = 
f (t )dt
where f(x) is Continuous probability function

By def. f(x) =
d
( F ( x ))
dx
General Properties;
Let X be a random variable with p.d.f f(x)
The mean of X,

   xf ( x)dx

The variance of X denoted by


   ( x   ) f ( x)dx   x2 f ( x)dx   2
2
2


The standard deviation of X denoted by


  x   
2
f ( x)dx
Results:
If X is a continuous random variable and Y=ax+b then
E(Y)=aE(X)+b
And V(X)= a Var(Y)
Var(X+k)=Var(X)
2
V(kX)=



Median:
k 2 Var(X)
xf ( x)dx
In case of continuous distribution, Mediam is a point which
divides the entire distribution into two equal parts.
If X is defined from a to b,M is median then
M
b
1
f
(
x
)
dx

f
(
x
)
dx

a
M
2
Mode:
Mode is the value of x for which f(x) is maxium.
Mode is given by
f ( x)  0 f ( x)  0 fora  x  b
CHEBYSHEV’S INEQUALITY
: "The probability that the outcome of an experiment with the random variable
will fall more than standard deviations beyond the mean of , , is less than
."
Or: "The proportion of the total area under the pdf of outside of standard
deviations from the mean is at most
."
Proof
Let be the sample space for a random variable, , and let
stand for the
pdf of . Let , and partition , such that for every sample point in
Then
.
Clearly
Since the term that evaluates to the variance in
subtracted on the right-hand side.
has been
For any sample point in
Notice that the direction of the inequality changes since
squaring causes the right-hand expression to become
positive.
And for any sample point in
So, for any sample point in
, and so
or
, it can be said that
Dividing each side of the inequality by
Or, in other terms
results in
UNIT –II
Binomial distribution:
Binomial and Poisson distributions are related to
discrete random variables and normal distribution is related
to continuous random variables.
In many cores, it is desirable to have situations called
repeated trials. For this, we develop a model that is useful
in representing the probability distributions pertaining to
the no. of occurrence of an event in repeated trials of an
experiment.
Binomial distribution is discovered by James
Bernoulli
Bernoulli Trials:
If there are n trials of an experiment in which each
trial has only two mutually exclusive, equally likely and
independent outcomes, then they are called Bernoulli trials,
Let us denote the two outcomes by success and failure.
The Bernoulli distribution is a discrete distribution
having two possible outcomes labeled by
and
in
which
("success") occurs with probability and
("failure") occurs with probability
, where
. It
therefore has probability density function
which can also be written
the binomial distribution
The Binomial Distribution is one of the discrete
probability distribution. It is used when there are
exactly two mutually exclusive outcomes of a trial.
These outcomes are appropriately labeled Success
and Failure. The Binomial Distribution is used to
obtain the probability of observing r successes in n
trials, with the probability of success on a single trial
denoted by p.
Formula:
P(X = r) = nCr p r (1-p)n-r
where,
n = Number of events.
r = Number of successful events.
p = Probability of success on a single trial.
nCr = ( n! / (n-r)! ) / r!
1-p = Probability of failure.
Example: Toss a coin for 12 times. What is the
probability of getting exactly 7 heads.
Step 1: Here,
Number of trials n = 12
Number of success r = 7 (since we define
getting a head as success)
Probability of success on any single trial p
= 0.5
Example: Toss a coin for 12 times. What is the
probability of getting exactly 7 heads.
Step 1: Here,
Number of trials n = 12
Number of success r = 7 (since we define
getting a head as success)
Probability of success on any single trial p
= 0.5
Step 2: To Calculate nCr formula is used.
nCr = ( n! / (n-r)! ) / r!
= ( 12! / (12-7)! ) / 7!
= ( 12! / 5! ) / 7!
= ( 479001600 / 120 ) / 5040
= ( 3991680 / 5040 )
= 792
Step 3: Find pr.
pr = 0.57
= 0.0078125
Step 4: To Find (1-p)n-r Calculate 1-p and n-r.
1-p = 1-0.5 = 0.5
n-r = 12-7 = 5
Step 5: Find (1-p)n-r.
= 0.55 = 0.03125
Step 6: Solve P(X = r) = nCr p r (1-p)n-r
= 792 × 0.0078125 × 0.03125
= 0.193359375
The probability of getting exactly 7 heads is 0.19
Step 2: To Calculate nCr formula is used.
nCr = ( n! / (n-r)! ) / r!
= ( 12! / (12-7)! ) / 7!
= ( 12! / 5! ) / 7!
= ( 479001600 / 120 ) / 5040
= ( 3991680 / 5040 )
= 792
Step 3: Find pr.
pr = 0.57
= 0.0078125
Step 4: To Find (1-p)n-r Calculate 1-p and n-r.
1-p = 1-0.5 = 0.5
n-r = 12-7 = 5
Step 5: Find (1-p)n-r.
= 0.55 = 0.03125
Step 6: Solve P(X = r) = nCr p r (1-p)n-r
= 792 × 0.0078125 × 0.03125
= 0.193359375
The probability of getting exactly 7 heads is 0.19
Mean and variance
If X ~ B(n, p) (that is, X is a binomially distributed random
variable), then the expected value of X is
and the variance is
This fact is easily proven as follows. Suppose first that we have a
single Bernoulli trial. There are two possible outcomes: 1 and 0,
the first occurring with probability p and the second having
probability 1 − p. The expected value in this trial will be equal to μ
= 1 · p + 0 · (1−p) = p. The variance in this trial is calculated
similarly: σ2 = (1−p)2·p + (0−p)2·(1−p) = p(1 − p).
The generic binomial distribution is a sum of n independent
Bernoulli trials. The mean and the variance of such distributions
are equal to the sums of means and variances of each individual
trial:
Mode and median
Usually the mode of a binomial B(n, p) distribution is equal to
⌊(n + 1)p⌋, where ⌊ ⌋ is the floor function. However when
(n + 1)p is an integer and p is neither 0 nor 1, then the distribution
has two modes: (n + 1)p and (n + 1)p − 1. When p is equal to 0 or
1, the mode will be 0 and n correspondingly. These cases can be
summarized as follows:
In general, there is no single formula to find the median for a
binomial distribution, and it may even be non-unique. However
several special results have been established:

If np is an integer, then the mean, median, and mode
coincide.
Any median m must lie within the interval ⌊np⌋ ≤ m ≤ ⌈np⌉.
POISSON DISTRIBUTION
In probability theory and statistics, the Poisson
distribution is a discrete probability distribution
that expresses the probability of a number of events
occurring in a fixed period of time if these events
occur with a known average rate and independently
of the time since the last event. The Poisson
distribution can also be used for the number of
events in other specified intervals such as distance,
area or volume.
The distribution was discovered by Siméon-Denis
Poisson (1781–1840) and published, together with
his probability theory, in 1838 in his work
Recherches sur la probabilité des jugements en
matières criminelles et matière civile ("Research on
the Probability of Judgments in Criminal and Civil
Matters"). The work focused on certain random
variables N that count, among other things, a
number of discrete occurrences (sometimes called
"arrivals") that take place during a time-interval of
given length. If the expected number of occurrences
in this interval is λ, then the probability that there
are exactly k occurrences (k being a non-negative
integer, k = 0, 1, 2, ...) is equal to
where

e is the base of the natural logarithm (e =
2.71828...)


k is the number of occurrences of an event - the
probability of which is given by the function
k! is the factorial of k
λ is a positive real number, equal to the expected
number of occurrences that occur during the given
interval. For instance, if the events occur on average
4 times per minute, and you are interested in the
number of events occurring in a 10 minute interval,
you would use as model a Poisson distribution with

λ = 10*4 = 40.
As a function of k, this is the probability mass
function. The Poisson distribution can be derived as a
limiting case of the binomial distribution.
The Poisson distribution can be applied to systems
with a large number of possible events, each of
which is rare. A classic example is the nuclear decay
of atoms.
The Poisson distribution is sometimes called a Poissonian,
analogous to the term Gaussian
The parameter λ is not only the mean number of
occurrences , but also its variance . Thus, the
number of observed occurrences fluctuates about its
mean λ with a standard deviation
Datat sometimes arise as the number of occurrences ( ) of an
event per unit time or space, e.g., the number of yeast cells per cm2
on a microscope slide. Under certain conditions (see below), the
random variable is said to follow a Poisson distribution, which,
as a count, is type of discrete distribution. Occurrences are
sometimes called arrivals when they take place in a fixed time
interval.
The Poisson distribution was discovered in 1838 by Simeon-Denis
Poisson as an approximation to the binomial distribution, when the
probability of success is small and the number of trials is large.
The Poisson distribution is called the law of small numbers
because Poisson events occur rarely even though there are many
opportunities for these evens to occur.
Poisson Experiment
The number of occurrences of an event per unit time or space will
have a Poisson distribution if:



the rate of occurrence is constant over time or space;
past occurrences do not influence the likelihood of future
occurrences;
simultaneous occurrences are nearly impossible.
Poisson Distribution
In the binomial distribution the probability P(r) of r
successes is given by P(r)=
e  r
x!
. This is known as Poisson
distribution.
Put r=0,1,2,……etc., the probabilities of 0,1,2,…..successes
 2 
 r 
are given by e ,  e , e ,............ e
2
r!


# The sum of these probabilities is unity.
A random variable X is said to follow a Poisson distribution if
it assumes only nonnegative values and its probability is given

 
e
by p(x,  ) =P(X=x)=
x!
x
;x=0,1,2,…….
O , otherwise.
^ Poisson Moments
The mean of a Poisson random variable is , i.e.,
. The
mean is a rate, e.g., a temporal rate for time events. For example, if
0.5 phone calls per hour are received on a home phone during the
day, then the mean number of phone calls between 9 A.M. and 5
P.M. is
.
The Poisson has the interesting property that the variance is also ,
i.e.,
. Thus, unlike the normal distribution, the
variance of a Poisson random variable depends on the mean.
The standard deviation is given by   
The mode for this distribution is(1)  1and if  is an intezer.
(2)integral part of  if  is not
an intezer.
Poisson Approximation to the Binomial
Consider a binomial distribution consisting of trials with
probability of success
, i.e,.
for
. If is sufficiently
large, then the binomial probability:
,
i.e., the binomial probability is approximately equal to the
corresponding Poisson probability. Note that
, the mean of
the binomial distribution.
Thus, binomial probabilities, which are hard to compute for large
, can be approximated by corresponding Poisson probabilies. For
example, suppose 10,000 soldiers are screened for a rare blood
disease (
). We want the probability at least 10 soldiers test
positive for the disease, i.e, we want
, where
. This is difficult to compute using the
binomial distribution, but much easier for the Poisson with
.
Related distributions.


If and are independent, and Y = X1 + X2, then
the distribution of X1 conditional on Y = y is a
binomial. Specifically, . More generally, if X1,
X2,..., Xn are Poisson random variables with
parameters λ1, λ2,..., λn then
 The Poisson distribution can be derived as a
limiting case to the binomial distribution as the
number of trials goes to infinity and the
expected number of successes remains fixed.
Therefore it can be used as an approximation of
the binomial distribution if n is sufficiently large
and p is sufficiently small. There is a rule of
thumb stating that the Poisson distribution is a
good approximation of the binomial distribution
if n is at least 20 and p is smaller than or equal
to 0.05. According to this rule the approximation
is excellent if n ≥ 100 and np ≤ 10.For
sufficiently large values of λ, (say λ>1000), the
normal distribution with mean λ, and variance λ,
is an excellent approximation to the Poisson
distribution. If λ is greater than about 10, then
the normal distribution is a good approximation
if an appropriate continuity correction is
performed, i.e., P(X ≤ x), where (lower-case) x
is a non-negative integer, is replaced by
P(X ≤ x + 0.5).
If the number of arrivals in a given time follows
the Poisson distribution, with mean = λ, then
the lengths of the inter-arrival times follow the
Exponential distribution, with rate 1 / λ.
Occurrence
The Poisson distribution arises in connection with
Poisson processes. It applies to various phenomena
of discrete nature (that is, those that may happen 0,
1, 2, 3, ... times during a given period of time or in a
given area) whenever the probability of the
phenomenon happening is constant in time or space.
Examples of events that may be modelled as a
Poisson distribution include:
The number of soldiers killed by horse-kicks
each year in each corps in the Prussian cavalry.
This example was made famous by a book of
Ladislaus Josephovich Bortkiewicz (1868–1931).
 The number of phone calls at a call center per
minute.
 The number of times a web server is accessed
per minute.
 The number of mutations in a given stretch of
DNA after a certain amount of radiation.

[Note: the intervals between successive Poisson
events are reciprocally-related, following the
Exponential distribution. For example, the lifetime of
a lightbulb, or waiting time between buses.]
The law of rare events
In several of the above examples—for example, the
number of mutations in a given sequence of DNA—
the events being counted are actually the
outcomes of discrete trials, and would more
precisely be modelled using the binomial
distribution. However, the binomial distribution
with parameters n and λ/n, i.e., the probability
distribution of the number of successes in n
trials, with probability λ/n of success on each
trial, approaches the Poisson distribution with
expected value λ as n approaches infinity. This
provides a means by which to approximate
random variables using the Poisson distribution
rather than the more-cumbersome binomial
distribution.
This limit is sometimes known as the law of
rare events, since each of the individual
Bernoulli events each rarely trigger. The name
may be misleading because the total count of
success events in a Poisson process need not
be rare if the parameter λ is not small. For
example, the number of telephone calls to a
busy switchboard in one hour follows a Poisson
distribution with the events appearing frequent
to the operator, but they are rare from the
point of the average member of the population
who is very unlikely to make a call to that
switchboard in that hour.
The "law of small numbers"

The word law is sometimes used as a synonym
of probability distribution, and convergence in
law means convergence in distribution.
Accordingly, the Poisson distribution is
sometimes called the law of small numbers
because it is the probability distribution of the
number of occurrences of an event that happens
rarely but has very many opportunities to
happen
Examples for rare events:
1. No. of printing mistakes per page.
2. No. of accidents on highway.
3. No. of bad cheques at a bank.
4. No. of blind persons
5. No. of noble prize winners
6. No. of Bharath rattans.
Normal Probability Distribution
Normal distribution:
Normal distribution is one of the most widely used
continuous probability distribution in applications of
statistical methods. It is of tremendous importance in the
analysis and evaluation of every aspect of experimental
date in science and medicine.
Def: Normal distribution is the probability distribution of a
continuous random variable x, known as normal random
variables or normal variate.
It is also called Gaussian distribution.
In this the variate can take all values with in a given
range.Examples are measurements (e.g., weight, time, etc.) can be
well approximated by the normal distribution.
It is another limiting form of binomial distribution for large values
of n when neither p nor q is very small. The normal distribution is
derived from the binomial distribution
By increasing the number of trials indefinitely.
A randon variable is said to be a normal distribution ,if its
probability distribution is given by
f ( x; , ) 
1
e
 2

( x   )2
2 2
  x  ,     ,  0
Where  =Mean and  =Standard deviation are two parameters
of the normal distribution.
Mean of Normal distribution
The mean of the normal distribution with b,  as parameters is b.
Variance of normal distribution is  2
The standard deviation of the normal distribution is

Mode
Mode is the value of x for which f(x) is maximum.i.e;the mode
is the solution of f(x)=0 and f  ( x)  0
f ( x)  
x
. f ( x)

f ( x)  
f ( x)  0  x    0  x  
1
0
2
 2
Hence x=  is the mode.
Median
M
We know that

1
2

f ( x)dx 

Using this we have Median =
Hence for the Normal Distribution Mean = Mode = Median.
Hence the distribution is symmetrical.
Chief characteristics:
1. The graph of the normal distribution y=f(x) in x-y plane
is known as the normal curve.
2. The curve is bell shaped and symmetrical about the line
x= 
3. Area under the normal curve is unity i.e., it represents
total population.
4. Mean = median =mode.
5.x- axis is an asymptote to the curve and the points of
inflexion of the curve at x=   
7. Since mean =  , the line x =  -- divides the total area
into two equal parts.
8. No portion of the curve lies below x – axis.
9.The prob. that the normal vaiate x with mean  – and
c.d. between    and    is 68.26%
10.Area of the normal curve between   2 and   2
is95.44%
11. Area of the normal curve between   3 and   3
is99.73%
Def:
The normal distribution for   0 AND  1
known as STANDARD NORMAL DISTRIBUTION.
is
Importance and applications of normal distribution:
Normal distribution plays a very important role in statistical
theory because of the following reasons
1. Most of the distributions for example, binomial, Poisson
etc can be approximated by Normal distribution..
2. Since it is a limiting case of Binomial distribution for
exceptionally large numbers, it is applicable to many
applied problems in kinetic theory of gases and fluctuations
in the magnitude of an electric current.
3. It is variable is normally distributed, it can sometimes be
brought to normal form by simple transformation of the variable.
4. The proofs of all the tests of significance in sampling are based
on the fundamental assumption that the population from which the
samples have been drawn are normal.
5. Normal distribution finds large applications in statistical quality
Control.
6. Many of the distributions of sample statistic, the distributions of
sample mean, sample variance etc tend to normality for large
samples and as such they can best be studied with the help of
normal curve.
Area property:
x
By taking z 

the standard normal curve is formed.
The probability that the normal variate x with mean – and s.d.
lies, between two specific values x1 and x2 with x1  x2 , can be
obtained using area under the standard normal curve as follows.
1.Put
2.To find
(i) If
x
z
, forx1 , x2 , findz1 , z2

P( x1  x  x2 )  P( z1  z  z2 )
z1 , z2
are positive or negative then
P( z1  z  z2 )  A( z2 )  A(z1)
(ii)If
z1  oand , z2  o
P( x1  x  x2 )
 A( z2 )  A( z1 )
P( z  z1 )
if , z1  0, then
P( z  z1 )  0.5  A( z1 )
if , z1  0, then
P( z  z1 )  0.5  A( z1 )
then
(4)
To find
P( z  z1 )
if , z1  0, then
P( z  z1 )  0.5  A( z1 )
if , z1  0, then
P( z  z1 )  0.5  A( z1 )
Def: A normal random variable with mean   0 and variance  2  1 is called standard normal variable. Its probability density
function is given by
1 z2/2
 ( s) 
e   z  
2
Def: The cumulative distribution fn. of a standard normal random
variable is
1
F ( z )  p( Z  z ) 
2

 z /2
 e dz
2

Normal Distribution as Approximation to Binomial
Distribution
Normal Distribution provides a valuable approximation to
Binomial when the sample sizes are large and the probability of
successes and failures is not close to zero.
Poisson Approximation to Binomial Distribution
Poisson provides an approximation to Binomial Distribution when
the sample sizes are large and the probability of successes or
failures is close to zero.
Normal Approximation to Poisson
The Poisson can be approximated fairly well by Normal
Distribution when λ is large.
Normal approximation is binomial distribution:
When n is very large, it is very difficult to calculate the
probabilities by using binominal distribution. Normal distribution
is a limiting case of binomial distribution under the following
conditions.
(1) N, the no. of trials very large n --(2) neither p nor q is very small.
For a b.d. E(x) = np var(x) = npq.
Then the standard normal variate
Z
x u


x  np
npq
tends to the distribution of standard normal variable
given by
z
1
( z ) 
e 2   z  
2
2
If p  q , and for large n, we can approximate binomial curve by
normal curve. Here the interval becomes (x-1/2,x+1/2)
Note: If X is a poisson variable with mean  then the standard
normal variable Z= x   and the probability can be calculated as

explained above
Poisson distribution approaches the normal distribution as   
A uniform distribution, sometimes also known as a
rectangular distribution, is a distribution that has constant
probability.
The probability density function and cumulative distribution
function for a continuous uniform distribution on the interval
are
(1)
(2)
These can be written in terms of the Heaviside step function
as
(3)
(4)
the latter of which simplifies to the expected
.
for
The continuous distribution is implemented as
UniformDistribution[a, b].
For a continuous uniform distribution, the characteristic
function is
(5)
If
and
, the characteristic function simplifies to
(6)
(7)
The moment-generating function is
(8)
(9)
(10)
and
(11)
(12)
The moment-generating function is not differentiable at zero,
but the moments can be calculated by differentiating and
then taking
. The raw moments are given analytically by
(13)
(14)
(15)
The first few are therefore given explicitly by
(16)
(17)
(18)
(19)
The central moments are given analytically by
(20)
(21)
(22)
The first few are therefore given explicitly by
(23)
(24)
(25)
(26)
The mean,
variance,
skewness, and
kurtosis excess
are therefore
Exponential
distribution
Inion (a.k.a. negative
exponential
distribution) is a
family of continuous
probability
distributions. It
describes the tim
between events in a
Poisson process, i.e.
a process in which
events occur
continuously and
independently at a
constant average rate.
Note that the exponential distribution is not the same as the class of
exponential families of distributions, which is a large class of
probability distributions that includes the exponential distribution
as one of its members, but also includes the normal distribution,
binomial distribution, Poisson, and many others.

Characterization
Probability density function
The probability density function (pdf) of an exponential
distribution is
Here λ > 0 is the parameter of the distribution, often called the rate
parameter. The distribution is supported on the interval [0, ∞). If a
random variable X has this distribution, we write X ~ Exp(λ).
A commonly used alternative parameterization is to define the
probability density function (pdf) of an exponential distribution as
where β > 0 is a scale parameter of the distribution and is the
reciprocal of the rate parameter, λ, defined above. In this
specification, β is a survival parameter in the sense that if a
random variable X is the duration of time that a given biological or
mechanical system manages to survive and X ~ Exponential(β)
then E[X] = β. That is to say, the expected duration of survival of
the system is β units of time. The parameterisation involving the
"rate" parameter arises in the context of events arriving at a rate λ,
when the time between events (which might be modelled using an
exponential distribution) has a mean of β = λ−1.
The alternative specification is sometimes more convenient than
the one given above, and some authors will use it as a standard
definition. This alternative specification is not used here.
Unfortunately this gives rise to a notational ambiguity. In general,
the reader must check which of these two specifications is being
used if an author writes "X ~ Exponential(λ)", since either the
notation in the previous (using λ) or the notation in this section
(here, using β to avoid confusion) could be intended.
Occurrence and applications
The exponential distribution occurs naturally when describing the
lengths of the inter-arrival times in a homogeneous Poisson
process.
The exponential distribution may be viewed as a continuous
counterpart of the geometric distribution, which describes the
number of Bernoulli trials necessary for a discrete process to
change state. In contrast, the exponential distribution describes the
time for a continuous process to change state.
In real-world scenarios, the assumption of a constant rate (or
probability per unit time) is rarely satisfied. For example, the rate
of incoming phone calls differs according to the time of day. But if
we focus on a time interval during which the rate is roughly
constant, such as from 2 to 4 p.m. during work days, the
exponential distribution can be used as a good approximate model
for the time until the next phone call arrives. Similar caveats apply
to the following examples which yield approximately
exponentially distributed variables:



The time until a radioactive particle decays, or the time
between clicks of a geiger counter
The time it takes before your next telephone call
The time until default (on payment to company debt holders)
in reduced form credit risk modeling
Exponential variables can also be used to model situations where
certain events occur with a constant probability per unit length,
such as the distance between mutations on a DNA strand, or
between roadkills on a given road.
Properties
Mean, variance, and median
The mean or expected value of an exponentially distributed
random variable X with rate parameter λ is given by
In light of the examples given above, this makes sense: if you
receive phone calls at an average rate of 2 per hour, then you can
expect to wait half an hour for every call.
The variance of X is given by
The median of X is given by
where ln refers to the natural logarithm. Thus the absolute
difference between the mean and median is.
Independent events
The standard definition says:
Two events A and B are independent if and only if Pr(A ∩ B)
= Pr(A)Pr(B).
Here A ∩ B is the intersection of A and B, that is, it is the event that
both events A and B occur.
More generally, any collection of events—possibly more than just
two of them—are mutually independent if and only if for every
finite subset A1, ..., An of the collection we have
variables X and Y are independent if and only if for every a and b,
the events {X ≤ a} and {Y ≤ b} are independent events as defined
above. Mathematically, this can be described as follows:
The random variables X and Y with distribution functions FX(x)
and FY(y), and probability densities ƒX(x) and ƒY(y), are
independent if and only if the combined random variable (X, Y) has
a joint cumulative distribution function
or equivalently, a joint density
.
Similar expressions characterise independence more generally for
more than two random variables.
An arbitrary collection of random variables – possibly more than
just two of them — is independent precisely if for any finite
collection X1, ..., Xn and any finite set of numbers a1, ..., an, the
events {X1 ≤ a1}, ..., {Xn ≤ an} are independent events as defined
above.
If any two of a collection of random variables are independent,
they may nonetheless fail to be mutually independent; this is called
pairwise independence.
If X and Y are independent, then the expectation operator E has the
property
and for the variance we have
so the covariance cov(X, Y) is zero. (The converse of these, i.e. the
proposition that if two random variables have a covariance of 0
they must be independent, is not true. See uncorrelated.)
Two independent random variables X and Y have the property that
the characteristic function of their sum is the product of their
marginal characteristic functions:
but the reverse implication is not trues
Conditionally independent random variables
Conditional independence
Intuitively, two random variables X and Y are conditionally
independent given Z if, once Z is known, the value of Y does not
add any additional information about X. For instance, two
measurements X and Y of the same underlying quantity Z are not
independent, but they are conditionally independent given Z
(unless the errors in the two measurements are somehow
connected).
The formal definition of conditional independence is based on the
idea of conditional distributions. If X, Y, and Z are discrete random
variables, then we define X and Y to be conditionally independent
given Z if
for all x, y and z such that P(Z = z) > 0. On the other hand, if the
random variables are continuous and have a joint probability
density function p, then X and Y are conditionally independent
given Z if
for all real numbers x, y and z such that pZ(z) > 0.
If X and Y are conditionally independent given Z, then
for any x, y and z with P(Z = z) > 0. That is, the conditional
distribution for X given Y and Z is the same as that given Z alone.
A similar equation holds for the conditional probability density
functions in the continuous case.
Independence can be seen as a special kind of conditional
independence, since probability can be seen as a kind of
conditional probability given no events.