Download In addition to the many formal applications of probability theory, the

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project

Document related concepts

Law of large numbers wikipedia, lookup

Central limit theorem wikipedia, lookup

History of statistics wikipedia, lookup

Inductive probability wikipedia, lookup

Birthday problem wikipedia, lookup

Foundations of statistics wikipedia, lookup

Infinite monkey theorem wikipedia, lookup

Expected value wikipedia, lookup

Negative binomial distribution wikipedia, lookup

Transcript
2 Interpretations of Probability
In addition to the many formal applications of probability theory, the concept of probability
enters our everyday life and conversations.
 It probably will rain tomorrow.
 It is very likely that the bus will arrive late.
 The chances are good he will win the game.
Despite the fact that the concept of probability is such a common and natural part of our
experience, no single scientific interpretation of the term probability is accepted by all
statisticians.
Frequency interpretation: In many problems, the probability that some specific outcome of a
process will be obtained can be interpreted to mean the relative frequency with which that
outcome would be obtained if the process were repeated a large number of times under
similar conditions.
Example: the probability of obtaining a head when a coin is tossed is approximately 1/2.
Difficulties: Large number, similar conditions.
The frequency interpretation applies only to a problem in which there can be, at least in
principle, a large number of similar repetitions of a certain process. Many important
problems are not of this type. For example, the probability that a particular medical
research project will lead to the development of a new treatment for a certain disease
within a specified period of time.
Remark: The mathematical theory of probability can be usefully applied, regardless of which
interpretation of probability is used in a particular problem. The theories and techniques that
will be presented in this book have served as valuable guides and tools in almost all
aspects of the design and analysis of effective experimentation.
2 Experiments and Events
The term “experiment” is used in probability theory to describe virtually every process
whose outcome is not known in advance with certainty. Here are some examples.
(1) In an experiment in which a coin is to be tossed 10 times, the experimenter might
want to determine the probability that at least four heads will be obtained.
(2) In an experiment in which the air temperature at a certain location is to be observed
every day at noon for 90 successive days, a person might want to determine the
probability that the average temperature during this period will be less than some specified
value.
(3) In evaluating an industrial research and development project at a certain time, a
person might want to determine the probability that the project will result in the successful
development of a new product within a specified number of months.
The interesting feature of an experiment is that each of its possible outcomes can be
specified before the experiment is performed, and probabilities can be assigned to various
combinations of outcomes that are of interest.
3 Set Theory
3.1 The sample space
The collection of all possible outcomes of an experiment is called the sample space of the
experiment. Each possible outcome is called a point, or an element, of the sample space.
The sample space can be classified into two types: countable and uncountable. A sample
space is countable if there is a one-to-one correspondence between the elements of A
and the set of natural numbers {1, 2, 3, . . .}. Otherwise, it is uncountable.
3.2 Relations of set theory
Let S denote the sample space of some experiment. Then each possible outcome s of
the experiment is said to be a member of the space S, or to belong to the space S.
The statement that s is a member of S is denoted symbolically by the relation s 2 S.
An event C is a certain subset of possible outcomes in the space S. We say the event
C occurs if the outcome of the experiment is in the set C.
Example 3.1 A six-sided die is rolled. The sample space is S
Let A = {2,
4, 6} denote the event A that an even number is obtained, and B = {4, 5, 6} denote the
event that a number greater than 3 is obtained. We say A occurs if we get the number
2 in a roll.
= {1, 2, 3, 4, 5, 6}.
It is said that an event A is contained in another event B i f every outcome that belongs
to the subset defining the event A also belongs to the subset defining the event B. The
relation is denoted by A ½ B. Mathematically, we have
A ½ B , x 2 A ) x 2 B
(containment)
If two events are so related that A ½ B a nd B ½ A , then A = B, which means
that A and B
m ust contain exactly the same sample points. Mathematically, we have
A = B , A ½ B and B ½ A (equality)
A subset of S that contains no outcomes is called the empty set, or null set, and it is
denoted by the symbol ;. For each event, it is true that ; ½ A ½ S .
3.3 Operations of set theory
Given any two events (or sets) A a nd B, we have the following elementary set
operations:
Union: The union of A a nd B, written A [ B , is the set of elements that belong to
either A or B or both:
A [ B = {x : x 2 A or x 2 B}
Intersection: The intersection of A and B, written A \ B, is the set of elements that
belong to both A and B:
A \ B = {x : x 2 A and x 2 B}
Complements: The complement of A, written A c , is the set of all elements that are not in A:
Ac = {x : x  A}
Disjoint events
It is said that two events A and B are disjoint, or mutually exclusive, if
no outcomes in common. It follows that A and B are disjoint if and only
The events A1, . . . ,An or the events A1,A2, . . . are disjoint if for every i
that Ai and Aj are disjoint, that is, Ai \ Aj = ; for all i 6= j.
Example 3.2 (Event operations) Consider the experiment of selecting a
A and B have
if A \ B = ; .
6= j, we have
card at random
from a standard deck and noting its suit: clubs (C), diamond (D), hearts (H), or
spades (S). The sample space is
and some possible events are
S = {C,D,H, S},
A = {C,D}, and B = {D,H, S}.
From these events we can form
A [ B = {C,D,H, S}, A \ B = {D}, and Ac = {H, S}.
Furthermore, notice that A [ B = S and ( A [ B ) c = ;.
4 The Definition of Probability
4.1 Axioms and Basic Theorems
In order to satisfy the mathematical definition of probability, P(A), the probability assigned
to event A, must satisfy three specific axioms.
Axiom 1 For every event A, P(A) ¸ 0. In words, the probability of every event must be
nonnegative.
Axiom 2 P(S) = 1. If an event is certain to occur, then the probability of that event is
1.
Axiom 3 For every infinite sequence of disjoint events A1,A2, . . .,
The mathematical definition of probability can now be given as follows: A probability
distribution, or simply a probability, on a sample space S is a specification of numbers
P(A) that satisfy Axioms 1, 2, and 3.
Some Theorems
Theorem 4.1 P(;) = 0.
Theorem 4.2 For every finite sequence of n disjoint events A1, . . . ,An,
Theorem 4.3 For every event A, P(Ac) = 1 − P (A).
Theorem 4.4 If A ½ B, then P(A) ≤ P (B).
Theorem 4.5 For any event A, 0 ≤ P (A) ≤ 1 .
Theorem 4.6 For every two events A and B,
P(A [ B) = P(A) + P(B) − P(AB).
5 Finite sample spaces
5.1 Requirements of probabilities
In an experiment of which the sample space contains only a finite number of points s1, . .
. , s n, a probability distribution on S is specified by assigning a probability pi to each point
si 2 S. In order to satisfy the axioms for a probability distribution, the numbers p1, . . . , pn
must satisfy the following two conditions:
pi ¸ 0 , for i = 1, . . . , n
and
The probability of each even A c an then be found by adding the probabilities pi of all
outcomes si that belong to A.
Example 5.1 Consider an experiment in which five fibers having different lengths are
subject to a testing process to learn which fiber will break first. Suppose that the lengths
of the five fibers are 1,2,3,4, and 5 inches, respectively. Suppose also that the probability
that any given fiber will be the first to break is proportional to the length of that fiber.
A sample space S c ontaining n o utcomes s1, . . . , sn is called a simple sample space if
the probability assigned to each of the outcomes s1, . . . , sn is 1/n. If an event A i n this
simple sample space contains exactly m o utcomes, then
5 Finite sample spaces
Example 5.2 (Tossing coins) Suppose that three fair coins are tossed simultaneously. We
shall determine the probability of obtaining exactly two heads.
6 Counting Methods
In many experiments, the number of outcomes in S i s so large that a complete listing of
these outcomes is too expensive, too slow, or too likely to be incorrect to be useful. In
such an experiment, it is convenient to have a method of determining the total number of
outcomes in the space S a nd in various events in S w ithout compiling a list of all these
outcomes.
Multiplication rule
For example, suppose that there are three different routes from city A to city B and five
different routes from city B t o city C. Then the number of different routes from A to C
that pass through B is 3 × 5 = 15.
6.2 Permutations
Suppose that k c ards are to be selected one at a time and removed from a deck of n
c ards (k = 1, 2, . . . , n). Then each possible outcome of this experiment will be a
permutation of k cards from the deck, and the total number of these permutations will be
Pn,k = n(n − 1) · · · (n − k + 1). This number Pn,k is called the number of permutations of n
e lements taken k a t a time.
If k = n, Pn , n = n(n − 1) · · · 1 = n!. Note that 0! = 1.
Example 6.1 (Arranging books) Suppose that six different books are to be arranged on a
shelf. The number of possible permutations of the books is 6! = 720.
Example 6.2 (Ordered sampling without replacement) Suppose that a club consists of 25
members, and that a manager and a secretary are to be chosen from the membership.
The number of all possible choices is
P25,2 = 25 × 24 = 600.
Example 6.3 (Ordered sampling with replacement) Consider the following experiment: A
box contains n balls numbered
1, . . . , n.
First, one ball is selected at random from the box
and its number is noted. This ball is then put back in the box and another ball is
selected. As many balls can be selected in this way. This process is called ordered
sampling with replacement. The sample space of this experiment will contain all vectors of
the form (x1, . . . , xk), where xi is the outcome of the ith selection. The total number of
vectors in S is nk. The probability of the event that each of the k balls that are selected
will have a different number is
Example 6.4
(The birthday problem) The problem is to determine the probability p that at least two
people in a group of k people (2 ≤ k ≤ 365) will have the same birthday. Under
appropriate assumptions (independence and uniformity),the probability that all k persons will
have different birthdays is P365,k / 3 65k. The probability that at least two of the people will
have the same birthday is
7 Combinatorial Methods
7.1 Combinations
Suppose that there is a set of n d istinct elements from which it is desired to choose a
subset containing k e lements (1 ≤ k ≤ n ). The number of different subsets that can be
chosen is
The number Cn,k is also denoted by the symbol
. When this notation is used, this
number is called a binomial coefficient because it appears in the binomial theorem.
Binomial theorem. For all numbers x and y and each positive integer n,
Examples
Example 7.1 (Unordered sampling without replacement) Suppose that a class contains 15
boys and 30 girls, and that 10 students are to be selected for a special assignment. We
shall determine the probability that exactly 3 boys will be selected.
The desired probability is
In summary, we have the following table for counting.
Example 7.3 (Unordered sampling with replacement) A restaurant has n items on its
menu. During a particular day, k customers will arrive and each one will choose one item.
The manger wants to count how many different collections of customer choices are possible
without regard to the order in which the choices are made.
Conditional Probability
A major use of probability in statistical inference is the updating of probabilities when
certain events are observed. The updated probability of event A after we learned that
event B has occurred is the conditional probability of A given B. The notation for this
conditional probability is P(A | B).
Example 1.1 Rolling Dice Suppose that two dices were rolled and it was observed that the
sum T of the two numbers was odd. We shall determine the probability that T was less
than 8.
1.1 The multiplication rule for conditional probability
The definition of conditional probability implies the following multiplication rule,
P(AB) = P(B)P(A | B)
Or
P(AB) = P(A)P(B | A).
Example 1.3 Suppose that two balls are to be selected at random, without replacement,
from a box containing r red balls and b blue balls. The probability that the first ball will
be red and the second ball will be blue is
2 Independent Events
Definition 2.1 Two events are independent if P(AB) = P(A)P(B).
Suppose that P(A) > 0 and P(B) > 0, then it follows from the definitions of
independence and conditional probability that A and B are independent if and only if:
P(A | B) = P(A) and
P(B | A) = P(B),
that is, learning that B (A) occurred does not change the probability of A (B).
Example 2.1 Machine operations. Suppose that two machines 1 and 2 in a factory are
operated independently of each other. Let A be the event that machine 1 will become
inoperative during a given 8-hour period; let B be the event that machine 2 will become
inoperative during the same period; and suppose that P(A) = 1/3 and P(B) = 1/4. What is
the probability that both machines will become inoperative during the period.
Theorem 2.1 If two events A and B are independent, then the events A and B
independent.
c
are also
Definition 2.2 The k events A1, . . . , Ak are independent if, for every subset
these events (j = 2, 3, . . . , k), P(Ai1 \ · · · \ Aij ) = P(Ai1 ) · · · P(Aij ).
In particular, in order for three events A, B, and C
relations must be satisfied:
Ai1 , . . . , Aij
of j of
to be independent, the following four
P(AB) = P(A)P(B)
P(AC) = P(A)P(C)
P(BC) = P(B)P(C)
P(ABC) = P(A)P(B)P(C).
Example 2.3
Suppose that a machine produces a defective item with probability p (0 < p < 1) and
produces a nondefective item with probability q = 1 - p. Suppose further that six items
produced by the machine are selected at random and inspected, and that the outcomes for
these six items are independent. The probability that at least one of the six items will be
defective is
p = 1 - P(D1D2D3D4D5D6) = 1 - q 6 ,
The probability that exactly two of six items are defective is:
Example 2.5 Suppose that the probability p that the machine produces a defective item
takes one of two values, either 0.01 or 0.4, the first corresponding to normal operation
and the sec-ond corresponding to a need for maintenance. Let B1 be the event that p =
0.01, and let B2 be the event that p = 0.4. If we knew that B1 had occurred, then we would
proceed under the assumption that the events D1, D2, . . . were independent with P(Di | B1) = 0.01
for all i. Let A be the event that we observe exactly two defectives in a random sample
of six items. Then
1 Random variables and discrete distributions
Definition 1.1 Consider an experiment for which the sample space is defined by S. A
random variable is a function from a sample space into the real numbers. In other words,
in a particular experiment a random variable X would be some function that assigns a real
number X(s) to each possible outcome s 2 S.
Example 1.1 Consider an experiment in which a coin is tossed 10 times. In this
experiment the sample space can be regarded as the set of outcomes consisting of the
210 different sequences of 10 heads and tails that are possible, and the random variable X
could be the number of heads obtained on the 10 tosses. For each possible sequence s
consisting of 10 heads and tails, this random variable would assign a number X(s) equal
to the number of heads in the sequence. Thus, if s is the sequence HHTTTHTTTH, then
X(s) = 4.
The distribution of a random variable
When a probability distribution has been specified on the sample space of an experiment,
we can determine a probability distribution for the possible values of each random variable
X. Let A be a subset of the real line, and let P (X 2 A) denote the probability that the
value of X will belong to the subset A. Then P (X 2 A) is equal to the probability that
the outcome s of the experiment will be such that X(s) 2 A. In symbols,
P(X 2 A) = P {s : X (s) 2 A}.
Example 1.3
Consider again an experiment in which a coin is tossed 10 times, and let X be the
number of heads that are obtained. In this experiment the possible values of X are 0, 1, 2,
. . ., 10.
Then:
for x =
0, 1, . . . , 10.
Discrete distributions (Discrete
Random Variable)
It is said that a random variable X has a discrete distribution or that X is a discrete
random variable if X can take only a finite number k of different values x1, . . . , xk or, at
most, an infinite sequence of different values x1, x2, . . ..
Continuous distributions (Continuous
Random Variable)
Random variables that can take every value in an interval are called continuous random
variables.
Probability Distribution Function (abbreviated p.d.f.)
If a random variable X has a discrete distribution, the probability distribution function
(abbreviated p.f.) of X is defined as the function f such that for every real number x,
f (x) = P (X = x). For every point x that is not one of the possible values of X, it is
evident that f (x) = 0. Also, if the sequence x1, x2, . . . includes all the possible values of
X, then
The uniform distribution on integers
Suppose that the value of a random variable X is equally likely to be each of the k
integers 1, 2, . . . , k. Then the p.d.f. of X is as follows:
The Binomial distribution
Suppose that a certain machine produces a defective item with a probability of p (0 < p
< 1) and produces a nondefective item with a probability of q = 1 − p. Suppose further
that n independent items produced by the machine are examined, and let X denote the
number of these items that are defective. Then the random variable X will have a discrete
distribution, and the possible values of X will be 0,1, . . ., n. The probability of X = x is
2 Continuous Distributions
The probability density function It is said that a random variable X has a continuous
distribution or that X is a continuous random variable if there exists a nonnegative function
f, defined on the real line, such that for every subset A of the real line, the probability
that X takes a value in A is the integral of f over the set A. In this course, we shall
be concerned primarily with subsets that are intervals. For an interval (a, b],
The function f is called the probability density function of X. Every p.d.f. f must satisfy
the following two requirements:
f (x) ¸ 0, for all
x,
and
The uniform distribution on an interval
Let a and b be two given real numbers such that a < b, and consider an
which a point X is selected from the interval S = {x : a · x · b} in such a
probability that X will belong to each subinterval of S is proportional to the
subinterval. This distribution of the random variable X is called the uniform
the interval [a, b].
experiment in
way that the
length of that
distribution on
Here X represents the outcome of an experiment that is often described by saying that a
point is chosen at random from the interval [a, b].
For example, if a random variable X has a uniform distribution on the interval
the pdf of X is
Example 2.2
[−1, 4],
then
Suppose that the pdf of a certain random variable X has the following form
3 The Distribution Function
Definition and Basic Properties The cumulative distribution function (cdf) F of a random
variable X is a function defined for each real number x as follows:
F (x) = P (X  x), −1 < x < 1.
It should be emphasized that the cdf is defined in this way for every random variable X,
regardless of whether the distribution of X is discrete, continuous, or mixed. The cdf of
every random variable X must have the following three properties:
Property 3.1 The function F(x) is nondecreasing as x increases; that is, if x1 < x2, then
F(x1)  F(x2).
1 The Expectation of a Random Variable
Expectation for a discrete distribution
Suppose that a random variable X has a discrete distribution for which the pf is f. The
expectation of X, denoted by E(X), is a number defined as follows:
Example 1.1 Suppose that a random variable X can have only the four different values 2, 0, 1, and 4, and that P(X = −2) = 0.1, P(X = 0) = 0.4, P(X = 1) = 0.3, and
P(X = 4) = 0.2. Then
E(X)
=
−2(0.1)
+
0(0.4)
+
1(0.3)
+
4(0.2)
=
0.9.
Expectation for a continuous distribution
If a random variable X has a continuous distribution for which the pdf is f, then the
expectation E(X) is defined as follows:
Example 1.2 Suppose that the pdf of a random variable X is
Interpretation of the expectation
The number E(X) is also called the expected value of X or the mean of X; and the
terms expectation, expected value, and mean can be used interchangeably.
3 Variance
Definitions of the variance and the standard deviation
Suppose that X is a random variable with mean µ = E(X). The variance of X, denoted
by Var(X), is defined as follows:
Var(X) = E [(X − µ)2].
5 The Median
Definition 5.1 For each random variable X, a median of the distribution of X is defined as
a number m such that P (X  m) ¸ 1/2 and P (X ¸ m) ¸ 1/2.
Gaussian pdf
If X represents the sum of N independent random components, and if each component
makes only a small contribution to the sum, then the CDF of X approaches a Gaussian CDF
as N becomes large, regardless of the distribution of the individual components.
Gaussian pdf
Random Processes (Stochastic Processes)
Random variable that changes with time.
Example:. The set of voltages generated by thermal electron motion in a large number of identical resistors.
Ergodicity
A Random Process is Ergodic if all time averages of sample functions equal the corresponding ensemble averages.
All ensemble averages of an Ergodic process are independent of time.