* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download p(E|S)
Survey
Document related concepts
Transcript
Discrete Mathematics and Its Applications Sixth Edition By Kenneth Rosen Chapter 6 Discrete Probability 歐亞書局 6.1 An Introduction to Discrete Probability 6.2 Probability Theory 6.3 Bayes’ Theorem 6.4 Expected Value and Variance 歐亞書局 P. 1 6.1 An Introduction to Discrete Probability • Gambling – Tossing a die • Finite Probability – We will restrict to experiments that have finitely many outcomes • Experiment: a procedure that yields one of a given set of possible outcomes • Sample space: the set of possible outcomes • Event: a subset of the sample space • Definition 1: If S is a finite sample space of equally likely outcomes, and E is an event (a subset of S), then the probability of E is p(E)=|E|/|S|. – – – – – Ex.1-8 Lotteries Poker Sampling without replacement Sampling with replacement • The probability of combinations of events – Theorem 1: Let E be an event in a sample space S. The probability of the event Ē, the complementary event of E, is given by: p(Ē) = 1 – p(E). • Proof • Ex.8 • Theorem 2: Let E1 and E2 be events in the sample space S. Then, p(E1E2)=p(E1)+p(E2)-p(E1E2). – Ex.9 • Probabilistic reasoning – Determining which of two events is more likely – Ex.10 6.2 Probability Thoery • Assigning probabilities – We assign a probability p(s) to each outcome s • 0<=p(s)<=1 for each sS (S: finite or countable number of outcomes) • sSp(s)=1 – For n possible outcomes x1, x2, …, xn • 0<=p(xi)<=1 for i=1,2,…,n • i=1..np(xi)=1 – P: probability distribution • Ex.1 • Definition 1: Suppose that S is a set with n elements. The uniform distribution assigns the probability 1/n to each element of S. • Definition 2: The probability of the event E is the sum of the probabilities of the outcomes in E. That is, p(E)= sEp(s) – Selecting at random – Ex.2 • Combinations of events – The same as Sec.6.1 • p(Ē) = 1 – p(E) • p(E1E2)=p(E1)+p(E2)-p(E1E2) – Theorem 1: If E1, E2, … is a sequence of pairwise disjoint events in a sample space S, then p(iEi)=ip(Ei). Conditional Probability • Definition 3: Let E and F be events with p(F)>0. The conditional probability of E given F, denoted by p(E|F), is defined as p(E|F)=p(EF)/p(F). – Ex.3-4 • Independence – P(E|F)=p(E) • Definition 4: The events E and F are independent if and only if p(EF)=p(E)p(F) – Ex.5-7 Bernoulli Trials and the Binomial Distribution • Bernoulli trial: each performance of an experiment with two possible outcomes – A success, or a failure – Ex.8 • Theorem 2: The probability of exactly k successes in n independent Bernoulli trials, with probability of success p and probability of failure q=1-p, is C(n,k)pkqn-k. – Denoted by b(k;n,p), binomial ditribution – Ex.9 Random Variable • Definition 5: A random variable is a function from the sample space of an experiment to the set of real numbers. – Not a variable, not random – Ex.10 • Definition 6: The distribution of a random variable X on a sample space S is the set of pairs (r, p(X=r)) for all rS, where p(X=r) is the probability that X takes the value r. – Ex. 11-12 The Birthday Problem • The smallest number of people needed in a room so that it’s more likely than not that at lease two of them have the same day of year as their birthday – Similar to hashing functions – Ex.13 (The birthday problem) – Ex.14 (Probability of a collision in hashing functions) Monte Carlo Algorithms • Deterministic vs. probabilistic algorithms • Monte Carlo algorithms – Always produce answers to problems, but a small probability remains that the answers may be incorrect • This probability decreases rapidly when the algorithm carries out sufficient computation • Ex. 15-16 The Probabilistic Method • Theorem 3: (The Probabilistic Method) If the probability that an element of a set S does not have a particular property is less than 1, there exists an element in S with this property. – Nonconstructive existence proof • Theorem 4: If k is an integer with k>=2, then R(k,k)>=2k/2. – Proof 6.3 Bayes’ Theorem • Ex.1 • Theorem 1: (Bayes’ Theorem) Suppose that E and F are events from a sample space S such that p(E)0 and p(F)0. Then p(F|E) = p(E|F)p(F)/(p(E|F)p(F)+p(E|~F)p(~F)). – Proof – Ex.2 • Theorem 2: (Generalized Bayes’ Theorem) Suppose that E is an event from a sample space S and F1, F2, …, Fn are mutually exclusive events such that i=1..nFi=S. Assume that p(E)0 and p(Fi)0 for i=1,2,…,n. Then p(Fj|E) = p(E|Fj)p(Fj)/i=1..np(E|Fi)p(Fi). Bayesian Spam Filters • Bayesian spam filters – – – – B: set of spam messages, G: set of good messages nB(w), nG(w) p(w)=nB(w)/|B|, q(w)= nG(w)/|G| P(S|E) = p(E|S)p(S)/(p(E|S)p(S)+p(E|~S)p(~S)). • Assume p(S)=p(~S)=1/2 • P(S|E)=p(E|S)/(p(E|S)+p(E|~S)) – P(E|S)=p(w), p(E|~S)=q(w) • r(w)=p(w)/(p(w)+q(w)) • Ex.3-4 – r(w1,w2,…,wk)=i=1..kp(wi)/(i=1..kp(wi)+i=1..kq(wi)) Thanks for Your Attention!