Survey
* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project
Probabilistic Graphical Models Inference Overview Conditional Probability Queries Daphne Koller Inference in a PGM • PGM encodes P(X1,…,Xn) • Can be used to answer any query over joint distribution Daphne Koller Conditional Probability Queries • Evidence: E = e • Query: a subset of variables Y • Task: compute P(Y | E=e) • Applications – Medical/fault diagnosis – Pedigree analysis Daphne Koller NP-Hardness The following are all NP-hard • Given a PGM P, a variable X and a value xVal(X), compute P(X=x) – Or even decide if P(X=x) > 0 • Let < 0.5. Given a PGM P, a variable X and a value xVal(X), and observation eVal(E), find a number p that has |P(X=x|E=e) – p| < Daphne Koller Sum-Product C I D G S L H J Daphne Koller Sum-Product A D B C Daphne Koller Evidence: Reduced Factors A D B C Daphne Koller Evidence: Reduced Factors C P(J,I=i, H=h) = I D G S L H J Daphne Koller Sum-Product Compute and renormalize Daphne Koller Algorithms: Conditional Probability • Push summations into factor product – Variable elimination • Message passing over a graph – Belief propagation – Variational approximations • Random sampling instantiations – Markov chain Monte Carlo (MCMC) – Importance sampling Daphne Koller Summary • Conditional probability queries of subset of variables given evidence on others • Summing over factor product • Evidence simply reduces the factors • Many exact and approximate algorithms Daphne Koller END END END Daphne Koller