* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Download Bayesian Statistical Inference Introductory Toy Example Suppose
Particle filter wikipedia , lookup
Inductive probability wikipedia , lookup
Density of states wikipedia , lookup
Bootstrapping (statistics) wikipedia , lookup
Confidence interval wikipedia , lookup
Misuse of statistics wikipedia , lookup
History of statistics wikipedia , lookup
Bayesian inference wikipedia , lookup
The classical (“frequentist”) approach involves drawing a random sample 35393 Seminar (Statistics) X1, . . . , Xn Whiteboard Lecture where Bayesian Statistical Inference Xi = Introductory Toy Example Suppose that we’re interested in } = probability that randomly chosen person is left-handed. ⇢ 1 if ith person is left-handed 0 otherwise and estimate } via the sample proportion b= } Pn i=1 Xi n . For large n we can get an approximate 95% confidence interval for } via e.g. b ± 1.96 } s b }(1 n b }) . A Bayesian approach treats 0 } 1 as a random variable. The Bayesian analyst then puts a prior distribution on } expressing his/her belief before seeing data. A density function that looks like this (and has mean 0.1) is p(}) = 342}(1 })17, 0 < } < 1, the Beta(2,18) density. Based on my belief about } I will specify a prior that looks like: My model for the data and its dependence on } is ⇢ }, Xi = 1 p(Xi|}) = 1 }, Xi = 0 independently for 1 i n. We can write this neatly as 0 mean=0.1 1 p(Xi|}) = }Xi(1 By independence })1 Xi , 1 i n. p(X1, . . . , Xn|}) = n Y }Xi(1 })1 Xi . p(}|X1, . . . , Xn) = i=1 = p(}, X1, . . . , Xn) p(X1, . . . , Xn) p(X1, . . . , Xn|}) p(}) p(X1, . . . , Xn) / p(X1, . . . , Xn|})p(}) n Y = }Xi (1 })1 Xi i=1 ⇥342}(1 / }1+ which is the Beta 2 + density function. Pn i=1 Xi Pn i=1 })17 (1 })17+n Xi, 18 + n I now update my belief via the posterior density: Now collect data on class participants: p(}|X1, . . . , Xn). n X i=1 Posterior is n = Xi = Pn i=1 Xi Pn i=1 Xi Beta( , ). The most common single number summary is: bBayes = E(}|X1, . . . , Xn) = } . Compare this with the frequentist answer: b= } The posterior density looks like: = . A common interval answer in Bayesian statistics is a 95% credible interval posterior density function area of each tail is 0.025 0.95 L U (L, U ) is the 95% credible interval for }. These days we can do the required calculations quickly in R . . . Until 1990 problems like this made practical Bayesian analysis very difficult. The last 25 years has seen a revolution in Bayesian analysis, fuelled mainly by Markov Chain Monte Carlo (MCMC) methodology. PROBLEM! Most practical Bayesian inference problems have integrals that cannot be solved analytically. Bayes estimates and 95% credible intervals require quadrature (e.g. trapezoidal rule). But quadrature is hard to impossible in higher dimensions. A Short History of Bayesian Inference • mid 1700s: Bayes Theorem established by Reverend Thomas Bayes. • next 250 years: Lots of philosophical discussion and debate on Bayesian versus frequentist inference. But most practical statistics was frequentist. • 1990: MCMC introduced to Statistics in Journal of the American Statistical Association paper by A.E. Gelfand & A.F.M. Smith. • mid 1990s: First ‘professional’ MCMC software package started. Named BUGS (Bayesian inference Using Gibbs Sampling). However, clumsy to use. • 2005: R package BRugs released. It allows script-based Bayesian analyses, run from inside R. • 2015: Even better R packages being developed such as rstan, but not compatible with current lab computers. ⌥ ⌥ ⌥