Download Introduction to Statistical Inference and Learning

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Mathematical physics wikipedia , lookup

Artificial intelligence wikipedia , lookup

Psychometrics wikipedia , lookup

Information theory wikipedia , lookup

Expectation–maximization algorithm wikipedia , lookup

Statistical mechanics wikipedia , lookup

Theoretical computer science wikipedia , lookup

Machine learning wikipedia , lookup

Pattern recognition wikipedia , lookup

Transcript
Introduction to Statistical Inference and Learning
Instructors: Roi Weiss, Boaz Nadler
In this course we will cover the basic concepts underlying modern data analysis, machine
learning and statistical inference. Subject to time constraints, topics covered will include
1. Basic probability, inequalities, classical distributions
2. Basic information theory, entropy, KL divergence
3. Parameter estimation, maximum likelihood, Bayesian approaches, MAP, MMSE, Empirical
risk minimization, consistency, Cramer-Rao lower bound (information inequality)
4. Parametric and non-parametric models
5. density estimation, kernel smoothing
6. The bias-variance tradeoff and the curse of dimensionality in high dimensional problems
7. Hypothesis testing and the Neyman-Pearson lemma
8. Principal Component Analysis, dimensionality reduction
9. Latent variable models, Latent Dirichlet Allocation, Hidden Markov Models
10. Sparsity and compressed sensing
11. Supervised and unsupervised ensemble learning
Suggested Reading:
1. L. Wasserman, All of statistics, Springer. (also take a look at all of non-parametric statistics
by the same author.)
2. T. Hastie, R. Tibshirani and J. Friedman, The elements of statistical learning, Springer.
3. V. Vapnik, The nature of statistical learning theory, Springer, 2nd edition.
4. K. Knight, Mathematical Statistics.
5. C. Bishop, Pattern Recognition and Machine Learning.
6. Thomas and Cover, Elements of Information Theory.
Grading:
Throughout the course we will give 4-5 homework exercises. It is mandatory to submit at least
75% of them.
There will also be a final exam.
The grade in the course consists of a weighted average: 60% exam + 40% HW exercises.
1