Survey
* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project
Fall 2017 Course announcement: MATH 583: Prediction and Statistical Learning MWF 10-10:50 room ENGA 0208 Instructor: David Olive Text: James, G., Witten, D., Hastie, T., and Tibshirani, R. (2013), An Introduction to Statistical Learning With Applications in R, Springer, New York, NY. ISBN: 978-1-46147137-0 ≈ $60 at www.addall.com and www.amazon.com. The library has the text online and the last 2 authors may have the pdf file available online. Also see online notes Olive, D.J. (2017), Prediction and Statistical Learning, (http://lagrange.math.siu.edu/Olive/slearnbk.htm), which will be provided free. (This course will eventually become Math 586: Statistical Computing and Learning, but I am not teaching the Statistical Computing part, which would require a computer programming class like CS 202.) The prerequisite for this class is Math 483 or equivalent (calculus based course in probability and statistics, CS 480 may be adequate) and Math 221 or equivalent (Linear Algebra). A course in regression such as Math 484, 473 or 485 is recommended (CS 473 or 438 may be adequate). Statistical Learning could be defined as the statistical analysis of multivariate data. Machine learning, data mining, big data, analytics, business analytics, data analytics, and predictive analytics are synonymous terms. The techniques are useful for Data Science and Statistics, the science of extracting information from data. The R software will be used. Often there is a response variable Y of interest, and a p × 1 vector x = (x1 , ..., xp)T used to predict Y . The data are (Y1 , x1 ), ..., (Yn, xn ). Multiple linear regression (MLR) and classification will be covered. Standard regression methods such as least squares (OLS) and linear discriminant analysis need n/p large. Statistical learning methods are often useful when n/p is not large. MLR methods will include OLS forward selection, lasso, relaxed lasso, ridge regression, partial least squares, and principal component regression. Prediction intervals will be used to compare these methods, and prediction regions will be used for bootstrap hypothesis tests. Classification will consider linear discriminant analysis (LDA), lasso for LDA, K-nearest neighbors, classification trees, and support vector machines. These are supervised learning methods. Unsupervised learning methods include clustering methods and principal component analysis. There may be overlap with Machine Learning courses such as CS437, 533, and 586. Final ?day, December ?, ?. The grading and schedule below are tentative. (Drop days are Friday, Oct. 27, Sunday Oct. 29.) 3 homeworks may be turned in one class period late (ie on Friday) with no penalty. A fourth late will be accepted with 25% penalty. 2 quizzes may be taken late before the next class period (ie on Monday). Two sheets of notes are allowed on quizzes more for exams. A calculator is permitted. Grading: 1 HW and 1 Quiz will be dropped. 1 HW exam1 final min. grade A D 300 100 300 points 900-1000 550-699 exam 2 Quizzes 100 min. grade B points 800-899 100 exam 3 total min. grade C 100 1000 points 700-799 O1.1 refers to Olive (2017, ch. 1, section 1). The other numbers are for James et al. (2013). Week of Aug 21 Aug 28 Sept 4 Sept 11 Sept 18 Sept 25 Oct 2 Oct 9 Oct 16 Oct 23 Nov 30 Nov 6 Nov 13 Nov 20 Nov 27 Dec 4 MON no class lab? no class 5.2,O2.3 6.1, 6.2, 6.3, 10.2 6.3, 10.2, 5.1 7.2 no class 7.7 4.3, 4.4 lab? 8.2, 9.1, 9.2 9.2, 9.3 10.3, HW11 WED FRI 1, 2.1, 2.2, O1.1, O1.2 O1.2, O1.3 O1.3, O1.4, HW1 O2.1, Q1 O2.1, HW2 5.2,O2.3, Q2 3.1,3.2, HW3 3.2,3.3, Q3 6.1, 6.2, 6.3,10.2 Exam 1 6.4, HW4 7.1, Q4 7.3, HW5 7.4, 7.5, Q5 7.6, HW6 7.7, Q6 4.1, 4.2 Exam 2 4.3, HW7 4.4, Q7 4.5, HW8 4.5, 7.7, Q8 8.1,HW9 8.1,Q9 8.2, HW10 9.1,Q10 no class no class 9.4,9.5 Exam 3 10.4, Q11 review Lorayne and Lucas (2000), The Memory Book is useful for memorization. Other references: Hastie, T., Tibshirani, R., and Friedman, J. (2009), The Elements of Statistical Learning: Data Mining, Inference and Prediction, 2nd ed., Springer, New York, NY. Hastie, T., Tibshirani, R., and Wainwright, M. (2015), Statistical Learning with Sparsity: the Lasso and Generalizations, CRC Press Taylor & Francis, Boca Raton, FL. Kuhn, M., and Johnson, K. (2013), Applied Predictive Modeling, Springer, New York, NY. Abu-Mostafa, Y.S., Magdon-Ismail, M., and Lin, H.T. (2012), Learning From Data, AMLBook, Pasadena, CA. Berk, R.A. (2008), Statistical Learning From a Regression Perspective, Springer, New York, NY. Witten, I.A., and Frank, E. (2005), Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed., Elsevier, San Francisco, CA. 2