Download Math583

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the work of artificial intelligence, which forms the content of this project

Document related concepts
no text concepts found
Transcript
Fall 2017 Course announcement: MATH 583: Prediction and Statistical Learning
MWF 10-10:50 room ENGA 0208
Instructor: David Olive
Text: James, G., Witten, D., Hastie, T., and Tibshirani, R. (2013), An Introduction to
Statistical Learning With Applications in R, Springer, New York, NY. ISBN: 978-1-46147137-0 ≈ $60 at www.addall.com and www.amazon.com. The library has the text online
and the last 2 authors may have the pdf file available online. Also see online notes Olive,
D.J. (2017), Prediction and Statistical Learning, (http://lagrange.math.siu.edu/Olive/slearnbk.htm),
which will be provided free. (This course will eventually become Math 586: Statistical
Computing and Learning, but I am not teaching the Statistical Computing part, which
would require a computer programming class like CS 202.)
The prerequisite for this class is Math 483 or equivalent (calculus based course in
probability and statistics, CS 480 may be adequate) and Math 221 or equivalent (Linear
Algebra). A course in regression such as Math 484, 473 or 485 is recommended (CS 473
or 438 may be adequate).
Statistical Learning could be defined as the statistical analysis of multivariate data.
Machine learning, data mining, big data, analytics, business analytics, data analytics, and
predictive analytics are synonymous terms. The techniques are useful for Data Science
and Statistics, the science of extracting information from data. The R software will be
used. Often there is a response variable Y of interest, and a p × 1 vector x = (x1 , ..., xp)T
used to predict Y . The data are (Y1 , x1 ), ..., (Yn, xn ). Multiple linear regression (MLR)
and classification will be covered. Standard regression methods such as least squares
(OLS) and linear discriminant analysis need n/p large. Statistical learning methods are
often useful when n/p is not large. MLR methods will include OLS forward selection,
lasso, relaxed lasso, ridge regression, partial least squares, and principal component regression. Prediction intervals will be used to compare these methods, and prediction
regions will be used for bootstrap hypothesis tests. Classification will consider linear discriminant analysis (LDA), lasso for LDA, K-nearest neighbors, classification trees, and
support vector machines. These are supervised learning methods. Unsupervised learning
methods include clustering methods and principal component analysis.
There may be overlap with Machine Learning courses such as CS437, 533, and 586.
Final ?day, December ?, ?.
The grading and schedule below are tentative. (Drop days are Friday, Oct. 27, Sunday
Oct. 29.)
3 homeworks may be turned in one class period late (ie on Friday) with no penalty.
A fourth late will be accepted with 25% penalty. 2 quizzes may be taken late before the
next class period (ie on Monday). Two sheets of notes are allowed on quizzes more for
exams. A calculator is permitted.
Grading: 1 HW and 1 Quiz will be dropped.
1
HW
exam1
final
min. grade
A
D
300
100
300
points
900-1000
550-699
exam 2
Quizzes
100
min. grade
B
points
800-899
100
exam 3
total
min. grade
C
100
1000
points
700-799
O1.1 refers to Olive (2017, ch. 1, section 1). The other numbers are for James et al.
(2013).
Week of
Aug 21
Aug 28
Sept 4
Sept 11
Sept 18
Sept 25
Oct 2
Oct 9
Oct 16
Oct 23
Nov 30
Nov 6
Nov 13
Nov 20
Nov 27
Dec 4
MON
no class
lab?
no class
5.2,O2.3
6.1, 6.2, 6.3, 10.2
6.3, 10.2, 5.1
7.2
no class
7.7
4.3,
4.4
lab?
8.2,
9.1, 9.2
9.2, 9.3
10.3, HW11
WED
FRI
1, 2.1, 2.2, O1.1, O1.2 O1.2, O1.3
O1.3, O1.4, HW1
O2.1, Q1
O2.1, HW2
5.2,O2.3, Q2
3.1,3.2, HW3
3.2,3.3, Q3
6.1, 6.2, 6.3,10.2
Exam 1
6.4, HW4
7.1, Q4
7.3, HW5
7.4, 7.5, Q5
7.6, HW6
7.7, Q6
4.1, 4.2
Exam 2
4.3, HW7
4.4, Q7
4.5, HW8
4.5, 7.7, Q8
8.1,HW9
8.1,Q9
8.2, HW10
9.1,Q10
no class
no class
9.4,9.5
Exam 3
10.4, Q11
review
Lorayne and Lucas (2000), The Memory Book is useful for memorization.
Other references:
Hastie, T., Tibshirani, R., and Friedman, J. (2009), The Elements of Statistical Learning: Data Mining, Inference and Prediction, 2nd ed., Springer, New York, NY.
Hastie, T., Tibshirani, R., and Wainwright, M. (2015), Statistical Learning with Sparsity: the Lasso and Generalizations, CRC Press Taylor & Francis, Boca Raton, FL.
Kuhn, M., and Johnson, K. (2013), Applied Predictive Modeling, Springer, New York,
NY.
Abu-Mostafa, Y.S., Magdon-Ismail, M., and Lin, H.T. (2012), Learning From Data,
AMLBook, Pasadena, CA.
Berk, R.A. (2008), Statistical Learning From a Regression Perspective, Springer, New
York, NY.
Witten, I.A., and Frank, E. (2005), Data Mining: Practical Machine Learning Tools
and Techniques, 2nd ed., Elsevier, San Francisco, CA.
2
Related documents