Statistical learning theory 2022
Содержание
[убрать]General Information
Lectures: Friday 16h20 -- 17h40, Bruno Bauwens, Maxim Kaledin
Seminars: 09.09 -- 01.10 Saturday 14:40 -- 16:00, starting from 07.10 Friday 18h10 -- 19h30, Artur Goldman,
For discussions of the materials, join the telegram group
The course is similar to last year.
Homeworks
Email to brbauwens-at-gmail.com. Start the subject line with SLT-HW.
Deadline before the lecture, every other lecture.
17 Sept: see problem lists 1 and 2
1 Oct: see problem lists 3 and 4
14 Oct: see problem lists 5 and 6
04 Nov: see problem list 7 and 8
28 Nov: see problem lists 9 and 10
02 Dec: see problem lists 11 and 12
Course materials
Video | Summary | Slides | Lecture notes | Problem list | Solutions |
---|---|---|---|---|---|
Part 1. Online learning | |||||
02 Sept | Philosophy. The online mistake bound model. Weighted majority and perceptron algorithms movies | sl01 | ch00 ch01 | list 1 update 05.09 | |
09 Sept | The perceptron algorithm in the agnostic setting. Kernels. The standard optimal algorithm. | sl02 | ch02 ch03 | list 2 | |
16 Sept | Prediction with expert advice and the exponentially weighted majority algorithm. Recap probability theory. | sl03 | ch04 ch05 | list 3 | |
Part 2. Distribution independent risk bounds | |||||
23 Sept | Sample complexity in the realizable setting, simple examples and bounds using VC-dimension | sl04 | ch06 | list 4 | |
30 Sept | Growth functions, VC-dimension and the characterization of sample comlexity with VC-dimensions | sl05 | ch07 ch08 | list 5 | |
07 Oct | Risk decomposition and the fundamental theorem of statistical learning theory | sl06 | ch09 | list 6 | |
14 Oct | Bounded differences inequality, Rademacher complexity, symmetrization, contraction lemma | sl07 | ch10 ch11 | list 7 | |
Part 3. Margin risk bounds with applications | |||||
21 Oct | Simple regression, support vector machines, margin risk bounds, and neural nets | sl08 | ch12 ch13 | list 8 | |
04 Nov | Kernels: RKHS, representer theorem, risk bounds | sl09 | ch14 | list 9 | |
11 Nov | AdaBoost and the margin hypothesis | sl10 | Mohri et al, chapt 7 | list 10 | |
18 Nov | Implicit regularization of stochastic gradient descent in neural nets | list 11 | |||
Part 4. Other topics | |||||
25 Nov | Regression I: classic noise assumption, sub-Guassian and sub-exponential noise | list 12 | |||
02 Dec | Regression II: Ridge and Lasso regression | list 13 | |||
09 Dec | Multiarmed bandids | list 14 | |||
16 Dec | Colloquium |
The lectures in October and November are based on the book: Foundations of machine learning 2nd ed, Mehryar Mohri, Afshin Rostamizadeh, and Ameet Talwalker, 2018. This book can be downloaded from Library Genesis (the link changes sometimes and sometimes vpn is needed).
Problems exam
Dates, problems TBA
During the exam
-- You may consult notes, books and search on the internet
-- You may not interact with other humans (e.g. by phone, forums, etc)
Office hours
Person | Monday | Tuesday | Wednesday | Thursday | Friday | |
---|---|---|---|---|---|---|
Bruno Bauwens | 14h--20h | |||||
Maxim Kaledin |
It is always good to send an email in advance. Questions and feedback are welcome.