Previously offered in in 2022W1 and (with the name 532S) in 2021W2; this instance will be broadly similar.
Italicized entries are tentative. The book acronyms are described here.
Date | Topic | Supplements | ||
---|---|---|---|---|
Tu | Sep 5 | No class: Imagine Day | ||
Th | Sep 7 | Course intro, ERM | SSBD 1-2; MRT 2 | |
F | Sep 8 | Assignment 1 posted: pdf, tex | ||
Tu | Sep 12 | Class canceled: sick | ||
Th | Sep 14 | Uniform convergence with finite classes | SSBD 2-4; MRT 2 | |
M | Sep 18 | Assignment 1 due at noon | ||
M | Sep 18 | Drop deadline | ||
Tu | Sep 19 | Concentration inequalities | SSBD B; MRT D Zhang 2; Wainwright 2 | |
Th | Sep 21 | PAC learning; covering numbers | SSBD 3, MRT 2 Bach 4.4.4, Zhang 3.4/4/5 (much more detailed) | |
Sa | Sep 23 | Assignment 2 posted: pdf, tex | ||
Tu | Sep 26 | Rademacher complexity | MRT 3; SSBD 26; Bach 4.5; Zhang 6 | |
Th | Sep 28 | More Rademacher (same notes) | ||
Tu | Oct 3 | VC dimension | ||
Th | Oct 5 | No Free Lunch | ||
Tu | Oct 10 | |||
W | Oct 11 | Assignment 2 due at midnight | ||
Th | Oct 12 | No class: UBC follows a Monday schedule | ||
Tu | Oct 17 | |||
Th | Oct 19 | |||
Tu | Oct 24 | |||
Th | Oct 26 | |||
F | Oct 27 | Withdrawal deadline | ||
Tu | Oct 31 | |||
Th | Nov 2 | |||
Tu | Nov 7 | |||
Th | Nov 9 | |||
Tu | Nov 14 | No class: midterm break | ||
Th | Nov 16 | |||
M | Nov 21 | |||
W | Nov 23 | |||
Tu | Nov 28 | |||
Th | Nov 30 | |||
Tu | Dec 5 | |||
Th | Dec 7 | |||
? | Dec ?? | Final exam (in person, handwritten) — date and time TBA, sometime Dec 11-22 |
The course meets in person in Swing 210, with possible rare exceptions (e.g. if I get sick but can still teach, I'll move it online). Note that this room does not have a recording setup.
Grading scheme: 70% assignments, 30% final.
There will be four or five written assignments through the term; answers should be written in LaTeX, and handed in on Gradescope. There will also be a small number (one or two) of assignments that involve reading a paper, reacting to it, and poking at it slightly further; details to come.
The brief idea of the course: when should we expect machine learning algorithms to work? What kinds of assumptions do we need to be able to be able to rigorously prove that they will work?
Definitely covered: PAC learning, VC dimension, Rademacher complexity, concentration inequalities, margin bounds, stability. Also, most of: PAC-Bayes, analysis of kernel methods, limitations of uniform convergence, analyzing deep nets via neural tangent kernels, provable gaps between kernel methods and deep learning, online learning, feasibility of private learning, compression-based bounds.
There are no formal prerequisites. I will roughly assume:
Books that the course will definitely pull from:
New books where I may or may not pull from sections, TBD:
Some other points of view you might like:
If you need to refresh your linear algebra or other areas of math:
Measure-theoretic probability is not required for this course, but there are instances and related areas where it could be helpful:
Similar courses: