Instructor: Jatin Batra
Schedule: Thursdays, 2:00–3:30 PM
Room: A-238
Email: jatinbatra50@gmail.com
Textbook: Learning theory from first principles — Francis Bach (Chapter 9 and relevant parts of Chapters 3, 4, 5, 7, 8, 12, 14).

Course Description

This mini-course serves as a take-off point for research in theory of deep learning, with neural networks as the prototypical setup, more specifically, the focus will be on implicit bias of overparameterized models. Generalization power of modern deep learning cannot be easily explained as the number of pa- rameters is often of the same order as the number of training samples. In this part, we will initiate an understanding of implicit bias which captures additional properties enjoyed by natural training procedures for overpa- rameterized models that are not captured explicitly, and which could hold the key to generalization.

Weekly Outline

Week Topic Notes
01 Implicit bias of Gradient Flow in Overparameterized Linear Regression Notes (PDF)
02 A Glimpse of Double Descent Notes (PDF)
03 Mickey Mouse Proof of Double Descent Notes (PDF)
04 Topic Notes (PDF)
05 Topic Notes (PDF)
06 Topic Notes (PDF)
07 Topic Notes (PDF)
08 Topic Notes (PDF)
09 Topic Notes (PDF)
10 Topic Notes (PDF)
11 Topic Notes (PDF)
12 Topic Notes (PDF)
13 Topic Notes (PDF)
14 Topic Notes (PDF)
15 Topic Notes (PDF)