Schedule
Week 1: Basics
Week 2: Approximation Theory
1.11:
Empirical observations of over-parameterized neural networks
Piecewise constant function approximation, bump function
Universal approximation by Stone-Weierstrass
1.13:
Readings:
Week 3: Non-convex Optimization
1.18:
1.20:
Landscape analysis introduction
Unique local minimum
Locally optimizable function, second order stationary point, strict saddle point
Reading:
Week 4: Non-convex Optimization
Week 5: Neural Tangent Kernel
Week 6: Kernel (Cont'd), Implicit Regularization
2.8:
2.10:
Gradient flow automatically balances layers
Margin and smoothed margin
gradient flow maximizes the margin of linear predictor
Readings:
Week 7: Implicit Regularization (Cont'd), Generalization
Week 8: Generalization
Week 9: Generalization
Week 10: Course Project Presentations
|