Lecture 1. Intro and theory for shallow networks
Perceptron convergence theorem
Universal approximation theorem
Approximation rates for shallow neural networks
Barron spaces
Advantages of additional hidden layers
Deep ReLU networks
Misclassification error for image deformation models
Optimization in machine learning
Weight balancing phenomenon
Analysis of dropout
Benign overfitting
Grokking