Feature Learning and Scaling Laws in Two-layer Neural Networks: A high dimensional analysis
Murat A. Erdogdu, University of Toronto
E18-304
Abstract: This talk will focus on gradient-based optimization of two-layer neural networks. We consider a high-dimensional setting where the number of samples and the input dimension are both large and show that, under different model assumptions, neural networks learn useful features and adapt to the model more efficiently than classical methods. Further, we derive scaling laws of the learning dynamics for the gradient descent, highlighting the power-law dependencies on the optimization time, and the model width. Bio: Murat A. Erdogdu…