Wei Hu will present his Pre FPO 'Understanding Deep Learning via Analyzing Dynamics of Gradient Descent' on Tuesday, April 13, 2021 at 4pm via Zoom.

Wei Hu will present his Pre FPO on Tuesday, April 13, 2021 at 4pm via Zoom. Zoom link: https://princeton.zoom.us/j/2180408086 Committee: Examiners: Sanjeev Arora (advisor), Mark Braverman, Chi Jin (ECE) Readers: Elad Hazan, Jason D. Lee (ECE) All are welcome to attend. Title: Understanding Deep Learning via Analyzing Dynamics of Gradient Descent Abstract: The phenomenal successes of deep learning build upon the mysterious abilities of gradient-based optimization algorithms. Not only can these algorithms often successfully optimize complicated non-convex training objectives, but the solutions found can also generalize remarkably well to unseen test data despite significant over-parameterization of the models. Classical approaches in optimization and learning theory that treat empirical risk minimization as a black box are insufficient to explain these mysteries in modern deep learning. In this talk, I will illustrate how we can make progress towards understanding optimization and generalization in deep learning by a more refined approach that opens the black box and analyzes the dynamics taken by the optimizer. In particular, I will present several results focusing on analyzing the dynamics of the gradient descent algorithm, including: (i) solving low-rank matrix completion via deep linear neural networks, (ii) positive and negative convergence results which reveal the optimal depth-width tradeoff for efficiently training deep linear neural networks, and (iii) the connection between wide neural networks and neural tangent kernels, and its theoretical and practical implications.
participants (1)
-
jfarquer@cs.princeton.edu