Please note, the time listed in the original subject line of the announcement was incorrect. This general exam will take place at 3:00 PM today, per the body text.

 

From: talks-bounces@lists.cs.princeton.edu <talks-bounces@lists.cs.princeton.edu> On Behalf Of Louis W. Riehl
Sent: Tuesday, January 17, 2023 3:00 PM
To: talks <talks@lists.cs.princeton.edu>
Subject: [talks] Abhishek Panigrahi will present his General Exam "Demystifying Gradient Descent in modern Deep Learning: Implicit training biases and Modular Generalization" on Wednesday, January 18, 2023 at 10:00 AM 3:00 PM over Zoom.

 

Abhishek Panigrahi will present his General Exam "Demystifying Gradient Descent in modern Deep Learning: Implicit training biases and Modular Generalization" on Wednesday, January 18, 2023 at 3:00 PM over Zoom. 

 

 

Committee Members: Sanjeev Arora (advisor), Elad Hazan, Danqi Chen

 

Abstract:

Modern deep learning involves training large scale neural networks, which comes at the cost of deciding the best training recipe. Traditional machine learning fails to explain the hidden mechanisms of such models, owing to the high non-convexity of the model landscape. My research focuses on the training-time interplay between different training algorithms and the architecture that drives the generalizability of these models. In this talk, I will focus on Gradient Descent and its two novel mechanisms: (a)  the Edge of Stability in Deep Learning, where the interplay between learning rate and the model landscape leads to implicit regularization of hessian flatness during training,  and (b) modular skill acquisition that drives generalization during language-model fine-tuning.

 

Reading List:

 

Everyone is invited to attend the talk, and those faculty wishing to remain for the oral exam following are welcome to do so.

 

Louis Riehl
Graduate Administrator
Computer Science Department, CS213
Princeton University
(609) 258-8014