Picture for Sanjeev Arora

Sanjeev Arora

Understanding Gradient Descent on Edge of Stability in Deep Learning

Add code
May 19, 2022
Figure 1 for Understanding Gradient Descent on Edge of Stability in Deep Learning
Figure 2 for Understanding Gradient Descent on Edge of Stability in Deep Learning
Figure 3 for Understanding Gradient Descent on Edge of Stability in Deep Learning
Figure 4 for Understanding Gradient Descent on Edge of Stability in Deep Learning
Viaarxiv icon

Adaptive Gradient Methods with Local Guarantees

Add code
Mar 05, 2022
Figure 1 for Adaptive Gradient Methods with Local Guarantees
Figure 2 for Adaptive Gradient Methods with Local Guarantees
Figure 3 for Adaptive Gradient Methods with Local Guarantees
Figure 4 for Adaptive Gradient Methods with Local Guarantees
Viaarxiv icon

Understanding Contrastive Learning Requires Incorporating Inductive Biases

Add code
Feb 28, 2022
Figure 1 for Understanding Contrastive Learning Requires Incorporating Inductive Biases
Figure 2 for Understanding Contrastive Learning Requires Incorporating Inductive Biases
Figure 3 for Understanding Contrastive Learning Requires Incorporating Inductive Biases
Figure 4 for Understanding Contrastive Learning Requires Incorporating Inductive Biases
Viaarxiv icon

Evaluating Gradient Inversion Attacks and Defenses in Federated Learning

Add code
Nov 30, 2021
Figure 1 for Evaluating Gradient Inversion Attacks and Defenses in Federated Learning
Figure 2 for Evaluating Gradient Inversion Attacks and Defenses in Federated Learning
Figure 3 for Evaluating Gradient Inversion Attacks and Defenses in Federated Learning
Figure 4 for Evaluating Gradient Inversion Attacks and Defenses in Federated Learning
Viaarxiv icon

On Predicting Generalization using GANs

Add code
Nov 28, 2021
Figure 1 for On Predicting Generalization using GANs
Figure 2 for On Predicting Generalization using GANs
Figure 3 for On Predicting Generalization using GANs
Figure 4 for On Predicting Generalization using GANs
Viaarxiv icon

Gradient Descent on Two-layer Nets: Margin Maximization and Simplicity Bias

Add code
Nov 09, 2021
Figure 1 for Gradient Descent on Two-layer Nets: Margin Maximization and Simplicity Bias
Figure 2 for Gradient Descent on Two-layer Nets: Margin Maximization and Simplicity Bias
Figure 3 for Gradient Descent on Two-layer Nets: Margin Maximization and Simplicity Bias
Figure 4 for Gradient Descent on Two-layer Nets: Margin Maximization and Simplicity Bias
Viaarxiv icon

What Happens after SGD Reaches Zero Loss? --A Mathematical Framework

Add code
Oct 13, 2021
Figure 1 for What Happens after SGD Reaches Zero Loss? --A Mathematical Framework
Viaarxiv icon

Rip van Winkle's Razor: A Simple Estimate of Overfit to Test Data

Add code
Feb 25, 2021
Figure 1 for Rip van Winkle's Razor: A Simple Estimate of Overfit to Test Data
Figure 2 for Rip van Winkle's Razor: A Simple Estimate of Overfit to Test Data
Figure 3 for Rip van Winkle's Razor: A Simple Estimate of Overfit to Test Data
Figure 4 for Rip van Winkle's Razor: A Simple Estimate of Overfit to Test Data
Viaarxiv icon

On the Validity of Modeling SGD with Stochastic Differential Equations

Add code
Feb 24, 2021
Figure 1 for On the Validity of Modeling SGD with Stochastic Differential Equations
Figure 2 for On the Validity of Modeling SGD with Stochastic Differential Equations
Figure 3 for On the Validity of Modeling SGD with Stochastic Differential Equations
Figure 4 for On the Validity of Modeling SGD with Stochastic Differential Equations
Viaarxiv icon

Why Are Convolutional Nets More Sample-Efficient than Fully-Connected Nets?

Add code
Oct 16, 2020
Figure 1 for Why Are Convolutional Nets More Sample-Efficient than Fully-Connected Nets?
Figure 2 for Why Are Convolutional Nets More Sample-Efficient than Fully-Connected Nets?
Viaarxiv icon