Picture for Daniel M. Roy

Daniel M. Roy

University of Toronto

Limitations of Information-Theoretic Generalization Bounds for Gradient Descent Methods in Stochastic Convex Optimization

Add code
Dec 27, 2022
Viaarxiv icon

Pruning's Effect on Generalization Through the Lens of Training and Regularization

Add code
Oct 25, 2022
Figure 1 for Pruning's Effect on Generalization Through the Lens of Training and Regularization
Figure 2 for Pruning's Effect on Generalization Through the Lens of Training and Regularization
Figure 3 for Pruning's Effect on Generalization Through the Lens of Training and Regularization
Figure 4 for Pruning's Effect on Generalization Through the Lens of Training and Regularization
Viaarxiv icon

Statistical Inference with Stochastic Gradient Algorithms

Add code
Jul 25, 2022
Figure 1 for Statistical Inference with Stochastic Gradient Algorithms
Figure 2 for Statistical Inference with Stochastic Gradient Algorithms
Figure 3 for Statistical Inference with Stochastic Gradient Algorithms
Figure 4 for Statistical Inference with Stochastic Gradient Algorithms
Viaarxiv icon

Understanding Generalization via Leave-One-Out Conditional Mutual Information

Add code
Jun 29, 2022
Figure 1 for Understanding Generalization via Leave-One-Out Conditional Mutual Information
Figure 2 for Understanding Generalization via Leave-One-Out Conditional Mutual Information
Viaarxiv icon

The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization

Add code
Jun 06, 2022
Figure 1 for The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization
Figure 2 for The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization
Figure 3 for The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization
Figure 4 for The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization
Viaarxiv icon

Adaptively Exploiting d-Separators with Causal Bandits

Add code
Feb 10, 2022
Figure 1 for Adaptively Exploiting d-Separators with Causal Bandits
Figure 2 for Adaptively Exploiting d-Separators with Causal Bandits
Figure 3 for Adaptively Exploiting d-Separators with Causal Bandits
Viaarxiv icon

Towards a Unified Information-Theoretic Framework for Generalization

Add code
Nov 17, 2021
Figure 1 for Towards a Unified Information-Theoretic Framework for Generalization
Viaarxiv icon

Minimax Optimal Quantile and Semi-Adversarial Regret via Root-Logarithmic Regularizers

Add code
Nov 07, 2021
Figure 1 for Minimax Optimal Quantile and Semi-Adversarial Regret via Root-Logarithmic Regularizers
Figure 2 for Minimax Optimal Quantile and Semi-Adversarial Regret via Root-Logarithmic Regularizers
Viaarxiv icon

The Future is Log-Gaussian: ResNets and Their Infinite-Depth-and-Width Limit at Initialization

Add code
Jun 07, 2021
Figure 1 for The Future is Log-Gaussian: ResNets and Their Infinite-Depth-and-Width Limit at Initialization
Figure 2 for The Future is Log-Gaussian: ResNets and Their Infinite-Depth-and-Width Limit at Initialization
Figure 3 for The Future is Log-Gaussian: ResNets and Their Infinite-Depth-and-Width Limit at Initialization
Figure 4 for The Future is Log-Gaussian: ResNets and Their Infinite-Depth-and-Width Limit at Initialization
Viaarxiv icon

NUQSGD: Provably Communication-efficient Data-parallel SGD via Nonuniform Quantization

Add code
May 01, 2021
Figure 1 for NUQSGD: Provably Communication-efficient Data-parallel SGD via Nonuniform Quantization
Figure 2 for NUQSGD: Provably Communication-efficient Data-parallel SGD via Nonuniform Quantization
Figure 3 for NUQSGD: Provably Communication-efficient Data-parallel SGD via Nonuniform Quantization
Figure 4 for NUQSGD: Provably Communication-efficient Data-parallel SGD via Nonuniform Quantization
Viaarxiv icon