Picture for Zachary Charles

Zachary Charles

Improving the convergence of SGD through adaptive batch sizes

Add code
Oct 18, 2019
Figure 1 for Improving the convergence of SGD through adaptive batch sizes
Figure 2 for Improving the convergence of SGD through adaptive batch sizes
Figure 3 for Improving the convergence of SGD through adaptive batch sizes
Figure 4 for Improving the convergence of SGD through adaptive batch sizes
Viaarxiv icon

DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation

Add code
Jul 29, 2019
Figure 1 for DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation
Figure 2 for DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation
Figure 3 for DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation
Figure 4 for DETOX: A Redundancy-based Framework for Faster and More Robust Gradient Aggregation
Viaarxiv icon

Convergence and Margin of Adversarial Training on Separable Data

Add code
May 22, 2019
Figure 1 for Convergence and Margin of Adversarial Training on Separable Data
Figure 2 for Convergence and Margin of Adversarial Training on Separable Data
Figure 3 for Convergence and Margin of Adversarial Training on Separable Data
Figure 4 for Convergence and Margin of Adversarial Training on Separable Data
Viaarxiv icon

Does Data Augmentation Lead to Positive Margin?

Add code
May 08, 2019
Figure 1 for Does Data Augmentation Lead to Positive Margin?
Figure 2 for Does Data Augmentation Lead to Positive Margin?
Figure 3 for Does Data Augmentation Lead to Positive Margin?
Figure 4 for Does Data Augmentation Lead to Positive Margin?
Viaarxiv icon

ErasureHead: Distributed Gradient Descent without Delays Using Approximate Gradient Coding

Add code
Jan 28, 2019
Figure 1 for ErasureHead: Distributed Gradient Descent without Delays Using Approximate Gradient Coding
Figure 2 for ErasureHead: Distributed Gradient Descent without Delays Using Approximate Gradient Coding
Figure 3 for ErasureHead: Distributed Gradient Descent without Delays Using Approximate Gradient Coding
Figure 4 for ErasureHead: Distributed Gradient Descent without Delays Using Approximate Gradient Coding
Viaarxiv icon

A Geometric Perspective on the Transferability of Adversarial Directions

Add code
Nov 08, 2018
Figure 1 for A Geometric Perspective on the Transferability of Adversarial Directions
Figure 2 for A Geometric Perspective on the Transferability of Adversarial Directions
Figure 3 for A Geometric Perspective on the Transferability of Adversarial Directions
Figure 4 for A Geometric Perspective on the Transferability of Adversarial Directions
Viaarxiv icon

ATOMO: Communication-efficient Learning via Atomic Sparsification

Add code
Jun 24, 2018
Figure 1 for ATOMO: Communication-efficient Learning via Atomic Sparsification
Figure 2 for ATOMO: Communication-efficient Learning via Atomic Sparsification
Figure 3 for ATOMO: Communication-efficient Learning via Atomic Sparsification
Figure 4 for ATOMO: Communication-efficient Learning via Atomic Sparsification
Viaarxiv icon

DRACO: Byzantine-resilient Distributed Training via Redundant Gradients

Add code
Jun 22, 2018
Figure 1 for DRACO: Byzantine-resilient Distributed Training via Redundant Gradients
Figure 2 for DRACO: Byzantine-resilient Distributed Training via Redundant Gradients
Figure 3 for DRACO: Byzantine-resilient Distributed Training via Redundant Gradients
Figure 4 for DRACO: Byzantine-resilient Distributed Training via Redundant Gradients
Viaarxiv icon

Gradient Coding via the Stochastic Block Model

Add code
May 25, 2018
Figure 1 for Gradient Coding via the Stochastic Block Model
Figure 2 for Gradient Coding via the Stochastic Block Model
Figure 3 for Gradient Coding via the Stochastic Block Model
Viaarxiv icon

Subspace Clustering with Missing and Corrupted Data

Add code
Jan 15, 2018
Figure 1 for Subspace Clustering with Missing and Corrupted Data
Viaarxiv icon