Picture for Shivaram Venkataraman

Shivaram Venkataraman

GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval

Add code
Jun 25, 2024
Figure 1 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Figure 2 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Figure 3 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Figure 4 for GraphSnapShot: Graph Machine Learning Acceleration with Fast Storage and Retrieval
Viaarxiv icon

CHAI: Clustered Head Attention for Efficient LLM Inference

Add code
Mar 12, 2024
Figure 1 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 2 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 3 for CHAI: Clustered Head Attention for Efficient LLM Inference
Figure 4 for CHAI: Clustered Head Attention for Efficient LLM Inference
Viaarxiv icon

Decoding Speculative Decoding

Add code
Feb 02, 2024
Figure 1 for Decoding Speculative Decoding
Figure 2 for Decoding Speculative Decoding
Figure 3 for Decoding Speculative Decoding
Figure 4 for Decoding Speculative Decoding
Viaarxiv icon

PolyThrottle: Energy-efficient Neural Network Inference on Edge Devices

Add code
Oct 30, 2023
Figure 1 for PolyThrottle: Energy-efficient Neural Network Inference on Edge Devices
Figure 2 for PolyThrottle: Energy-efficient Neural Network Inference on Edge Devices
Figure 3 for PolyThrottle: Energy-efficient Neural Network Inference on Edge Devices
Figure 4 for PolyThrottle: Energy-efficient Neural Network Inference on Edge Devices
Viaarxiv icon

Does compressing activations help model parallel training?

Add code
Jan 06, 2023
Figure 1 for Does compressing activations help model parallel training?
Figure 2 for Does compressing activations help model parallel training?
Figure 3 for Does compressing activations help model parallel training?
Figure 4 for Does compressing activations help model parallel training?
Viaarxiv icon

BagPipe: Accelerating Deep Recommendation Model Training

Add code
Feb 24, 2022
Figure 1 for BagPipe: Accelerating Deep Recommendation Model Training
Figure 2 for BagPipe: Accelerating Deep Recommendation Model Training
Figure 3 for BagPipe: Accelerating Deep Recommendation Model Training
Figure 4 for BagPipe: Accelerating Deep Recommendation Model Training
Viaarxiv icon

Marius++: Large-Scale Training of Graph Neural Networks on a Single Machine

Add code
Feb 04, 2022
Figure 1 for Marius++: Large-Scale Training of Graph Neural Networks on a Single Machine
Figure 2 for Marius++: Large-Scale Training of Graph Neural Networks on a Single Machine
Figure 3 for Marius++: Large-Scale Training of Graph Neural Networks on a Single Machine
Figure 4 for Marius++: Large-Scale Training of Graph Neural Networks on a Single Machine
Viaarxiv icon

Doing More by Doing Less: How Structured Partial Backpropagation Improves Deep Learning Clusters

Add code
Nov 20, 2021
Figure 1 for Doing More by Doing Less: How Structured Partial Backpropagation Improves Deep Learning Clusters
Figure 2 for Doing More by Doing Less: How Structured Partial Backpropagation Improves Deep Learning Clusters
Figure 3 for Doing More by Doing Less: How Structured Partial Backpropagation Improves Deep Learning Clusters
Figure 4 for Doing More by Doing Less: How Structured Partial Backpropagation Improves Deep Learning Clusters
Viaarxiv icon

KAISA: An Adaptive Second-order Optimizer Framework for Deep Neural Networks

Add code
Jul 04, 2021
Figure 1 for KAISA: An Adaptive Second-order Optimizer Framework for Deep Neural Networks
Figure 2 for KAISA: An Adaptive Second-order Optimizer Framework for Deep Neural Networks
Figure 3 for KAISA: An Adaptive Second-order Optimizer Framework for Deep Neural Networks
Figure 4 for KAISA: An Adaptive Second-order Optimizer Framework for Deep Neural Networks
Viaarxiv icon

On the Utility of Gradient Compression in Distributed Training Systems

Add code
Mar 03, 2021
Figure 1 for On the Utility of Gradient Compression in Distributed Training Systems
Figure 2 for On the Utility of Gradient Compression in Distributed Training Systems
Figure 3 for On the Utility of Gradient Compression in Distributed Training Systems
Figure 4 for On the Utility of Gradient Compression in Distributed Training Systems
Viaarxiv icon