Picture for Hector Yuen

Hector Yuen

Supporting Massive DLRM Inference Through Software Defined Memory

Add code
Nov 08, 2021
Figure 1 for Supporting Massive DLRM Inference Through Software Defined Memory
Figure 2 for Supporting Massive DLRM Inference Through Software Defined Memory
Figure 3 for Supporting Massive DLRM Inference Through Software Defined Memory
Figure 4 for Supporting Massive DLRM Inference Through Software Defined Memory
Viaarxiv icon

Low-Precision Hardware Architectures Meet Recommendation Model Inference at Scale

Add code
May 26, 2021
Figure 1 for Low-Precision Hardware Architectures Meet Recommendation Model Inference at Scale
Figure 2 for Low-Precision Hardware Architectures Meet Recommendation Model Inference at Scale
Figure 3 for Low-Precision Hardware Architectures Meet Recommendation Model Inference at Scale
Figure 4 for Low-Precision Hardware Architectures Meet Recommendation Model Inference at Scale
Viaarxiv icon

Post-Training 4-bit Quantization on Embedding Tables

Add code
Nov 05, 2019
Figure 1 for Post-Training 4-bit Quantization on Embedding Tables
Figure 2 for Post-Training 4-bit Quantization on Embedding Tables
Figure 3 for Post-Training 4-bit Quantization on Embedding Tables
Figure 4 for Post-Training 4-bit Quantization on Embedding Tables
Viaarxiv icon

A Study of BFLOAT16 for Deep Learning Training

Add code
Jun 13, 2019
Figure 1 for A Study of BFLOAT16 for Deep Learning Training
Figure 2 for A Study of BFLOAT16 for Deep Learning Training
Figure 3 for A Study of BFLOAT16 for Deep Learning Training
Figure 4 for A Study of BFLOAT16 for Deep Learning Training
Viaarxiv icon

Deep Learning Inference in Facebook Data Centers: Characterization, Performance Optimizations and Hardware Implications

Add code
Nov 29, 2018
Figure 1 for Deep Learning Inference in Facebook Data Centers: Characterization, Performance Optimizations and Hardware Implications
Figure 2 for Deep Learning Inference in Facebook Data Centers: Characterization, Performance Optimizations and Hardware Implications
Figure 3 for Deep Learning Inference in Facebook Data Centers: Characterization, Performance Optimizations and Hardware Implications
Figure 4 for Deep Learning Inference in Facebook Data Centers: Characterization, Performance Optimizations and Hardware Implications
Viaarxiv icon