Picture for Minsoo Rhu

Minsoo Rhu

PreSto: An In-Storage Data Preprocessing System for Training Recommendation Models

Add code
Jun 11, 2024
Viaarxiv icon

ElasticRec: A Microservice-based Model Serving Architecture Enabling Elastic Resource Scaling for Recommendation Models

Add code
Jun 11, 2024
Viaarxiv icon

LazyDP: Co-Designing Algorithm-Software for Scalable Training of Differentially Private Recommendation Models

Add code
Apr 12, 2024
Figure 1 for LazyDP: Co-Designing Algorithm-Software for Scalable Training of Differentially Private Recommendation Models
Figure 2 for LazyDP: Co-Designing Algorithm-Software for Scalable Training of Differentially Private Recommendation Models
Figure 3 for LazyDP: Co-Designing Algorithm-Software for Scalable Training of Differentially Private Recommendation Models
Figure 4 for LazyDP: Co-Designing Algorithm-Software for Scalable Training of Differentially Private Recommendation Models
Viaarxiv icon

Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference

Add code
Aug 23, 2023
Figure 1 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Figure 2 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Figure 3 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Figure 4 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Viaarxiv icon

Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations

Add code
Feb 23, 2023
Figure 1 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Figure 2 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Figure 3 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Figure 4 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Viaarxiv icon

GPU-based Private Information Retrieval for On-Device Machine Learning Inference

Add code
Jan 27, 2023
Figure 1 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Figure 2 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Figure 3 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Figure 4 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Viaarxiv icon

DiVa: An Accelerator for Differentially Private Machine Learning

Add code
Aug 26, 2022
Figure 1 for DiVa: An Accelerator for Differentially Private Machine Learning
Figure 2 for DiVa: An Accelerator for Differentially Private Machine Learning
Figure 3 for DiVa: An Accelerator for Differentially Private Machine Learning
Figure 4 for DiVa: An Accelerator for Differentially Private Machine Learning
Viaarxiv icon

SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures

Add code
May 10, 2022
Figure 1 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Figure 2 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Figure 3 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Figure 4 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Viaarxiv icon

Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards

Add code
May 10, 2022
Figure 1 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Figure 2 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Figure 3 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Figure 4 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Viaarxiv icon

GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks

Add code
Mar 02, 2022
Figure 1 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Figure 2 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Figure 3 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Figure 4 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Viaarxiv icon