Alert button
Picture for Minsoo Rhu

Minsoo Rhu

Alert button

LazyDP: Co-Designing Algorithm-Software for Scalable Training of Differentially Private Recommendation Models

Add code
Bookmark button
Alert button
Apr 12, 2024
Juntaek Lim, Youngeun Kwon, Ranggi Hwang, Kiwan Maeng, G. Edward Suh, Minsoo Rhu

Viaarxiv icon

Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference

Add code
Bookmark button
Alert button
Aug 23, 2023
Ranggi Hwang, Jianyu Wei, Shijie Cao, Changho Hwang, Xiaohu Tang, Ting Cao, Mao Yang, Minsoo Rhu

Figure 1 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Figure 2 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Figure 3 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Figure 4 for Pre-gated MoE: An Algorithm-System Co-Design for Fast and Scalable Mixture-of-Expert Inference
Viaarxiv icon

Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations

Add code
Bookmark button
Alert button
Feb 23, 2023
Yujeong Choi, John Kim, Minsoo Rhu

Figure 1 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Figure 2 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Figure 3 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Figure 4 for Hera: A Heterogeneity-Aware Multi-Tenant Inference Server for Personalized Recommendations
Viaarxiv icon

GPU-based Private Information Retrieval for On-Device Machine Learning Inference

Add code
Bookmark button
Alert button
Jan 27, 2023
Maximilian Lam, Jeff Johnson, Wenjie Xiong, Kiwan Maeng, Udit Gupta, Minsoo Rhu, Hsien-Hsin S. Lee, Vijay Janapa Reddi, Gu-Yeon Wei, David Brooks, Edward Suh

Figure 1 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Figure 2 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Figure 3 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Figure 4 for GPU-based Private Information Retrieval for On-Device Machine Learning Inference
Viaarxiv icon

DiVa: An Accelerator for Differentially Private Machine Learning

Add code
Bookmark button
Alert button
Aug 26, 2022
Beomsik Park, Ranggi Hwang, Dongho Yoon, Yoonhyuk Choi, Minsoo Rhu

Figure 1 for DiVa: An Accelerator for Differentially Private Machine Learning
Figure 2 for DiVa: An Accelerator for Differentially Private Machine Learning
Figure 3 for DiVa: An Accelerator for Differentially Private Machine Learning
Figure 4 for DiVa: An Accelerator for Differentially Private Machine Learning
Viaarxiv icon

SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures

Add code
Bookmark button
Alert button
May 10, 2022
Yunjae Lee, Jinha Chung, Minsoo Rhu

Figure 1 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Figure 2 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Figure 3 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Figure 4 for SmartSAGE: Training Large-scale Graph Neural Networks using In-Storage Processing Architectures
Viaarxiv icon

Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards

Add code
Bookmark button
Alert button
May 10, 2022
Youngeun Kwon, Minsoo Rhu

Figure 1 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Figure 2 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Figure 3 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Figure 4 for Training Personalized Recommendation Systems from (GPU) Scratch: Look Forward not Backwards
Viaarxiv icon

GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks

Add code
Bookmark button
Alert button
Mar 02, 2022
Minhoo Kang, Ranggi Hwang, Jiwon Lee, Dongyun Kam, Youngjoo Lee, Minsoo Rhu

Figure 1 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Figure 2 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Figure 3 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Figure 4 for GROW: A Row-Stationary Sparse-Dense GEMM Accelerator for Memory-Efficient Graph Convolutional Neural Networks
Viaarxiv icon

PARIS and ELSA: An Elastic Scheduling Algorithm for Reconfigurable Multi-GPU Inference Servers

Add code
Bookmark button
Alert button
Feb 27, 2022
Yunseong Kim, Yujeong Choi, Minsoo Rhu

Figure 1 for PARIS and ELSA: An Elastic Scheduling Algorithm for Reconfigurable Multi-GPU Inference Servers
Figure 2 for PARIS and ELSA: An Elastic Scheduling Algorithm for Reconfigurable Multi-GPU Inference Servers
Figure 3 for PARIS and ELSA: An Elastic Scheduling Algorithm for Reconfigurable Multi-GPU Inference Servers
Figure 4 for PARIS and ELSA: An Elastic Scheduling Algorithm for Reconfigurable Multi-GPU Inference Servers
Viaarxiv icon