Picture for Yanqi Zhou

Yanqi Zhou

Mixture-of-Experts with Expert Choice Routing

Add code
Feb 18, 2022
Figure 1 for Mixture-of-Experts with Expert Choice Routing
Figure 2 for Mixture-of-Experts with Expert Choice Routing
Figure 3 for Mixture-of-Experts with Expert Choice Routing
Figure 4 for Mixture-of-Experts with Expert Choice Routing
Viaarxiv icon

LaMDA: Language Models for Dialog Applications

Add code
Feb 10, 2022
Figure 1 for LaMDA: Language Models for Dialog Applications
Figure 2 for LaMDA: Language Models for Dialog Applications
Figure 3 for LaMDA: Language Models for Dialog Applications
Figure 4 for LaMDA: Language Models for Dialog Applications
Viaarxiv icon

GLaM: Efficient Scaling of Language Models with Mixture-of-Experts

Add code
Dec 13, 2021
Figure 1 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 2 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 3 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 4 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Viaarxiv icon

A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules

Add code
Dec 07, 2021
Figure 1 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Figure 2 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Figure 3 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Figure 4 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Viaarxiv icon

Do Transformer Modifications Transfer Across Implementations and Applications?

Add code
Feb 23, 2021
Figure 1 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 2 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 3 for Do Transformer Modifications Transfer Across Implementations and Applications?
Viaarxiv icon

Rethinking Co-design of Neural Architectures and Hardware Accelerators

Add code
Feb 17, 2021
Figure 1 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Figure 2 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Figure 3 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Figure 4 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Viaarxiv icon

Apollo: Transferable Architecture Exploration

Add code
Feb 02, 2021
Figure 1 for Apollo: Transferable Architecture Exploration
Figure 2 for Apollo: Transferable Architecture Exploration
Figure 3 for Apollo: Transferable Architecture Exploration
Figure 4 for Apollo: Transferable Architecture Exploration
Viaarxiv icon

Transferable Graph Optimizers for ML Compilers

Add code
Oct 21, 2020
Figure 1 for Transferable Graph Optimizers for ML Compilers
Figure 2 for Transferable Graph Optimizers for ML Compilers
Figure 3 for Transferable Graph Optimizers for ML Compilers
Figure 4 for Transferable Graph Optimizers for ML Compilers
Viaarxiv icon

A Learned Performance Model for the Tensor Processing Unit

Add code
Aug 03, 2020
Figure 1 for A Learned Performance Model for the Tensor Processing Unit
Figure 2 for A Learned Performance Model for the Tensor Processing Unit
Figure 3 for A Learned Performance Model for the Tensor Processing Unit
Figure 4 for A Learned Performance Model for the Tensor Processing Unit
Viaarxiv icon

ODE-CNN: Omnidirectional Depth Extension Networks

Add code
Jul 03, 2020
Figure 1 for ODE-CNN: Omnidirectional Depth Extension Networks
Figure 2 for ODE-CNN: Omnidirectional Depth Extension Networks
Figure 3 for ODE-CNN: Omnidirectional Depth Extension Networks
Figure 4 for ODE-CNN: Omnidirectional Depth Extension Networks
Viaarxiv icon