Alert button
Picture for Anjali Sridhar

Anjali Sridhar

Alert button

Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference

Add code
Bookmark button
Alert button
Mar 10, 2023
Haiyang Huang, Newsha Ardalani, Anna Sun, Liu Ke, Hsien-Hsin S. Lee, Anjali Sridhar, Shruti Bhosale, Carole-Jean Wu, Benjamin Lee

Figure 1 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 2 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 3 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Figure 4 for Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference
Viaarxiv icon

OPT: Open Pre-trained Transformer Language Models

Add code
Bookmark button
Alert button
May 05, 2022
Susan Zhang, Stephen Roller, Naman Goyal, Mikel Artetxe, Moya Chen, Shuohui Chen, Christopher Dewan, Mona Diab, Xian Li, Xi Victoria Lin, Todor Mihaylov, Myle Ott, Sam Shleifer, Kurt Shuster, Daniel Simig, Punit Singh Koura, Anjali Sridhar, Tianlu Wang, Luke Zettlemoyer

Figure 1 for OPT: Open Pre-trained Transformer Language Models
Figure 2 for OPT: Open Pre-trained Transformer Language Models
Figure 3 for OPT: Open Pre-trained Transformer Language Models
Figure 4 for OPT: Open Pre-trained Transformer Language Models
Viaarxiv icon