Alert button
Picture for Yanqi Zhou

Yanqi Zhou

Alert button

LaMDA: Language Models for Dialog Applications

Add code
Bookmark button
Alert button
Jan 21, 2022
Romal Thoppilan, Daniel De Freitas, Jamie Hall, Noam Shazeer, Apoorv Kulshreshtha, Heng-Tze Cheng, Alicia Jin, Taylor Bos, Leslie Baker, Yu Du, YaGuang Li, Hongrae Lee, Huaixiu Steven Zheng, Amin Ghafouri, Marcelo Menegali, Yanping Huang, Maxim Krikun, Dmitry Lepikhin, James Qin, Dehao Chen, Yuanzhong Xu, Zhifeng Chen, Adam Roberts, Maarten Bosma, Yanqi Zhou, Chung-Ching Chang, Igor Krivokon, Will Rusch, Marc Pickett, Kathleen Meier-Hellstern, Meredith Ringel Morris, Tulsee Doshi, Renelito Delos Santos, Toju Duke, Johnny Soraker, Ben Zevenbergen, Vinodkumar Prabhakaran, Mark Diaz, Ben Hutchinson, Kristen Olson, Alejandra Molina, Erin Hoffman-John, Josh Lee, Lora Aroyo, Ravi Rajakumar, Alena Butryna, Matthew Lamm, Viktoriya Kuzmina, Joe Fenton, Aaron Cohen, Rachel Bernstein, Ray Kurzweil, Blaise Aguera-Arcas, Claire Cui, Marian Croak, Ed Chi, Quoc Le

Figure 1 for LaMDA: Language Models for Dialog Applications
Figure 2 for LaMDA: Language Models for Dialog Applications
Figure 3 for LaMDA: Language Models for Dialog Applications
Figure 4 for LaMDA: Language Models for Dialog Applications
Viaarxiv icon

GLaM: Efficient Scaling of Language Models with Mixture-of-Experts

Add code
Bookmark button
Alert button
Dec 13, 2021
Nan Du, Yanping Huang, Andrew M. Dai, Simon Tong, Dmitry Lepikhin, Yuanzhong Xu, Maxim Krikun, Yanqi Zhou, Adams Wei Yu, Orhan Firat, Barret Zoph, Liam Fedus, Maarten Bosma, Zongwei Zhou, Tao Wang, Yu Emma Wang, Kellie Webster, Marie Pellat, Kevin Robinson, Kathy Meier-Hellstern, Toju Duke, Lucas Dixon, Kun Zhang, Quoc V Le, Yonghui Wu, Zhifeng Chen, Claire Cui

Figure 1 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 2 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 3 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 4 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Viaarxiv icon

A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules

Add code
Bookmark button
Alert button
Dec 07, 2021
Xinfeng Xie, Prakash Prabhu, Ulysse Beaugnon, Phitchaya Mangpo Phothilimthana, Sudip Roy, Azalia Mirhoseini, Eugene Brevdo, James Laudon, Yanqi Zhou

Figure 1 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Figure 2 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Figure 3 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Figure 4 for A Transferable Approach for Partitioning Machine Learning Models on Multi-Chip-Modules
Viaarxiv icon

Do Transformer Modifications Transfer Across Implementations and Applications?

Add code
Bookmark button
Alert button
Feb 23, 2021
Sharan Narang, Hyung Won Chung, Yi Tay, William Fedus, Thibault Fevry, Michael Matena, Karishma Malkan, Noah Fiedel, Noam Shazeer, Zhenzhong Lan, Yanqi Zhou, Wei Li, Nan Ding, Jake Marcus, Adam Roberts, Colin Raffel

Figure 1 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 2 for Do Transformer Modifications Transfer Across Implementations and Applications?
Figure 3 for Do Transformer Modifications Transfer Across Implementations and Applications?
Viaarxiv icon

Rethinking Co-design of Neural Architectures and Hardware Accelerators

Add code
Bookmark button
Alert button
Feb 17, 2021
Yanqi Zhou, Xuanyi Dong, Berkin Akin, Mingxing Tan, Daiyi Peng, Tianjian Meng, Amir Yazdanbakhsh, Da Huang, Ravi Narayanaswami, James Laudon

Figure 1 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Figure 2 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Figure 3 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Figure 4 for Rethinking Co-design of Neural Architectures and Hardware Accelerators
Viaarxiv icon

Apollo: Transferable Architecture Exploration

Add code
Bookmark button
Alert button
Feb 02, 2021
Amir Yazdanbakhsh, Christof Angermueller, Berkin Akin, Yanqi Zhou, Albin Jones, Milad Hashemi, Kevin Swersky, Satrajit Chatterjee, Ravi Narayanaswami, James Laudon

Figure 1 for Apollo: Transferable Architecture Exploration
Figure 2 for Apollo: Transferable Architecture Exploration
Figure 3 for Apollo: Transferable Architecture Exploration
Figure 4 for Apollo: Transferable Architecture Exploration
Viaarxiv icon

Transferable Graph Optimizers for ML Compilers

Add code
Bookmark button
Alert button
Oct 21, 2020
Yanqi Zhou, Sudip Roy, Amirali Abdolrashidi, Daniel Wong, Peter Ma, Qiumin Xu, Hanxiao Liu, Mangpo Phitchaya Phothilimtha, Shen Wang, Anna Goldie, Azalia Mirhoseini, James Laudon

Figure 1 for Transferable Graph Optimizers for ML Compilers
Figure 2 for Transferable Graph Optimizers for ML Compilers
Figure 3 for Transferable Graph Optimizers for ML Compilers
Figure 4 for Transferable Graph Optimizers for ML Compilers
Viaarxiv icon

A Learned Performance Model for the Tensor Processing Unit

Add code
Bookmark button
Alert button
Aug 03, 2020
Samuel J. Kaufman, Phitchaya Mangpo Phothilimthana, Yanqi Zhou, Mike Burrows

Figure 1 for A Learned Performance Model for the Tensor Processing Unit
Figure 2 for A Learned Performance Model for the Tensor Processing Unit
Figure 3 for A Learned Performance Model for the Tensor Processing Unit
Figure 4 for A Learned Performance Model for the Tensor Processing Unit
Viaarxiv icon

ODE-CNN: Omnidirectional Depth Extension Networks

Add code
Bookmark button
Alert button
Jul 03, 2020
Xinjing Cheng, Peng Wang, Yanqi Zhou, Chenye Guan, Ruigang Yang

Figure 1 for ODE-CNN: Omnidirectional Depth Extension Networks
Figure 2 for ODE-CNN: Omnidirectional Depth Extension Networks
Figure 3 for ODE-CNN: Omnidirectional Depth Extension Networks
Figure 4 for ODE-CNN: Omnidirectional Depth Extension Networks
Viaarxiv icon

Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer

Add code
Bookmark button
Alert button
Oct 24, 2019
Colin Raffel, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, Yanqi Zhou, Wei Li, Peter J. Liu

Figure 1 for Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Figure 2 for Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Figure 3 for Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Figure 4 for Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer
Viaarxiv icon