Alert button
Picture for Barret Zoph

Barret Zoph

Alert button

PaLM: Scaling Language Modeling with Pathways

Add code
Bookmark button
Alert button
Apr 07, 2022
Aakanksha Chowdhery, Sharan Narang, Jacob Devlin, Maarten Bosma, Gaurav Mishra, Adam Roberts, Paul Barham, Hyung Won Chung, Charles Sutton, Sebastian Gehrmann, Parker Schuh, Kensen Shi, Sasha Tsvyashchenko, Joshua Maynez, Abhishek Rao, Parker Barnes, Yi Tay, Noam Shazeer, Vinodkumar Prabhakaran, Emily Reif, Nan Du, Ben Hutchinson, Reiner Pope, James Bradbury, Jacob Austin, Michael Isard, Guy Gur-Ari, Pengcheng Yin, Toju Duke, Anselm Levskaya, Sanjay Ghemawat, Sunipa Dev, Henryk Michalewski, Xavier Garcia, Vedant Misra, Kevin Robinson, Liam Fedus, Denny Zhou, Daphne Ippolito, David Luan, Hyeontaek Lim, Barret Zoph, Alexander Spiridonov, Ryan Sepassi, David Dohan, Shivani Agrawal, Mark Omernick, Andrew M. Dai, Thanumalayan Sankaranarayana Pillai, Marie Pellat, Aitor Lewkowycz, Erica Moreira, Rewon Child, Oleksandr Polozov, Katherine Lee, Zongwei Zhou, Xuezhi Wang, Brennan Saeta, Mark Diaz, Orhan Firat, Michele Catasta, Jason Wei, Kathy Meier-Hellstern, Douglas Eck, Jeff Dean, Slav Petrov, Noah Fiedel

Figure 1 for PaLM: Scaling Language Modeling with Pathways
Figure 2 for PaLM: Scaling Language Modeling with Pathways
Figure 3 for PaLM: Scaling Language Modeling with Pathways
Figure 4 for PaLM: Scaling Language Modeling with Pathways
Viaarxiv icon

Designing Effective Sparse Expert Models

Add code
Bookmark button
Alert button
Feb 17, 2022
Barret Zoph, Irwan Bello, Sameer Kumar, Nan Du, Yanping Huang, Jeff Dean, Noam Shazeer, William Fedus

Figure 1 for Designing Effective Sparse Expert Models
Figure 2 for Designing Effective Sparse Expert Models
Figure 3 for Designing Effective Sparse Expert Models
Figure 4 for Designing Effective Sparse Expert Models
Viaarxiv icon

GLaM: Efficient Scaling of Language Models with Mixture-of-Experts

Add code
Bookmark button
Alert button
Dec 13, 2021
Nan Du, Yanping Huang, Andrew M. Dai, Simon Tong, Dmitry Lepikhin, Yuanzhong Xu, Maxim Krikun, Yanqi Zhou, Adams Wei Yu, Orhan Firat, Barret Zoph, Liam Fedus, Maarten Bosma, Zongwei Zhou, Tao Wang, Yu Emma Wang, Kellie Webster, Marie Pellat, Kevin Robinson, Kathy Meier-Hellstern, Toju Duke, Lucas Dixon, Kun Zhang, Quoc V Le, Yonghui Wu, Zhifeng Chen, Claire Cui

Figure 1 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 2 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 3 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 4 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Viaarxiv icon

Multi-Task Self-Training for Learning General Representations

Add code
Bookmark button
Alert button
Aug 25, 2021
Golnaz Ghiasi, Barret Zoph, Ekin D. Cubuk, Quoc V. Le, Tsung-Yi Lin

Figure 1 for Multi-Task Self-Training for Learning General Representations
Figure 2 for Multi-Task Self-Training for Learning General Representations
Figure 3 for Multi-Task Self-Training for Learning General Representations
Figure 4 for Multi-Task Self-Training for Learning General Representations
Viaarxiv icon

Simple Training Strategies and Model Scaling for Object Detection

Add code
Bookmark button
Alert button
Jun 30, 2021
Xianzhi Du, Barret Zoph, Wei-Chih Hung, Tsung-Yi Lin

Figure 1 for Simple Training Strategies and Model Scaling for Object Detection
Figure 2 for Simple Training Strategies and Model Scaling for Object Detection
Figure 3 for Simple Training Strategies and Model Scaling for Object Detection
Figure 4 for Simple Training Strategies and Model Scaling for Object Detection
Viaarxiv icon

Revisiting ResNets: Improved Training and Scaling Strategies

Add code
Bookmark button
Alert button
Mar 13, 2021
Irwan Bello, William Fedus, Xianzhi Du, Ekin D. Cubuk, Aravind Srinivas, Tsung-Yi Lin, Jonathon Shlens, Barret Zoph

Figure 1 for Revisiting ResNets: Improved Training and Scaling Strategies
Figure 2 for Revisiting ResNets: Improved Training and Scaling Strategies
Figure 3 for Revisiting ResNets: Improved Training and Scaling Strategies
Figure 4 for Revisiting ResNets: Improved Training and Scaling Strategies
Viaarxiv icon

Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity

Add code
Bookmark button
Alert button
Jan 11, 2021
William Fedus, Barret Zoph, Noam Shazeer

Figure 1 for Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
Figure 2 for Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
Figure 3 for Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
Figure 4 for Switch Transformers: Scaling to Trillion Parameter Models with Simple and Efficient Sparsity
Viaarxiv icon

Simple Copy-Paste is a Strong Data Augmentation Method for Instance Segmentation

Add code
Bookmark button
Alert button
Dec 13, 2020
Golnaz Ghiasi, Yin Cui, Aravind Srinivas, Rui Qian, Tsung-Yi Lin, Ekin D. Cubuk, Quoc V. Le, Barret Zoph

Figure 1 for Simple Copy-Paste is a Strong Data Augmentation Method for Instance Segmentation
Figure 2 for Simple Copy-Paste is a Strong Data Augmentation Method for Instance Segmentation
Figure 3 for Simple Copy-Paste is a Strong Data Augmentation Method for Instance Segmentation
Figure 4 for Simple Copy-Paste is a Strong Data Augmentation Method for Instance Segmentation
Viaarxiv icon

Does Data Augmentation Benefit from Split BatchNorms

Add code
Bookmark button
Alert button
Oct 15, 2020
Amil Merchant, Barret Zoph, Ekin Dogus Cubuk

Figure 1 for Does Data Augmentation Benefit from Split BatchNorms
Figure 2 for Does Data Augmentation Benefit from Split BatchNorms
Figure 3 for Does Data Augmentation Benefit from Split BatchNorms
Figure 4 for Does Data Augmentation Benefit from Split BatchNorms
Viaarxiv icon

Rethinking Pre-training and Self-training

Add code
Bookmark button
Alert button
Jun 11, 2020
Barret Zoph, Golnaz Ghiasi, Tsung-Yi Lin, Yin Cui, Hanxiao Liu, Ekin D. Cubuk, Quoc V. Le

Figure 1 for Rethinking Pre-training and Self-training
Figure 2 for Rethinking Pre-training and Self-training
Figure 3 for Rethinking Pre-training and Self-training
Figure 4 for Rethinking Pre-training and Self-training
Viaarxiv icon