Alert button
Picture for Cliff Young

Cliff Young

Alert button

TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings

Apr 20, 2023
Norman P. Jouppi, George Kurian, Sheng Li, Peter Ma, Rahul Nagarajan, Lifeng Nai, Nishant Patil, Suvinay Subramanian, Andy Swing, Brian Towles, Cliff Young, Xiang Zhou, Zongwei Zhou, David Patterson

Figure 1 for TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings
Figure 2 for TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings
Figure 3 for TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings
Figure 4 for TPU v4: An Optically Reconfigurable Supercomputer for Machine Learning with Hardware Support for Embeddings
Viaarxiv icon

MegaBlocks: Efficient Sparse Training with Mixture-of-Experts

Nov 29, 2022
Trevor Gale, Deepak Narayanan, Cliff Young, Matei Zaharia

Figure 1 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Figure 2 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Figure 3 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Figure 4 for MegaBlocks: Efficient Sparse Training with Mixture-of-Experts
Viaarxiv icon

Exploring the limits of Concurrency in ML Training on Google TPUs

Nov 07, 2020
Sameer Kumar, James Bradbury, Cliff Young, Yu Emma Wang, Anselm Levskaya, Blake Hechtman, Dehao Chen, HyoukJoong Lee, Mehmet Deveci, Naveen Kumar, Pankaj Kanwar, Shibo Wang, Skye Wanderman-Milne, Steve Lacy, Tao Wang, Tayo Oguntebi, Yazhou Zu, Yuanzhong Xu, Andy Swing

Figure 1 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 2 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 3 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 4 for Exploring the limits of Concurrency in ML Training on Google TPUs
Viaarxiv icon

Sparse GPU Kernels for Deep Learning

Jun 18, 2020
Trevor Gale, Matei Zaharia, Cliff Young, Erich Elsen

Figure 1 for Sparse GPU Kernels for Deep Learning
Figure 2 for Sparse GPU Kernels for Deep Learning
Figure 3 for Sparse GPU Kernels for Deep Learning
Figure 4 for Sparse GPU Kernels for Deep Learning
Viaarxiv icon

Bit-Parallel Vector Composability for Neural Acceleration

Apr 11, 2020
Soroush Ghodrati, Hardik Sharma, Cliff Young, Nam Sung Kim, Hadi Esmaeilzadeh

Figure 1 for Bit-Parallel Vector Composability for Neural Acceleration
Figure 2 for Bit-Parallel Vector Composability for Neural Acceleration
Figure 3 for Bit-Parallel Vector Composability for Neural Acceleration
Figure 4 for Bit-Parallel Vector Composability for Neural Acceleration
Viaarxiv icon

MLPerf Training Benchmark

Oct 30, 2019
Peter Mattson, Christine Cheng, Cody Coleman, Greg Diamos, Paulius Micikevicius, David Patterson, Hanlin Tang, Gu-Yeon Wei, Peter Bailis, Victor Bittorf, David Brooks, Dehao Chen, Debojyoti Dutta, Udit Gupta, Kim Hazelwood, Andrew Hock, Xinyuan Huang, Bill Jia, Daniel Kang, David Kanter, Naveen Kumar, Jeffery Liao, Guokai Ma, Deepak Narayanan, Tayo Oguntebi, Gennady Pekhimenko, Lillian Pentecost, Vijay Janapa Reddi, Taylor Robie, Tom St. John, Carole-Jean Wu, Lingjie Xu, Cliff Young, Matei Zaharia

Figure 1 for MLPerf Training Benchmark
Figure 2 for MLPerf Training Benchmark
Figure 3 for MLPerf Training Benchmark
Figure 4 for MLPerf Training Benchmark
Viaarxiv icon

Mesh-TensorFlow: Deep Learning for Supercomputers

Nov 05, 2018
Noam Shazeer, Youlong Cheng, Niki Parmar, Dustin Tran, Ashish Vaswani, Penporn Koanantakool, Peter Hawkins, HyoukJoong Lee, Mingsheng Hong, Cliff Young, Ryan Sepassi, Blake Hechtman

Figure 1 for Mesh-TensorFlow: Deep Learning for Supercomputers
Figure 2 for Mesh-TensorFlow: Deep Learning for Supercomputers
Figure 3 for Mesh-TensorFlow: Deep Learning for Supercomputers
Viaarxiv icon