Alert button
Picture for Yu Emma Wang

Yu Emma Wang

Alert button

Hadamard Domain Training with Integers for Class Incremental Quantized Learning

Add code
Bookmark button
Alert button
Oct 05, 2023
Martin Schiemer, Clemens JS Schaefer, Jayden Parker Vap, Mark James Horeni, Yu Emma Wang, Juan Ye, Siddharth Joshi

Viaarxiv icon

Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization

Add code
Bookmark button
Alert button
Jun 08, 2023
Clemens JS Schaefer, Navid Lambert-Shirzad, Xiaofan Zhang, Chiachen Chou, Tom Jablin, Jian Li, Elfie Guo, Caitlin Stanton, Siddharth Joshi, Yu Emma Wang

Figure 1 for Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization
Figure 2 for Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization
Figure 3 for Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization
Figure 4 for Augmenting Hessians with Inter-Layer Dependencies for Mixed-Precision Post-Training Quantization
Viaarxiv icon

Mixed Precision Post Training Quantization of Neural Networks with Sensitivity Guided Search

Add code
Bookmark button
Alert button
Feb 07, 2023
Clemens JS Schaefer, Elfie Guo, Caitlin Stanton, Xiaofan Zhang, Tom Jablin, Navid Lambert-Shirzad, Jian Li, Chiachen Chou, Siddharth Joshi, Yu Emma Wang

Figure 1 for Mixed Precision Post Training Quantization of Neural Networks with Sensitivity Guided Search
Figure 2 for Mixed Precision Post Training Quantization of Neural Networks with Sensitivity Guided Search
Figure 3 for Mixed Precision Post Training Quantization of Neural Networks with Sensitivity Guided Search
Figure 4 for Mixed Precision Post Training Quantization of Neural Networks with Sensitivity Guided Search
Viaarxiv icon

AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language Models

Add code
Bookmark button
Alert button
Jan 21, 2022
Xiaofan Zhang, Zongwei Zhou, Deming Chen, Yu Emma Wang

Figure 1 for AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language Models
Figure 2 for AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language Models
Figure 3 for AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language Models
Figure 4 for AutoDistill: an End-to-End Framework to Explore and Distill Hardware-Efficient Language Models
Viaarxiv icon

GLaM: Efficient Scaling of Language Models with Mixture-of-Experts

Add code
Bookmark button
Alert button
Dec 13, 2021
Nan Du, Yanping Huang, Andrew M. Dai, Simon Tong, Dmitry Lepikhin, Yuanzhong Xu, Maxim Krikun, Yanqi Zhou, Adams Wei Yu, Orhan Firat, Barret Zoph, Liam Fedus, Maarten Bosma, Zongwei Zhou, Tao Wang, Yu Emma Wang, Kellie Webster, Marie Pellat, Kevin Robinson, Kathy Meier-Hellstern, Toju Duke, Lucas Dixon, Kun Zhang, Quoc V Le, Yonghui Wu, Zhifeng Chen, Claire Cui

Figure 1 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 2 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 3 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Figure 4 for GLaM: Efficient Scaling of Language Models with Mixture-of-Experts
Viaarxiv icon

Exploring the limits of Concurrency in ML Training on Google TPUs

Add code
Bookmark button
Alert button
Nov 07, 2020
Sameer Kumar, James Bradbury, Cliff Young, Yu Emma Wang, Anselm Levskaya, Blake Hechtman, Dehao Chen, HyoukJoong Lee, Mehmet Deveci, Naveen Kumar, Pankaj Kanwar, Shibo Wang, Skye Wanderman-Milne, Steve Lacy, Tao Wang, Tayo Oguntebi, Yazhou Zu, Yuanzhong Xu, Andy Swing

Figure 1 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 2 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 3 for Exploring the limits of Concurrency in ML Training on Google TPUs
Figure 4 for Exploring the limits of Concurrency in ML Training on Google TPUs
Viaarxiv icon

Exploiting Parallelism Opportunities with Deep Learning Frameworks

Add code
Bookmark button
Alert button
Aug 13, 2019
Yu Emma Wang, Carole-Jean Wu, Xiaodong Wang, Kim Hazelwood, David Brooks

Figure 1 for Exploiting Parallelism Opportunities with Deep Learning Frameworks
Figure 2 for Exploiting Parallelism Opportunities with Deep Learning Frameworks
Figure 3 for Exploiting Parallelism Opportunities with Deep Learning Frameworks
Figure 4 for Exploiting Parallelism Opportunities with Deep Learning Frameworks
Viaarxiv icon

Benchmarking TPU, GPU, and CPU Platforms for Deep Learning

Add code
Bookmark button
Alert button
Aug 06, 2019
Yu Emma Wang, Gu-Yeon Wei, David Brooks

Figure 1 for Benchmarking TPU, GPU, and CPU Platforms for Deep Learning
Figure 2 for Benchmarking TPU, GPU, and CPU Platforms for Deep Learning
Figure 3 for Benchmarking TPU, GPU, and CPU Platforms for Deep Learning
Figure 4 for Benchmarking TPU, GPU, and CPU Platforms for Deep Learning
Viaarxiv icon