Alert button
Picture for Yuzhen Huang

Yuzhen Huang

Alert button

Compression Represents Intelligence Linearly

Add code
Bookmark button
Alert button
Apr 15, 2024
Yuzhen Huang, Jinghan Zhang, Zifei Shan, Junxian He

Viaarxiv icon

C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models

Add code
Bookmark button
Alert button
May 17, 2023
Yuzhen Huang, Yuzhuo Bai, Zhihao Zhu, Junlei Zhang, Jinghan Zhang, Tangjun Su, Junteng Liu, Chuancheng Lv, Yikai Zhang, Jiayi Lei, Yao Fu, Maosong Sun, Junxian He

Figure 1 for C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models
Figure 2 for C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models
Figure 3 for C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models
Figure 4 for C-Eval: A Multi-Level Multi-Discipline Chinese Evaluation Suite for Foundation Models
Viaarxiv icon

Pre-train and Search: Efficient Embedding Table Sharding with Pre-trained Neural Cost Models

Add code
Bookmark button
Alert button
May 03, 2023
Daochen Zha, Louis Feng, Liang Luo, Bhargav Bhushanam, Zirui Liu, Yusuo Hu, Jade Nie, Yuzhen Huang, Yuandong Tian, Arun Kejariwal, Xia Hu

Figure 1 for Pre-train and Search: Efficient Embedding Table Sharding with Pre-trained Neural Cost Models
Figure 2 for Pre-train and Search: Efficient Embedding Table Sharding with Pre-trained Neural Cost Models
Figure 3 for Pre-train and Search: Efficient Embedding Table Sharding with Pre-trained Neural Cost Models
Figure 4 for Pre-train and Search: Efficient Embedding Table Sharding with Pre-trained Neural Cost Models
Viaarxiv icon

Fast Distributed Training of Deep Neural Networks: Dynamic Communication Thresholding for Model and Data Parallelism

Add code
Bookmark button
Alert button
Oct 18, 2020
Vipul Gupta, Dhruv Choudhary, Ping Tak Peter Tang, Xiaohan Wei, Xing Wang, Yuzhen Huang, Arun Kejariwal, Kannan Ramchandran, Michael W. Mahoney

Figure 1 for Fast Distributed Training of Deep Neural Networks: Dynamic Communication Thresholding for Model and Data Parallelism
Figure 2 for Fast Distributed Training of Deep Neural Networks: Dynamic Communication Thresholding for Model and Data Parallelism
Figure 3 for Fast Distributed Training of Deep Neural Networks: Dynamic Communication Thresholding for Model and Data Parallelism
Figure 4 for Fast Distributed Training of Deep Neural Networks: Dynamic Communication Thresholding for Model and Data Parallelism
Viaarxiv icon

TensorOpt: Exploring the Tradeoffs in Distributed DNN Training with Auto-Parallelism

Add code
Bookmark button
Alert button
Apr 16, 2020
Zhenkun Cai, Kaihao Ma, Xiao Yan, Yidi Wu, Yuzhen Huang, James Cheng, Teng Su, Fan Yu

Figure 1 for TensorOpt: Exploring the Tradeoffs in Distributed DNN Training with Auto-Parallelism
Figure 2 for TensorOpt: Exploring the Tradeoffs in Distributed DNN Training with Auto-Parallelism
Figure 3 for TensorOpt: Exploring the Tradeoffs in Distributed DNN Training with Auto-Parallelism
Figure 4 for TensorOpt: Exploring the Tradeoffs in Distributed DNN Training with Auto-Parallelism
Viaarxiv icon