Alert button
Picture for Binhang Yuan

Binhang Yuan

Alert button

Exploring the Robustness of Decentralized Training for Large Language Models

Dec 01, 2023
Lin Lu, Chenxi Dai, Wangcheng Tao, Binhang Yuan, Yanan Sun, Pan Zhou

Viaarxiv icon

Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time

Oct 26, 2023
Zichang Liu, Jue Wang, Tri Dao, Tianyi Zhou, Binhang Yuan, Zhao Song, Anshumali Shrivastava, Ce Zhang, Yuandong Tian, Christopher Re, Beidi Chen

Figure 1 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 2 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 3 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 4 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Viaarxiv icon

Serving Deep Learning Model in Relational Databases

Oct 10, 2023
Alexandre Eichenberger, Qi Lin, Saif Masood, Hong Min, Alexander Sim, Jie Wang, Yida Wang, Kesheng Wu, Binhang Yuan, Lixi Zhou, Jia Zou

Figure 1 for Serving Deep Learning Model in Relational Databases
Figure 2 for Serving Deep Learning Model in Relational Databases
Figure 3 for Serving Deep Learning Model in Relational Databases
Figure 4 for Serving Deep Learning Model in Relational Databases
Viaarxiv icon

Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning

Jun 07, 2023
Yuxin Tang, Zhimin Ding, Dimitrije Jankov, Binhang Yuan, Daniel Bourgeois, Chris Jermaine

Figure 1 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Figure 2 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Figure 3 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Figure 4 for Auto-Differentiation of Relational Computations for Very Large Scale Machine Learning
Viaarxiv icon

High-throughput Generative Inference of Large Language Models with a Single GPU

Mar 13, 2023
Ying Sheng, Lianmin Zheng, Binhang Yuan, Zhuohan Li, Max Ryabinin, Daniel Y. Fu, Zhiqiang Xie, Beidi Chen, Clark Barrett, Joseph E. Gonzalez, Percy Liang, Christopher Ré, Ion Stoica, Ce Zhang

Figure 1 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 2 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 3 for High-throughput Generative Inference of Large Language Models with a Single GPU
Figure 4 for High-throughput Generative Inference of Large Language Models with a Single GPU
Viaarxiv icon

Holistic Evaluation of Language Models

Nov 16, 2022
Percy Liang, Rishi Bommasani, Tony Lee, Dimitris Tsipras, Dilara Soylu, Michihiro Yasunaga, Yian Zhang, Deepak Narayanan, Yuhuai Wu, Ananya Kumar, Benjamin Newman, Binhang Yuan, Bobby Yan, Ce Zhang, Christian Cosgrove, Christopher D. Manning, Christopher Ré, Diana Acosta-Navas, Drew A. Hudson, Eric Zelikman, Esin Durmus, Faisal Ladhak, Frieda Rong, Hongyu Ren, Huaxiu Yao, Jue Wang, Keshav Santhanam, Laurel Orr, Lucia Zheng, Mert Yuksekgonul, Mirac Suzgun, Nathan Kim, Neel Guha, Niladri Chatterji, Omar Khattab, Peter Henderson, Qian Huang, Ryan Chi, Sang Michael Xie, Shibani Santurkar, Surya Ganguli, Tatsunori Hashimoto, Thomas Icard, Tianyi Zhang, Vishrav Chaudhary, William Wang, Xuechen Li, Yifan Mai, Yuhui Zhang, Yuta Koreeda

Figure 1 for Holistic Evaluation of Language Models
Figure 2 for Holistic Evaluation of Language Models
Figure 3 for Holistic Evaluation of Language Models
Figure 4 for Holistic Evaluation of Language Models
Viaarxiv icon

Stochastic Gradient Descent without Full Data Shuffle

Jun 12, 2022
Lijie Xu, Shuang Qiu, Binhang Yuan, Jiawei Jiang, Cedric Renggli, Shaoduo Gan, Kaan Kara, Guoliang Li, Ji Liu, Wentao Wu, Jieping Ye, Ce Zhang

Figure 1 for Stochastic Gradient Descent without Full Data Shuffle
Figure 2 for Stochastic Gradient Descent without Full Data Shuffle
Figure 3 for Stochastic Gradient Descent without Full Data Shuffle
Figure 4 for Stochastic Gradient Descent without Full Data Shuffle
Viaarxiv icon

Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees

Jun 02, 2022
Jue Wang, Binhang Yuan, Luka Rimanic, Yongjun He, Tri Dao, Beidi Chen, Christopher Re, Ce Zhang

Figure 1 for Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees
Figure 2 for Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees
Figure 3 for Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees
Figure 4 for Fine-tuning Language Models over Slow Networks using Activation Compression with Guarantees
Viaarxiv icon

Decentralized Training of Foundation Models in Heterogeneous Environments

Jun 02, 2022
Binhang Yuan, Yongjun He, Jared Quincy Davis, Tianyi Zhang, Tri Dao, Beidi Chen, Percy Liang, Christopher Re, Ce Zhang

Figure 1 for Decentralized Training of Foundation Models in Heterogeneous Environments
Figure 2 for Decentralized Training of Foundation Models in Heterogeneous Environments
Figure 3 for Decentralized Training of Foundation Models in Heterogeneous Environments
Figure 4 for Decentralized Training of Foundation Models in Heterogeneous Environments
Viaarxiv icon

Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters

Nov 23, 2021
Xiangru Lian, Binhang Yuan, Xuefeng Zhu, Yulong Wang, Yongjun He, Honghuan Wu, Lei Sun, Haodong Lyu, Chengjun Liu, Xing Dong, Yiqiao Liao, Mingnan Luo, Congfei Zhang, Jingru Xie, Haonan Li, Lei Chen, Renjie Huang, Jianying Lin, Chengchun Shu, Xuezhong Qiu, Zhishan Liu, Dongying Kong, Lei Yuan, Hai Yu, Sen Yang, Ce Zhang, Ji Liu

Figure 1 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Figure 2 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Figure 3 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Figure 4 for Persia: An Open, Hybrid System Scaling Deep Learning-based Recommenders up to 100 Trillion Parameters
Viaarxiv icon