Alert button
Picture for Beidi Chen

Beidi Chen

Alert button

TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding

Add code
Bookmark button
Alert button
Apr 18, 2024
Hanshi Sun, Zhuoming Chen, Xinyu Yang, Yuandong Tian, Beidi Chen

Viaarxiv icon

Megalodon: Efficient LLM Pretraining and Inference with Unlimited Context Length

Add code
Bookmark button
Alert button
Apr 16, 2024
Xuezhe Ma, Xiaomeng Yang, Wenhan Xiong, Beidi Chen, Lili Yu, Hao Zhang, Jonathan May, Luke Zettlemoyer, Omer Levy, Chunting Zhou

Viaarxiv icon

Prompt-prompted Mixture of Experts for Efficient LLM Generation

Add code
Bookmark button
Alert button
Apr 05, 2024
Harry Dong, Beidi Chen, Yuejie Chi

Viaarxiv icon

LLM Inference Unveiled: Survey and Roofline Model Insights

Add code
Bookmark button
Alert button
Mar 11, 2024
Zhihang Yuan, Yuzhang Shang, Yang Zhou, Zhen Dong, Zhe Zhou, Chenhao Xue, Bingzhe Wu, Zhikai Li, Qingyi Gu, Yong Jae Lee, Yan Yan, Beidi Chen, Guangyu Sun, Kurt Keutzer

Viaarxiv icon

GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection

Add code
Bookmark button
Alert button
Mar 06, 2024
Jiawei Zhao, Zhenyu Zhang, Beidi Chen, Zhangyang Wang, Anima Anandkumar, Yuandong Tian

Figure 1 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Figure 2 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Figure 3 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Figure 4 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Viaarxiv icon

Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding

Add code
Bookmark button
Alert button
Mar 05, 2024
Zhenyu Zhang, Runjin Chen, Shiwei Liu, Zhewei Yao, Olatunji Ruwase, Beidi Chen, Xiaoxia Wu, Zhangyang Wang

Figure 1 for Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding
Figure 2 for Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding
Figure 3 for Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding
Figure 4 for Found in the Middle: How Language Models Use Long Contexts Better via Plug-and-Play Positional Encoding
Viaarxiv icon

Sequoia: Scalable, Robust, and Hardware-aware Speculative Decoding

Add code
Bookmark button
Alert button
Feb 29, 2024
Zhuoming Chen, Avner May, Ruslan Svirschevski, Yuhsun Huang, Max Ryabinin, Zhihao Jia, Beidi Chen

Viaarxiv icon