Alert button
Picture for Zechun Liu

Zechun Liu

Alert button

MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

Feb 22, 2024
Zechun Liu, Changsheng Zhao, Forrest Iandola, Chen Lai, Yuandong Tian, Igor Fedorov, Yunyang Xiong, Ernie Chang, Yangyang Shi, Raghuraman Krishnamoorthi, Liangzhen Lai, Vikas Chandra

Viaarxiv icon

On The Open Prompt Challenge In Conditional Audio Generation

Nov 01, 2023
Ernie Chang, Sidd Srinivasan, Mahi Luthra, Pin-Jie Lin, Varun Nagaraja, Forrest Iandola, Zechun Liu, Zhaoheng Ni, Changsheng Zhao, Yangyang Shi, Vikas Chandra

Viaarxiv icon

MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning

Oct 26, 2023
Jun Chen, Deyao Zhu, Xiaoqian Shen, Xiang Li, Zechun Liu, Pengchuan Zhang, Raghuraman Krishnamoorthi, Vikas Chandra, Yunyang Xiong, Mohamed Elhoseiny

Viaarxiv icon

LLM-FP4: 4-Bit Floating-Point Quantized Transformers

Oct 25, 2023
Shih-yang Liu, Zechun Liu, Xijie Huang, Pingcheng Dong, Kwang-Ting Cheng

Viaarxiv icon

Efficient Quantization-aware Training with Adaptive Coreset Selection

Jun 12, 2023
Xijie Huang, Zechun Liu, Shih-Yang Liu, Kwang-Ting Cheng

Figure 1 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Figure 2 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Figure 3 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Figure 4 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Viaarxiv icon

Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts

Jun 08, 2023
Ganesh Jawahar, Haichuan Yang, Yunyang Xiong, Zechun Liu, Dilin Wang, Fei Sun, Meng Li, Aasish Pappu, Barlas Oguz, Muhammad Abdul-Mageed, Laks V. S. Lakshmanan, Raghuraman Krishnamoorthi, Vikas Chandra

Figure 1 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Figure 2 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Figure 3 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Figure 4 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Viaarxiv icon

Binary and Ternary Natural Language Generation

Jun 02, 2023
Zechun Liu, Barlas Oguz, Aasish Pappu, Yangyang Shi, Raghuraman Krishnamoorthi

Figure 1 for Binary and Ternary Natural Language Generation
Figure 2 for Binary and Ternary Natural Language Generation
Figure 3 for Binary and Ternary Natural Language Generation
Figure 4 for Binary and Ternary Natural Language Generation
Viaarxiv icon

LLM-QAT: Data-Free Quantization Aware Training for Large Language Models

May 29, 2023
Zechun Liu, Barlas Oguz, Changsheng Zhao, Ernie Chang, Pierre Stock, Yashar Mehdad, Yangyang Shi, Raghuraman Krishnamoorthi, Vikas Chandra

Figure 1 for LLM-QAT: Data-Free Quantization Aware Training for Large Language Models
Figure 2 for LLM-QAT: Data-Free Quantization Aware Training for Large Language Models
Figure 3 for LLM-QAT: Data-Free Quantization Aware Training for Large Language Models
Figure 4 for LLM-QAT: Data-Free Quantization Aware Training for Large Language Models
Viaarxiv icon

EBSR: Enhanced Binary Neural Network for Image Super-Resolution

Mar 22, 2023
Renjie Wei, Shuwen Zhang, Zechun Liu, Meng Li, Yuchen Fan, Runsheng Wang, Ru Huang

Figure 1 for EBSR: Enhanced Binary Neural Network for Image Super-Resolution
Figure 2 for EBSR: Enhanced Binary Neural Network for Image Super-Resolution
Figure 3 for EBSR: Enhanced Binary Neural Network for Image Super-Resolution
Figure 4 for EBSR: Enhanced Binary Neural Network for Image Super-Resolution
Viaarxiv icon