Picture for Zechun Liu

Zechun Liu

RoLoRA: Fine-tuning Rotated Outlier-free LLMs for Effective Weight-Activation Quantization

Add code
Jul 10, 2024
Viaarxiv icon

SpinQuant: LLM quantization with learned rotations

Add code
May 28, 2024
Viaarxiv icon

An Introduction to Vision-Language Modeling

Add code
May 27, 2024
Viaarxiv icon

MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

Add code
Feb 22, 2024
Figure 1 for MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases
Figure 2 for MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases
Figure 3 for MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases
Figure 4 for MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases
Viaarxiv icon

On The Open Prompt Challenge In Conditional Audio Generation

Add code
Nov 01, 2023
Figure 1 for On The Open Prompt Challenge In Conditional Audio Generation
Figure 2 for On The Open Prompt Challenge In Conditional Audio Generation
Figure 3 for On The Open Prompt Challenge In Conditional Audio Generation
Figure 4 for On The Open Prompt Challenge In Conditional Audio Generation
Viaarxiv icon

MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning

Add code
Oct 26, 2023
Figure 1 for MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning
Figure 2 for MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning
Figure 3 for MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning
Figure 4 for MiniGPT-v2: large language model as a unified interface for vision-language multi-task learning
Viaarxiv icon

LLM-FP4: 4-Bit Floating-Point Quantized Transformers

Add code
Oct 25, 2023
Figure 1 for LLM-FP4: 4-Bit Floating-Point Quantized Transformers
Figure 2 for LLM-FP4: 4-Bit Floating-Point Quantized Transformers
Figure 3 for LLM-FP4: 4-Bit Floating-Point Quantized Transformers
Figure 4 for LLM-FP4: 4-Bit Floating-Point Quantized Transformers
Viaarxiv icon

Efficient Quantization-aware Training with Adaptive Coreset Selection

Add code
Jun 12, 2023
Figure 1 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Figure 2 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Figure 3 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Figure 4 for Efficient Quantization-aware Training with Adaptive Coreset Selection
Viaarxiv icon

Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts

Add code
Jun 08, 2023
Figure 1 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Figure 2 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Figure 3 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Figure 4 for Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts
Viaarxiv icon

Binary and Ternary Natural Language Generation

Add code
Jun 02, 2023
Figure 1 for Binary and Ternary Natural Language Generation
Figure 2 for Binary and Ternary Natural Language Generation
Figure 3 for Binary and Ternary Natural Language Generation
Figure 4 for Binary and Ternary Natural Language Generation
Viaarxiv icon