Picture for Zhuocheng Gong

Zhuocheng Gong

Mixture-of-Modules: Reinventing Transformers as Dynamic Assemblies of Modules

Add code
Jul 09, 2024
Viaarxiv icon

Parallel Decoding via Hidden Transfer for Lossless Large Language Model Acceleration

Add code
Apr 18, 2024
Figure 1 for Parallel Decoding via Hidden Transfer for Lossless Large Language Model Acceleration
Figure 2 for Parallel Decoding via Hidden Transfer for Lossless Large Language Model Acceleration
Figure 3 for Parallel Decoding via Hidden Transfer for Lossless Large Language Model Acceleration
Figure 4 for Parallel Decoding via Hidden Transfer for Lossless Large Language Model Acceleration
Viaarxiv icon

What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation

Add code
Mar 11, 2024
Figure 1 for What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation
Figure 2 for What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation
Figure 3 for What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation
Figure 4 for What Makes Quantization for Large Language Models Hard? An Empirical Study from the Lens of Perturbation
Viaarxiv icon

Improving Input-label Mapping with Demonstration Replay for In-context Learning

Add code
Oct 30, 2023
Figure 1 for Improving Input-label Mapping with Demonstration Replay for In-context Learning
Figure 2 for Improving Input-label Mapping with Demonstration Replay for In-context Learning
Figure 3 for Improving Input-label Mapping with Demonstration Replay for In-context Learning
Figure 4 for Improving Input-label Mapping with Demonstration Replay for In-context Learning
Viaarxiv icon

PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models

Add code
May 30, 2023
Figure 1 for PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models
Figure 2 for PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models
Figure 3 for PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models
Figure 4 for PreQuant: A Task-agnostic Quantization Approach for Pre-trained Language Models
Viaarxiv icon