Picture for Daning Cheng

Daning Cheng

Rethinking Parameter Sharing as Graph Coloring for Structured Compression

Add code
Nov 10, 2025
Viaarxiv icon

Can the capability of Large Language Models be described by human ability? A Meta Study

Add code
Apr 13, 2025
Figure 1 for Can the capability of Large Language Models be described by human ability? A Meta Study
Figure 2 for Can the capability of Large Language Models be described by human ability? A Meta Study
Figure 3 for Can the capability of Large Language Models be described by human ability? A Meta Study
Figure 4 for Can the capability of Large Language Models be described by human ability? A Meta Study
Viaarxiv icon

A General Error-Theoretical Analysis Framework for Constructing Compression Strategies

Add code
Feb 19, 2025
Viaarxiv icon

Compression for Better: A General and Stable Lossless Compression Framework

Add code
Dec 09, 2024
Viaarxiv icon

Lossless Model Compression via Joint Low-Rank Factorization Optimization

Add code
Dec 09, 2024
Viaarxiv icon

FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantization

Add code
Dec 09, 2024
Figure 1 for FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantization
Figure 2 for FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantization
Figure 3 for FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantization
Figure 4 for FP=xINT:A Low-Bit Series Expansion Algorithm for Post-Training Quantization
Viaarxiv icon

Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss

Add code
Jul 20, 2022
Figure 1 for Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss
Figure 2 for Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss
Figure 3 for Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss
Figure 4 for Mixed-Precision Inference Quantization: Radically Towards Faster inference speed, Lower Storage requirement, and Lower Loss
Viaarxiv icon

Quantization in Layer's Input is Matter

Add code
Feb 10, 2022
Viaarxiv icon