Picture for Decheng Wu

Decheng Wu

Sherry: Hardware-Efficient 1.25-Bit Ternary Quantization via Fine-grained Sparsification

Add code
Jan 12, 2026
Viaarxiv icon

Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought

Add code
May 21, 2025
Viaarxiv icon

Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent

Add code
Nov 05, 2024
Figure 1 for Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent
Figure 2 for Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent
Figure 3 for Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent
Figure 4 for Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent
Viaarxiv icon

EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs

Add code
Mar 05, 2024
Figure 1 for EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs
Figure 2 for EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs
Figure 3 for EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs
Figure 4 for EasyQuant: An Efficient Data-free Quantization Algorithm for LLMs
Viaarxiv icon