Picture for Yunhe Wang

Yunhe Wang

and Other Contributors

Revealing the Power of Post-Training for Small Language Models via Knowledge Distillation

Add code
Sep 30, 2025
Viaarxiv icon

Rethinking 1-bit Optimization Leveraging Pre-trained Large Language Models

Add code
Aug 09, 2025
Viaarxiv icon

EAQuant: Enhancing Post-Training Quantization for MoE Models via Expert-Aware Optimization

Add code
Jun 16, 2025
Viaarxiv icon

Pangu DeepDiver: Adaptive Search Intensity Scaling via Open-Web Reinforcement Learning

Add code
May 30, 2025
Viaarxiv icon

Pangu Embedded: An Efficient Dual-system LLM Reasoner with Metacognition

Add code
May 29, 2025
Viaarxiv icon

SlimLLM: Accurate Structured Pruning for Large Language Models

Add code
May 28, 2025
Figure 1 for SlimLLM: Accurate Structured Pruning for Large Language Models
Figure 2 for SlimLLM: Accurate Structured Pruning for Large Language Models
Figure 3 for SlimLLM: Accurate Structured Pruning for Large Language Models
Figure 4 for SlimLLM: Accurate Structured Pruning for Large Language Models
Viaarxiv icon

Pangu Pro MoE: Mixture of Grouped Experts for Efficient Sparsity

Add code
May 28, 2025
Viaarxiv icon

Pangu Light: Weight Re-Initialization for Pruning and Accelerating LLMs

Add code
May 26, 2025
Viaarxiv icon

Pangu Ultra MoE: How to Train Your Big MoE on Ascend NPUs

Add code
May 07, 2025
Viaarxiv icon

DSPO: Direct Semantic Preference Optimization for Real-World Image Super-Resolution

Add code
Apr 21, 2025
Viaarxiv icon