Picture for Muhan Zhang

Muhan Zhang

Rethinking Diffusion Models with Symmetries through Canonicalization with Applications to Molecular Graph Generation

Add code
Feb 16, 2026
Viaarxiv icon

GREPO: A Benchmark for Graph Neural Networks on Repository-Level Bug Localization

Add code
Feb 14, 2026
Viaarxiv icon

SHINE: A Scalable In-Context Hypernetwork for Mapping Context to LoRA in a Single Pass

Add code
Feb 06, 2026
Viaarxiv icon

LiteToken: Removing Intermediate Merge Residues From BPE Tokenizers

Add code
Feb 04, 2026
Viaarxiv icon

Proof-RM: A Scalable and Generalizable Reward Model for Math Proof

Add code
Feb 02, 2026
Viaarxiv icon

Breaking the Blocks: Continuous Low-Rank Decomposed Scaling for Unified LLM Quantization and Adaptation

Add code
Jan 30, 2026
Viaarxiv icon

Knowledge is Not Enough: Injecting RL Skills for Continual Adaptation

Add code
Jan 16, 2026
Viaarxiv icon

SubTokenTest: A Practical Benchmark for Real-World Sub-token Understanding

Add code
Jan 14, 2026
Viaarxiv icon

What Affects the Effective Depth of Large Language Models?

Add code
Dec 16, 2025
Figure 1 for What Affects the Effective Depth of Large Language Models?
Figure 2 for What Affects the Effective Depth of Large Language Models?
Figure 3 for What Affects the Effective Depth of Large Language Models?
Figure 4 for What Affects the Effective Depth of Large Language Models?
Viaarxiv icon

Diffusion As Self-Distillation: End-to-End Latent Diffusion In One Model

Add code
Nov 18, 2025
Figure 1 for Diffusion As Self-Distillation: End-to-End Latent Diffusion In One Model
Figure 2 for Diffusion As Self-Distillation: End-to-End Latent Diffusion In One Model
Figure 3 for Diffusion As Self-Distillation: End-to-End Latent Diffusion In One Model
Figure 4 for Diffusion As Self-Distillation: End-to-End Latent Diffusion In One Model
Viaarxiv icon