Picture for Dawei Yang

Dawei Yang

R3-VAE: Reference Vector-Guided Rating Residual Quantization VAE for Generative Recommendation

Add code
Apr 14, 2026
Viaarxiv icon

MoBiE: Efficient Inference of Mixture of Binary Experts under Post-Training Quantization

Add code
Apr 08, 2026
Viaarxiv icon

SAES-SVD: Self-Adaptive Suppression of Accumulated and Local Errors for SVD-based LLM Compression

Add code
Feb 03, 2026
Viaarxiv icon

NLI:Non-uniform Linear Interpolation Approximation of Nonlinear Operations for Efficient LLMs Inference

Add code
Feb 03, 2026
Viaarxiv icon

OTARo: Once Tuning for All Precisions toward Robust On-Device LLMs

Add code
Nov 17, 2025
Viaarxiv icon

FQ-PETR: Fully Quantized Position Embedding Transformation for Multi-View 3D Object Detection

Add code
Nov 14, 2025
Viaarxiv icon

VAEVQ: Enhancing Discrete Visual Tokenization through Variational Modeling

Add code
Nov 10, 2025
Viaarxiv icon

PCDVQ: Enhancing Vector Quantization for Large Language Models via Polar Coordinate Decoupling

Add code
Jun 05, 2025
Viaarxiv icon

RWKVQuant: Quantizing the RWKV Family with Proxy Guided Hybrid of Scalar and Vector Quantization

Add code
May 02, 2025
Figure 1 for RWKVQuant: Quantizing the RWKV Family with Proxy Guided Hybrid of Scalar and Vector Quantization
Figure 2 for RWKVQuant: Quantizing the RWKV Family with Proxy Guided Hybrid of Scalar and Vector Quantization
Figure 3 for RWKVQuant: Quantizing the RWKV Family with Proxy Guided Hybrid of Scalar and Vector Quantization
Figure 4 for RWKVQuant: Quantizing the RWKV Family with Proxy Guided Hybrid of Scalar and Vector Quantization
Viaarxiv icon

MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance

Add code
May 02, 2025
Figure 1 for MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance
Figure 2 for MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance
Figure 3 for MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance
Figure 4 for MoEQuant: Enhancing Quantization for Mixture-of-Experts Large Language Models via Expert-Balanced Sampling and Affinity Guidance
Viaarxiv icon