Picture for Huanrui Yang

Huanrui Yang

MSQ: Memory-Efficient Bit Sparsification Quantization

Add code
Jul 30, 2025
Viaarxiv icon

Is Attention Required for Transformer Inference? Explore Function-preserving Attention Replacement

Add code
May 29, 2025
Viaarxiv icon

SAFER: Sharpness Aware layer-selective Finetuning for Enhanced Robustness in vision transformers

Add code
Jan 02, 2025
Figure 1 for SAFER: Sharpness Aware layer-selective Finetuning for Enhanced Robustness in vision transformers
Figure 2 for SAFER: Sharpness Aware layer-selective Finetuning for Enhanced Robustness in vision transformers
Figure 3 for SAFER: Sharpness Aware layer-selective Finetuning for Enhanced Robustness in vision transformers
Figure 4 for SAFER: Sharpness Aware layer-selective Finetuning for Enhanced Robustness in vision transformers
Viaarxiv icon

Taming Sensitive Weights : Noise Perturbation Fine-tuning for Robust LLM Quantization

Add code
Dec 08, 2024
Viaarxiv icon

Personalized Multimodal Large Language Models: A Survey

Add code
Dec 03, 2024
Viaarxiv icon

A Survey of Small Language Models

Add code
Oct 25, 2024
Figure 1 for A Survey of Small Language Models
Figure 2 for A Survey of Small Language Models
Figure 3 for A Survey of Small Language Models
Viaarxiv icon

PAT: Pruning-Aware Tuning for Large Language Models

Add code
Aug 27, 2024
Figure 1 for PAT: Pruning-Aware Tuning for Large Language Models
Figure 2 for PAT: Pruning-Aware Tuning for Large Language Models
Figure 3 for PAT: Pruning-Aware Tuning for Large Language Models
Figure 4 for PAT: Pruning-Aware Tuning for Large Language Models
Viaarxiv icon

Criticality Leveraged Adversarial Training (CLAT) for Boosted Performance via Parameter Efficiency

Add code
Aug 19, 2024
Figure 1 for Criticality Leveraged Adversarial Training (CLAT) for Boosted Performance via Parameter Efficiency
Figure 2 for Criticality Leveraged Adversarial Training (CLAT) for Boosted Performance via Parameter Efficiency
Figure 3 for Criticality Leveraged Adversarial Training (CLAT) for Boosted Performance via Parameter Efficiency
Figure 4 for Criticality Leveraged Adversarial Training (CLAT) for Boosted Performance via Parameter Efficiency
Viaarxiv icon

FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models

Add code
Aug 15, 2024
Figure 1 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Figure 2 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Figure 3 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Figure 4 for FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models
Viaarxiv icon

Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance

Add code
Jul 17, 2024
Figure 1 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Figure 2 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Figure 3 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Figure 4 for Sharpness-diversity tradeoff: improving flat ensembles with SharpBalance
Viaarxiv icon