Low Rank Compression


ReCalKV: Low-Rank KV Cache Compression via Head Reordering and Offline Calibration

Add code
May 30, 2025
Viaarxiv icon

Smooth Model Compression without Fine-Tuning

Add code
May 30, 2025
Viaarxiv icon

FLAT-LLM: Fine-grained Low-rank Activation Space Transformation for Large Language Model Compression

Add code
May 29, 2025
Viaarxiv icon

Compressing Sine-Activated Low-Rank Adapters through Post-Training Quantization

Add code
May 28, 2025
Viaarxiv icon

TuneComp: Joint Fine-tuning and Compression for Large Foundation Models

Add code
May 27, 2025
Viaarxiv icon

DenseLoRA: Dense Low-Rank Adaptation of Large Language Models

Add code
May 27, 2025
Viaarxiv icon

ResSVD: Residual Compensated SVD for Large Language Model Compression

Add code
May 26, 2025
Viaarxiv icon

Tensorization is a powerful but underexplored tool for compression and interpretability of neural networks

Add code
May 26, 2025
Viaarxiv icon

Efficient Speech Translation through Model Compression and Knowledge Distillation

Add code
May 26, 2025
Viaarxiv icon

MiniLongBench: The Low-cost Long Context Understanding Benchmark for Large Language Models

Add code
May 26, 2025
Viaarxiv icon