Network Pruning


Network pruning is a popular approach to reduce a heavy network to obtain a lightweight form by removing redundancy in the heavy network. In this approach, a complex over-parameterized network is first trained, then pruned based on some criteria, and finally fine-tuned to achieve comparable performance with reduced parameters.

Dynamic UGV-UAV Cooperative Path Planning in Uncertain Environments

Add code
Apr 28, 2026
Viaarxiv icon

Rethinking Layer Redundancy in Large Language Models: Calibration Objectives and Search for Depth Pruning

Add code
Apr 27, 2026
Viaarxiv icon

Supernodes and Halos: Loss-Critical Hubs in LLM Feed-Forward Layers

Add code
Apr 26, 2026
Viaarxiv icon

Quantization robustness from dense representations of sparse functions in high-capacity kernel associative memory

Add code
Apr 22, 2026
Viaarxiv icon

Towards Adaptive Continual Model Merging via Manifold-Aware Expert Evolution

Add code
Apr 24, 2026
Viaarxiv icon

HGQ-LUT: Fast LUT-Aware Training and Efficient Architectures for DNN Inference

Add code
Apr 24, 2026
Viaarxiv icon

FlashFPS: Efficient Farthest Point Sampling for Large-Scale Point Clouds via Pruning and Caching

Add code
Apr 20, 2026
Viaarxiv icon

Soft Label Pruning and Quantization for Large-Scale Dataset Distillation

Add code
Apr 20, 2026
Viaarxiv icon

Fairness is Not Flat: Geometric Phase Transitions Against Shortcut Learning

Add code
Apr 13, 2026
Viaarxiv icon

A Comparative Study of CNN Optimization Methods for Edge AI: Exploring the Role of Early Exits

Add code
Apr 16, 2026
Viaarxiv icon