Picture for Xiaowen Chu

Xiaowen Chu

Can Compressed LLMs Truly Act? An Empirical Evaluation of Agentic Capabilities in LLM Compression

Add code
May 26, 2025
Viaarxiv icon

FlowKV: Enhancing Multi-Turn Conversational Coherence in LLMs via Isolated Key-Value Cache Management

Add code
May 21, 2025
Viaarxiv icon

Jupiter: Fast and Resource-Efficient Collaborative Inference of Generative LLMs on Edge Devices

Add code
Apr 11, 2025
Viaarxiv icon

MRD-RAG: Enhancing Medical Diagnosis with Multi-Round Retrieval-Augmented Generation

Add code
Apr 10, 2025
Viaarxiv icon

The Lottery LLM Hypothesis, Rethinking What Abilities Should LLM Compression Preserve?

Add code
Feb 24, 2025
Viaarxiv icon

Perovskite-LLM: Knowledge-Enhanced Large Language Models for Perovskite Solar Cell Research

Add code
Feb 18, 2025
Viaarxiv icon

SphereFusion: Efficient Panorama Depth Estimation via Gated Fusion

Add code
Feb 09, 2025
Viaarxiv icon

Mediator: Memory-efficient LLM Merging with Less Parameter Conflicts and Uncertainty Based Routing

Add code
Feb 06, 2025
Viaarxiv icon

Can LLMs Maintain Fundamental Abilities under KV Cache Compression?

Add code
Feb 04, 2025
Figure 1 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 2 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 3 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Figure 4 for Can LLMs Maintain Fundamental Abilities under KV Cache Compression?
Viaarxiv icon

FSMoE: A Flexible and Scalable Training System for Sparse Mixture-of-Experts Models

Add code
Jan 18, 2025
Viaarxiv icon