Picture for Zhenglun Kong

Zhenglun Kong

Token Transforming: A Unified and Training-Free Token Compression Framework for Vision Transformer Acceleration

Add code
Jun 06, 2025
Viaarxiv icon

Enabling Flexible Multi-LLM Integration for Scalable Knowledge Aggregation

Add code
May 28, 2025
Viaarxiv icon

ACE: Exploring Activation Cosine Similarity and Variance for Accurate and Calibration-Efficient LLM Pruning

Add code
May 28, 2025
Viaarxiv icon

Token Reduction Should Go Beyond Efficiency in Generative Models -- From Vision, Language to Multimodality

Add code
May 23, 2025
Viaarxiv icon

Structured Agent Distillation for Large Language Model

Add code
May 20, 2025
Viaarxiv icon

TxAgent: An AI Agent for Therapeutic Reasoning Across a Universe of Tools

Add code
Mar 14, 2025
Viaarxiv icon

RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation

Add code
Jan 08, 2025
Figure 1 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Figure 2 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Figure 3 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Figure 4 for RoRA: Efficient Fine-Tuning of LLM with Reliability Optimization for Rank Adaptation
Viaarxiv icon

Fully Open Source Moxin-7B Technical Report

Add code
Dec 08, 2024
Figure 1 for Fully Open Source Moxin-7B Technical Report
Figure 2 for Fully Open Source Moxin-7B Technical Report
Figure 3 for Fully Open Source Moxin-7B Technical Report
Figure 4 for Fully Open Source Moxin-7B Technical Report
Viaarxiv icon

Fast and Memory-Efficient Video Diffusion Using Streamlined Inference

Add code
Nov 02, 2024
Figure 1 for Fast and Memory-Efficient Video Diffusion Using Streamlined Inference
Figure 2 for Fast and Memory-Efficient Video Diffusion Using Streamlined Inference
Figure 3 for Fast and Memory-Efficient Video Diffusion Using Streamlined Inference
Figure 4 for Fast and Memory-Efficient Video Diffusion Using Streamlined Inference
Viaarxiv icon

Pruning Foundation Models for High Accuracy without Retraining

Add code
Oct 21, 2024
Figure 1 for Pruning Foundation Models for High Accuracy without Retraining
Figure 2 for Pruning Foundation Models for High Accuracy without Retraining
Figure 3 for Pruning Foundation Models for High Accuracy without Retraining
Viaarxiv icon