Picture for Zhisong Zhang

Zhisong Zhang

VScan: Rethinking Visual Token Reduction for Efficient Large Vision-Language Models

Add code
May 28, 2025
Viaarxiv icon

InComeS: Integrating Compression and Selection Mechanisms into LLMs for Efficient Model Editing

Add code
May 28, 2025
Viaarxiv icon

WebCoT: Enhancing Web Agent Reasoning by Reconstructing Chain-of-Thought in Reflection, Branching, and Rollback

Add code
May 26, 2025
Viaarxiv icon

UNCLE: Uncertainty Expressions in Long-Form Generation

Add code
May 22, 2025
Viaarxiv icon

Recall with Reasoning: Chain-of-Thought Distillation for Mamba's Long-Context Memory and Extrapolation

Add code
May 06, 2025
Viaarxiv icon

WebEvolver: Enhancing Web Agent Self-Improvement with Coevolving World Model

Add code
Apr 23, 2025
Viaarxiv icon

Enhancing Web Agents with Explicit Rollback Mechanisms

Add code
Apr 16, 2025
Viaarxiv icon

Structured Extraction of Process Structure Properties Relationships in Materials Science

Add code
Apr 04, 2025
Viaarxiv icon

A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression

Add code
Dec 23, 2024
Figure 1 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 2 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 3 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 4 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Viaarxiv icon

Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models

Add code
Dec 21, 2024
Figure 1 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 2 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 3 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 4 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Viaarxiv icon