Picture for Zhisong Zhang

Zhisong Zhang

Enhancing Web Agents with Explicit Rollback Mechanisms

Add code
Apr 16, 2025
Viaarxiv icon

Structured Extraction of Process Structure Properties Relationships in Materials Science

Add code
Apr 04, 2025
Viaarxiv icon

A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression

Add code
Dec 23, 2024
Figure 1 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 2 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 3 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Figure 4 for A Silver Bullet or a Compromise for Full Attention? A Comprehensive Study of Gist Token-based Context Compression
Viaarxiv icon

Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models

Add code
Dec 21, 2024
Figure 1 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 2 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 3 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Figure 4 for Attention Entropy is a Key Factor: An Analysis of Parallel Context Encoding with Full-attention-based Pre-trained Language Models
Viaarxiv icon

Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens

Add code
Nov 26, 2024
Figure 1 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Figure 2 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Figure 3 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Figure 4 for Low-Bit Quantization Favors Undertrained LLMs: Scaling Laws for Quantized LLMs with 100T Training Tokens
Viaarxiv icon

LoGU: Long-form Generation with Uncertainty Expressions

Add code
Oct 18, 2024
Figure 1 for LoGU: Long-form Generation with Uncertainty Expressions
Figure 2 for LoGU: Long-form Generation with Uncertainty Expressions
Figure 3 for LoGU: Long-form Generation with Uncertainty Expressions
Figure 4 for LoGU: Long-form Generation with Uncertainty Expressions
Viaarxiv icon

Atomic Calibration of LLMs in Long-Form Generations

Add code
Oct 17, 2024
Viaarxiv icon

On the Transformations across Reward Model, Parameter Update, and In-Context Prompt

Add code
Jun 24, 2024
Viaarxiv icon

On the Worst Prompt Performance of Large Language Models

Add code
Jun 08, 2024
Figure 1 for On the Worst Prompt Performance of Large Language Models
Figure 2 for On the Worst Prompt Performance of Large Language Models
Figure 3 for On the Worst Prompt Performance of Large Language Models
Figure 4 for On the Worst Prompt Performance of Large Language Models
Viaarxiv icon

Self-playing Adversarial Language Game Enhances LLM Reasoning

Add code
Apr 16, 2024
Viaarxiv icon