Sentence Compression


Sentence compression is the task of reducing the length of text by removing non-essential content while preserving important facts and grammaticality.

Zonkey: A Hierarchical Diffusion Language Model with Differentiable Tokenization and Probabilistic Attention

Add code
Jan 29, 2026
Viaarxiv icon

Reducing Tokenization Premiums for Low-Resource Languages

Add code
Jan 19, 2026
Viaarxiv icon

GenProve: Learning to Generate Text with Fine-Grained Provenance

Add code
Jan 08, 2026
Viaarxiv icon

Visual Merit or Linguistic Crutch? A Close Look at DeepSeek-OCR

Add code
Jan 08, 2026
Viaarxiv icon

SAP: Syntactic Attention Pruning for Transformer-based Language Models

Add code
Dec 22, 2025
Viaarxiv icon

SkipKV: Selective Skipping of KV Generation and Storage for Efficient Inference with Large Reasoning Models

Add code
Dec 08, 2025
Figure 1 for SkipKV: Selective Skipping of KV Generation and Storage for Efficient Inference with Large Reasoning Models
Figure 2 for SkipKV: Selective Skipping of KV Generation and Storage for Efficient Inference with Large Reasoning Models
Figure 3 for SkipKV: Selective Skipping of KV Generation and Storage for Efficient Inference with Large Reasoning Models
Figure 4 for SkipKV: Selective Skipping of KV Generation and Storage for Efficient Inference with Large Reasoning Models
Viaarxiv icon

Sentence-Anchored Gist Compression for Long-Context LLMs

Add code
Nov 11, 2025
Figure 1 for Sentence-Anchored Gist Compression for Long-Context LLMs
Figure 2 for Sentence-Anchored Gist Compression for Long-Context LLMs
Figure 3 for Sentence-Anchored Gist Compression for Long-Context LLMs
Figure 4 for Sentence-Anchored Gist Compression for Long-Context LLMs
Viaarxiv icon

SABlock: Semantic-Aware KV Cache Eviction with Adaptive Compression Block Size

Add code
Oct 26, 2025
Figure 1 for SABlock: Semantic-Aware KV Cache Eviction with Adaptive Compression Block Size
Figure 2 for SABlock: Semantic-Aware KV Cache Eviction with Adaptive Compression Block Size
Figure 3 for SABlock: Semantic-Aware KV Cache Eviction with Adaptive Compression Block Size
Figure 4 for SABlock: Semantic-Aware KV Cache Eviction with Adaptive Compression Block Size
Viaarxiv icon

Protein Structure Tokenization via Geometric Byte Pair Encoding

Add code
Nov 13, 2025
Figure 1 for Protein Structure Tokenization via Geometric Byte Pair Encoding
Figure 2 for Protein Structure Tokenization via Geometric Byte Pair Encoding
Figure 3 for Protein Structure Tokenization via Geometric Byte Pair Encoding
Figure 4 for Protein Structure Tokenization via Geometric Byte Pair Encoding
Viaarxiv icon

Frustratingly Easy Task-aware Pruning for Large Language Models

Add code
Oct 26, 2025
Figure 1 for Frustratingly Easy Task-aware Pruning for Large Language Models
Figure 2 for Frustratingly Easy Task-aware Pruning for Large Language Models
Figure 3 for Frustratingly Easy Task-aware Pruning for Large Language Models
Figure 4 for Frustratingly Easy Task-aware Pruning for Large Language Models
Viaarxiv icon