Picture for Dianhai Yu

Dianhai Yu

Boosting Document Parsing Efficiency and Performance with Coarse-to-Fine Visual Processing

Add code
Mar 25, 2026
Viaarxiv icon

Sparse Layer Sharpness-Aware Minimization for Efficient Fine-Tuning

Add code
Feb 10, 2026
Viaarxiv icon

ERNIE 5.0 Technical Report

Add code
Feb 04, 2026
Viaarxiv icon

PaddleOCR-VL-1.5: Towards a Multi-Task 0.9B VLM for Robust In-the-Wild Document Parsing

Add code
Jan 29, 2026
Viaarxiv icon

PaddleOCR-VL: Boosting Multilingual Document Parsing via a 0.9B Ultra-Compact Vision-Language Model

Add code
Oct 16, 2025
Viaarxiv icon

Low-Precision Training of Large Language Models: Methods, Challenges, and Opportunities

Add code
May 02, 2025
Figure 1 for Low-Precision Training of Large Language Models: Methods, Challenges, and Opportunities
Figure 2 for Low-Precision Training of Large Language Models: Methods, Challenges, and Opportunities
Figure 3 for Low-Precision Training of Large Language Models: Methods, Challenges, and Opportunities
Figure 4 for Low-Precision Training of Large Language Models: Methods, Challenges, and Opportunities
Viaarxiv icon

FlashMask: Efficient and Rich Mask Extension of FlashAttention

Add code
Oct 02, 2024
Figure 1 for FlashMask: Efficient and Rich Mask Extension of FlashAttention
Figure 2 for FlashMask: Efficient and Rich Mask Extension of FlashAttention
Figure 3 for FlashMask: Efficient and Rich Mask Extension of FlashAttention
Figure 4 for FlashMask: Efficient and Rich Mask Extension of FlashAttention
Viaarxiv icon

NACL: A General and Effective KV Cache Eviction Framework for LLMs at Inference Time

Add code
Aug 07, 2024
Figure 1 for NACL: A General and Effective KV Cache Eviction Framework for LLMs at Inference Time
Figure 2 for NACL: A General and Effective KV Cache Eviction Framework for LLMs at Inference Time
Figure 3 for NACL: A General and Effective KV Cache Eviction Framework for LLMs at Inference Time
Figure 4 for NACL: A General and Effective KV Cache Eviction Framework for LLMs at Inference Time
Viaarxiv icon

A Framework for Cost-Effective and Self-Adaptive LLM Shaking and Recovery Mechanism

Add code
Mar 12, 2024
Figure 1 for A Framework for Cost-Effective and Self-Adaptive LLM Shaking and Recovery Mechanism
Figure 2 for A Framework for Cost-Effective and Self-Adaptive LLM Shaking and Recovery Mechanism
Figure 3 for A Framework for Cost-Effective and Self-Adaptive LLM Shaking and Recovery Mechanism
Figure 4 for A Framework for Cost-Effective and Self-Adaptive LLM Shaking and Recovery Mechanism
Viaarxiv icon

Spectral Heterogeneous Graph Convolutions via Positive Noncommutative Polynomials

Add code
May 31, 2023
Figure 1 for Spectral Heterogeneous Graph Convolutions via Positive Noncommutative Polynomials
Figure 2 for Spectral Heterogeneous Graph Convolutions via Positive Noncommutative Polynomials
Figure 3 for Spectral Heterogeneous Graph Convolutions via Positive Noncommutative Polynomials
Figure 4 for Spectral Heterogeneous Graph Convolutions via Positive Noncommutative Polynomials
Viaarxiv icon