Picture for Xunhao Lai

Xunhao Lai

Model Merging in Pre-training of Large Language Models

Add code
May 17, 2025
Viaarxiv icon

FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference

Add code
Feb 28, 2025
Figure 1 for FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference
Figure 2 for FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference
Figure 3 for FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference
Figure 4 for FlexPrefill: A Context-Aware Sparse Attention Mechanism for Efficient Long-Sequence Inference
Viaarxiv icon

Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?

Add code
Oct 02, 2024
Figure 1 for Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
Figure 2 for Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
Figure 3 for Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
Figure 4 for Fira: Can We Achieve Full-rank Training of LLMs Under Low-rank Constraint?
Viaarxiv icon