Picture for Yujie Wang

Yujie Wang

Training-free and Adaptive Sparse Attention for Efficient Long Video Generation

Add code
Feb 28, 2025
Figure 1 for Training-free and Adaptive Sparse Attention for Efficient Long Video Generation
Figure 2 for Training-free and Adaptive Sparse Attention for Efficient Long Video Generation
Figure 3 for Training-free and Adaptive Sparse Attention for Efficient Long Video Generation
Figure 4 for Training-free and Adaptive Sparse Attention for Efficient Long Video Generation
Viaarxiv icon

Sparse Brains are Also Adaptive Brains: Cognitive-Load-Aware Dynamic Activation for LLMs

Add code
Feb 26, 2025
Figure 1 for Sparse Brains are Also Adaptive Brains: Cognitive-Load-Aware Dynamic Activation for LLMs
Figure 2 for Sparse Brains are Also Adaptive Brains: Cognitive-Load-Aware Dynamic Activation for LLMs
Figure 3 for Sparse Brains are Also Adaptive Brains: Cognitive-Load-Aware Dynamic Activation for LLMs
Figure 4 for Sparse Brains are Also Adaptive Brains: Cognitive-Load-Aware Dynamic Activation for LLMs
Viaarxiv icon

Pioneer: Physics-informed Riemannian Graph ODE for Entropy-increasing Dynamics

Add code
Feb 05, 2025
Figure 1 for Pioneer: Physics-informed Riemannian Graph ODE for Entropy-increasing Dynamics
Figure 2 for Pioneer: Physics-informed Riemannian Graph ODE for Entropy-increasing Dynamics
Figure 3 for Pioneer: Physics-informed Riemannian Graph ODE for Entropy-increasing Dynamics
Figure 4 for Pioneer: Physics-informed Riemannian Graph ODE for Entropy-increasing Dynamics
Viaarxiv icon

Towards Effective Graph Rationalization via Boosting Environment Diversity

Add code
Dec 17, 2024
Figure 1 for Towards Effective Graph Rationalization via Boosting Environment Diversity
Figure 2 for Towards Effective Graph Rationalization via Boosting Environment Diversity
Figure 3 for Towards Effective Graph Rationalization via Boosting Environment Diversity
Figure 4 for Towards Effective Graph Rationalization via Boosting Environment Diversity
Viaarxiv icon

Data-Centric and Heterogeneity-Adaptive Sequence Parallelism for Efficient LLM Training

Add code
Dec 02, 2024
Figure 1 for Data-Centric and Heterogeneity-Adaptive Sequence Parallelism for Efficient LLM Training
Figure 2 for Data-Centric and Heterogeneity-Adaptive Sequence Parallelism for Efficient LLM Training
Figure 3 for Data-Centric and Heterogeneity-Adaptive Sequence Parallelism for Efficient LLM Training
Figure 4 for Data-Centric and Heterogeneity-Adaptive Sequence Parallelism for Efficient LLM Training
Viaarxiv icon

Absolute State-wise Constrained Policy Optimization: High-Probability State-wise Constraints Satisfaction

Add code
Oct 02, 2024
Figure 1 for Absolute State-wise Constrained Policy Optimization: High-Probability State-wise Constraints Satisfaction
Figure 2 for Absolute State-wise Constrained Policy Optimization: High-Probability State-wise Constraints Satisfaction
Figure 3 for Absolute State-wise Constrained Policy Optimization: High-Probability State-wise Constraints Satisfaction
Figure 4 for Absolute State-wise Constrained Policy Optimization: High-Probability State-wise Constraints Satisfaction
Viaarxiv icon

Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management

Add code
Sep 05, 2024
Figure 1 for Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management
Figure 2 for Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management
Figure 3 for Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management
Figure 4 for Efficient Multi-Task Large Model Training via Data Heterogeneity-aware Model Management
Viaarxiv icon

First Activations Matter: Training-Free Methods for Dynamic Activation in Large Language Models

Add code
Aug 21, 2024
Viaarxiv icon

MOYU: A Theoretical Study on Massive Over-activation Yielded Uplifts in LLMs

Add code
Jun 18, 2024
Figure 1 for MOYU: A Theoretical Study on Massive Over-activation Yielded Uplifts in LLMs
Figure 2 for MOYU: A Theoretical Study on Massive Over-activation Yielded Uplifts in LLMs
Viaarxiv icon

QQQ: Quality Quattuor-Bit Quantization for Large Language Models

Add code
Jun 14, 2024
Figure 1 for QQQ: Quality Quattuor-Bit Quantization for Large Language Models
Figure 2 for QQQ: Quality Quattuor-Bit Quantization for Large Language Models
Figure 3 for QQQ: Quality Quattuor-Bit Quantization for Large Language Models
Figure 4 for QQQ: Quality Quattuor-Bit Quantization for Large Language Models
Viaarxiv icon