Picture for Nan Duan

Nan Duan

Step-Video-TI2V Technical Report: A State-of-the-Art Text-Driven Image-to-Video Generation Model

Add code
Mar 14, 2025
Viaarxiv icon

DSV: Exploiting Dynamic Sparsity to Accelerate Large-Scale Video DiT Training

Add code
Feb 11, 2025
Viaarxiv icon

Taming Teacher Forcing for Masked Autoregressive Video Generation

Add code
Jan 21, 2025
Viaarxiv icon

Automated Proof Generation for Rust Code via Self-Evolution

Add code
Oct 21, 2024
Figure 1 for Automated Proof Generation for Rust Code via Self-Evolution
Figure 2 for Automated Proof Generation for Rust Code via Self-Evolution
Figure 3 for Automated Proof Generation for Rust Code via Self-Evolution
Figure 4 for Automated Proof Generation for Rust Code via Self-Evolution
Viaarxiv icon

Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation

Add code
Oct 21, 2024
Figure 1 for Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation
Figure 2 for Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation
Figure 3 for Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation
Figure 4 for Alchemy: Amplifying Theorem-Proving Capability through Symbolic Mutation
Viaarxiv icon

AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition

Add code
Aug 21, 2024
Figure 1 for AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition
Figure 2 for AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition
Figure 3 for AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition
Figure 4 for AutoDirector: Online Auto-scheduling Agents for Multi-sensory Composition
Viaarxiv icon

Rho-1: Not All Tokens Are What You Need

Add code
Apr 11, 2024
Figure 1 for Rho-1: Not All Tokens Are What You Need
Figure 2 for Rho-1: Not All Tokens Are What You Need
Figure 3 for Rho-1: Not All Tokens Are What You Need
Figure 4 for Rho-1: Not All Tokens Are What You Need
Viaarxiv icon

LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models

Add code
Apr 03, 2024
Figure 1 for LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models
Figure 2 for LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models
Figure 3 for LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models
Figure 4 for LVLM-Intrepret: An Interpretability Tool for Large Vision-Language Models
Viaarxiv icon

Exploring the Mystery of Influential Data for Mathematical Reasoning

Add code
Apr 01, 2024
Figure 1 for Exploring the Mystery of Influential Data for Mathematical Reasoning
Figure 2 for Exploring the Mystery of Influential Data for Mathematical Reasoning
Figure 3 for Exploring the Mystery of Influential Data for Mathematical Reasoning
Figure 4 for Exploring the Mystery of Influential Data for Mathematical Reasoning
Viaarxiv icon

PPTC-R benchmark: Towards Evaluating the Robustness of Large Language Models for PowerPoint Task Completion

Add code
Mar 06, 2024
Viaarxiv icon