Picture for Yuandong Tian

Yuandong Tian

TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding

Add code
Apr 18, 2024
Viaarxiv icon

GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection

Add code
Mar 06, 2024
Figure 1 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Figure 2 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Figure 3 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Figure 4 for GaLore: Memory-Efficient LLM Training by Gradient Low-Rank Projection
Viaarxiv icon

MobileLLM: Optimizing Sub-billion Parameter Language Models for On-Device Use Cases

Add code
Feb 22, 2024
Viaarxiv icon

Beyond A*: Better Planning with Transformers via Search Dynamics Bootstrapping

Add code
Feb 21, 2024
Viaarxiv icon

Diffusion World Model

Add code
Feb 11, 2024
Figure 1 for Diffusion World Model
Figure 2 for Diffusion World Model
Figure 3 for Diffusion World Model
Figure 4 for Diffusion World Model
Viaarxiv icon

TravelPlanner: A Benchmark for Real-World Planning with Language Agents

Add code
Feb 05, 2024
Figure 1 for TravelPlanner: A Benchmark for Real-World Planning with Language Agents
Figure 2 for TravelPlanner: A Benchmark for Real-World Planning with Language Agents
Figure 3 for TravelPlanner: A Benchmark for Real-World Planning with Language Agents
Figure 4 for TravelPlanner: A Benchmark for Real-World Planning with Language Agents
Viaarxiv icon

H-GAP: Humanoid Control with a Generalist Planner

Add code
Dec 05, 2023
Figure 1 for H-GAP: Humanoid Control with a Generalist Planner
Figure 2 for H-GAP: Humanoid Control with a Generalist Planner
Figure 3 for H-GAP: Humanoid Control with a Generalist Planner
Figure 4 for H-GAP: Humanoid Control with a Generalist Planner
Viaarxiv icon

Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time

Add code
Oct 26, 2023
Figure 1 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 2 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 3 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Figure 4 for Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Viaarxiv icon

End-to-end Story Plot Generator

Add code
Oct 13, 2023
Viaarxiv icon

Learning Personalized Story Evaluation

Add code
Oct 10, 2023
Viaarxiv icon