Picture for Seohong Park

Seohong Park

Decoupled Q-Chunking

Add code
Dec 12, 2025
Viaarxiv icon

Scalable Offline Model-Based RL with Action Chunks

Add code
Dec 08, 2025
Viaarxiv icon

Transitive RL: Value Learning via Divide and Conquer

Add code
Oct 26, 2025
Figure 1 for Transitive RL: Value Learning via Divide and Conquer
Figure 2 for Transitive RL: Value Learning via Divide and Conquer
Figure 3 for Transitive RL: Value Learning via Divide and Conquer
Figure 4 for Transitive RL: Value Learning via Divide and Conquer
Viaarxiv icon

Intention-Conditioned Flow Occupancy Models

Add code
Jun 10, 2025
Viaarxiv icon

Horizon Reduction Makes RL Scalable

Add code
Jun 08, 2025
Figure 1 for Horizon Reduction Makes RL Scalable
Figure 2 for Horizon Reduction Makes RL Scalable
Figure 3 for Horizon Reduction Makes RL Scalable
Figure 4 for Horizon Reduction Makes RL Scalable
Viaarxiv icon

Diffusion Guidance Is a Controllable Policy Improvement Operator

Add code
May 29, 2025
Figure 1 for Diffusion Guidance Is a Controllable Policy Improvement Operator
Figure 2 for Diffusion Guidance Is a Controllable Policy Improvement Operator
Figure 3 for Diffusion Guidance Is a Controllable Policy Improvement Operator
Figure 4 for Diffusion Guidance Is a Controllable Policy Improvement Operator
Viaarxiv icon

Flow Q-Learning

Add code
Feb 04, 2025
Viaarxiv icon

GHIL-Glue: Hierarchical Control with Filtered Subgoal Images

Add code
Oct 26, 2024
Figure 1 for GHIL-Glue: Hierarchical Control with Filtered Subgoal Images
Figure 2 for GHIL-Glue: Hierarchical Control with Filtered Subgoal Images
Figure 3 for GHIL-Glue: Hierarchical Control with Filtered Subgoal Images
Figure 4 for GHIL-Glue: Hierarchical Control with Filtered Subgoal Images
Viaarxiv icon

OGBench: Benchmarking Offline Goal-Conditioned RL

Add code
Oct 26, 2024
Viaarxiv icon

Unsupervised-to-Online Reinforcement Learning

Add code
Aug 27, 2024
Viaarxiv icon