Picture for Anurag Ajay

Anurag Ajay

Learning Multimodal Behaviors from Scratch with Diffusion Policy Gradient

Add code
Jun 02, 2024
Viaarxiv icon

An Introduction to Vision-Language Modeling

Add code
May 27, 2024
Viaarxiv icon

Compositional Foundation Models for Hierarchical Planning

Add code
Sep 21, 2023
Figure 1 for Compositional Foundation Models for Hierarchical Planning
Figure 2 for Compositional Foundation Models for Hierarchical Planning
Figure 3 for Compositional Foundation Models for Hierarchical Planning
Figure 4 for Compositional Foundation Models for Hierarchical Planning
Viaarxiv icon

Parallel $Q$-Learning: Scaling Off-policy Reinforcement Learning under Massively Parallel Simulation

Add code
Jul 24, 2023
Viaarxiv icon

Statistical Learning under Heterogenous Distribution Shift

Add code
Feb 27, 2023
Figure 1 for Statistical Learning under Heterogenous Distribution Shift
Figure 2 for Statistical Learning under Heterogenous Distribution Shift
Figure 3 for Statistical Learning under Heterogenous Distribution Shift
Figure 4 for Statistical Learning under Heterogenous Distribution Shift
Viaarxiv icon

Is Conditional Generative Modeling all you need for Decision-Making?

Add code
Dec 07, 2022
Figure 1 for Is Conditional Generative Modeling all you need for Decision-Making?
Figure 2 for Is Conditional Generative Modeling all you need for Decision-Making?
Figure 3 for Is Conditional Generative Modeling all you need for Decision-Making?
Figure 4 for Is Conditional Generative Modeling all you need for Decision-Making?
Viaarxiv icon

Distributionally Adaptive Meta Reinforcement Learning

Add code
Oct 06, 2022
Figure 1 for Distributionally Adaptive Meta Reinforcement Learning
Figure 2 for Distributionally Adaptive Meta Reinforcement Learning
Figure 3 for Distributionally Adaptive Meta Reinforcement Learning
Figure 4 for Distributionally Adaptive Meta Reinforcement Learning
Viaarxiv icon

Offline RL Policies Should be Trained to be Adaptive

Add code
Jul 05, 2022
Figure 1 for Offline RL Policies Should be Trained to be Adaptive
Figure 2 for Offline RL Policies Should be Trained to be Adaptive
Figure 3 for Offline RL Policies Should be Trained to be Adaptive
Figure 4 for Offline RL Policies Should be Trained to be Adaptive
Viaarxiv icon

Overcoming the Spectral Bias of Neural Value Approximation

Add code
Jun 09, 2022
Figure 1 for Overcoming the Spectral Bias of Neural Value Approximation
Figure 2 for Overcoming the Spectral Bias of Neural Value Approximation
Figure 3 for Overcoming the Spectral Bias of Neural Value Approximation
Figure 4 for Overcoming the Spectral Bias of Neural Value Approximation
Viaarxiv icon

OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning

Add code
Oct 27, 2020
Figure 1 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 2 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 3 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Figure 4 for OPAL: Offline Primitive Discovery for Accelerating Offline Reinforcement Learning
Viaarxiv icon