Picture for Chongjie Zhang

Chongjie Zhang

The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning

Add code
Feb 27, 2023
Figure 1 for The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning
Figure 2 for The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning
Figure 3 for The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning
Figure 4 for The Provable Benefits of Unsupervised Data Sharing for Offline Reinforcement Learning
Viaarxiv icon

A Survey on Transformers in Reinforcement Learning

Add code
Jan 08, 2023
Viaarxiv icon

Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery

Add code
Dec 02, 2022
Figure 1 for Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery
Figure 2 for Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery
Figure 3 for Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery
Figure 4 for Flow to Control: Offline Reinforcement Learning with Lossless Primitive Discovery
Viaarxiv icon

Low-Rank Modular Reinforcement Learning via Muscle Synergy

Add code
Oct 26, 2022
Figure 1 for Low-Rank Modular Reinforcement Learning via Muscle Synergy
Figure 2 for Low-Rank Modular Reinforcement Learning via Muscle Synergy
Figure 3 for Low-Rank Modular Reinforcement Learning via Muscle Synergy
Figure 4 for Low-Rank Modular Reinforcement Learning via Muscle Synergy
Viaarxiv icon

Non-Linear Coordination Graphs

Add code
Oct 26, 2022
Figure 1 for Non-Linear Coordination Graphs
Figure 2 for Non-Linear Coordination Graphs
Figure 3 for Non-Linear Coordination Graphs
Figure 4 for Non-Linear Coordination Graphs
Viaarxiv icon

CUP: Critic-Guided Policy Reuse

Add code
Oct 15, 2022
Figure 1 for CUP: Critic-Guided Policy Reuse
Figure 2 for CUP: Critic-Guided Policy Reuse
Figure 3 for CUP: Critic-Guided Policy Reuse
Figure 4 for CUP: Critic-Guided Policy Reuse
Viaarxiv icon

On the Role of Discount Factor in Offline Reinforcement Learning

Add code
Jun 15, 2022
Figure 1 for On the Role of Discount Factor in Offline Reinforcement Learning
Figure 2 for On the Role of Discount Factor in Offline Reinforcement Learning
Figure 3 for On the Role of Discount Factor in Offline Reinforcement Learning
Figure 4 for On the Role of Discount Factor in Offline Reinforcement Learning
Viaarxiv icon

RORL: Robust Offline Reinforcement Learning via Conservative Smoothing

Add code
Jun 06, 2022
Figure 1 for RORL: Robust Offline Reinforcement Learning via Conservative Smoothing
Figure 2 for RORL: Robust Offline Reinforcement Learning via Conservative Smoothing
Figure 3 for RORL: Robust Offline Reinforcement Learning via Conservative Smoothing
Figure 4 for RORL: Robust Offline Reinforcement Learning via Conservative Smoothing
Viaarxiv icon

Latent-Variable Advantage-Weighted Policy Optimization for Offline RL

Add code
Mar 16, 2022
Figure 1 for Latent-Variable Advantage-Weighted Policy Optimization for Offline RL
Figure 2 for Latent-Variable Advantage-Weighted Policy Optimization for Offline RL
Figure 3 for Latent-Variable Advantage-Weighted Policy Optimization for Offline RL
Figure 4 for Latent-Variable Advantage-Weighted Policy Optimization for Offline RL
Viaarxiv icon

Multi-Agent Policy Transfer via Task Relationship Modeling

Add code
Mar 09, 2022
Figure 1 for Multi-Agent Policy Transfer via Task Relationship Modeling
Figure 2 for Multi-Agent Policy Transfer via Task Relationship Modeling
Figure 3 for Multi-Agent Policy Transfer via Task Relationship Modeling
Figure 4 for Multi-Agent Policy Transfer via Task Relationship Modeling
Viaarxiv icon