Picture for Junyu Zhang

Junyu Zhang

An Improved Finite-time Analysis of Temporal Difference Learning with Deep Neural Networks

Add code
May 07, 2024
Viaarxiv icon

Leveraging Hyperbolic Embeddings for Coarse-to-Fine Robot Design

Add code
Nov 02, 2023
Viaarxiv icon

Synthesizing Physically Plausible Human Motions in 3D Scenes

Add code
Aug 17, 2023
Figure 1 for Synthesizing Physically Plausible Human Motions in 3D Scenes
Figure 2 for Synthesizing Physically Plausible Human Motions in 3D Scenes
Figure 3 for Synthesizing Physically Plausible Human Motions in 3D Scenes
Figure 4 for Synthesizing Physically Plausible Human Motions in 3D Scenes
Viaarxiv icon

Offline Meta Reinforcement Learning with In-Distribution Online Adaptation

Add code
Jun 01, 2023
Figure 1 for Offline Meta Reinforcement Learning with In-Distribution Online Adaptation
Figure 2 for Offline Meta Reinforcement Learning with In-Distribution Online Adaptation
Figure 3 for Offline Meta Reinforcement Learning with In-Distribution Online Adaptation
Figure 4 for Offline Meta Reinforcement Learning with In-Distribution Online Adaptation
Viaarxiv icon

Symmetry-Aware Robot Design with Structured Subgroups

Add code
May 31, 2023
Figure 1 for Symmetry-Aware Robot Design with Structured Subgroups
Figure 2 for Symmetry-Aware Robot Design with Structured Subgroups
Figure 3 for Symmetry-Aware Robot Design with Structured Subgroups
Figure 4 for Symmetry-Aware Robot Design with Structured Subgroups
Viaarxiv icon

Provably Efficient Gauss-Newton Temporal Difference Learning Method with Function Approximation

Add code
Feb 25, 2023
Figure 1 for Provably Efficient Gauss-Newton Temporal Difference Learning Method with Function Approximation
Figure 2 for Provably Efficient Gauss-Newton Temporal Difference Learning Method with Function Approximation
Figure 3 for Provably Efficient Gauss-Newton Temporal Difference Learning Method with Function Approximation
Figure 4 for Provably Efficient Gauss-Newton Temporal Difference Learning Method with Function Approximation
Viaarxiv icon

A Near-Optimal Primal-Dual Method for Off-Policy Learning in CMDP

Add code
Jul 13, 2022
Figure 1 for A Near-Optimal Primal-Dual Method for Off-Policy Learning in CMDP
Viaarxiv icon

On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control

Add code
Jun 15, 2021
Figure 1 for On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control
Figure 2 for On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control
Figure 3 for On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control
Figure 4 for On the Sample Complexity and Metastability of Heavy-tailed Policy Search in Continuous Control
Viaarxiv icon

MARL with General Utilities via Decentralized Shadow Reward Actor-Critic

Add code
May 29, 2021
Figure 1 for MARL with General Utilities via Decentralized Shadow Reward Actor-Critic
Figure 2 for MARL with General Utilities via Decentralized Shadow Reward Actor-Critic
Figure 3 for MARL with General Utilities via Decentralized Shadow Reward Actor-Critic
Figure 4 for MARL with General Utilities via Decentralized Shadow Reward Actor-Critic
Viaarxiv icon

On the Convergence and Sample Efficiency of Variance-Reduced Policy Gradient Method

Add code
Feb 17, 2021
Figure 1 for On the Convergence and Sample Efficiency of Variance-Reduced Policy Gradient Method
Viaarxiv icon