Picture for Simon S. Du

Simon S. Du

Frank

On the Power of Multitask Representation Learning in Linear MDP

Add code
Jun 15, 2021
Figure 1 for On the Power of Multitask Representation Learning in Linear MDP
Figure 2 for On the Power of Multitask Representation Learning in Linear MDP
Viaarxiv icon

Provable Adaptation across Multiway Domains via Representation Learning

Add code
Jun 12, 2021
Figure 1 for Provable Adaptation across Multiway Domains via Representation Learning
Figure 2 for Provable Adaptation across Multiway Domains via Representation Learning
Figure 3 for Provable Adaptation across Multiway Domains via Representation Learning
Figure 4 for Provable Adaptation across Multiway Domains via Representation Learning
Viaarxiv icon

Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret

Add code
Apr 22, 2021
Figure 1 for Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret
Figure 2 for Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret
Viaarxiv icon

Nearly Horizon-Free Offline Reinforcement Learning

Add code
Mar 25, 2021
Figure 1 for Nearly Horizon-Free Offline Reinforcement Learning
Figure 2 for Nearly Horizon-Free Offline Reinforcement Learning
Viaarxiv icon

Bilinear Classes: A Structural Framework for Provable Generalization in RL

Add code
Mar 19, 2021
Figure 1 for Bilinear Classes: A Structural Framework for Provable Generalization in RL
Figure 2 for Bilinear Classes: A Structural Framework for Provable Generalization in RL
Viaarxiv icon

Improved Corruption Robust Algorithms for Episodic Reinforcement Learning

Add code
Mar 08, 2021
Viaarxiv icon

Variance-Aware Confidence Set: Variance-Dependent Bound for Linear Bandits and Horizon-Free Bound for Linear Mixture MDP

Add code
Feb 19, 2021
Viaarxiv icon

Randomized Exploration is Near-Optimal for Tabular MDP

Add code
Feb 19, 2021
Figure 1 for Randomized Exploration is Near-Optimal for Tabular MDP
Figure 2 for Randomized Exploration is Near-Optimal for Tabular MDP
Figure 3 for Randomized Exploration is Near-Optimal for Tabular MDP
Viaarxiv icon

Provably Efficient Policy Gradient Methods for Two-Player Zero-Sum Markov Games

Add code
Feb 17, 2021
Viaarxiv icon

Fine-Grained Gap-Dependent Bounds for Tabular MDPs via Adaptive Multi-Step Bootstrap

Add code
Feb 09, 2021
Figure 1 for Fine-Grained Gap-Dependent Bounds for Tabular MDPs via Adaptive Multi-Step Bootstrap
Figure 2 for Fine-Grained Gap-Dependent Bounds for Tabular MDPs via Adaptive Multi-Step Bootstrap
Viaarxiv icon