Alert button
Picture for Wen Sun

Wen Sun

Alert button

The Benefits of Being Distributional: Small-Loss Bounds for Reinforcement Learning

Add code
Bookmark button
Alert button
May 25, 2023
Kaiwen Wang, Kevin Zhou, Runzhe Wu, Nathan Kallus, Wen Sun

Figure 1 for The Benefits of Being Distributional: Small-Loss Bounds for Reinforcement Learning
Figure 2 for The Benefits of Being Distributional: Small-Loss Bounds for Reinforcement Learning
Viaarxiv icon

Provable Offline Reinforcement Learning with Human Feedback

Add code
Bookmark button
Alert button
May 24, 2023
Wenhao Zhan, Masatoshi Uehara, Nathan Kallus, Jason D. Lee, Wen Sun

Viaarxiv icon

Distributional Offline Policy Evaluation with Predictive Error Guarantees

Add code
Bookmark button
Alert button
Feb 19, 2023
Runzhe Wu, Masatoshi Uehara, Wen Sun

Figure 1 for Distributional Offline Policy Evaluation with Predictive Error Guarantees
Figure 2 for Distributional Offline Policy Evaluation with Predictive Error Guarantees
Figure 3 for Distributional Offline Policy Evaluation with Predictive Error Guarantees
Figure 4 for Distributional Offline Policy Evaluation with Predictive Error Guarantees
Viaarxiv icon

Multi-task Representation Learning for Pure Exploration in Linear Bandits

Add code
Bookmark button
Alert button
Feb 09, 2023
Yihan Du, Longbo Huang, Wen Sun

Figure 1 for Multi-task Representation Learning for Pure Exploration in Linear Bandits
Viaarxiv icon

Near-Minimax-Optimal Risk-Sensitive Reinforcement Learning with CVaR

Add code
Bookmark button
Alert button
Feb 07, 2023
Kaiwen Wang, Nathan Kallus, Wen Sun

Viaarxiv icon

Refined Value-Based Offline RL under Realizability and Partial Coverage

Add code
Bookmark button
Alert button
Feb 05, 2023
Masatoshi Uehara, Nathan Kallus, Jason D. Lee, Wen Sun

Figure 1 for Refined Value-Based Offline RL under Realizability and Partial Coverage
Viaarxiv icon

Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient

Add code
Bookmark button
Alert button
Oct 13, 2022
Yuda Song, Yifei Zhou, Ayush Sekhari, J. Andrew Bagnell, Akshay Krishnamurthy, Wen Sun

Figure 1 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Figure 2 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Figure 3 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Figure 4 for Hybrid RL: Using Both Offline and Online Data Can Make RL Efficient
Viaarxiv icon

Sample-efficient Safe Learning for Online Nonlinear Control with Control Barrier Functions

Add code
Bookmark button
Alert button
Jul 29, 2022
Wenhao Luo, Wen Sun, Ashish Kapoor

Figure 1 for Sample-efficient Safe Learning for Online Nonlinear Control with Control Barrier Functions
Figure 2 for Sample-efficient Safe Learning for Online Nonlinear Control with Control Barrier Functions
Figure 3 for Sample-efficient Safe Learning for Online Nonlinear Control with Control Barrier Functions
Viaarxiv icon

Future-Dependent Value-Based Off-Policy Evaluation in POMDPs

Add code
Bookmark button
Alert button
Jul 26, 2022
Masatoshi Uehara, Haruka Kiyohara, Andrew Bennett, Victor Chernozhukov, Nan Jiang, Nathan Kallus, Chengchun Shi, Wen Sun

Figure 1 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Figure 2 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Figure 3 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Figure 4 for Future-Dependent Value-Based Off-Policy Evaluation in POMDPs
Viaarxiv icon

PAC Reinforcement Learning for Predictive State Representations

Add code
Bookmark button
Alert button
Jul 15, 2022
Wenhao Zhan, Masatoshi Uehara, Wen Sun, Jason D. Lee

Figure 1 for PAC Reinforcement Learning for Predictive State Representations
Figure 2 for PAC Reinforcement Learning for Predictive State Representations
Figure 3 for PAC Reinforcement Learning for Predictive State Representations
Figure 4 for PAC Reinforcement Learning for Predictive State Representations
Viaarxiv icon