Alert button
Picture for Wenhao Yang

Wenhao Yang

Alert button

Estimation and Inference in Distributional Reinforcement Learning

Add code
Bookmark button
Alert button
Sep 29, 2023
Liangyu Zhang, Yang Peng, Jiadong Liang, Wenhao Yang, Zhihua Zhang

Figure 1 for Estimation and Inference in Distributional Reinforcement Learning
Figure 2 for Estimation and Inference in Distributional Reinforcement Learning
Figure 3 for Estimation and Inference in Distributional Reinforcement Learning
Figure 4 for Estimation and Inference in Distributional Reinforcement Learning
Viaarxiv icon

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice

Add code
Bookmark button
Alert button
May 22, 2023
Toshinori Kitamura, Tadashi Kozuno, Yunhao Tang, Nino Vieillard, Michal Valko, Wenhao Yang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári, Wataru Kumagai, Yutaka Matsuo

Figure 1 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 2 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 3 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 4 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Viaarxiv icon

Non-stationary Projection-free Online Learning with Dynamic and Adaptive Regret Guarantees

Add code
Bookmark button
Alert button
May 19, 2023
Yibo Wang, Wenhao Yang, Wei Jiang, Shiyin Lu, Bing Wang, Haihong Tang, Yuanyu Wan, Lijun Zhang

Figure 1 for Non-stationary Projection-free Online Learning with Dynamic and Adaptive Regret Guarantees
Figure 2 for Non-stationary Projection-free Online Learning with Dynamic and Adaptive Regret Guarantees
Figure 3 for Non-stationary Projection-free Online Learning with Dynamic and Adaptive Regret Guarantees
Figure 4 for Non-stationary Projection-free Online Learning with Dynamic and Adaptive Regret Guarantees
Viaarxiv icon

Semi-Infinitely Constrained Markov Decision Processes and Efficient Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 29, 2023
Liangyu Zhang, Yang Peng, Wenhao Yang, Zhihua Zhang

Figure 1 for Semi-Infinitely Constrained Markov Decision Processes and Efficient Reinforcement Learning
Figure 2 for Semi-Infinitely Constrained Markov Decision Processes and Efficient Reinforcement Learning
Figure 3 for Semi-Infinitely Constrained Markov Decision Processes and Efficient Reinforcement Learning
Figure 4 for Semi-Infinitely Constrained Markov Decision Processes and Efficient Reinforcement Learning
Viaarxiv icon

Avoiding Model Estimation in Robust Markov Decision Processes with a Generative Model

Add code
Bookmark button
Alert button
Feb 02, 2023
Wenhao Yang, Han Wang, Tadashi Kozuno, Scott M. Jordan, Zhihua Zhang

Figure 1 for Avoiding Model Estimation in Robust Markov Decision Processes with a Generative Model
Figure 2 for Avoiding Model Estimation in Robust Markov Decision Processes with a Generative Model
Figure 3 for Avoiding Model Estimation in Robust Markov Decision Processes with a Generative Model
Figure 4 for Avoiding Model Estimation in Robust Markov Decision Processes with a Generative Model
Viaarxiv icon

Statistical Estimation of Confounded Linear MDPs: An Instrumental Variable Approach

Add code
Bookmark button
Alert button
Sep 12, 2022
Miao Lu, Wenhao Yang, Liangyu Zhang, Zhihua Zhang

Viaarxiv icon

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
Bookmark button
Alert button
May 27, 2022
Tadashi Kozuno, Wenhao Yang, Nino Vieillard, Toshinori Kitamura, Yunhao Tang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Michal Valko, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári

Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

Pluralistic Image Completion with Probabilistic Mixture-of-Experts

Add code
Bookmark button
Alert button
May 18, 2022
Xiaobo Xia, Wenhao Yang, Jie Ren, Yewen Li, Yibing Zhan, Bo Han, Tongliang Liu

Figure 1 for Pluralistic Image Completion with Probabilistic Mixture-of-Experts
Figure 2 for Pluralistic Image Completion with Probabilistic Mixture-of-Experts
Figure 3 for Pluralistic Image Completion with Probabilistic Mixture-of-Experts
Figure 4 for Pluralistic Image Completion with Probabilistic Mixture-of-Experts
Viaarxiv icon

Federated Reinforcement Learning with Environment Heterogeneity

Add code
Bookmark button
Alert button
Apr 06, 2022
Hao Jin, Yang Peng, Wenhao Yang, Shusen Wang, Zhihua Zhang

Figure 1 for Federated Reinforcement Learning with Environment Heterogeneity
Figure 2 for Federated Reinforcement Learning with Environment Heterogeneity
Figure 3 for Federated Reinforcement Learning with Environment Heterogeneity
Figure 4 for Federated Reinforcement Learning with Environment Heterogeneity
Viaarxiv icon

Polyak-Ruppert Averaged Q-Leaning is Statistically Efficient

Add code
Bookmark button
Alert button
Jan 23, 2022
Xiang Li, Wenhao Yang, Jiadong Liang, Zhihua Zhang, Michael I. Jordan

Figure 1 for Polyak-Ruppert Averaged Q-Leaning is Statistically Efficient
Viaarxiv icon