Alert button
Picture for Jincheng Mei

Jincheng Mei

Alert button

Stochastic Gradient Succeeds for Bandits

Add code
Bookmark button
Alert button
Feb 27, 2024
Jincheng Mei, Zixin Zhong, Bo Dai, Alekh Agarwal, Csaba Szepesvari, Dale Schuurmans

Viaarxiv icon

Beyond Expectations: Learning with Stochastic Dominance Made Practical

Add code
Bookmark button
Alert button
Feb 05, 2024
Shicong Cen, Jincheng Mei, Hanjun Dai, Dale Schuurmans, Yuejie Chi, Bo Dai

Viaarxiv icon

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice

Add code
Bookmark button
Alert button
May 22, 2023
Toshinori Kitamura, Tadashi Kozuno, Yunhao Tang, Nino Vieillard, Michal Valko, Wenhao Yang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári, Wataru Kumagai, Yutaka Matsuo

Figure 1 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 2 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 3 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 4 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Viaarxiv icon

The Role of Baselines in Policy Gradient Optimization

Add code
Bookmark button
Alert button
Jan 16, 2023
Jincheng Mei, Wesley Chung, Valentin Thomas, Bo Dai, Csaba Szepesvari, Dale Schuurmans

Figure 1 for The Role of Baselines in Policy Gradient Optimization
Figure 2 for The Role of Baselines in Policy Gradient Optimization
Figure 3 for The Role of Baselines in Policy Gradient Optimization
Viaarxiv icon

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
Bookmark button
Alert button
May 27, 2022
Tadashi Kozuno, Wenhao Yang, Nino Vieillard, Toshinori Kitamura, Yunhao Tang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Michal Valko, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári

Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

Understanding the Effect of Stochasticity in Policy Optimization

Add code
Bookmark button
Alert button
Oct 29, 2021
Jincheng Mei, Bo Dai, Chenjun Xiao, Csaba Szepesvari, Dale Schuurmans

Figure 1 for Understanding the Effect of Stochasticity in Policy Optimization
Figure 2 for Understanding the Effect of Stochasticity in Policy Optimization
Figure 3 for Understanding the Effect of Stochasticity in Policy Optimization
Viaarxiv icon

Leveraging Non-uniformity in First-order Non-convex Optimization

Add code
Bookmark button
Alert button
May 13, 2021
Jincheng Mei, Yue Gao, Bo Dai, Csaba Szepesvari, Dale Schuurmans

Figure 1 for Leveraging Non-uniformity in First-order Non-convex Optimization
Figure 2 for Leveraging Non-uniformity in First-order Non-convex Optimization
Figure 3 for Leveraging Non-uniformity in First-order Non-convex Optimization
Figure 4 for Leveraging Non-uniformity in First-order Non-convex Optimization
Viaarxiv icon

On the Optimality of Batch Policy Optimization Algorithms

Add code
Bookmark button
Alert button
Apr 06, 2021
Chenjun Xiao, Yifan Wu, Tor Lattimore, Bo Dai, Jincheng Mei, Lihong Li, Csaba Szepesvari, Dale Schuurmans

Figure 1 for On the Optimality of Batch Policy Optimization Algorithms
Figure 2 for On the Optimality of Batch Policy Optimization Algorithms
Viaarxiv icon

Beyond Prioritized Replay: Sampling States in Model-Based RL via Simulated Priorities

Add code
Bookmark button
Alert button
Jul 19, 2020
Jincheng Mei, Yangchen Pan, Martha White, Amir-massoud Farahmand, Hengshuai Yao

Figure 1 for Beyond Prioritized Replay: Sampling States in Model-Based RL via Simulated Priorities
Figure 2 for Beyond Prioritized Replay: Sampling States in Model-Based RL via Simulated Priorities
Figure 3 for Beyond Prioritized Replay: Sampling States in Model-Based RL via Simulated Priorities
Figure 4 for Beyond Prioritized Replay: Sampling States in Model-Based RL via Simulated Priorities
Viaarxiv icon

On the Global Convergence Rates of Softmax Policy Gradient Methods

Add code
Bookmark button
Alert button
May 13, 2020
Jincheng Mei, Chenjun Xiao, Csaba Szepesvari, Dale Schuurmans

Figure 1 for On the Global Convergence Rates of Softmax Policy Gradient Methods
Figure 2 for On the Global Convergence Rates of Softmax Policy Gradient Methods
Figure 3 for On the Global Convergence Rates of Softmax Policy Gradient Methods
Figure 4 for On the Global Convergence Rates of Softmax Policy Gradient Methods
Viaarxiv icon