Alert button
Picture for Xun Yu Zhou

Xun Yu Zhou

Alert button

Learning Merton's Strategies in an Incomplete Market: Recursive Entropy Regularization and Biased Gaussian Exploration

Add code
Bookmark button
Alert button
Dec 19, 2023
Min Dai, Yuchao Dong, Yanwei Jia, Xun Yu Zhou

Viaarxiv icon

Variable Clustering via Distributionally Robust Nodewise Regression

Add code
Bookmark button
Alert button
Dec 21, 2022
Kaizheng Wang, Xiao Xu, Xun Yu Zhou

Figure 1 for Variable Clustering via Distributionally Robust Nodewise Regression
Figure 2 for Variable Clustering via Distributionally Robust Nodewise Regression
Figure 3 for Variable Clustering via Distributionally Robust Nodewise Regression
Figure 4 for Variable Clustering via Distributionally Robust Nodewise Regression
Viaarxiv icon

Square-root regret bounds for continuous-time episodic Markov decision processes

Add code
Bookmark button
Alert button
Oct 03, 2022
Xuefeng Gao, Xun Yu Zhou

Figure 1 for Square-root regret bounds for continuous-time episodic Markov decision processes
Viaarxiv icon

Choquet regularization for reinforcement learning

Add code
Bookmark button
Alert button
Aug 17, 2022
Xia Han, Ruodu Wang, Xun Yu Zhou

Figure 1 for Choquet regularization for reinforcement learning
Figure 2 for Choquet regularization for reinforcement learning
Viaarxiv icon

q-Learning in Continuous Time

Add code
Bookmark button
Alert button
Jul 02, 2022
Yanwei Jia, Xun Yu Zhou

Figure 1 for q-Learning in Continuous Time
Figure 2 for q-Learning in Continuous Time
Figure 3 for q-Learning in Continuous Time
Figure 4 for q-Learning in Continuous Time
Viaarxiv icon

Logarithmic regret bounds for continuous-time average-reward Markov decision processes

Add code
Bookmark button
Alert button
May 24, 2022
Xuefeng Gao, Xun Yu Zhou

Viaarxiv icon

Policy Gradient and Actor-Critic Learning in Continuous Time and Space: Theory and Algorithms

Add code
Bookmark button
Alert button
Nov 22, 2021
Yanwei Jia, Xun Yu Zhou

Figure 1 for Policy Gradient and Actor-Critic Learning in Continuous Time and Space: Theory and Algorithms
Figure 2 for Policy Gradient and Actor-Critic Learning in Continuous Time and Space: Theory and Algorithms
Viaarxiv icon

Policy Evaluation and Temporal-Difference Learning in Continuous Time and Space: A Martingale Approach

Add code
Bookmark button
Alert button
Aug 15, 2021
Yanwei Jia, Xun Yu Zhou

Figure 1 for Policy Evaluation and Temporal-Difference Learning in Continuous Time and Space: A Martingale Approach
Figure 2 for Policy Evaluation and Temporal-Difference Learning in Continuous Time and Space: A Martingale Approach
Figure 3 for Policy Evaluation and Temporal-Difference Learning in Continuous Time and Space: A Martingale Approach
Figure 4 for Policy Evaluation and Temporal-Difference Learning in Continuous Time and Space: A Martingale Approach
Viaarxiv icon

Simulated annealing from continuum to discretization: a convergence analysis via the Eyring--Kramers law

Add code
Bookmark button
Alert button
Feb 09, 2021
Wenpin Tang, Xun Yu Zhou

Figure 1 for Simulated annealing from continuum to discretization: a convergence analysis via the Eyring--Kramers law
Viaarxiv icon