Alert button
Picture for Liyu Chen

Liyu Chen

Alert button

$\mathbf{(N,K)}$-Puzzle: A Cost-Efficient Testbed for Benchmarking Reinforcement Learning Algorithms in Generative Language Model

Add code
Bookmark button
Alert button
Mar 11, 2024
Yufeng Zhang, Liyu Chen, Boyi Liu, Yingxiang Yang, Qiwen Cui, Yunzhe Tao, Hongxia Yang

Figure 1 for $\mathbf{(N,K)}$-Puzzle: A Cost-Efficient Testbed for Benchmarking Reinforcement Learning Algorithms in Generative Language Model
Figure 2 for $\mathbf{(N,K)}$-Puzzle: A Cost-Efficient Testbed for Benchmarking Reinforcement Learning Algorithms in Generative Language Model
Figure 3 for $\mathbf{(N,K)}$-Puzzle: A Cost-Efficient Testbed for Benchmarking Reinforcement Learning Algorithms in Generative Language Model
Figure 4 for $\mathbf{(N,K)}$-Puzzle: A Cost-Efficient Testbed for Benchmarking Reinforcement Learning Algorithms in Generative Language Model
Viaarxiv icon

$\mathcal{B}$-Coder: Value-Based Deep Reinforcement Learning for Program Synthesis

Add code
Bookmark button
Alert button
Oct 04, 2023
Zishun Yu, Yunzhe Tao, Liyu Chen, Tao Sun, Hongxia Yang

Figure 1 for $\mathcal{B}$-Coder: Value-Based Deep Reinforcement Learning for Program Synthesis
Figure 2 for $\mathcal{B}$-Coder: Value-Based Deep Reinforcement Learning for Program Synthesis
Figure 3 for $\mathcal{B}$-Coder: Value-Based Deep Reinforcement Learning for Program Synthesis
Figure 4 for $\mathcal{B}$-Coder: Value-Based Deep Reinforcement Learning for Program Synthesis
Viaarxiv icon

Layered State Discovery for Incremental Autonomous Exploration

Add code
Bookmark button
Alert button
Feb 07, 2023
Liyu Chen, Andrea Tirinzoni, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Layered State Discovery for Incremental Autonomous Exploration
Figure 2 for Layered State Discovery for Incremental Autonomous Exploration
Viaarxiv icon

Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path

Add code
Bookmark button
Alert button
Oct 10, 2022
Liyu Chen, Andrea Tirinzoni, Matteo Pirotta, Alessandro Lazaric

Figure 1 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Figure 2 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Figure 3 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Viaarxiv icon

Follow-the-Perturbed-Leader for Adversarial Markov Decision Processes with Bandit Feedback

Add code
Bookmark button
Alert button
May 26, 2022
Yan Dai, Haipeng Luo, Liyu Chen

Figure 1 for Follow-the-Perturbed-Leader for Adversarial Markov Decision Processes with Bandit Feedback
Viaarxiv icon

Near-Optimal Goal-Oriented Reinforcement Learning in Non-Stationary Environments

Add code
Bookmark button
Alert button
May 25, 2022
Liyu Chen, Haipeng Luo

Viaarxiv icon

Policy Learning and Evaluation with Randomized Quasi-Monte Carlo

Add code
Bookmark button
Alert button
Feb 21, 2022
Sebastien M. R. Arnold, Pierre L'Ecuyer, Liyu Chen, Yi-fan Chen, Fei Sha

Figure 1 for Policy Learning and Evaluation with Randomized Quasi-Monte Carlo
Figure 2 for Policy Learning and Evaluation with Randomized Quasi-Monte Carlo
Figure 3 for Policy Learning and Evaluation with Randomized Quasi-Monte Carlo
Figure 4 for Policy Learning and Evaluation with Randomized Quasi-Monte Carlo
Viaarxiv icon

Policy Optimization for Stochastic Shortest Path

Add code
Bookmark button
Alert button
Feb 07, 2022
Liyu Chen, Haipeng Luo, Aviv Rosenberg

Viaarxiv icon

Learning Infinite-Horizon Average-Reward Markov Decision Processes with Constraints

Add code
Bookmark button
Alert button
Jan 31, 2022
Liyu Chen, Rahul Jain, Haipeng Luo

Viaarxiv icon