Alert button
Picture for Ping-Chun Hsieh

Ping-Chun Hsieh

Alert button

Image Deraining via Self-supervised Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 27, 2024
He-Hao Liao, Yan-Tsung Peng, Wen-Tao Chu, Ping-Chun Hsieh, Chung-Chi Tsai

Viaarxiv icon

Offline Imitation of Badminton Player Behavior via Experiential Contexts and Brownian Motion

Add code
Bookmark button
Alert button
Mar 19, 2024
Kuang-Da Wang, Wei-Yao Wang, Ping-Chun Hsieh, Wen-Chih Peng

Figure 1 for Offline Imitation of Badminton Player Behavior via Experiential Contexts and Brownian Motion
Figure 2 for Offline Imitation of Badminton Player Behavior via Experiential Contexts and Brownian Motion
Figure 3 for Offline Imitation of Badminton Player Behavior via Experiential Contexts and Brownian Motion
Figure 4 for Offline Imitation of Badminton Player Behavior via Experiential Contexts and Brownian Motion
Viaarxiv icon

PPO-Clip Attains Global Optimality: Towards Deeper Understandings of Clipping

Add code
Bookmark button
Alert button
Dec 19, 2023
Nai-Chieh Huang, Ping-Chun Hsieh, Kuo-Hao Ho, I-Chen Wu

Viaarxiv icon

Accelerated Policy Gradient: On the Nesterov Momentum for Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 18, 2023
Yen-Ju Chen, Nai-Chieh Huang, Ping-Chun Hsieh

Viaarxiv icon

Value-Biased Maximum Likelihood Estimation for Model-based Reinforcement Learning in Discounted Linear MDPs

Add code
Bookmark button
Alert button
Oct 17, 2023
Yu-Heng Hung, Ping-Chun Hsieh, Akshay Mete, P. R. Kumar

Viaarxiv icon

Towards Human-Like RL: Taming Non-Naturalistic Behavior in Deep RL via Adaptive Behavioral Costs in 3D Games

Add code
Bookmark button
Alert button
Sep 27, 2023
Kuo-Hao Ho, Ping-Chun Hsieh, Chiu-Chou Lin, You-Ren Luo, Feng-Jian Wang, I-Chen Wu

Viaarxiv icon

Coordinate Ascent for Off-Policy RL with Global Convergence Guarantees

Add code
Bookmark button
Alert button
Dec 10, 2022
Hsin-En Su, Yen-Ju Chen, Ping-Chun Hsieh, Xi Liu

Figure 1 for Coordinate Ascent for Off-Policy RL with Global Convergence Guarantees
Figure 2 for Coordinate Ascent for Off-Policy RL with Global Convergence Guarantees
Figure 3 for Coordinate Ascent for Off-Policy RL with Global Convergence Guarantees
Figure 4 for Coordinate Ascent for Off-Policy RL with Global Convergence Guarantees
Viaarxiv icon

Q-Pensieve: Boosting Sample Efficiency of Multi-Objective RL Through Memory Sharing of Q-Snapshots

Add code
Bookmark button
Alert button
Dec 06, 2022
Wei Hung, Bo-Kai Huang, Ping-Chun Hsieh, Xi Liu

Figure 1 for Q-Pensieve: Boosting Sample Efficiency of Multi-Objective RL Through Memory Sharing of Q-Snapshots
Figure 2 for Q-Pensieve: Boosting Sample Efficiency of Multi-Objective RL Through Memory Sharing of Q-Snapshots
Figure 3 for Q-Pensieve: Boosting Sample Efficiency of Multi-Objective RL Through Memory Sharing of Q-Snapshots
Figure 4 for Q-Pensieve: Boosting Sample Efficiency of Multi-Objective RL Through Memory Sharing of Q-Snapshots
Viaarxiv icon

Neural Frank-Wolfe Policy Optimization for Region-of-Interest Intra-Frame Coding with HEVC/H.265

Add code
Bookmark button
Alert button
Sep 27, 2022
Yung-Han Ho, Chia-Hao Kao, Wen-Hsiao Peng, Ping-Chun Hsieh

Figure 1 for Neural Frank-Wolfe Policy Optimization for Region-of-Interest Intra-Frame Coding with HEVC/H.265
Figure 2 for Neural Frank-Wolfe Policy Optimization for Region-of-Interest Intra-Frame Coding with HEVC/H.265
Figure 3 for Neural Frank-Wolfe Policy Optimization for Region-of-Interest Intra-Frame Coding with HEVC/H.265
Figure 4 for Neural Frank-Wolfe Policy Optimization for Region-of-Interest Intra-Frame Coding with HEVC/H.265
Viaarxiv icon

Neural Contextual Bandits via Reward-Biased Maximum Likelihood Estimation

Add code
Bookmark button
Alert button
Mar 08, 2022
Yu-Heng Hung, Ping-Chun Hsieh

Figure 1 for Neural Contextual Bandits via Reward-Biased Maximum Likelihood Estimation
Figure 2 for Neural Contextual Bandits via Reward-Biased Maximum Likelihood Estimation
Figure 3 for Neural Contextual Bandits via Reward-Biased Maximum Likelihood Estimation
Figure 4 for Neural Contextual Bandits via Reward-Biased Maximum Likelihood Estimation
Viaarxiv icon