Alert button
Picture for Zhaopeng Meng

Zhaopeng Meng

Alert button

Qibo: A Large Language Model for Traditional Chinese Medicine

Add code
Bookmark button
Alert button
Mar 24, 2024
Heyi Zhang, Xin Wang, Zhaopeng Meng, Yongzhe Jia, Dawei Xu

Viaarxiv icon

Learning with Noisy Labels Using Collaborative Sample Selection and Contrastive Semi-Supervised Learning

Add code
Bookmark button
Alert button
Oct 24, 2023
Qing Miao, Xiaohe Wu, Chao Xu, Yanli Ji, Wangmeng Zuo, Yiwen Guo, Zhaopeng Meng

Viaarxiv icon

Ensemble-based Offline-to-Online Reinforcement Learning: From Pessimistic Learning to Optimistic Exploration

Add code
Bookmark button
Alert button
Jun 12, 2023
Kai Zhao, Yi Ma, Jinyi Liu, Yan Zheng, Zhaopeng Meng

Figure 1 for Ensemble-based Offline-to-Online Reinforcement Learning: From Pessimistic Learning to Optimistic Exploration
Figure 2 for Ensemble-based Offline-to-Online Reinforcement Learning: From Pessimistic Learning to Optimistic Exploration
Figure 3 for Ensemble-based Offline-to-Online Reinforcement Learning: From Pessimistic Learning to Optimistic Exploration
Figure 4 for Ensemble-based Offline-to-Online Reinforcement Learning: From Pessimistic Learning to Optimistic Exploration
Viaarxiv icon

HIPODE: Enhancing Offline Reinforcement Learning with High-Quality Synthetic Data from a Policy-Decoupled Approach

Add code
Bookmark button
Alert button
Jun 10, 2023
Shixi Lian, Yi Ma, Jinyi Liu, Yan Zheng, Zhaopeng Meng

Figure 1 for HIPODE: Enhancing Offline Reinforcement Learning with High-Quality Synthetic Data from a Policy-Decoupled Approach
Figure 2 for HIPODE: Enhancing Offline Reinforcement Learning with High-Quality Synthetic Data from a Policy-Decoupled Approach
Figure 3 for HIPODE: Enhancing Offline Reinforcement Learning with High-Quality Synthetic Data from a Policy-Decoupled Approach
Figure 4 for HIPODE: Enhancing Offline Reinforcement Learning with High-Quality Synthetic Data from a Policy-Decoupled Approach
Viaarxiv icon

In-Sample Policy Iteration for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 09, 2023
Xiaohan Hu, Yi Ma, Chenjun Xiao, Yan Zheng, Zhaopeng Meng

Figure 1 for In-Sample Policy Iteration for Offline Reinforcement Learning
Figure 2 for In-Sample Policy Iteration for Offline Reinforcement Learning
Figure 3 for In-Sample Policy Iteration for Offline Reinforcement Learning
Figure 4 for In-Sample Policy Iteration for Offline Reinforcement Learning
Viaarxiv icon

ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation

Add code
Bookmark button
Alert button
Oct 26, 2022
Pengyi Li, Hongyao Tang, Jianye Hao, Yan Zheng, Xian Fu, Zhaopeng Meng

Figure 1 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Figure 2 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Figure 3 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Figure 4 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Viaarxiv icon

PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations

Add code
Bookmark button
Alert button
Apr 06, 2022
Tong Sang, Hongyao Tang, Yi Ma, Jianye Hao, Yan Zheng, Zhaopeng Meng, Boyan Li, Zhen Wang

Figure 1 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Figure 2 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Figure 3 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Figure 4 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Viaarxiv icon

Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 19, 2021
Tong Sang, Hongyao Tang, Jianye Hao, Yan Zheng, Zhaopeng Meng

Figure 1 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 2 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 3 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 4 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Viaarxiv icon

Exploration in Deep Reinforcement Learning: A Comprehensive Survey

Add code
Bookmark button
Alert button
Sep 15, 2021
Tianpei Yang, Hongyao Tang, Chenjia Bai, Jinyi Liu, Jianye Hao, Zhaopeng Meng, Peng Liu

Figure 1 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Figure 2 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Figure 3 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Figure 4 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Viaarxiv icon

HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation

Add code
Bookmark button
Alert button
Sep 12, 2021
Boyan Li, Hongyao Tang, Yan Zheng, Jianye Hao, Pengyi Li, Zhen Wang, Zhaopeng Meng, Li Wang

Figure 1 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Figure 2 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Figure 3 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Figure 4 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Viaarxiv icon