Alert button
Picture for Jianye Hao

Jianye Hao

Alert button

ED2: An Environment Dynamics Decomposition Framework for World Model Construction

Add code
Bookmark button
Alert button
Dec 06, 2021
Cong Wang, Tianpei Yang, Jianye Hao, Yan Zheng, Hongyao Tang, Fazl Barez, Jinyi Liu, Jiajie Peng, Haiyin Piao, Zhixiao Sun

Figure 1 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Figure 2 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Figure 3 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Figure 4 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Viaarxiv icon

Learning State Representations via Retracing in Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 24, 2021
Changmin Yu, Dong Li, Jianye Hao, Jun Wang, Neil Burgess

Figure 1 for Learning State Representations via Retracing in Reinforcement Learning
Figure 2 for Learning State Representations via Retracing in Reinforcement Learning
Figure 3 for Learning State Representations via Retracing in Reinforcement Learning
Figure 4 for Learning State Representations via Retracing in Reinforcement Learning
Viaarxiv icon

Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Nov 19, 2021
Tong Sang, Hongyao Tang, Jianye Hao, Yan Zheng, Zhaopeng Meng

Figure 1 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 2 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 3 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 4 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Viaarxiv icon

Lifelong Reinforcement Learning with Temporal Logic Formulas and Reward Machines

Add code
Bookmark button
Alert button
Nov 18, 2021
Xuejing Zheng, Chao Yu, Chen Chen, Jianye Hao, Hankz Hankui Zhuo

Figure 1 for Lifelong Reinforcement Learning with Temporal Logic Formulas and Reward Machines
Figure 2 for Lifelong Reinforcement Learning with Temporal Logic Formulas and Reward Machines
Figure 3 for Lifelong Reinforcement Learning with Temporal Logic Formulas and Reward Machines
Figure 4 for Lifelong Reinforcement Learning with Temporal Logic Formulas and Reward Machines
Viaarxiv icon

SEIHAI: A Sample-efficient Hierarchical AI for the MineRL Competition

Add code
Bookmark button
Alert button
Nov 17, 2021
Hangyu Mao, Chao Wang, Xiaotian Hao, Yihuan Mao, Yiming Lu, Chengjie Wu, Jianye Hao, Dong Li, Pingzhong Tang

Figure 1 for SEIHAI: A Sample-efficient Hierarchical AI for the MineRL Competition
Figure 2 for SEIHAI: A Sample-efficient Hierarchical AI for the MineRL Competition
Figure 3 for SEIHAI: A Sample-efficient Hierarchical AI for the MineRL Competition
Figure 4 for SEIHAI: A Sample-efficient Hierarchical AI for the MineRL Competition
Viaarxiv icon

Dynamic Bottleneck for Robust Self-Supervised Exploration

Add code
Bookmark button
Alert button
Oct 25, 2021
Chenjia Bai, Lingxiao Wang, Lei Han, Animesh Garg, Jianye Hao, Peng Liu, Zhaoran Wang

Figure 1 for Dynamic Bottleneck for Robust Self-Supervised Exploration
Figure 2 for Dynamic Bottleneck for Robust Self-Supervised Exploration
Figure 3 for Dynamic Bottleneck for Robust Self-Supervised Exploration
Figure 4 for Dynamic Bottleneck for Robust Self-Supervised Exploration
Viaarxiv icon

Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning

Add code
Bookmark button
Alert button
Oct 09, 2021
Danruo Deng, Guangyong Chen, Jianye Hao, Qiong Wang, Pheng-Ann Heng

Figure 1 for Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning
Figure 2 for Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning
Figure 3 for Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning
Figure 4 for Flattening Sharpness for Dynamic Gradient Projection Memory Benefits Continual Learning
Viaarxiv icon

Ranking Cost: Building An Efficient and Scalable Circuit Routing Planner with Evolution-Based Optimization

Add code
Bookmark button
Alert button
Oct 08, 2021
Shiyu Huang, Bin Wang, Dong Li, Jianye Hao, Ting Chen, Jun Zhu

Figure 1 for Ranking Cost: Building An Efficient and Scalable Circuit Routing Planner with Evolution-Based Optimization
Figure 2 for Ranking Cost: Building An Efficient and Scalable Circuit Routing Planner with Evolution-Based Optimization
Figure 3 for Ranking Cost: Building An Efficient and Scalable Circuit Routing Planner with Evolution-Based Optimization
Figure 4 for Ranking Cost: Building An Efficient and Scalable Circuit Routing Planner with Evolution-Based Optimization
Viaarxiv icon

Exploration in Deep Reinforcement Learning: A Comprehensive Survey

Add code
Bookmark button
Alert button
Sep 15, 2021
Tianpei Yang, Hongyao Tang, Chenjia Bai, Jinyi Liu, Jianye Hao, Zhaopeng Meng, Peng Liu

Figure 1 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Figure 2 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Figure 3 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Figure 4 for Exploration in Deep Reinforcement Learning: A Comprehensive Survey
Viaarxiv icon

HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation

Add code
Bookmark button
Alert button
Sep 12, 2021
Boyan Li, Hongyao Tang, Yan Zheng, Jianye Hao, Pengyi Li, Zhen Wang, Zhaopeng Meng, Li Wang

Figure 1 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Figure 2 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Figure 3 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Figure 4 for HyAR: Addressing Discrete-Continuous Action Reinforcement Learning via Hybrid Action Representation
Viaarxiv icon