Picture for Hongyao Tang

Hongyao Tang

MFE-ETP: A Comprehensive Evaluation Benchmark for Multi-modal Foundation Models on Embodied Task Planning

Add code
Jul 06, 2024
Figure 1 for MFE-ETP: A Comprehensive Evaluation Benchmark for Multi-modal Foundation Models on Embodied Task Planning
Figure 2 for MFE-ETP: A Comprehensive Evaluation Benchmark for Multi-modal Foundation Models on Embodied Task Planning
Figure 3 for MFE-ETP: A Comprehensive Evaluation Benchmark for Multi-modal Foundation Models on Embodied Task Planning
Figure 4 for MFE-ETP: A Comprehensive Evaluation Benchmark for Multi-modal Foundation Models on Embodied Task Planning
Viaarxiv icon

Bridging Evolutionary Algorithms and Reinforcement Learning: A Comprehensive Survey

Add code
Jan 22, 2024
Figure 1 for Bridging Evolutionary Algorithms and Reinforcement Learning: A Comprehensive Survey
Figure 2 for Bridging Evolutionary Algorithms and Reinforcement Learning: A Comprehensive Survey
Figure 3 for Bridging Evolutionary Algorithms and Reinforcement Learning: A Comprehensive Survey
Figure 4 for Bridging Evolutionary Algorithms and Reinforcement Learning: A Comprehensive Survey
Viaarxiv icon

The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting

Add code
Mar 02, 2023
Figure 1 for The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting
Figure 2 for The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting
Figure 3 for The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting
Figure 4 for The Ladder in Chaos: A Simple and Effective Improvement to General DRL Algorithms by Policy Path Trimming and Boosting
Viaarxiv icon

State-Aware Proximal Pessimistic Algorithms for Offline Reinforcement Learning

Add code
Nov 28, 2022
Figure 1 for State-Aware Proximal Pessimistic Algorithms for Offline Reinforcement Learning
Figure 2 for State-Aware Proximal Pessimistic Algorithms for Offline Reinforcement Learning
Figure 3 for State-Aware Proximal Pessimistic Algorithms for Offline Reinforcement Learning
Figure 4 for State-Aware Proximal Pessimistic Algorithms for Offline Reinforcement Learning
Viaarxiv icon

ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation

Add code
Oct 26, 2022
Figure 1 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Figure 2 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Figure 3 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Figure 4 for ERL-Re$^2$: Efficient Evolutionary Reinforcement Learning with Shared State Representation and Individual Policy Representation
Viaarxiv icon

Towards A Unified Policy Abstraction Theory and Representation Learning Approach in Markov Decision Processes

Add code
Sep 16, 2022
Figure 1 for Towards A Unified Policy Abstraction Theory and Representation Learning Approach in Markov Decision Processes
Figure 2 for Towards A Unified Policy Abstraction Theory and Representation Learning Approach in Markov Decision Processes
Figure 3 for Towards A Unified Policy Abstraction Theory and Representation Learning Approach in Markov Decision Processes
Figure 4 for Towards A Unified Policy Abstraction Theory and Representation Learning Approach in Markov Decision Processes
Viaarxiv icon

PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations

Add code
Apr 06, 2022
Figure 1 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Figure 2 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Figure 3 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Figure 4 for PAnDR: Fast Adaptation to New Environments from Offline Experiences via Decoupling Policy and Environment Representations
Viaarxiv icon

PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration

Add code
Mar 16, 2022
Figure 1 for PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
Figure 2 for PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
Figure 3 for PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
Figure 4 for PMIC: Improving Multi-Agent Reinforcement Learning with Progressive Mutual Information Collaboration
Viaarxiv icon

ED2: An Environment Dynamics Decomposition Framework for World Model Construction

Add code
Dec 06, 2021
Figure 1 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Figure 2 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Figure 3 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Figure 4 for ED2: An Environment Dynamics Decomposition Framework for World Model Construction
Viaarxiv icon

Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning

Add code
Nov 19, 2021
Figure 1 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 2 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 3 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Figure 4 for Uncertainty-aware Low-Rank Q-Matrix Estimation for Deep Reinforcement Learning
Viaarxiv icon