Picture for Zhang-Wei Hong

Zhang-Wei Hong

ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization

Add code
Oct 17, 2024
Figure 1 for ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization
Figure 2 for ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization
Figure 3 for ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization
Figure 4 for ORSO: Accelerating Reward Design via Online Reward Selection and Policy Optimization
Viaarxiv icon

Random Latent Exploration for Deep Reinforcement Learning

Add code
Jul 18, 2024
Figure 1 for Random Latent Exploration for Deep Reinforcement Learning
Figure 2 for Random Latent Exploration for Deep Reinforcement Learning
Figure 3 for Random Latent Exploration for Deep Reinforcement Learning
Figure 4 for Random Latent Exploration for Deep Reinforcement Learning
Viaarxiv icon

ROER: Regularized Optimal Experience Replay

Add code
Jul 04, 2024
Figure 1 for ROER: Regularized Optimal Experience Replay
Figure 2 for ROER: Regularized Optimal Experience Replay
Figure 3 for ROER: Regularized Optimal Experience Replay
Figure 4 for ROER: Regularized Optimal Experience Replay
Viaarxiv icon

Text-to-Drive: Diverse Driving Behavior Synthesis via Large Language Models

Add code
Jun 06, 2024
Figure 1 for Text-to-Drive: Diverse Driving Behavior Synthesis via Large Language Models
Figure 2 for Text-to-Drive: Diverse Driving Behavior Synthesis via Large Language Models
Figure 3 for Text-to-Drive: Diverse Driving Behavior Synthesis via Large Language Models
Figure 4 for Text-to-Drive: Diverse Driving Behavior Synthesis via Large Language Models
Viaarxiv icon

Curiosity-driven Red-teaming for Large Language Models

Add code
Feb 29, 2024
Figure 1 for Curiosity-driven Red-teaming for Large Language Models
Figure 2 for Curiosity-driven Red-teaming for Large Language Models
Figure 3 for Curiosity-driven Red-teaming for Large Language Models
Figure 4 for Curiosity-driven Red-teaming for Large Language Models
Viaarxiv icon

Neuro-Inspired Fragmentation and Recall to Overcome Catastrophic Forgetting in Curiosity

Add code
Oct 26, 2023
Figure 1 for Neuro-Inspired Fragmentation and Recall to Overcome Catastrophic Forgetting in Curiosity
Figure 2 for Neuro-Inspired Fragmentation and Recall to Overcome Catastrophic Forgetting in Curiosity
Figure 3 for Neuro-Inspired Fragmentation and Recall to Overcome Catastrophic Forgetting in Curiosity
Figure 4 for Neuro-Inspired Fragmentation and Recall to Overcome Catastrophic Forgetting in Curiosity
Viaarxiv icon

Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets

Add code
Oct 12, 2023
Figure 1 for Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets
Figure 2 for Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets
Figure 3 for Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets
Figure 4 for Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets
Viaarxiv icon

Parallel $Q$-Learning: Scaling Off-policy Reinforcement Learning under Massively Parallel Simulation

Add code
Jul 24, 2023
Viaarxiv icon

Neuro-Inspired Efficient Map Building via Fragmentation and Recall

Add code
Jul 11, 2023
Viaarxiv icon

TGRL: An Algorithm for Teacher Guided Reinforcement Learning

Add code
Jul 06, 2023
Viaarxiv icon