Picture for Pratap Tokekar

Pratap Tokekar

University of Maryland, College Park

Option Discovery Using LLM-guided Semantic Hierarchical Reinforcement Learning

Add code
Mar 24, 2025
Figure 1 for Option Discovery Using LLM-guided Semantic Hierarchical Reinforcement Learning
Figure 2 for Option Discovery Using LLM-guided Semantic Hierarchical Reinforcement Learning
Figure 3 for Option Discovery Using LLM-guided Semantic Hierarchical Reinforcement Learning
Figure 4 for Option Discovery Using LLM-guided Semantic Hierarchical Reinforcement Learning
Viaarxiv icon

Learning Multi-Robot Coordination through Locality-Based Factorized Multi-Agent Actor-Critic Algorithm

Add code
Mar 24, 2025
Figure 1 for Learning Multi-Robot Coordination through Locality-Based Factorized Multi-Agent Actor-Critic Algorithm
Figure 2 for Learning Multi-Robot Coordination through Locality-Based Factorized Multi-Agent Actor-Critic Algorithm
Figure 3 for Learning Multi-Robot Coordination through Locality-Based Factorized Multi-Agent Actor-Critic Algorithm
Figure 4 for Learning Multi-Robot Coordination through Locality-Based Factorized Multi-Agent Actor-Critic Algorithm
Viaarxiv icon

PEnGUiN: Partially Equivariant Graph NeUral Networks for Sample Efficient MARL

Add code
Mar 19, 2025
Figure 1 for PEnGUiN: Partially Equivariant Graph NeUral Networks for Sample Efficient MARL
Figure 2 for PEnGUiN: Partially Equivariant Graph NeUral Networks for Sample Efficient MARL
Figure 3 for PEnGUiN: Partially Equivariant Graph NeUral Networks for Sample Efficient MARL
Figure 4 for PEnGUiN: Partially Equivariant Graph NeUral Networks for Sample Efficient MARL
Viaarxiv icon

VARP: Reinforcement Learning from Vision-Language Model Feedback with Agent Regularized Preferences

Add code
Mar 18, 2025
Viaarxiv icon

Sketch-to-Skill: Bootstrapping Robot Learning with Human Drawn Trajectory Sketches

Add code
Mar 14, 2025
Figure 1 for Sketch-to-Skill: Bootstrapping Robot Learning with Human Drawn Trajectory Sketches
Figure 2 for Sketch-to-Skill: Bootstrapping Robot Learning with Human Drawn Trajectory Sketches
Figure 3 for Sketch-to-Skill: Bootstrapping Robot Learning with Human Drawn Trajectory Sketches
Figure 4 for Sketch-to-Skill: Bootstrapping Robot Learning with Human Drawn Trajectory Sketches
Viaarxiv icon

CAML: Collaborative Auxiliary Modality Learning for Multi-Agent Systems

Add code
Feb 25, 2025
Viaarxiv icon

AUKT: Adaptive Uncertainty-Guided Knowledge Transfer with Conformal Prediction

Add code
Feb 25, 2025
Figure 1 for AUKT: Adaptive Uncertainty-Guided Knowledge Transfer with Conformal Prediction
Figure 2 for AUKT: Adaptive Uncertainty-Guided Knowledge Transfer with Conformal Prediction
Figure 3 for AUKT: Adaptive Uncertainty-Guided Knowledge Transfer with Conformal Prediction
Figure 4 for AUKT: Adaptive Uncertainty-Guided Knowledge Transfer with Conformal Prediction
Viaarxiv icon

When to Localize? A POMDP Approach

Add code
Nov 13, 2024
Figure 1 for When to Localize? A POMDP Approach
Figure 2 for When to Localize? A POMDP Approach
Figure 3 for When to Localize? A POMDP Approach
Figure 4 for When to Localize? A POMDP Approach
Viaarxiv icon

When to Localize? A Risk-Constrained Reinforcement Learning Approach

Add code
Nov 05, 2024
Figure 1 for When to Localize? A Risk-Constrained Reinforcement Learning Approach
Figure 2 for When to Localize? A Risk-Constrained Reinforcement Learning Approach
Figure 3 for When to Localize? A Risk-Constrained Reinforcement Learning Approach
Figure 4 for When to Localize? A Risk-Constrained Reinforcement Learning Approach
Viaarxiv icon

On the Sample Complexity of a Policy Gradient Algorithm with Occupancy Approximation for General Utility Reinforcement Learning

Add code
Oct 05, 2024
Figure 1 for On the Sample Complexity of a Policy Gradient Algorithm with Occupancy Approximation for General Utility Reinforcement Learning
Figure 2 for On the Sample Complexity of a Policy Gradient Algorithm with Occupancy Approximation for General Utility Reinforcement Learning
Figure 3 for On the Sample Complexity of a Policy Gradient Algorithm with Occupancy Approximation for General Utility Reinforcement Learning
Figure 4 for On the Sample Complexity of a Policy Gradient Algorithm with Occupancy Approximation for General Utility Reinforcement Learning
Viaarxiv icon