Picture for Satinder Singh

Satinder Singh

Hierarchical Reinforcement Learning in Complex 3D Environments

Add code
Feb 28, 2023
Figure 1 for Hierarchical Reinforcement Learning in Complex 3D Environments
Figure 2 for Hierarchical Reinforcement Learning in Complex 3D Environments
Figure 3 for Hierarchical Reinforcement Learning in Complex 3D Environments
Figure 4 for Hierarchical Reinforcement Learning in Complex 3D Environments
Viaarxiv icon

ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs

Add code
Feb 02, 2023
Figure 1 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 2 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 3 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 4 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Viaarxiv icon

Composing Task Knowledge with Modular Successor Feature Approximators

Add code
Jan 28, 2023
Figure 1 for Composing Task Knowledge with Modular Successor Feature Approximators
Figure 2 for Composing Task Knowledge with Modular Successor Feature Approximators
Figure 3 for Composing Task Knowledge with Modular Successor Feature Approximators
Figure 4 for Composing Task Knowledge with Modular Successor Feature Approximators
Viaarxiv icon

Optimistic Meta-Gradients

Add code
Jan 09, 2023
Figure 1 for Optimistic Meta-Gradients
Figure 2 for Optimistic Meta-Gradients
Figure 3 for Optimistic Meta-Gradients
Figure 4 for Optimistic Meta-Gradients
Viaarxiv icon

POMRL: No-Regret Learning-to-Plan with Increasing Horizons

Add code
Dec 30, 2022
Viaarxiv icon

Discovering Evolution Strategies via Meta-Black-Box Optimization

Add code
Nov 25, 2022
Figure 1 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Figure 2 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Figure 3 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Figure 4 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Viaarxiv icon

Planning to the Information Horizon of BAMDPs via Epistemic State Abstraction

Add code
Oct 30, 2022
Viaarxiv icon

In-context Reinforcement Learning with Algorithm Distillation

Add code
Oct 25, 2022
Figure 1 for In-context Reinforcement Learning with Algorithm Distillation
Figure 2 for In-context Reinforcement Learning with Algorithm Distillation
Figure 3 for In-context Reinforcement Learning with Algorithm Distillation
Figure 4 for In-context Reinforcement Learning with Algorithm Distillation
Viaarxiv icon

In-Context Policy Iteration

Add code
Oct 07, 2022
Figure 1 for In-Context Policy Iteration
Figure 2 for In-Context Policy Iteration
Figure 3 for In-Context Policy Iteration
Figure 4 for In-Context Policy Iteration
Viaarxiv icon

Meta-Gradients in Non-Stationary Environments

Add code
Sep 13, 2022
Figure 1 for Meta-Gradients in Non-Stationary Environments
Figure 2 for Meta-Gradients in Non-Stationary Environments
Figure 3 for Meta-Gradients in Non-Stationary Environments
Figure 4 for Meta-Gradients in Non-Stationary Environments
Viaarxiv icon