Picture for Sebastian Flennerhag

Sebastian Flennerhag

Vision-Language Models as a Source of Rewards

Add code
Dec 14, 2023
Figure 1 for Vision-Language Models as a Source of Rewards
Figure 2 for Vision-Language Models as a Source of Rewards
Figure 3 for Vision-Language Models as a Source of Rewards
Figure 4 for Vision-Language Models as a Source of Rewards
Viaarxiv icon

Optimism and Adaptivity in Policy Optimization

Add code
Jun 18, 2023
Figure 1 for Optimism and Adaptivity in Policy Optimization
Figure 2 for Optimism and Adaptivity in Policy Optimization
Figure 3 for Optimism and Adaptivity in Policy Optimization
Figure 4 for Optimism and Adaptivity in Policy Optimization
Viaarxiv icon

Discovering Attention-Based Genetic Algorithms via Meta-Black-Box Optimization

Add code
Apr 08, 2023
Figure 1 for Discovering Attention-Based Genetic Algorithms via Meta-Black-Box Optimization
Figure 2 for Discovering Attention-Based Genetic Algorithms via Meta-Black-Box Optimization
Figure 3 for Discovering Attention-Based Genetic Algorithms via Meta-Black-Box Optimization
Figure 4 for Discovering Attention-Based Genetic Algorithms via Meta-Black-Box Optimization
Viaarxiv icon

ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs

Add code
Feb 02, 2023
Figure 1 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 2 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 3 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 4 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Viaarxiv icon

Optimistic Meta-Gradients

Add code
Jan 09, 2023
Figure 1 for Optimistic Meta-Gradients
Figure 2 for Optimistic Meta-Gradients
Figure 3 for Optimistic Meta-Gradients
Figure 4 for Optimistic Meta-Gradients
Viaarxiv icon

Discovering Evolution Strategies via Meta-Black-Box Optimization

Add code
Nov 25, 2022
Figure 1 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Figure 2 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Figure 3 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Figure 4 for Discovering Evolution Strategies via Meta-Black-Box Optimization
Viaarxiv icon

Meta-Gradients in Non-Stationary Environments

Add code
Sep 13, 2022
Figure 1 for Meta-Gradients in Non-Stationary Environments
Figure 2 for Meta-Gradients in Non-Stationary Environments
Figure 3 for Meta-Gradients in Non-Stationary Environments
Figure 4 for Meta-Gradients in Non-Stationary Environments
Viaarxiv icon

Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality

Add code
May 26, 2022
Figure 1 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 2 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 3 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 4 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Viaarxiv icon

Introducing Symmetries to Black Box Meta Reinforcement Learning

Add code
Sep 22, 2021
Figure 1 for Introducing Symmetries to Black Box Meta Reinforcement Learning
Figure 2 for Introducing Symmetries to Black Box Meta Reinforcement Learning
Figure 3 for Introducing Symmetries to Black Box Meta Reinforcement Learning
Figure 4 for Introducing Symmetries to Black Box Meta Reinforcement Learning
Viaarxiv icon

Bootstrapped Meta-Learning

Add code
Sep 09, 2021
Figure 1 for Bootstrapped Meta-Learning
Figure 2 for Bootstrapped Meta-Learning
Figure 3 for Bootstrapped Meta-Learning
Figure 4 for Bootstrapped Meta-Learning
Viaarxiv icon