Alert button
Picture for Yannick Schroecker

Yannick Schroecker

Alert button

Vision-Language Models as a Source of Rewards

Dec 14, 2023
Kate Baumli, Satinder Baveja, Feryal Behbahani, Harris Chan, Gheorghe Comanici, Sebastian Flennerhag, Maxime Gazeau, Kristian Holsheimer, Dan Horgan, Michael Laskin, Clare Lyle, Hussain Masoom, Kay McKinney, Volodymyr Mnih, Alexander Neitz, Fabio Pardo, Jack Parker-Holder, John Quan, Tim Rocktäschel, Himanshu Sahni, Tom Schaul, Yannick Schroecker, Stephen Spencer, Richie Steigerwald, Luyu Wang, Lei Zhang

Viaarxiv icon

Structured State Space Models for In-Context Reinforcement Learning

Mar 09, 2023
Chris Lu, Yannick Schroecker, Albert Gu, Emilio Parisotto, Jakob Foerster, Satinder Singh, Feryal Behbahani

Figure 1 for Structured State Space Models for In-Context Reinforcement Learning
Figure 2 for Structured State Space Models for In-Context Reinforcement Learning
Figure 3 for Structured State Space Models for In-Context Reinforcement Learning
Figure 4 for Structured State Space Models for In-Context Reinforcement Learning
Viaarxiv icon

Human-Timescale Adaptation in an Open-Ended Task Space

Jan 18, 2023
Adaptive Agent Team, Jakob Bauer, Kate Baumli, Satinder Baveja, Feryal Behbahani, Avishkar Bhoopchand, Nathalie Bradley-Schmieg, Michael Chang, Natalie Clay, Adrian Collister, Vibhavari Dasagi, Lucy Gonzalez, Karol Gregor, Edward Hughes, Sheleem Kashem, Maria Loks-Thompson, Hannah Openshaw, Jack Parker-Holder, Shreya Pathak, Nicolas Perez-Nieves, Nemanja Rakicevic, Tim Rocktäschel, Yannick Schroecker, Jakub Sygnowski, Karl Tuyls, Sarah York, Alexander Zacherl, Lei Zhang

Figure 1 for Human-Timescale Adaptation in an Open-Ended Task Space
Figure 2 for Human-Timescale Adaptation in an Open-Ended Task Space
Figure 3 for Human-Timescale Adaptation in an Open-Ended Task Space
Figure 4 for Human-Timescale Adaptation in an Open-Ended Task Space
Viaarxiv icon

Meta-Gradients in Non-Stationary Environments

Sep 13, 2022
Jelena Luketina, Sebastian Flennerhag, Yannick Schroecker, David Abel, Tom Zahavy, Satinder Singh

Figure 1 for Meta-Gradients in Non-Stationary Environments
Figure 2 for Meta-Gradients in Non-Stationary Environments
Figure 3 for Meta-Gradients in Non-Stationary Environments
Figure 4 for Meta-Gradients in Non-Stationary Environments
Viaarxiv icon

Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality

May 26, 2022
Tom Zahavy, Yannick Schroecker, Feryal Behbahani, Kate Baumli, Sebastian Flennerhag, Shaobo Hou, Satinder Singh

Figure 1 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 2 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 3 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 4 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Viaarxiv icon

Bootstrapped Meta-Learning

Sep 09, 2021
Sebastian Flennerhag, Yannick Schroecker, Tom Zahavy, Hado van Hasselt, David Silver, Satinder Singh

Figure 1 for Bootstrapped Meta-Learning
Figure 2 for Bootstrapped Meta-Learning
Figure 3 for Bootstrapped Meta-Learning
Figure 4 for Bootstrapped Meta-Learning
Viaarxiv icon

Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning

Feb 15, 2020
Yannick Schroecker, Charles Isbell

Figure 1 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Figure 2 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Figure 3 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Figure 4 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Viaarxiv icon

Active Learning within Constrained Environments through Imitation of an Expert Questioner

Jul 01, 2019
Kalesha Bullard, Yannick Schroecker, Sonia Chernova

Figure 1 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Figure 2 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Figure 3 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Figure 4 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Viaarxiv icon

Generative predecessor models for sample-efficient imitation learning

Apr 01, 2019
Yannick Schroecker, Mel Vecerik, Jonathan Scholz

Figure 1 for Generative predecessor models for sample-efficient imitation learning
Figure 2 for Generative predecessor models for sample-efficient imitation learning
Viaarxiv icon

Imitating Latent Policies from Observation

May 24, 2018
Ashley D. Edwards, Himanshu Sahni, Yannick Schroecker, Charles L. Isbell

Figure 1 for Imitating Latent Policies from Observation
Figure 2 for Imitating Latent Policies from Observation
Figure 3 for Imitating Latent Policies from Observation
Figure 4 for Imitating Latent Policies from Observation
Viaarxiv icon