Alert button
Picture for Yannick Schroecker

Yannick Schroecker

Alert button

Vision-Language Models as a Source of Rewards

Add code
Bookmark button
Alert button
Dec 14, 2023
Kate Baumli, Satinder Baveja, Feryal Behbahani, Harris Chan, Gheorghe Comanici, Sebastian Flennerhag, Maxime Gazeau, Kristian Holsheimer, Dan Horgan, Michael Laskin, Clare Lyle, Hussain Masoom, Kay McKinney, Volodymyr Mnih, Alexander Neitz, Fabio Pardo, Jack Parker-Holder, John Quan, Tim Rocktäschel, Himanshu Sahni, Tom Schaul, Yannick Schroecker, Stephen Spencer, Richie Steigerwald, Luyu Wang, Lei Zhang

Viaarxiv icon

Structured State Space Models for In-Context Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 09, 2023
Chris Lu, Yannick Schroecker, Albert Gu, Emilio Parisotto, Jakob Foerster, Satinder Singh, Feryal Behbahani

Figure 1 for Structured State Space Models for In-Context Reinforcement Learning
Figure 2 for Structured State Space Models for In-Context Reinforcement Learning
Figure 3 for Structured State Space Models for In-Context Reinforcement Learning
Figure 4 for Structured State Space Models for In-Context Reinforcement Learning
Viaarxiv icon

Human-Timescale Adaptation in an Open-Ended Task Space

Add code
Bookmark button
Alert button
Jan 18, 2023
Adaptive Agent Team, Jakob Bauer, Kate Baumli, Satinder Baveja, Feryal Behbahani, Avishkar Bhoopchand, Nathalie Bradley-Schmieg, Michael Chang, Natalie Clay, Adrian Collister, Vibhavari Dasagi, Lucy Gonzalez, Karol Gregor, Edward Hughes, Sheleem Kashem, Maria Loks-Thompson, Hannah Openshaw, Jack Parker-Holder, Shreya Pathak, Nicolas Perez-Nieves, Nemanja Rakicevic, Tim Rocktäschel, Yannick Schroecker, Jakub Sygnowski, Karl Tuyls, Sarah York, Alexander Zacherl, Lei Zhang

Figure 1 for Human-Timescale Adaptation in an Open-Ended Task Space
Figure 2 for Human-Timescale Adaptation in an Open-Ended Task Space
Figure 3 for Human-Timescale Adaptation in an Open-Ended Task Space
Figure 4 for Human-Timescale Adaptation in an Open-Ended Task Space
Viaarxiv icon

Meta-Gradients in Non-Stationary Environments

Add code
Bookmark button
Alert button
Sep 13, 2022
Jelena Luketina, Sebastian Flennerhag, Yannick Schroecker, David Abel, Tom Zahavy, Satinder Singh

Figure 1 for Meta-Gradients in Non-Stationary Environments
Figure 2 for Meta-Gradients in Non-Stationary Environments
Figure 3 for Meta-Gradients in Non-Stationary Environments
Figure 4 for Meta-Gradients in Non-Stationary Environments
Viaarxiv icon

Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality

Add code
Bookmark button
Alert button
May 26, 2022
Tom Zahavy, Yannick Schroecker, Feryal Behbahani, Kate Baumli, Sebastian Flennerhag, Shaobo Hou, Satinder Singh

Figure 1 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 2 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 3 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Figure 4 for Discovering Policies with DOMiNO: Diversity Optimization Maintaining Near Optimality
Viaarxiv icon

Bootstrapped Meta-Learning

Add code
Bookmark button
Alert button
Sep 09, 2021
Sebastian Flennerhag, Yannick Schroecker, Tom Zahavy, Hado van Hasselt, David Silver, Satinder Singh

Figure 1 for Bootstrapped Meta-Learning
Figure 2 for Bootstrapped Meta-Learning
Figure 3 for Bootstrapped Meta-Learning
Figure 4 for Bootstrapped Meta-Learning
Viaarxiv icon

Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 15, 2020
Yannick Schroecker, Charles Isbell

Figure 1 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Figure 2 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Figure 3 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Figure 4 for Universal Value Density Estimation for Imitation Learning and Goal-Conditioned Reinforcement Learning
Viaarxiv icon

Active Learning within Constrained Environments through Imitation of an Expert Questioner

Add code
Bookmark button
Alert button
Jul 01, 2019
Kalesha Bullard, Yannick Schroecker, Sonia Chernova

Figure 1 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Figure 2 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Figure 3 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Figure 4 for Active Learning within Constrained Environments through Imitation of an Expert Questioner
Viaarxiv icon

Generative predecessor models for sample-efficient imitation learning

Add code
Bookmark button
Alert button
Apr 01, 2019
Yannick Schroecker, Mel Vecerik, Jonathan Scholz

Figure 1 for Generative predecessor models for sample-efficient imitation learning
Figure 2 for Generative predecessor models for sample-efficient imitation learning
Viaarxiv icon

Imitating Latent Policies from Observation

Add code
Bookmark button
Alert button
May 24, 2018
Ashley D. Edwards, Himanshu Sahni, Yannick Schroecker, Charles L. Isbell

Figure 1 for Imitating Latent Policies from Observation
Figure 2 for Imitating Latent Policies from Observation
Figure 3 for Imitating Latent Policies from Observation
Figure 4 for Imitating Latent Policies from Observation
Viaarxiv icon