Alert button
Picture for Lars Buesing

Lars Buesing

Alert button

Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?

Jan 13, 2022
Nenad Tomasev, Ioana Bica, Brian McWilliams, Lars Buesing, Razvan Pascanu, Charles Blundell, Jovana Mitrovic

Figure 1 for Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?
Figure 2 for Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?
Figure 3 for Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?
Figure 4 for Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?
Viaarxiv icon

Counterfactual Credit Assignment in Model-Free Reinforcement Learning

Nov 18, 2020
Thomas Mesnard, Théophane Weber, Fabio Viola, Shantanu Thakoor, Alaa Saade, Anna Harutyunyan, Will Dabney, Tom Stepleton, Nicolas Heess, Arthur Guez, Marcus Hutter, Lars Buesing, Rémi Munos

Figure 1 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Figure 2 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Figure 3 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Figure 4 for Counterfactual Credit Assignment in Model-Free Reinforcement Learning
Viaarxiv icon

On the role of planning in model-based deep reinforcement learning

Nov 08, 2020
Jessica B. Hamrick, Abram L. Friesen, Feryal Behbahani, Arthur Guez, Fabio Viola, Sims Witherspoon, Thomas Anthony, Lars Buesing, Petar Veličković, Théophane Weber

Figure 1 for On the role of planning in model-based deep reinforcement learning
Figure 2 for On the role of planning in model-based deep reinforcement learning
Figure 3 for On the role of planning in model-based deep reinforcement learning
Figure 4 for On the role of planning in model-based deep reinforcement learning
Viaarxiv icon

Representation Learning via Invariant Causal Mechanisms

Oct 15, 2020
Jovana Mitrovic, Brian McWilliams, Jacob Walker, Lars Buesing, Charles Blundell

Figure 1 for Representation Learning via Invariant Causal Mechanisms
Figure 2 for Representation Learning via Invariant Causal Mechanisms
Figure 3 for Representation Learning via Invariant Causal Mechanisms
Figure 4 for Representation Learning via Invariant Causal Mechanisms
Viaarxiv icon

Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban

Oct 03, 2020
Peter Karkus, Mehdi Mirza, Arthur Guez, Andrew Jaegle, Timothy Lillicrap, Lars Buesing, Nicolas Heess, Theophane Weber

Figure 1 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Figure 2 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Figure 3 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Figure 4 for Beyond Tabula-Rasa: a Modular Reinforcement Learning Approach for Physically Embedded 3D Sokoban
Viaarxiv icon

Physically Embedded Planning Problems: New Challenges for Reinforcement Learning

Sep 11, 2020
Mehdi Mirza, Andrew Jaegle, Jonathan J. Hunt, Arthur Guez, Saran Tunyasuvunakool, Alistair Muldal, Théophane Weber, Peter Karkus, Sébastien Racanière, Lars Buesing, Timothy Lillicrap, Nicolas Heess

Figure 1 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Figure 2 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Figure 3 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Figure 4 for Physically Embedded Planning Problems: New Challenges for Reinforcement Learning
Viaarxiv icon

Pointer Graph Networks

Jun 11, 2020
Petar Veličković, Lars Buesing, Matthew C. Overlan, Razvan Pascanu, Oriol Vinyals, Charles Blundell

Figure 1 for Pointer Graph Networks
Figure 2 for Pointer Graph Networks
Figure 3 for Pointer Graph Networks
Figure 4 for Pointer Graph Networks
Viaarxiv icon

Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning

Apr 23, 2020
Giambattista Parascandolo, Lars Buesing, Josh Merel, Leonard Hasenclever, John Aslanides, Jessica B. Hamrick, Nicolas Heess, Alexander Neitz, Theophane Weber

Figure 1 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Figure 2 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Figure 3 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Figure 4 for Divide-and-Conquer Monte Carlo Tree Search For Goal-Directed Planning
Viaarxiv icon

Value-driven Hindsight Modelling

Feb 19, 2020
Arthur Guez, Fabio Viola, Théophane Weber, Lars Buesing, Steven Kapturowski, Doina Precup, David Silver, Nicolas Heess

Figure 1 for Value-driven Hindsight Modelling
Figure 2 for Value-driven Hindsight Modelling
Figure 3 for Value-driven Hindsight Modelling
Figure 4 for Value-driven Hindsight Modelling
Viaarxiv icon

Causally Correct Partial Models for Reinforcement Learning

Feb 07, 2020
Danilo J. Rezende, Ivo Danihelka, George Papamakarios, Nan Rosemary Ke, Ray Jiang, Theophane Weber, Karol Gregor, Hamza Merzic, Fabio Viola, Jane Wang, Jovana Mitrovic, Frederic Besse, Ioannis Antonoglou, Lars Buesing

Figure 1 for Causally Correct Partial Models for Reinforcement Learning
Figure 2 for Causally Correct Partial Models for Reinforcement Learning
Figure 3 for Causally Correct Partial Models for Reinforcement Learning
Figure 4 for Causally Correct Partial Models for Reinforcement Learning
Viaarxiv icon