Alert button
Picture for Finbarr Timbers

Finbarr Timbers

Alert button

Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 30, 2022
Julien Perolat, Bart de Vylder, Daniel Hennes, Eugene Tarassov, Florian Strub, Vincent de Boer, Paul Muller, Jerome T. Connor, Neil Burch, Thomas Anthony, Stephen McAleer, Romuald Elie, Sarah H. Cen, Zhe Wang, Audrunas Gruslys, Aleksandra Malysheva, Mina Khan, Sherjil Ozair, Finbarr Timbers, Toby Pohlen, Tom Eccles, Mark Rowland, Marc Lanctot, Jean-Baptiste Lespiau, Bilal Piot, Shayegan Omidshafiei, Edward Lockhart, Laurent Sifre, Nathalie Beauguerlange, Remi Munos, David Silver, Satinder Singh, Demis Hassabis, Karl Tuyls

Figure 1 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Figure 2 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Figure 3 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Figure 4 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Viaarxiv icon

Emergent Bartering Behaviour in Multi-Agent Reinforcement Learning

Add code
Bookmark button
Alert button
May 13, 2022
Michael Bradley Johanson, Edward Hughes, Finbarr Timbers, Joel Z. Leibo

Figure 1 for Emergent Bartering Behaviour in Multi-Agent Reinforcement Learning
Figure 2 for Emergent Bartering Behaviour in Multi-Agent Reinforcement Learning
Figure 3 for Emergent Bartering Behaviour in Multi-Agent Reinforcement Learning
Figure 4 for Emergent Bartering Behaviour in Multi-Agent Reinforcement Learning
Viaarxiv icon

Reward-Respecting Subtasks for Model-Based Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 09, 2022
Richard S. Sutton, Marlos C. Machado, G. Zacharias Holland, David Szepesvari, Finbarr Timbers, Brian Tanner, Adam White

Figure 1 for Reward-Respecting Subtasks for Model-Based Reinforcement Learning
Figure 2 for Reward-Respecting Subtasks for Model-Based Reinforcement Learning
Figure 3 for Reward-Respecting Subtasks for Model-Based Reinforcement Learning
Figure 4 for Reward-Respecting Subtasks for Model-Based Reinforcement Learning
Viaarxiv icon

Player of Games

Add code
Bookmark button
Alert button
Dec 06, 2021
Martin Schmid, Matej Moravcik, Neil Burch, Rudolf Kadlec, Josh Davidson, Kevin Waugh, Nolan Bard, Finbarr Timbers, Marc Lanctot, Zach Holland, Elnaz Davoodi, Alden Christianson, Michael Bowling

Figure 1 for Player of Games
Figure 2 for Player of Games
Figure 3 for Player of Games
Figure 4 for Player of Games
Viaarxiv icon

Solving Common-Payoff Games with Approximate Policy Iteration

Add code
Bookmark button
Alert button
Jan 11, 2021
Samuel Sokota, Edward Lockhart, Finbarr Timbers, Elnaz Davoodi, Ryan D'Orazio, Neil Burch, Martin Schmid, Michael Bowling, Marc Lanctot

Figure 1 for Solving Common-Payoff Games with Approximate Policy Iteration
Figure 2 for Solving Common-Payoff Games with Approximate Policy Iteration
Figure 3 for Solving Common-Payoff Games with Approximate Policy Iteration
Figure 4 for Solving Common-Payoff Games with Approximate Policy Iteration
Viaarxiv icon

The Advantage Regret-Matching Actor-Critic

Add code
Bookmark button
Alert button
Aug 27, 2020
Audrūnas Gruslys, Marc Lanctot, Rémi Munos, Finbarr Timbers, Martin Schmid, Julien Perolat, Dustin Morrill, Vinicius Zambaldi, Jean-Baptiste Lespiau, John Schultz, Mohammad Gheshlaghi Azar, Michael Bowling, Karl Tuyls

Figure 1 for The Advantage Regret-Matching Actor-Critic
Figure 2 for The Advantage Regret-Matching Actor-Critic
Figure 3 for The Advantage Regret-Matching Actor-Critic
Figure 4 for The Advantage Regret-Matching Actor-Critic
Viaarxiv icon

Approximate exploitability: Learning a best response in large games

Add code
Bookmark button
Alert button
Apr 20, 2020
Finbarr Timbers, Edward Lockhart, Martin Schmid, Marc Lanctot, Michael Bowling

Figure 1 for Approximate exploitability: Learning a best response in large games
Figure 2 for Approximate exploitability: Learning a best response in large games
Figure 3 for Approximate exploitability: Learning a best response in large games
Figure 4 for Approximate exploitability: Learning a best response in large games
Viaarxiv icon

OpenSpiel: A Framework for Reinforcement Learning in Games

Add code
Bookmark button
Alert button
Oct 10, 2019
Marc Lanctot, Edward Lockhart, Jean-Baptiste Lespiau, Vinicius Zambaldi, Satyaki Upadhyay, Julien Pérolat, Sriram Srinivasan, Finbarr Timbers, Karl Tuyls, Shayegan Omidshafiei, Daniel Hennes, Dustin Morrill, Paul Muller, Timo Ewalds, Ryan Faulkner, János Kramár, Bart De Vylder, Brennan Saeta, James Bradbury, David Ding, Sebastian Borgeaud, Matthew Lai, Julian Schrittwieser, Thomas Anthony, Edward Hughes, Ivo Danihelka, Jonah Ryan-Davis

Figure 1 for OpenSpiel: A Framework for Reinforcement Learning in Games
Figure 2 for OpenSpiel: A Framework for Reinforcement Learning in Games
Figure 3 for OpenSpiel: A Framework for Reinforcement Learning in Games
Figure 4 for OpenSpiel: A Framework for Reinforcement Learning in Games
Viaarxiv icon