Alert button
Picture for Karl Tuyls

Karl Tuyls

Alert button

OpenSpiel: A Framework for Reinforcement Learning in Games

Add code
Bookmark button
Alert button
Aug 28, 2019
Marc Lanctot, Edward Lockhart, Jean-Baptiste Lespiau, Vinicius Zambaldi, Satyaki Upadhyay, Julien Pérolat, Sriram Srinivasan, Finbarr Timbers, Karl Tuyls, Shayegan Omidshafiei, Daniel Hennes, Dustin Morrill, Paul Muller, Timo Ewalds, Ryan Faulkner, János Kramár, Bart De Vylder, Brennan Saeta, James Bradbury, David Ding, Sebastian Borgeaud, Matthew Lai, Julian Schrittwieser, Thomas Anthony, Edward Hughes, Ivo Danihelka, Jonah Ryan-Davis

Figure 1 for OpenSpiel: A Framework for Reinforcement Learning in Games
Figure 2 for OpenSpiel: A Framework for Reinforcement Learning in Games
Figure 3 for OpenSpiel: A Framework for Reinforcement Learning in Games
Figure 4 for OpenSpiel: A Framework for Reinforcement Learning in Games
Viaarxiv icon

Neural Replicator Dynamics

Add code
Bookmark button
Alert button
Jun 01, 2019
Shayegan Omidshafiei, Daniel Hennes, Dustin Morrill, Remi Munos, Julien Perolat, Marc Lanctot, Audrunas Gruslys, Jean-Baptiste Lespiau, Karl Tuyls

Figure 1 for Neural Replicator Dynamics
Figure 2 for Neural Replicator Dynamics
Figure 3 for Neural Replicator Dynamics
Figure 4 for Neural Replicator Dynamics
Viaarxiv icon

Differentiable Game Mechanics

Add code
Bookmark button
Alert button
May 13, 2019
Alistair Letcher, David Balduzzi, Sebastien Racaniere, James Martens, Jakob Foerster, Karl Tuyls, Thore Graepel

Figure 1 for Differentiable Game Mechanics
Figure 2 for Differentiable Game Mechanics
Figure 3 for Differentiable Game Mechanics
Figure 4 for Differentiable Game Mechanics
Viaarxiv icon

Evolving Indoor Navigational Strategies Using Gated Recurrent Units In NEAT

Add code
Bookmark button
Alert button
Apr 12, 2019
James Butterworth, Rahul Savani, Karl Tuyls

Figure 1 for Evolving Indoor Navigational Strategies Using Gated Recurrent Units In NEAT
Figure 2 for Evolving Indoor Navigational Strategies Using Gated Recurrent Units In NEAT
Viaarxiv icon

Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent

Add code
Bookmark button
Alert button
Mar 21, 2019
Edward Lockhart, Marc Lanctot, Julien Pérolat, Jean-Baptiste Lespiau, Dustin Morrill, Finbarr Timbers, Karl Tuyls

Figure 1 for Computing Approximate Equilibria in Sequential Adversarial Games by Exploitability Descent
Viaarxiv icon

Fully Convolutional One-Shot Object Segmentation for Industrial Robotics

Add code
Bookmark button
Alert button
Mar 02, 2019
Benjamin Schnieders, Shan Luo, Gregory Palmer, Karl Tuyls

Figure 1 for Fully Convolutional One-Shot Object Segmentation for Industrial Robotics
Figure 2 for Fully Convolutional One-Shot Object Segmentation for Industrial Robotics
Figure 3 for Fully Convolutional One-Shot Object Segmentation for Industrial Robotics
Figure 4 for Fully Convolutional One-Shot Object Segmentation for Industrial Robotics
Viaarxiv icon

Robust temporal difference learning for critical domains

Add code
Bookmark button
Alert button
Jan 23, 2019
Richard Klima, Daan Bloembergen, Michael Kaisers, Karl Tuyls

Figure 1 for Robust temporal difference learning for critical domains
Figure 2 for Robust temporal difference learning for critical domains
Figure 3 for Robust temporal difference learning for critical domains
Figure 4 for Robust temporal difference learning for critical domains
Viaarxiv icon

Re-evaluating Evaluation

Add code
Bookmark button
Alert button
Oct 30, 2018
David Balduzzi, Karl Tuyls, Julien Perolat, Thore Graepel

Figure 1 for Re-evaluating Evaluation
Figure 2 for Re-evaluating Evaluation
Figure 3 for Re-evaluating Evaluation
Figure 4 for Re-evaluating Evaluation
Viaarxiv icon

Actor-Critic Policy Optimization in Partially Observable Multiagent Environments

Add code
Bookmark button
Alert button
Oct 21, 2018
Sriram Srinivasan, Marc Lanctot, Vinicius Zambaldi, Julien Perolat, Karl Tuyls, Remi Munos, Michael Bowling

Figure 1 for Actor-Critic Policy Optimization in Partially Observable Multiagent Environments
Figure 2 for Actor-Critic Policy Optimization in Partially Observable Multiagent Environments
Figure 3 for Actor-Critic Policy Optimization in Partially Observable Multiagent Environments
Figure 4 for Actor-Critic Policy Optimization in Partially Observable Multiagent Environments
Viaarxiv icon

Inequity aversion improves cooperation in intertemporal social dilemmas

Add code
Bookmark button
Alert button
Sep 27, 2018
Edward Hughes, Joel Z. Leibo, Matthew G. Phillips, Karl Tuyls, Edgar A. Duéñez-Guzmán, Antonio García Castañeda, Iain Dunning, Tina Zhu, Kevin R. McKee, Raphael Koster, Heather Roff, Thore Graepel

Figure 1 for Inequity aversion improves cooperation in intertemporal social dilemmas
Figure 2 for Inequity aversion improves cooperation in intertemporal social dilemmas
Figure 3 for Inequity aversion improves cooperation in intertemporal social dilemmas
Figure 4 for Inequity aversion improves cooperation in intertemporal social dilemmas
Viaarxiv icon