Alert button
Picture for Demis Hassabis

Demis Hassabis

Alert button

Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 30, 2022
Julien Perolat, Bart de Vylder, Daniel Hennes, Eugene Tarassov, Florian Strub, Vincent de Boer, Paul Muller, Jerome T. Connor, Neil Burch, Thomas Anthony, Stephen McAleer, Romuald Elie, Sarah H. Cen, Zhe Wang, Audrunas Gruslys, Aleksandra Malysheva, Mina Khan, Sherjil Ozair, Finbarr Timbers, Toby Pohlen, Tom Eccles, Mark Rowland, Marc Lanctot, Jean-Baptiste Lespiau, Bilal Piot, Shayegan Omidshafiei, Edward Lockhart, Laurent Sifre, Nathalie Beauguerlange, Remi Munos, David Silver, Satinder Singh, Demis Hassabis, Karl Tuyls

Figure 1 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Figure 2 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Figure 3 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Figure 4 for Mastering the Game of Stratego with Model-Free Multiagent Reinforcement Learning
Viaarxiv icon

Scaling Language Models: Methods, Analysis & Insights from Training Gopher

Add code
Bookmark button
Alert button
Dec 08, 2021
Jack W. Rae, Sebastian Borgeaud, Trevor Cai, Katie Millican, Jordan Hoffmann, Francis Song, John Aslanides, Sarah Henderson, Roman Ring, Susannah Young, Eliza Rutherford, Tom Hennigan, Jacob Menick, Albin Cassirer, Richard Powell, George van den Driessche, Lisa Anne Hendricks, Maribeth Rauh, Po-Sen Huang, Amelia Glaese, Johannes Welbl, Sumanth Dathathri, Saffron Huang, Jonathan Uesato, John Mellor, Irina Higgins, Antonia Creswell, Nat McAleese, Amy Wu, Erich Elsen, Siddhant Jayakumar, Elena Buchatskaya, David Budden, Esme Sutherland, Karen Simonyan, Michela Paganini, Laurent Sifre, Lena Martens, Xiang Lorraine Li, Adhiguna Kuncoro, Aida Nematzadeh, Elena Gribovskaya, Domenic Donato, Angeliki Lazaridou, Arthur Mensch, Jean-Baptiste Lespiau, Maria Tsimpoukelli, Nikolai Grigorev, Doug Fritz, Thibault Sottiaux, Mantas Pajarskas, Toby Pohlen, Zhitao Gong, Daniel Toyama, Cyprien de Masson d'Autume, Yujia Li, Tayfun Terzi, Vladimir Mikulik, Igor Babuschkin, Aidan Clark, Diego de Las Casas, Aurelia Guy, Chris Jones, James Bradbury, Matthew Johnson, Blake Hechtman, Laura Weidinger, Iason Gabriel, William Isaac, Ed Lockhart, Simon Osindero, Laura Rimell, Chris Dyer, Oriol Vinyals, Kareem Ayoub, Jeff Stanway, Lorrayne Bennett, Demis Hassabis, Koray Kavukcuoglu, Geoffrey Irving

Figure 1 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 2 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 3 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Figure 4 for Scaling Language Models: Methods, Analysis & Insights from Training Gopher
Viaarxiv icon

Acquisition of Chess Knowledge in AlphaZero

Add code
Bookmark button
Alert button
Nov 27, 2021
Thomas McGrath, Andrei Kapishnikov, Nenad Tomašev, Adam Pearce, Demis Hassabis, Been Kim, Ulrich Paquet, Vladimir Kramnik

Figure 1 for Acquisition of Chess Knowledge in AlphaZero
Figure 2 for Acquisition of Chess Knowledge in AlphaZero
Figure 3 for Acquisition of Chess Knowledge in AlphaZero
Figure 4 for Acquisition of Chess Knowledge in AlphaZero
Viaarxiv icon

Alchemy: A structured task distribution for meta-reinforcement learning

Add code
Bookmark button
Alert button
Feb 04, 2021
Jane X. Wang, Michael King, Nicolas Porcel, Zeb Kurth-Nelson, Tina Zhu, Charlie Deck, Peter Choy, Mary Cassin, Malcolm Reynolds, Francis Song, Gavin Buttimore, David P. Reichert, Neil Rabinowitz, Loic Matthey, Demis Hassabis, Alexander Lerchner, Matthew Botvinick

Figure 1 for Alchemy: A structured task distribution for meta-reinforcement learning
Figure 2 for Alchemy: A structured task distribution for meta-reinforcement learning
Figure 3 for Alchemy: A structured task distribution for meta-reinforcement learning
Figure 4 for Alchemy: A structured task distribution for meta-reinforcement learning
Viaarxiv icon

Game Plan: What AI can do for Football, and What Football can do for AI

Add code
Bookmark button
Alert button
Nov 18, 2020
Karl Tuyls, Shayegan Omidshafiei, Paul Muller, Zhe Wang, Jerome Connor, Daniel Hennes, Ian Graham, William Spearman, Tim Waskett, Dafydd Steele, Pauline Luc, Adria Recasens, Alexandre Galashov, Gregory Thornton, Romuald Elie, Pablo Sprechmann, Pol Moreno, Kris Cao, Marta Garnelo, Praneet Dutta, Michal Valko, Nicolas Heess, Alex Bridgland, Julien Perolat, Bart De Vylder, Ali Eslami, Mark Rowland, Andrew Jaegle, Remi Munos, Trevor Back, Razia Ahamed, Simon Bouton, Nathalie Beauguerlange, Jackson Broshear, Thore Graepel, Demis Hassabis

Figure 1 for Game Plan: What AI can do for Football, and What Football can do for AI
Figure 2 for Game Plan: What AI can do for Football, and What Football can do for AI
Figure 3 for Game Plan: What AI can do for Football, and What Football can do for AI
Figure 4 for Game Plan: What AI can do for Football, and What Football can do for AI
Viaarxiv icon

Assessing Game Balance with AlphaZero: Exploring Alternative Rule Sets in Chess

Add code
Bookmark button
Alert button
Sep 15, 2020
Nenad Tomašev, Ulrich Paquet, Demis Hassabis, Vladimir Kramnik

Figure 1 for Assessing Game Balance with AlphaZero: Exploring Alternative Rule Sets in Chess
Figure 2 for Assessing Game Balance with AlphaZero: Exploring Alternative Rule Sets in Chess
Figure 3 for Assessing Game Balance with AlphaZero: Exploring Alternative Rule Sets in Chess
Figure 4 for Assessing Game Balance with AlphaZero: Exploring Alternative Rule Sets in Chess
Viaarxiv icon

MEMO: A Deep Network for Flexible Combination of Episodic Memories

Add code
Bookmark button
Alert button
Jan 29, 2020
Andrea Banino, Adrià Puigdomènech Badia, Raphael Köster, Martin J. Chadwick, Vinicius Zambaldi, Demis Hassabis, Caswell Barry, Matthew Botvinick, Dharshan Kumaran, Charles Blundell

Figure 1 for MEMO: A Deep Network for Flexible Combination of Episodic Memories
Figure 2 for MEMO: A Deep Network for Flexible Combination of Episodic Memories
Figure 3 for MEMO: A Deep Network for Flexible Combination of Episodic Memories
Figure 4 for MEMO: A Deep Network for Flexible Combination of Episodic Memories
Viaarxiv icon

Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model

Add code
Bookmark button
Alert button
Nov 19, 2019
Julian Schrittwieser, Ioannis Antonoglou, Thomas Hubert, Karen Simonyan, Laurent Sifre, Simon Schmitt, Arthur Guez, Edward Lockhart, Demis Hassabis, Thore Graepel, Timothy Lillicrap, David Silver

Figure 1 for Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
Figure 2 for Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
Figure 3 for Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
Figure 4 for Mastering Atari, Go, Chess and Shogi by Planning with a Learned Model
Viaarxiv icon

Tackling Climate Change with Machine Learning

Add code
Bookmark button
Alert button
Jun 10, 2019
David Rolnick, Priya L. Donti, Lynn H. Kaack, Kelly Kochanski, Alexandre Lacoste, Kris Sankaran, Andrew Slavin Ross, Nikola Milojevic-Dupont, Natasha Jaques, Anna Waldman-Brown, Alexandra Luccioni, Tegan Maharaj, Evan D. Sherwin, S. Karthik Mukkavilli, Konrad P. Kording, Carla Gomes, Andrew Y. Ng, Demis Hassabis, John C. Platt, Felix Creutzig, Jennifer Chayes, Yoshua Bengio

Figure 1 for Tackling Climate Change with Machine Learning
Viaarxiv icon