Alert button
Picture for Michal Valko

Michal Valko

Alert button

Improved Sample Complexity for Incremental Autonomous Exploration in MDPs

Add code
Bookmark button
Alert button
Dec 29, 2020
Jean Tarbouriech, Matteo Pirotta, Michal Valko, Alessandro Lazaric

Figure 1 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Figure 2 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Figure 3 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Figure 4 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Viaarxiv icon

Game Plan: What AI can do for Football, and What Football can do for AI

Add code
Bookmark button
Alert button
Nov 18, 2020
Karl Tuyls, Shayegan Omidshafiei, Paul Muller, Zhe Wang, Jerome Connor, Daniel Hennes, Ian Graham, William Spearman, Tim Waskett, Dafydd Steele, Pauline Luc, Adria Recasens, Alexandre Galashov, Gregory Thornton, Romuald Elie, Pablo Sprechmann, Pol Moreno, Kris Cao, Marta Garnelo, Praneet Dutta, Michal Valko, Nicolas Heess, Alex Bridgland, Julien Perolat, Bart De Vylder, Ali Eslami, Mark Rowland, Andrew Jaegle, Remi Munos, Trevor Back, Razia Ahamed, Simon Bouton, Nathalie Beauguerlange, Jackson Broshear, Thore Graepel, Demis Hassabis

Figure 1 for Game Plan: What AI can do for Football, and What Football can do for AI
Figure 2 for Game Plan: What AI can do for Football, and What Football can do for AI
Figure 3 for Game Plan: What AI can do for Football, and What Football can do for AI
Figure 4 for Game Plan: What AI can do for Football, and What Football can do for AI
Viaarxiv icon

BYOL works even without batch statistics

Add code
Bookmark button
Alert button
Oct 20, 2020
Pierre H. Richemond, Jean-Bastien Grill, Florent Altché, Corentin Tallec, Florian Strub, Andrew Brock, Samuel Smith, Soham De, Razvan Pascanu, Bilal Piot, Michal Valko

Figure 1 for BYOL works even without batch statistics
Figure 2 for BYOL works even without batch statistics
Viaarxiv icon

Episodic Reinforcement Learning in Finite MDPs: Minimax Lower Bounds Revisited

Add code
Bookmark button
Alert button
Oct 07, 2020
Omar Darwiche Domingues, Pierre Ménard, Emilie Kaufmann, Michal Valko

Figure 1 for Episodic Reinforcement Learning in Finite MDPs: Minimax Lower Bounds Revisited
Figure 2 for Episodic Reinforcement Learning in Finite MDPs: Minimax Lower Bounds Revisited
Viaarxiv icon

Fast active learning for pure exploration in reinforcement learning

Add code
Bookmark button
Alert button
Jul 27, 2020
Pierre Ménard, Omar Darwiche Domingues, Anders Jonsson, Emilie Kaufmann, Edouard Leurent, Michal Valko

Figure 1 for Fast active learning for pure exploration in reinforcement learning
Viaarxiv icon

Monte-Carlo Tree Search as Regularized Policy Optimization

Add code
Bookmark button
Alert button
Jul 24, 2020
Jean-Bastien Grill, Florent Altché, Yunhao Tang, Thomas Hubert, Michal Valko, Ioannis Antonoglou, Rémi Munos

Figure 1 for Monte-Carlo Tree Search as Regularized Policy Optimization
Figure 2 for Monte-Carlo Tree Search as Regularized Policy Optimization
Figure 3 for Monte-Carlo Tree Search as Regularized Policy Optimization
Figure 4 for Monte-Carlo Tree Search as Regularized Policy Optimization
Viaarxiv icon

A Provably Efficient Sample Collection Strategy for Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 13, 2020
Jean Tarbouriech, Matteo Pirotta, Michal Valko, Alessandro Lazaric

Figure 1 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 2 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 3 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 4 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Viaarxiv icon

A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces

Add code
Bookmark button
Alert button
Jul 09, 2020
Omar Darwiche Domingues, Pierre Ménard, Matteo Pirotta, Emilie Kaufmann, Michal Valko

Figure 1 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 2 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 3 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 4 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Viaarxiv icon

Gamification of Pure Exploration for Linear Bandits

Add code
Bookmark button
Alert button
Jul 02, 2020
Rémy Degenne, Pierre Ménard, Xuedong Shang, Michal Valko

Figure 1 for Gamification of Pure Exploration for Linear Bandits
Figure 2 for Gamification of Pure Exploration for Linear Bandits
Figure 3 for Gamification of Pure Exploration for Linear Bandits
Figure 4 for Gamification of Pure Exploration for Linear Bandits
Viaarxiv icon

Sampling from a $k$-DPP without looking at all items

Add code
Bookmark button
Alert button
Jun 30, 2020
Daniele Calandriello, Michał Dereziński, Michal Valko

Figure 1 for Sampling from a $k$-DPP without looking at all items
Figure 2 for Sampling from a $k$-DPP without looking at all items
Figure 3 for Sampling from a $k$-DPP without looking at all items
Viaarxiv icon