Alert button
Picture for Ian Osband

Ian Osband

Alert button

Scalable Coordinated Exploration in Concurrent Reinforcement Learning

Add code
Bookmark button
Alert button
May 23, 2018
Maria Dimakopoulou, Ian Osband, Benjamin Van Roy

Figure 1 for Scalable Coordinated Exploration in Concurrent Reinforcement Learning
Figure 2 for Scalable Coordinated Exploration in Concurrent Reinforcement Learning
Figure 3 for Scalable Coordinated Exploration in Concurrent Reinforcement Learning
Viaarxiv icon

Noisy Networks for Exploration

Add code
Bookmark button
Alert button
Feb 15, 2018
Meire Fortunato, Mohammad Gheshlaghi Azar, Bilal Piot, Jacob Menick, Ian Osband, Alex Graves, Vlad Mnih, Remi Munos, Demis Hassabis, Olivier Pietquin, Charles Blundell, Shane Legg

Figure 1 for Noisy Networks for Exploration
Figure 2 for Noisy Networks for Exploration
Figure 3 for Noisy Networks for Exploration
Figure 4 for Noisy Networks for Exploration
Viaarxiv icon

Gaussian-Dirichlet Posterior Dominance in Sequential Learning

Add code
Bookmark button
Alert button
Feb 09, 2018
Ian Osband, Benjamin Van Roy

Figure 1 for Gaussian-Dirichlet Posterior Dominance in Sequential Learning
Figure 2 for Gaussian-Dirichlet Posterior Dominance in Sequential Learning
Viaarxiv icon

Deep Q-learning from Demonstrations

Add code
Bookmark button
Alert button
Nov 22, 2017
Todd Hester, Matej Vecerik, Olivier Pietquin, Marc Lanctot, Tom Schaul, Bilal Piot, Dan Horgan, John Quan, Andrew Sendonaris, Gabriel Dulac-Arnold, Ian Osband, John Agapiou, Joel Z. Leibo, Audrunas Gruslys

Figure 1 for Deep Q-learning from Demonstrations
Figure 2 for Deep Q-learning from Demonstrations
Figure 3 for Deep Q-learning from Demonstrations
Viaarxiv icon

A Tutorial on Thompson Sampling

Add code
Bookmark button
Alert button
Nov 19, 2017
Daniel Russo, Benjamin Van Roy, Abbas Kazerouni, Ian Osband, Zheng Wen

Figure 1 for A Tutorial on Thompson Sampling
Figure 2 for A Tutorial on Thompson Sampling
Figure 3 for A Tutorial on Thompson Sampling
Figure 4 for A Tutorial on Thompson Sampling
Viaarxiv icon

Minimax Regret Bounds for Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 01, 2017
Mohammad Gheshlaghi Azar, Ian Osband, Rémi Munos

Viaarxiv icon

On Optimistic versus Randomized Exploration in Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 13, 2017
Ian Osband, Benjamin Van Roy

Figure 1 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 2 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 3 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 4 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Viaarxiv icon

Why is Posterior Sampling Better than Optimism for Reinforcement Learning?

Add code
Bookmark button
Alert button
Jun 13, 2017
Ian Osband, Benjamin Van Roy

Figure 1 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 2 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 3 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 4 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Viaarxiv icon

On Lower Bounds for Regret in Reinforcement Learning

Add code
Bookmark button
Alert button
Aug 09, 2016
Ian Osband, Benjamin Van Roy

Figure 1 for On Lower Bounds for Regret in Reinforcement Learning
Viaarxiv icon

Posterior Sampling for Reinforcement Learning Without Episodes

Add code
Bookmark button
Alert button
Aug 09, 2016
Ian Osband, Benjamin Van Roy

Viaarxiv icon