Alert button
Picture for Benjamin Van Roy

Benjamin Van Roy

Alert button

Gaussian-Dirichlet Posterior Dominance in Sequential Learning

Add code
Bookmark button
Alert button
Feb 09, 2018
Ian Osband, Benjamin Van Roy

Figure 1 for Gaussian-Dirichlet Posterior Dominance in Sequential Learning
Figure 2 for Gaussian-Dirichlet Posterior Dominance in Sequential Learning
Viaarxiv icon

Coordinated Exploration in Concurrent Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 05, 2018
Maria Dimakopoulou, Benjamin Van Roy

Figure 1 for Coordinated Exploration in Concurrent Reinforcement Learning
Figure 2 for Coordinated Exploration in Concurrent Reinforcement Learning
Figure 3 for Coordinated Exploration in Concurrent Reinforcement Learning
Figure 4 for Coordinated Exploration in Concurrent Reinforcement Learning
Viaarxiv icon

Ensemble Sampling

Add code
Bookmark button
Alert button
Nov 22, 2017
Xiuyuan Lu, Benjamin Van Roy

Figure 1 for Ensemble Sampling
Figure 2 for Ensemble Sampling
Figure 3 for Ensemble Sampling
Viaarxiv icon

A Tutorial on Thompson Sampling

Add code
Bookmark button
Alert button
Nov 19, 2017
Daniel Russo, Benjamin Van Roy, Abbas Kazerouni, Ian Osband, Zheng Wen

Figure 1 for A Tutorial on Thompson Sampling
Figure 2 for A Tutorial on Thompson Sampling
Figure 3 for A Tutorial on Thompson Sampling
Figure 4 for A Tutorial on Thompson Sampling
Viaarxiv icon

Learning to Price with Reference Effects

Add code
Bookmark button
Alert button
Aug 29, 2017
Abbas Kazerouni, Benjamin Van Roy

Figure 1 for Learning to Price with Reference Effects
Figure 2 for Learning to Price with Reference Effects
Figure 3 for Learning to Price with Reference Effects
Viaarxiv icon

Learning to Optimize via Information-Directed Sampling

Add code
Bookmark button
Alert button
Jul 07, 2017
Daniel Russo, Benjamin Van Roy

Figure 1 for Learning to Optimize via Information-Directed Sampling
Figure 2 for Learning to Optimize via Information-Directed Sampling
Figure 3 for Learning to Optimize via Information-Directed Sampling
Figure 4 for Learning to Optimize via Information-Directed Sampling
Viaarxiv icon

On Optimistic versus Randomized Exploration in Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 13, 2017
Ian Osband, Benjamin Van Roy

Figure 1 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 2 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 3 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 4 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Viaarxiv icon

Why is Posterior Sampling Better than Optimism for Reinforcement Learning?

Add code
Bookmark button
Alert button
Jun 13, 2017
Ian Osband, Benjamin Van Roy

Figure 1 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 2 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 3 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 4 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Viaarxiv icon

Time-Sensitive Bandit Learning and Satisficing Thompson Sampling

Add code
Bookmark button
Alert button
Apr 28, 2017
Daniel Russo, David Tse, Benjamin Van Roy

Figure 1 for Time-Sensitive Bandit Learning and Satisficing Thompson Sampling
Viaarxiv icon

Conservative Contextual Linear Bandits

Add code
Bookmark button
Alert button
Mar 04, 2017
Abbas Kazerouni, Mohammad Ghavamzadeh, Yasin Abbasi-Yadkori, Benjamin Van Roy

Figure 1 for Conservative Contextual Linear Bandits
Viaarxiv icon