Picture for Benjamin Van Roy

Benjamin Van Roy

Stanford University Department of Electrical Engineering

Gaussian-Dirichlet Posterior Dominance in Sequential Learning

Add code
Feb 09, 2018
Figure 1 for Gaussian-Dirichlet Posterior Dominance in Sequential Learning
Figure 2 for Gaussian-Dirichlet Posterior Dominance in Sequential Learning
Viaarxiv icon

Coordinated Exploration in Concurrent Reinforcement Learning

Add code
Feb 05, 2018
Figure 1 for Coordinated Exploration in Concurrent Reinforcement Learning
Figure 2 for Coordinated Exploration in Concurrent Reinforcement Learning
Figure 3 for Coordinated Exploration in Concurrent Reinforcement Learning
Figure 4 for Coordinated Exploration in Concurrent Reinforcement Learning
Viaarxiv icon

Ensemble Sampling

Add code
Nov 22, 2017
Figure 1 for Ensemble Sampling
Figure 2 for Ensemble Sampling
Figure 3 for Ensemble Sampling
Viaarxiv icon

A Tutorial on Thompson Sampling

Add code
Nov 19, 2017
Figure 1 for A Tutorial on Thompson Sampling
Figure 2 for A Tutorial on Thompson Sampling
Figure 3 for A Tutorial on Thompson Sampling
Figure 4 for A Tutorial on Thompson Sampling
Viaarxiv icon

Learning to Price with Reference Effects

Add code
Aug 29, 2017
Figure 1 for Learning to Price with Reference Effects
Figure 2 for Learning to Price with Reference Effects
Figure 3 for Learning to Price with Reference Effects
Viaarxiv icon

Learning to Optimize via Information-Directed Sampling

Add code
Jul 07, 2017
Figure 1 for Learning to Optimize via Information-Directed Sampling
Figure 2 for Learning to Optimize via Information-Directed Sampling
Figure 3 for Learning to Optimize via Information-Directed Sampling
Figure 4 for Learning to Optimize via Information-Directed Sampling
Viaarxiv icon

On Optimistic versus Randomized Exploration in Reinforcement Learning

Add code
Jun 13, 2017
Figure 1 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 2 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 3 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Figure 4 for On Optimistic versus Randomized Exploration in Reinforcement Learning
Viaarxiv icon

Why is Posterior Sampling Better than Optimism for Reinforcement Learning?

Add code
Jun 13, 2017
Figure 1 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 2 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 3 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Figure 4 for Why is Posterior Sampling Better than Optimism for Reinforcement Learning?
Viaarxiv icon

Time-Sensitive Bandit Learning and Satisficing Thompson Sampling

Add code
Apr 28, 2017
Figure 1 for Time-Sensitive Bandit Learning and Satisficing Thompson Sampling
Viaarxiv icon

Conservative Contextual Linear Bandits

Add code
Mar 04, 2017
Figure 1 for Conservative Contextual Linear Bandits
Viaarxiv icon