Alert button
Picture for Alessandro Lazaric

Alessandro Lazaric

Alert button

INRIA Lille - Nord Europe

Upper-Confidence-Bound Algorithms for Active Learning in Multi-Armed Bandits

Add code
Bookmark button
Alert button
Jul 16, 2015
Alexandra Carpentier, Alessandro Lazaric, Mohammad Ghavamzadeh, Rémi Munos, Peter Auer, András Antos

Figure 1 for Upper-Confidence-Bound Algorithms for Active Learning in Multi-Armed Bandits
Viaarxiv icon

Best-Arm Identification in Linear Bandits

Add code
Bookmark button
Alert button
Nov 04, 2014
Marta Soare, Alessandro Lazaric, Rémi Munos

Figure 1 for Best-Arm Identification in Linear Bandits
Figure 2 for Best-Arm Identification in Linear Bandits
Viaarxiv icon

Online Stochastic Optimization under Correlated Bandit Feedback

Add code
Bookmark button
Alert button
May 19, 2014
Mohammad Gheshlaghi Azar, Alessandro Lazaric, Emma Brunskill

Figure 1 for Online Stochastic Optimization under Correlated Bandit Feedback
Figure 2 for Online Stochastic Optimization under Correlated Bandit Feedback
Figure 3 for Online Stochastic Optimization under Correlated Bandit Feedback
Viaarxiv icon

Sequential Transfer in Multi-armed Bandit with Finite Set of Models

Add code
Bookmark button
Alert button
Jul 25, 2013
Mohammad Gheshlaghi Azar, Alessandro Lazaric, Emma Brunskill

Figure 1 for Sequential Transfer in Multi-armed Bandit with Finite Set of Models
Figure 2 for Sequential Transfer in Multi-armed Bandit with Finite Set of Models
Figure 3 for Sequential Transfer in Multi-armed Bandit with Finite Set of Models
Figure 4 for Sequential Transfer in Multi-armed Bandit with Finite Set of Models
Viaarxiv icon

Regret Bounds for Reinforcement Learning with Policy Advice

Add code
Bookmark button
Alert button
Jul 17, 2013
Mohammad Gheshlaghi Azar, Alessandro Lazaric, Emma Brunskill

Figure 1 for Regret Bounds for Reinforcement Learning with Policy Advice
Viaarxiv icon

Risk-Aversion in Multi-armed Bandits

Add code
Bookmark button
Alert button
Jan 09, 2013
Amir Sani, Alessandro Lazaric, Rémi Munos

Figure 1 for Risk-Aversion in Multi-armed Bandits
Figure 2 for Risk-Aversion in Multi-armed Bandits
Figure 3 for Risk-Aversion in Multi-armed Bandits
Figure 4 for Risk-Aversion in Multi-armed Bandits
Viaarxiv icon

A Dantzig Selector Approach to Temporal Difference Learning

Add code
Bookmark button
Alert button
Jun 27, 2012
Matthieu Geist, Bruno Scherrer, Alessandro Lazaric, Mohammad Ghavamzadeh

Figure 1 for A Dantzig Selector Approach to Temporal Difference Learning
Figure 2 for A Dantzig Selector Approach to Temporal Difference Learning
Figure 3 for A Dantzig Selector Approach to Temporal Difference Learning
Figure 4 for A Dantzig Selector Approach to Temporal Difference Learning
Viaarxiv icon

Transfer from Multiple MDPs

Add code
Bookmark button
Alert button
Sep 01, 2011
Alessandro Lazaric, Marcello Restelli

Figure 1 for Transfer from Multiple MDPs
Figure 2 for Transfer from Multiple MDPs
Figure 3 for Transfer from Multiple MDPs
Figure 4 for Transfer from Multiple MDPs
Viaarxiv icon