Picture for Alessandro Lazaric

Alessandro Lazaric

INRIA Lille - Nord Europe

An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits

Add code
Oct 23, 2020
Figure 1 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Figure 2 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Figure 3 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Figure 4 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Viaarxiv icon

Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration

Add code
Aug 18, 2020
Figure 1 for Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration
Figure 2 for Provably Efficient Reward-Agnostic Navigation with Linear Value Iteration
Viaarxiv icon

Efficient Optimistic Exploration in Linear-Quadratic Regulators via Lagrangian Relaxation

Add code
Jul 13, 2020
Figure 1 for Efficient Optimistic Exploration in Linear-Quadratic Regulators via Lagrangian Relaxation
Figure 2 for Efficient Optimistic Exploration in Linear-Quadratic Regulators via Lagrangian Relaxation
Figure 3 for Efficient Optimistic Exploration in Linear-Quadratic Regulators via Lagrangian Relaxation
Figure 4 for Efficient Optimistic Exploration in Linear-Quadratic Regulators via Lagrangian Relaxation
Viaarxiv icon

A Provably Efficient Sample Collection Strategy for Reinforcement Learning

Add code
Jul 13, 2020
Figure 1 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 2 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 3 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 4 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Viaarxiv icon

Improved Analysis of UCRL2 with Empirical Bernstein Inequality

Add code
Jul 10, 2020
Figure 1 for Improved Analysis of UCRL2 with Empirical Bernstein Inequality
Viaarxiv icon

A Novel Confidence-Based Algorithm for Structured Bandits

Add code
May 23, 2020
Figure 1 for A Novel Confidence-Based Algorithm for Structured Bandits
Figure 2 for A Novel Confidence-Based Algorithm for Structured Bandits
Figure 3 for A Novel Confidence-Based Algorithm for Structured Bandits
Figure 4 for A Novel Confidence-Based Algorithm for Structured Bandits
Viaarxiv icon

Meta-learning with Stochastic Linear Bandits

Add code
May 18, 2020
Figure 1 for Meta-learning with Stochastic Linear Bandits
Figure 2 for Meta-learning with Stochastic Linear Bandits
Figure 3 for Meta-learning with Stochastic Linear Bandits
Figure 4 for Meta-learning with Stochastic Linear Bandits
Viaarxiv icon

Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization

Add code
May 06, 2020
Figure 1 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 2 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 3 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 4 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Viaarxiv icon

Active Model Estimation in Markov Decision Processes

Add code
Mar 06, 2020
Figure 1 for Active Model Estimation in Markov Decision Processes
Figure 2 for Active Model Estimation in Markov Decision Processes
Figure 3 for Active Model Estimation in Markov Decision Processes
Figure 4 for Active Model Estimation in Markov Decision Processes
Viaarxiv icon

Learning Near Optimal Policies with Low Inherent Bellman Error

Add code
Mar 05, 2020
Figure 1 for Learning Near Optimal Policies with Low Inherent Bellman Error
Viaarxiv icon