Picture for Alessandro Lazaric

Alessandro Lazaric

INRIA Lille - Nord Europe

Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization

Add code
May 06, 2020
Figure 1 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 2 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 3 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 4 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Viaarxiv icon

Active Model Estimation in Markov Decision Processes

Add code
Mar 06, 2020
Figure 1 for Active Model Estimation in Markov Decision Processes
Figure 2 for Active Model Estimation in Markov Decision Processes
Figure 3 for Active Model Estimation in Markov Decision Processes
Figure 4 for Active Model Estimation in Markov Decision Processes
Viaarxiv icon

Learning Near Optimal Policies with Low Inherent Bellman Error

Add code
Mar 05, 2020
Figure 1 for Learning Near Optimal Policies with Low Inherent Bellman Error
Viaarxiv icon

Near-linear Time Gaussian Process Optimization with Adaptive Batching and Resparsification

Add code
Feb 26, 2020
Figure 1 for Near-linear Time Gaussian Process Optimization with Adaptive Batching and Resparsification
Figure 2 for Near-linear Time Gaussian Process Optimization with Adaptive Batching and Resparsification
Figure 3 for Near-linear Time Gaussian Process Optimization with Adaptive Batching and Resparsification
Figure 4 for Near-linear Time Gaussian Process Optimization with Adaptive Batching and Resparsification
Viaarxiv icon

Adversarial Attacks on Linear Contextual Bandits

Add code
Feb 11, 2020
Figure 1 for Adversarial Attacks on Linear Contextual Bandits
Figure 2 for Adversarial Attacks on Linear Contextual Bandits
Figure 3 for Adversarial Attacks on Linear Contextual Bandits
Figure 4 for Adversarial Attacks on Linear Contextual Bandits
Viaarxiv icon

Improved Algorithms for Conservative Exploration in Bandits

Add code
Feb 08, 2020
Figure 1 for Improved Algorithms for Conservative Exploration in Bandits
Figure 2 for Improved Algorithms for Conservative Exploration in Bandits
Figure 3 for Improved Algorithms for Conservative Exploration in Bandits
Figure 4 for Improved Algorithms for Conservative Exploration in Bandits
Viaarxiv icon

Conservative Exploration in Reinforcement Learning

Add code
Feb 08, 2020
Figure 1 for Conservative Exploration in Reinforcement Learning
Figure 2 for Conservative Exploration in Reinforcement Learning
Figure 3 for Conservative Exploration in Reinforcement Learning
Figure 4 for Conservative Exploration in Reinforcement Learning
Viaarxiv icon

Concentration Inequalities for Multinoulli Random Variables

Add code
Jan 30, 2020
Viaarxiv icon

No-Regret Exploration in Goal-Oriented Reinforcement Learning

Add code
Jan 30, 2020
Figure 1 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 2 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 3 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 4 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Viaarxiv icon

Frequentist Regret Bounds for Randomized Least-Squares Value Iteration

Add code
Nov 01, 2019
Figure 1 for Frequentist Regret Bounds for Randomized Least-Squares Value Iteration
Viaarxiv icon