Picture for Alessandro Lazaric

Alessandro Lazaric

INRIA Lille - Nord Europe

Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees

Add code
Oct 24, 2022
Viaarxiv icon

Contextual bandits with concave rewards, and an application to fair ranking

Add code
Oct 18, 2022
Figure 1 for Contextual bandits with concave rewards, and an application to fair ranking
Figure 2 for Contextual bandits with concave rewards, and an application to fair ranking
Figure 3 for Contextual bandits with concave rewards, and an application to fair ranking
Figure 4 for Contextual bandits with concave rewards, and an application to fair ranking
Viaarxiv icon

Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path

Add code
Oct 10, 2022
Figure 1 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Figure 2 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Figure 3 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Viaarxiv icon

Linear Convergence of Natural Policy Gradient Methods with Log-Linear Policies

Add code
Oct 04, 2022
Viaarxiv icon

Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL

Add code
Mar 21, 2022
Figure 1 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Figure 2 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Figure 3 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Figure 4 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Viaarxiv icon

Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning

Add code
Feb 08, 2022
Figure 1 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Figure 2 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Figure 3 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Figure 4 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Viaarxiv icon

Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times

Add code
Jan 30, 2022
Figure 1 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 2 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 3 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 4 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Viaarxiv icon

Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations

Add code
Dec 14, 2021
Figure 1 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 2 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 3 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 4 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Viaarxiv icon

Differentially Private Exploration in Reinforcement Learning with Linear Representation

Add code
Dec 07, 2021
Figure 1 for Differentially Private Exploration in Reinforcement Learning with Linear Representation
Viaarxiv icon

Adaptive Multi-Goal Exploration

Add code
Nov 23, 2021
Figure 1 for Adaptive Multi-Goal Exploration
Figure 2 for Adaptive Multi-Goal Exploration
Figure 3 for Adaptive Multi-Goal Exploration
Viaarxiv icon