Alert button
Picture for Alessandro Lazaric

Alessandro Lazaric

Alert button

Linear Convergence of Natural Policy Gradient Methods with Log-Linear Policies

Add code
Bookmark button
Alert button
Oct 04, 2022
Rui Yuan, Simon S. Du, Robert M. Gower, Alessandro Lazaric, Lin Xiao

Figure 1 for Linear Convergence of Natural Policy Gradient Methods with Log-Linear Policies
Viaarxiv icon

Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL

Add code
Bookmark button
Alert button
Mar 21, 2022
Akram Erraqabi, Marlos C. Machado, Mingde Zhao, Sainbayar Sukhbaatar, Alessandro Lazaric, Ludovic Denoyer, Yoshua Bengio

Figure 1 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Figure 2 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Figure 3 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Figure 4 for Temporal Abstractions-Augmented Temporally Contrastive Learning: An Alternative to the Laplacian in RL
Viaarxiv icon

Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 08, 2022
Denis Yarats, David Brandfonbrener, Hao Liu, Michael Laskin, Pieter Abbeel, Alessandro Lazaric, Lerrel Pinto

Figure 1 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Figure 2 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Figure 3 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Figure 4 for Don't Change the Algorithm, Change the Data: Exploratory Data for Offline Reinforcement Learning
Viaarxiv icon

Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times

Add code
Bookmark button
Alert button
Jan 30, 2022
Daniele Calandriello, Luigi Carratino, Alessandro Lazaric, Michal Valko, Lorenzo Rosasco

Figure 1 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 2 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 3 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 4 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Viaarxiv icon

Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations

Add code
Bookmark button
Alert button
Dec 14, 2021
Evrard Garcelon, Vashist Avadhanula, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 2 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 3 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 4 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Viaarxiv icon

Differentially Private Exploration in Reinforcement Learning with Linear Representation

Add code
Bookmark button
Alert button
Dec 07, 2021
Paul Luyo, Evrard Garcelon, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Differentially Private Exploration in Reinforcement Learning with Linear Representation
Viaarxiv icon

Adaptive Multi-Goal Exploration

Add code
Bookmark button
Alert button
Nov 23, 2021
Jean Tarbouriech, Omar Darwiche Domingues, Pierre Ménard, Matteo Pirotta, Michal Valko, Alessandro Lazaric

Figure 1 for Adaptive Multi-Goal Exploration
Figure 2 for Adaptive Multi-Goal Exploration
Figure 3 for Adaptive Multi-Goal Exploration
Viaarxiv icon

Reinforcement Learning in Linear MDPs: Constant Regret and Representation Selection

Add code
Bookmark button
Alert button
Oct 27, 2021
Matteo Papini, Andrea Tirinzoni, Aldo Pacchiano, Marcello Restelli, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Reinforcement Learning in Linear MDPs: Constant Regret and Representation Selection
Figure 2 for Reinforcement Learning in Linear MDPs: Constant Regret and Representation Selection
Figure 3 for Reinforcement Learning in Linear MDPs: Constant Regret and Representation Selection
Figure 4 for Reinforcement Learning in Linear MDPs: Constant Regret and Representation Selection
Viaarxiv icon

Direct then Diffuse: Incremental Unsupervised Skill Discovery for State Covering and Goal Reaching

Add code
Bookmark button
Alert button
Oct 27, 2021
Pierre-Alexandre Kamienny, Jean Tarbouriech, Alessandro Lazaric, Ludovic Denoyer

Figure 1 for Direct then Diffuse: Incremental Unsupervised Skill Discovery for State Covering and Goal Reaching
Figure 2 for Direct then Diffuse: Incremental Unsupervised Skill Discovery for State Covering and Goal Reaching
Figure 3 for Direct then Diffuse: Incremental Unsupervised Skill Discovery for State Covering and Goal Reaching
Figure 4 for Direct then Diffuse: Incremental Unsupervised Skill Discovery for State Covering and Goal Reaching
Viaarxiv icon