Alert button
Picture for Matteo Pirotta

Matteo Pirotta

Alert button

Local Differentially Private Regret Minimization in Reinforcement Learning

Oct 15, 2020
Evrard Garcelon, Vianney Perchet, Ciara Pike-Burke, Matteo Pirotta

Figure 1 for Local Differentially Private Regret Minimization in Reinforcement Learning
Figure 2 for Local Differentially Private Regret Minimization in Reinforcement Learning
Figure 3 for Local Differentially Private Regret Minimization in Reinforcement Learning
Figure 4 for Local Differentially Private Regret Minimization in Reinforcement Learning
Viaarxiv icon

A Provably Efficient Sample Collection Strategy for Reinforcement Learning

Jul 13, 2020
Jean Tarbouriech, Matteo Pirotta, Michal Valko, Alessandro Lazaric

Figure 1 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 2 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 3 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 4 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Viaarxiv icon

Improved Analysis of UCRL2 with Empirical Bernstein Inequality

Jul 10, 2020
Ronan Fruit, Matteo Pirotta, Alessandro Lazaric

Figure 1 for Improved Analysis of UCRL2 with Empirical Bernstein Inequality
Viaarxiv icon

A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces

Jul 09, 2020
Omar Darwiche Domingues, Pierre Ménard, Matteo Pirotta, Emilie Kaufmann, Michal Valko

Figure 1 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 2 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 3 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 4 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Viaarxiv icon

Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization

May 06, 2020
Pierre-Alexandre Kamienny, Matteo Pirotta, Alessandro Lazaric, Thibault Lavril, Nicolas Usunier, Ludovic Denoyer

Figure 1 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 2 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 3 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 4 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Viaarxiv icon

Regret Bounds for Kernel-Based Reinforcement Learning

Apr 12, 2020
Omar Darwiche Domingues, Pierre Ménard, Matteo Pirotta, Emilie Kaufmann, Michal Valko

Figure 1 for Regret Bounds for Kernel-Based Reinforcement Learning
Figure 2 for Regret Bounds for Kernel-Based Reinforcement Learning
Figure 3 for Regret Bounds for Kernel-Based Reinforcement Learning
Viaarxiv icon

Active Model Estimation in Markov Decision Processes

Mar 06, 2020
Jean Tarbouriech, Shubhanshu Shekhar, Matteo Pirotta, Mohammad Ghavamzadeh, Alessandro Lazaric

Figure 1 for Active Model Estimation in Markov Decision Processes
Figure 2 for Active Model Estimation in Markov Decision Processes
Figure 3 for Active Model Estimation in Markov Decision Processes
Figure 4 for Active Model Estimation in Markov Decision Processes
Viaarxiv icon

Exploration-Exploitation in Constrained MDPs

Mar 04, 2020
Yonathan Efroni, Shie Mannor, Matteo Pirotta

Figure 1 for Exploration-Exploitation in Constrained MDPs
Viaarxiv icon

Adversarial Attacks on Linear Contextual Bandits

Feb 11, 2020
Evrard Garcelon, Baptiste Roziere, Laurent Meunier, Jean Tarbouriech, Olivier Teytaud, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Adversarial Attacks on Linear Contextual Bandits
Figure 2 for Adversarial Attacks on Linear Contextual Bandits
Figure 3 for Adversarial Attacks on Linear Contextual Bandits
Figure 4 for Adversarial Attacks on Linear Contextual Bandits
Viaarxiv icon

Improved Algorithms for Conservative Exploration in Bandits

Feb 08, 2020
Evrard Garcelon, Mohammad Ghavamzadeh, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Improved Algorithms for Conservative Exploration in Bandits
Figure 2 for Improved Algorithms for Conservative Exploration in Bandits
Figure 3 for Improved Algorithms for Conservative Exploration in Bandits
Figure 4 for Improved Algorithms for Conservative Exploration in Bandits
Viaarxiv icon