Alert button
Picture for Matteo Pirotta

Matteo Pirotta

Alert button

Conservative Exploration in Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 08, 2020
Evrard Garcelon, Mohammad Ghavamzadeh, Alessandro Lazaric, Matteo Pirotta

Figure 1 for Conservative Exploration in Reinforcement Learning
Figure 2 for Conservative Exploration in Reinforcement Learning
Figure 3 for Conservative Exploration in Reinforcement Learning
Figure 4 for Conservative Exploration in Reinforcement Learning
Viaarxiv icon

Concentration Inequalities for Multinoulli Random Variables

Add code
Bookmark button
Alert button
Jan 30, 2020
Jian Qian, Ronan Fruit, Matteo Pirotta, Alessandro Lazaric

Viaarxiv icon

No-Regret Exploration in Goal-Oriented Reinforcement Learning

Add code
Bookmark button
Alert button
Jan 30, 2020
Jean Tarbouriech, Evrard Garcelon, Michal Valko, Matteo Pirotta, Alessandro Lazaric

Figure 1 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 2 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 3 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 4 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Viaarxiv icon

Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning

Add code
Bookmark button
Alert button
Jan 13, 2020
Michiel van der Meer, Matteo Pirotta, Elia Bruni

Figure 1 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Figure 2 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Figure 3 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Figure 4 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Viaarxiv icon

Frequentist Regret Bounds for Randomized Least-Squares Value Iteration

Add code
Bookmark button
Alert button
Nov 01, 2019
Andrea Zanette, David Brandfonbrener, Matteo Pirotta, Alessandro Lazaric

Figure 1 for Frequentist Regret Bounds for Randomized Least-Squares Value Iteration
Viaarxiv icon

Smoothing Policies and Safe Policy Gradients

Add code
Bookmark button
Alert button
May 08, 2019
Matteo Papini, Matteo Pirotta, Marcello Restelli

Figure 1 for Smoothing Policies and Safe Policy Gradients
Figure 2 for Smoothing Policies and Safe Policy Gradients
Viaarxiv icon

Exploration Bonus for Regret Minimization in Undiscounted Discrete and Continuous Markov Decision Processes

Add code
Bookmark button
Alert button
Dec 11, 2018
Jian Qian, Ronan Fruit, Matteo Pirotta, Alessandro Lazaric

Viaarxiv icon

Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 06, 2018
Ronan Fruit, Matteo Pirotta, Alessandro Lazaric, Ronald Ortner

Figure 1 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Figure 2 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Figure 3 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Figure 4 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Viaarxiv icon

Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes

Add code
Bookmark button
Alert button
Jul 06, 2018
Ronan Fruit, Matteo Pirotta, Alessandro Lazaric

Figure 1 for Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes
Figure 2 for Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes
Figure 3 for Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes
Figure 4 for Near Optimal Exploration-Exploitation in Non-Communicating Markov Decision Processes
Viaarxiv icon