Picture for Matteo Pirotta

Matteo Pirotta

Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret

Add code
Apr 22, 2021
Figure 1 for Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret
Figure 2 for Stochastic Shortest Path: Minimax, Parameter-Free and Towards Horizon-Free Regret
Viaarxiv icon

Leveraging Good Representations in Linear Contextual Bandits

Add code
Apr 08, 2021
Figure 1 for Leveraging Good Representations in Linear Contextual Bandits
Figure 2 for Leveraging Good Representations in Linear Contextual Bandits
Figure 3 for Leveraging Good Representations in Linear Contextual Bandits
Figure 4 for Leveraging Good Representations in Linear Contextual Bandits
Viaarxiv icon

Homomorphically Encrypted Linear Contextual Bandit

Add code
Mar 17, 2021
Figure 1 for Homomorphically Encrypted Linear Contextual Bandit
Figure 2 for Homomorphically Encrypted Linear Contextual Bandit
Figure 3 for Homomorphically Encrypted Linear Contextual Bandit
Viaarxiv icon

Improved Sample Complexity for Incremental Autonomous Exploration in MDPs

Add code
Dec 29, 2020
Figure 1 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Figure 2 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Figure 3 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Figure 4 for Improved Sample Complexity for Incremental Autonomous Exploration in MDPs
Viaarxiv icon

An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits

Add code
Oct 23, 2020
Figure 1 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Figure 2 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Figure 3 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Figure 4 for An Asymptotically Optimal Primal-Dual Incremental Algorithm for Contextual Linear Bandits
Viaarxiv icon

Local Differentially Private Regret Minimization in Reinforcement Learning

Add code
Oct 15, 2020
Figure 1 for Local Differentially Private Regret Minimization in Reinforcement Learning
Figure 2 for Local Differentially Private Regret Minimization in Reinforcement Learning
Figure 3 for Local Differentially Private Regret Minimization in Reinforcement Learning
Figure 4 for Local Differentially Private Regret Minimization in Reinforcement Learning
Viaarxiv icon

A Provably Efficient Sample Collection Strategy for Reinforcement Learning

Add code
Jul 13, 2020
Figure 1 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 2 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 3 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Figure 4 for A Provably Efficient Sample Collection Strategy for Reinforcement Learning
Viaarxiv icon

Improved Analysis of UCRL2 with Empirical Bernstein Inequality

Add code
Jul 10, 2020
Figure 1 for Improved Analysis of UCRL2 with Empirical Bernstein Inequality
Viaarxiv icon

A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces

Add code
Jul 09, 2020
Figure 1 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 2 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 3 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Figure 4 for A Kernel-Based Approach to Non-Stationary Reinforcement Learning in Metric Spaces
Viaarxiv icon

Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization

Add code
May 06, 2020
Figure 1 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 2 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 3 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Figure 4 for Learning Adaptive Exploration Strategies in Dynamic Environments Through Informed Policy Regularization
Viaarxiv icon