Picture for Matteo Pirotta

Matteo Pirotta

Adversarial Attacks on Linear Contextual Bandits

Add code
Feb 11, 2020
Figure 1 for Adversarial Attacks on Linear Contextual Bandits
Figure 2 for Adversarial Attacks on Linear Contextual Bandits
Figure 3 for Adversarial Attacks on Linear Contextual Bandits
Figure 4 for Adversarial Attacks on Linear Contextual Bandits
Viaarxiv icon

Improved Algorithms for Conservative Exploration in Bandits

Add code
Feb 08, 2020
Figure 1 for Improved Algorithms for Conservative Exploration in Bandits
Figure 2 for Improved Algorithms for Conservative Exploration in Bandits
Figure 3 for Improved Algorithms for Conservative Exploration in Bandits
Figure 4 for Improved Algorithms for Conservative Exploration in Bandits
Viaarxiv icon

Conservative Exploration in Reinforcement Learning

Add code
Feb 08, 2020
Figure 1 for Conservative Exploration in Reinforcement Learning
Figure 2 for Conservative Exploration in Reinforcement Learning
Figure 3 for Conservative Exploration in Reinforcement Learning
Figure 4 for Conservative Exploration in Reinforcement Learning
Viaarxiv icon

Concentration Inequalities for Multinoulli Random Variables

Add code
Jan 30, 2020
Viaarxiv icon

No-Regret Exploration in Goal-Oriented Reinforcement Learning

Add code
Jan 30, 2020
Figure 1 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 2 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 3 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Figure 4 for No-Regret Exploration in Goal-Oriented Reinforcement Learning
Viaarxiv icon

Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning

Add code
Jan 13, 2020
Figure 1 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Figure 2 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Figure 3 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Figure 4 for Exploiting Language Instructions for Interpretable and Compositional Reinforcement Learning
Viaarxiv icon

Frequentist Regret Bounds for Randomized Least-Squares Value Iteration

Add code
Nov 01, 2019
Figure 1 for Frequentist Regret Bounds for Randomized Least-Squares Value Iteration
Viaarxiv icon

Smoothing Policies and Safe Policy Gradients

Add code
May 08, 2019
Figure 1 for Smoothing Policies and Safe Policy Gradients
Figure 2 for Smoothing Policies and Safe Policy Gradients
Viaarxiv icon

Exploration Bonus for Regret Minimization in Undiscounted Discrete and Continuous Markov Decision Processes

Add code
Dec 11, 2018
Viaarxiv icon

Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning

Add code
Jul 06, 2018
Figure 1 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Figure 2 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Figure 3 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Figure 4 for Efficient Bias-Span-Constrained Exploration-Exploitation in Reinforcement Learning
Viaarxiv icon