Picture for Matteo Pirotta

Matteo Pirotta

Simple Ingredients for Offline Reinforcement Learning

Add code
Mar 19, 2024
Figure 1 for Simple Ingredients for Offline Reinforcement Learning
Figure 2 for Simple Ingredients for Offline Reinforcement Learning
Figure 3 for Simple Ingredients for Offline Reinforcement Learning
Figure 4 for Simple Ingredients for Offline Reinforcement Learning
Viaarxiv icon

Layered State Discovery for Incremental Autonomous Exploration

Add code
Feb 07, 2023
Figure 1 for Layered State Discovery for Incremental Autonomous Exploration
Figure 2 for Layered State Discovery for Incremental Autonomous Exploration
Viaarxiv icon

On the Complexity of Representation Learning in Contextual Linear Bandits

Add code
Dec 19, 2022
Figure 1 for On the Complexity of Representation Learning in Contextual Linear Bandits
Viaarxiv icon

Improved Adaptive Algorithm for Scalable Active Learning with Weak Labeler

Add code
Nov 04, 2022
Figure 1 for Improved Adaptive Algorithm for Scalable Active Learning with Weak Labeler
Figure 2 for Improved Adaptive Algorithm for Scalable Active Learning with Weak Labeler
Figure 3 for Improved Adaptive Algorithm for Scalable Active Learning with Weak Labeler
Figure 4 for Improved Adaptive Algorithm for Scalable Active Learning with Weak Labeler
Viaarxiv icon

Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees

Add code
Oct 24, 2022
Figure 1 for Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Figure 2 for Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Figure 3 for Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Figure 4 for Scalable Representation Learning in Linear Contextual Bandits with Constant Regret Guarantees
Viaarxiv icon

Contextual bandits with concave rewards, and an application to fair ranking

Add code
Oct 18, 2022
Figure 1 for Contextual bandits with concave rewards, and an application to fair ranking
Figure 2 for Contextual bandits with concave rewards, and an application to fair ranking
Figure 3 for Contextual bandits with concave rewards, and an application to fair ranking
Figure 4 for Contextual bandits with concave rewards, and an application to fair ranking
Viaarxiv icon

Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path

Add code
Oct 10, 2022
Figure 1 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Figure 2 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Figure 3 for Reaching Goals is Hard: Settling the Sample Complexity of the Stochastic Shortest Path
Viaarxiv icon

Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations

Add code
Dec 14, 2021
Figure 1 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 2 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 3 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Figure 4 for Top $K$ Ranking for Multi-Armed Bandit with Noisy Evaluations
Viaarxiv icon

Privacy Amplification via Shuffling for Linear Contextual Bandits

Add code
Dec 11, 2021
Figure 1 for Privacy Amplification via Shuffling for Linear Contextual Bandits
Figure 2 for Privacy Amplification via Shuffling for Linear Contextual Bandits
Viaarxiv icon

Differentially Private Exploration in Reinforcement Learning with Linear Representation

Add code
Dec 07, 2021
Figure 1 for Differentially Private Exploration in Reinforcement Learning with Linear Representation
Viaarxiv icon