Picture for Claudio Gentile

Claudio Gentile

IMT

A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations

Add code
Nov 29, 2022
Figure 1 for A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations
Figure 2 for A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations
Figure 3 for A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations
Figure 4 for A Contextual Bandit Approach for Learning to Plan in Environments with Probabilistic Goal Configurations
Viaarxiv icon

Best of Both Worlds Model Selection

Add code
Jun 29, 2022
Viaarxiv icon

Achieving Minimax Rates in Pool-Based Batch Active Learning

Add code
Feb 11, 2022
Viaarxiv icon

Nonstochastic Bandits with Composite Anonymous Feedback

Add code
Dec 06, 2021
Figure 1 for Nonstochastic Bandits with Composite Anonymous Feedback
Viaarxiv icon

Batch Active Learning at Scale

Add code
Jul 29, 2021
Figure 1 for Batch Active Learning at Scale
Figure 2 for Batch Active Learning at Scale
Figure 3 for Batch Active Learning at Scale
Figure 4 for Batch Active Learning at Scale
Viaarxiv icon

Adapting to Misspecification in Contextual Bandits

Add code
Jul 12, 2021
Viaarxiv icon

On Learning to Rank Long Sequences with Contextual Bandits

Add code
Jun 07, 2021
Figure 1 for On Learning to Rank Long Sequences with Contextual Bandits
Figure 2 for On Learning to Rank Long Sequences with Contextual Bandits
Figure 3 for On Learning to Rank Long Sequences with Contextual Bandits
Figure 4 for On Learning to Rank Long Sequences with Contextual Bandits
Viaarxiv icon

Neural Active Learning with Performance Guarantees

Add code
Jun 06, 2021
Viaarxiv icon

Regret Bound Balancing and Elimination for Model Selection in Bandits and RL

Add code
Dec 24, 2020
Figure 1 for Regret Bound Balancing and Elimination for Model Selection in Bandits and RL
Figure 2 for Regret Bound Balancing and Elimination for Model Selection in Bandits and RL
Viaarxiv icon

Online Model Selection: a Rested Bandit Formulation

Add code
Dec 07, 2020
Figure 1 for Online Model Selection: a Rested Bandit Formulation
Viaarxiv icon