Picture for Branislav Kveton

Branislav Kveton

Adobe Research

Pessimistic Off-Policy Optimization for Learning to Rank

Add code
Jun 06, 2022
Figure 1 for Pessimistic Off-Policy Optimization for Learning to Rank
Figure 2 for Pessimistic Off-Policy Optimization for Learning to Rank
Figure 3 for Pessimistic Off-Policy Optimization for Learning to Rank
Figure 4 for Pessimistic Off-Policy Optimization for Learning to Rank
Viaarxiv icon

Generalizing Hierarchical Bayesian Bandits

Add code
May 30, 2022
Figure 1 for Generalizing Hierarchical Bayesian Bandits
Figure 2 for Generalizing Hierarchical Bayesian Bandits
Figure 3 for Generalizing Hierarchical Bayesian Bandits
Figure 4 for Generalizing Hierarchical Bayesian Bandits
Viaarxiv icon

Safe Exploration for Efficient Policy Evaluation and Comparison

Add code
Feb 26, 2022
Figure 1 for Safe Exploration for Efficient Policy Evaluation and Comparison
Figure 2 for Safe Exploration for Efficient Policy Evaluation and Comparison
Figure 3 for Safe Exploration for Efficient Policy Evaluation and Comparison
Figure 4 for Safe Exploration for Efficient Policy Evaluation and Comparison
Viaarxiv icon

Meta-Learning for Simple Regret Minimization

Add code
Feb 25, 2022
Figure 1 for Meta-Learning for Simple Regret Minimization
Figure 2 for Meta-Learning for Simple Regret Minimization
Figure 3 for Meta-Learning for Simple Regret Minimization
Figure 4 for Meta-Learning for Simple Regret Minimization
Viaarxiv icon

Deep Hierarchy in Bandits

Add code
Feb 03, 2022
Figure 1 for Deep Hierarchy in Bandits
Figure 2 for Deep Hierarchy in Bandits
Figure 3 for Deep Hierarchy in Bandits
Figure 4 for Deep Hierarchy in Bandits
Viaarxiv icon

IMO$^3$: Interactive Multi-Objective Off-Policy Optimization

Add code
Jan 25, 2022
Figure 1 for IMO$^3$: Interactive Multi-Objective Off-Policy Optimization
Figure 2 for IMO$^3$: Interactive Multi-Objective Off-Policy Optimization
Figure 3 for IMO$^3$: Interactive Multi-Objective Off-Policy Optimization
Figure 4 for IMO$^3$: Interactive Multi-Objective Off-Policy Optimization
Viaarxiv icon

Hierarchical Bayesian Bandits

Add code
Nov 12, 2021
Figure 1 for Hierarchical Bayesian Bandits
Figure 2 for Hierarchical Bayesian Bandits
Figure 3 for Hierarchical Bayesian Bandits
Viaarxiv icon

Safe Optimal Design with Applications in Policy Learning

Add code
Nov 08, 2021
Figure 1 for Safe Optimal Design with Applications in Policy Learning
Figure 2 for Safe Optimal Design with Applications in Policy Learning
Figure 3 for Safe Optimal Design with Applications in Policy Learning
Figure 4 for Safe Optimal Design with Applications in Policy Learning
Viaarxiv icon

Optimal Probing with Statistical Guarantees for Network Monitoring at Scale

Add code
Sep 16, 2021
Figure 1 for Optimal Probing with Statistical Guarantees for Network Monitoring at Scale
Figure 2 for Optimal Probing with Statistical Guarantees for Network Monitoring at Scale
Figure 3 for Optimal Probing with Statistical Guarantees for Network Monitoring at Scale
Figure 4 for Optimal Probing with Statistical Guarantees for Network Monitoring at Scale
Viaarxiv icon

No Regrets for Learning the Prior in Bandits

Add code
Jul 13, 2021
Figure 1 for No Regrets for Learning the Prior in Bandits
Figure 2 for No Regrets for Learning the Prior in Bandits
Figure 3 for No Regrets for Learning the Prior in Bandits
Figure 4 for No Regrets for Learning the Prior in Bandits
Viaarxiv icon