Alert button
Picture for Michal Valko

Michal Valko

Alert button

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
Bookmark button
Alert button
May 27, 2022
Tadashi Kozuno, Wenhao Yang, Nino Vieillard, Toshinori Kitamura, Yunhao Tang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Michal Valko, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári

Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses

Add code
Bookmark button
Alert button
May 16, 2022
Daniil Tiapkin, Denis Belomestny, Eric Moulines, Alexey Naumov, Sergey Samsonov, Yunhao Tang, Michal Valko, Pierre Menard

Figure 1 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Figure 2 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Figure 3 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Figure 4 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Viaarxiv icon

Marginalized Operators for Off-policy Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 30, 2022
Yunhao Tang, Mark Rowland, Rémi Munos, Michal Valko

Figure 1 for Marginalized Operators for Off-policy Reinforcement Learning
Figure 2 for Marginalized Operators for Off-policy Reinforcement Learning
Figure 3 for Marginalized Operators for Off-policy Reinforcement Learning
Figure 4 for Marginalized Operators for Off-policy Reinforcement Learning
Viaarxiv icon

Retrieval-Augmented Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 09, 2022
Anirudh Goyal, Abram L. Friesen, Andrea Banino, Theophane Weber, Nan Rosemary Ke, Adria Puigdomenech Badia, Arthur Guez, Mehdi Mirza, Peter C. Humphreys, Ksenia Konyushkova, Michal Valko, Simon Osindero, Timothy Lillicrap, Nicolas Heess, Charles Blundell

Figure 1 for Retrieval-Augmented Reinforcement Learning
Figure 2 for Retrieval-Augmented Reinforcement Learning
Figure 3 for Retrieval-Augmented Reinforcement Learning
Figure 4 for Retrieval-Augmented Reinforcement Learning
Viaarxiv icon

Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times

Add code
Bookmark button
Alert button
Jan 30, 2022
Daniele Calandriello, Luigi Carratino, Alessandro Lazaric, Michal Valko, Lorenzo Rosasco

Figure 1 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 2 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 3 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Figure 4 for Scaling Gaussian Process Optimization by Evaluating a Few Unique Candidates Multiple Times
Viaarxiv icon

Adaptive Multi-Goal Exploration

Add code
Bookmark button
Alert button
Nov 23, 2021
Jean Tarbouriech, Omar Darwiche Domingues, Pierre Ménard, Matteo Pirotta, Michal Valko, Alessandro Lazaric

Figure 1 for Adaptive Multi-Goal Exploration
Figure 2 for Adaptive Multi-Goal Exploration
Figure 3 for Adaptive Multi-Goal Exploration
Viaarxiv icon

Drop, Swap, and Generate: A Self-Supervised Approach for Generating Neural Activity

Add code
Bookmark button
Alert button
Nov 03, 2021
Ran Liu, Mehdi Azabou, Max Dabagia, Chi-Heng Lin, Mohammad Gheshlaghi Azar, Keith B. Hengen, Michal Valko, Eva L. Dyer

Viaarxiv icon

Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation

Add code
Bookmark button
Alert button
Jun 24, 2021
Yunhao Tang, Tadashi Kozuno, Mark Rowland, Rémi Munos, Michal Valko

Figure 1 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Figure 2 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Figure 3 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Figure 4 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Viaarxiv icon

Taylor Expansion of Discount Factors

Add code
Bookmark button
Alert button
Jun 14, 2021
Yunhao Tang, Mark Rowland, Rémi Munos, Michal Valko

Figure 1 for Taylor Expansion of Discount Factors
Figure 2 for Taylor Expansion of Discount Factors
Figure 3 for Taylor Expansion of Discount Factors
Figure 4 for Taylor Expansion of Discount Factors
Viaarxiv icon