Alert button
Picture for Yunhao Tang

Yunhao Tang

Alert button

BYOL-Explore: Exploration by Bootstrapped Prediction

Add code
Bookmark button
Alert button
Jun 16, 2022
Zhaohan Daniel Guo, Shantanu Thakoor, Miruna Pîslar, Bernardo Avila Pires, Florent Altché, Corentin Tallec, Alaa Saade, Daniele Calandriello, Jean-Bastien Grill, Yunhao Tang, Michal Valko, Rémi Munos, Mohammad Gheshlaghi Azar, Bilal Piot

Figure 1 for BYOL-Explore: Exploration by Bootstrapped Prediction
Figure 2 for BYOL-Explore: Exploration by Bootstrapped Prediction
Figure 3 for BYOL-Explore: Exploration by Bootstrapped Prediction
Figure 4 for BYOL-Explore: Exploration by Bootstrapped Prediction
Viaarxiv icon

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
Bookmark button
Alert button
May 27, 2022
Tadashi Kozuno, Wenhao Yang, Nino Vieillard, Toshinori Kitamura, Yunhao Tang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Michal Valko, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári

Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses

Add code
Bookmark button
Alert button
May 16, 2022
Daniil Tiapkin, Denis Belomestny, Eric Moulines, Alexey Naumov, Sergey Samsonov, Yunhao Tang, Michal Valko, Pierre Menard

Figure 1 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Figure 2 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Figure 3 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Figure 4 for From Dirichlet to Rubin: Optimistic Exploration in RL without Bonuses
Viaarxiv icon

Marginalized Operators for Off-policy Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 30, 2022
Yunhao Tang, Mark Rowland, Rémi Munos, Michal Valko

Figure 1 for Marginalized Operators for Off-policy Reinforcement Learning
Figure 2 for Marginalized Operators for Off-policy Reinforcement Learning
Figure 3 for Marginalized Operators for Off-policy Reinforcement Learning
Figure 4 for Marginalized Operators for Off-policy Reinforcement Learning
Viaarxiv icon

Biased Gradient Estimate with Drastic Variance Reduction for Meta Reinforcement Learning

Add code
Bookmark button
Alert button
Dec 14, 2021
Yunhao Tang

Figure 1 for Biased Gradient Estimate with Drastic Variance Reduction for Meta Reinforcement Learning
Figure 2 for Biased Gradient Estimate with Drastic Variance Reduction for Meta Reinforcement Learning
Viaarxiv icon

Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation

Add code
Bookmark button
Alert button
Jun 24, 2021
Yunhao Tang, Tadashi Kozuno, Mark Rowland, Rémi Munos, Michal Valko

Figure 1 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Figure 2 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Figure 3 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Figure 4 for Unifying Gradient Estimators for Meta-Reinforcement Learning via Off-Policy Evaluation
Viaarxiv icon

Taylor Expansion of Discount Factors

Add code
Bookmark button
Alert button
Jun 14, 2021
Yunhao Tang, Mark Rowland, Rémi Munos, Michal Valko

Figure 1 for Taylor Expansion of Discount Factors
Figure 2 for Taylor Expansion of Discount Factors
Figure 3 for Taylor Expansion of Discount Factors
Figure 4 for Taylor Expansion of Discount Factors
Viaarxiv icon

Unlocking Pixels for Reinforcement Learning via Implicit Attention

Add code
Bookmark button
Alert button
Mar 04, 2021
Krzysztof Choromanski, Deepali Jain, Jack Parker-Holder, Xingyou Song, Valerii Likhosherstov, Anirban Santara, Aldo Pacchiano, Yunhao Tang, Adrian Weller

Figure 1 for Unlocking Pixels for Reinforcement Learning via Implicit Attention
Figure 2 for Unlocking Pixels for Reinforcement Learning via Implicit Attention
Figure 3 for Unlocking Pixels for Reinforcement Learning via Implicit Attention
Figure 4 for Unlocking Pixels for Reinforcement Learning via Implicit Attention
Viaarxiv icon

Revisiting Peng's Q($λ$) for Modern Reinforcement Learning

Add code
Bookmark button
Alert button
Feb 27, 2021
Tadashi Kozuno, Yunhao Tang, Mark Rowland, Rémi Munos, Steven Kapturowski, Will Dabney, Michal Valko, David Abel

Figure 1 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Figure 2 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Figure 3 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Figure 4 for Revisiting Peng's Q($λ$) for Modern Reinforcement Learning
Viaarxiv icon