Alert button
Picture for Daniil Tiapkin

Daniil Tiapkin

Alert button

Incentivized Learning in Principal-Agent Bandit Games

Add code
Bookmark button
Alert button
Mar 06, 2024
Antoine Scheid, Daniil Tiapkin, Etienne Boursier, Aymeric Capitaine, El Mahdi El Mhamdi, Eric Moulines, Michael I. Jordan, Alain Durmus

Figure 1 for Incentivized Learning in Principal-Agent Bandit Games
Figure 2 for Incentivized Learning in Principal-Agent Bandit Games
Figure 3 for Incentivized Learning in Principal-Agent Bandit Games
Figure 4 for Incentivized Learning in Principal-Agent Bandit Games
Viaarxiv icon

Model-free Posterior Sampling via Learning Rate Randomization

Add code
Bookmark button
Alert button
Oct 27, 2023
Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Remi Munos, Alexey Naumov, Pierre Perrault, Michal Valko, Pierre Menard

Figure 1 for Model-free Posterior Sampling via Learning Rate Randomization
Figure 2 for Model-free Posterior Sampling via Learning Rate Randomization
Figure 3 for Model-free Posterior Sampling via Learning Rate Randomization
Figure 4 for Model-free Posterior Sampling via Learning Rate Randomization
Viaarxiv icon

Demonstration-Regularized RL

Add code
Bookmark button
Alert button
Oct 26, 2023
Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Alexey Naumov, Pierre Perrault, Michal Valko, Pierre Menard

Viaarxiv icon

Generative Flow Networks as Entropy-Regularized RL

Add code
Bookmark button
Alert button
Oct 23, 2023
Daniil Tiapkin, Nikita Morozov, Alexey Naumov, Dmitry Vetrov

Figure 1 for Generative Flow Networks as Entropy-Regularized RL
Figure 2 for Generative Flow Networks as Entropy-Regularized RL
Figure 3 for Generative Flow Networks as Entropy-Regularized RL
Figure 4 for Generative Flow Networks as Entropy-Regularized RL
Viaarxiv icon

Finite-Sample Analysis of the Temporal Difference Learning

Add code
Bookmark button
Alert button
Oct 22, 2023
Sergey Samsonov, Daniil Tiapkin, Alexey Naumov, Eric Moulines

Viaarxiv icon

Sharp Deviations Bounds for Dirichlet Weighted Sums with Application to analysis of Bayesian algorithms

Add code
Bookmark button
Alert button
Apr 06, 2023
Denis Belomestny, Pierre Menard, Alexey Naumov, Daniil Tiapkin, Michal Valko

Viaarxiv icon

Orthogonal Directions Constrained Gradient Method: from non-linear equality constraints to Stiefel manifold

Add code
Bookmark button
Alert button
Mar 16, 2023
Sholom Schechtman, Daniil Tiapkin, Michael Muehlebach, Eric Moulines

Figure 1 for Orthogonal Directions Constrained Gradient Method: from non-linear equality constraints to Stiefel manifold
Figure 2 for Orthogonal Directions Constrained Gradient Method: from non-linear equality constraints to Stiefel manifold
Figure 3 for Orthogonal Directions Constrained Gradient Method: from non-linear equality constraints to Stiefel manifold
Figure 4 for Orthogonal Directions Constrained Gradient Method: from non-linear equality constraints to Stiefel manifold
Viaarxiv icon

Fast Rates for Maximum Entropy Exploration

Add code
Bookmark button
Alert button
Mar 14, 2023
Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Remi Munos, Alexey Naumov, Pierre Perrault, Yunhao Tang, Michal Valko, Pierre Menard

Figure 1 for Fast Rates for Maximum Entropy Exploration
Figure 2 for Fast Rates for Maximum Entropy Exploration
Figure 3 for Fast Rates for Maximum Entropy Exploration
Figure 4 for Fast Rates for Maximum Entropy Exploration
Viaarxiv icon

Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees

Add code
Bookmark button
Alert button
Sep 28, 2022
Daniil Tiapkin, Denis Belomestny, Daniele Calandriello, Eric Moulines, Remi Munos, Alexey Naumov, Mark Rowland, Michal Valko, Pierre Menard

Figure 1 for Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees
Figure 2 for Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees
Figure 3 for Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees
Figure 4 for Optimistic Posterior Sampling for Reinforcement Learning with Few Samples and Tight Guarantees
Viaarxiv icon