Alert button
Picture for Csaba Szepesvári

Csaba Szepesvári

Alert button

Regret Minimization via Saddle Point Optimization

Add code
Bookmark button
Alert button
Mar 15, 2024
Johannes Kirschner, Seyed Alireza Bakhtiari, Kushagra Chandak, Volodymyr Tkachuk, Csaba Szepesvári

Figure 1 for Regret Minimization via Saddle Point Optimization
Figure 2 for Regret Minimization via Saddle Point Optimization
Viaarxiv icon

Switching the Loss Reduces the Cost in Batch Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 12, 2024
Alex Ayoub, Kaiwen Wang, Vincent Liu, Samuel Robertson, James McInerney, Dawen Liang, Nathan Kallus, Csaba Szepesvári

Figure 1 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 2 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Figure 3 for Switching the Loss Reduces the Cost in Batch Reinforcement Learning
Viaarxiv icon

Ensemble sampling for linear bandits: small ensembles suffice

Add code
Bookmark button
Alert button
Nov 14, 2023
David Janz, Alexander E. Litvak, Csaba Szepesvári

Viaarxiv icon

Exploration via linearly perturbed loss minimisation

Add code
Bookmark button
Alert button
Nov 13, 2023
David Janz, Shuai Liu, Alex Ayoub, Csaba Szepesvári

Figure 1 for Exploration via linearly perturbed loss minimisation
Figure 2 for Exploration via linearly perturbed loss minimisation
Viaarxiv icon

Stochastic Gradient Descent for Gaussian Processes Done Right

Add code
Bookmark button
Alert button
Oct 31, 2023
Jihao Andreas Lin, Shreyas Padhy, Javier Antorán, Austin Tripp, Alexander Terenin, Csaba Szepesvári, José Miguel Hernández-Lobato, David Janz

Viaarxiv icon

Online RL in Linearly $q^π$-Realizable MDPs Is as Easy as in Linear MDPs If You Learn What to Ignore

Add code
Bookmark button
Alert button
Oct 11, 2023
Gellért Weisz, András György, Csaba Szepesvári

Viaarxiv icon

The Optimal Approximation Factors in Misspecified Off-Policy Value Function Estimation

Add code
Bookmark button
Alert button
Jul 25, 2023
Philip Amortila, Nan Jiang, Csaba Szepesvári

Figure 1 for The Optimal Approximation Factors in Misspecified Off-Policy Value Function Estimation
Figure 2 for The Optimal Approximation Factors in Misspecified Off-Policy Value Function Estimation
Figure 3 for The Optimal Approximation Factors in Misspecified Off-Policy Value Function Estimation
Figure 4 for The Optimal Approximation Factors in Misspecified Off-Policy Value Function Estimation
Viaarxiv icon

Context-lumpable stochastic bandits

Add code
Bookmark button
Alert button
Jun 22, 2023
Chung-Wei Lee, Qinghua Liu, Yasin Abbasi-Yadkori, Chi Jin, Tor Lattimore, Csaba Szepesvári

Viaarxiv icon

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice

Add code
Bookmark button
Alert button
May 22, 2023
Toshinori Kitamura, Tadashi Kozuno, Yunhao Tang, Nino Vieillard, Michal Valko, Wenhao Yang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári, Wataru Kumagai, Yutaka Matsuo

Figure 1 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 2 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 3 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 4 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Viaarxiv icon