Alert button
Picture for Doina Precup

Doina Precup

Alert button

Finite time analysis of temporal difference learning with linear function approximation: Tail averaging and regularisation

Add code
Bookmark button
Alert button
Oct 12, 2022
Gandharv Patil, Prashanth L. A., Dheeraj Nagaraj, Doina Precup

Figure 1 for Finite time analysis of temporal difference learning with linear function approximation: Tail averaging and regularisation
Figure 2 for Finite time analysis of temporal difference learning with linear function approximation: Tail averaging and regularisation
Figure 3 for Finite time analysis of temporal difference learning with linear function approximation: Tail averaging and regularisation
Viaarxiv icon

Towards Safe Mechanical Ventilation Treatment Using Deep Offline Reinforcement Learning

Add code
Bookmark button
Alert button
Oct 05, 2022
Flemming Kondrup, Thomas Jiralerspong, Elaine Lau, Nathan de Lara, Jacob Shkrob, My Duc Tran, Doina Precup, Sumana Basu

Viaarxiv icon

Bayesian Q-learning With Imperfect Expert Demonstrations

Add code
Bookmark button
Alert button
Oct 01, 2022
Fengdi Che, Xiru Zhu, Doina Precup, David Meger, Gregory Dudek

Figure 1 for Bayesian Q-learning With Imperfect Expert Demonstrations
Figure 2 for Bayesian Q-learning With Imperfect Expert Demonstrations
Figure 3 for Bayesian Q-learning With Imperfect Expert Demonstrations
Figure 4 for Bayesian Q-learning With Imperfect Expert Demonstrations
Viaarxiv icon

Continuous MDP Homomorphisms and Homomorphic Policy Gradient

Add code
Bookmark button
Alert button
Sep 15, 2022
Sahand Rezaei-Shoshtari, Rosie Zhao, Prakash Panangaden, David Meger, Doina Precup

Figure 1 for Continuous MDP Homomorphisms and Homomorphic Policy Gradient
Figure 2 for Continuous MDP Homomorphisms and Homomorphic Policy Gradient
Figure 3 for Continuous MDP Homomorphisms and Homomorphic Policy Gradient
Figure 4 for Continuous MDP Homomorphisms and Homomorphic Policy Gradient
Viaarxiv icon

Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 16, 2022
Safa Alver, Doina Precup

Figure 1 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Figure 2 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Figure 3 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Figure 4 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Viaarxiv icon

Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification

Add code
Bookmark button
Alert button
May 19, 2022
Leo Schwinn, Leon Bungert, An Nguyen, René Raab, Falk Pulsmeyer, Doina Precup, Björn Eskofier, Dario Zanca

Figure 1 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Figure 2 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Figure 3 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Figure 4 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Viaarxiv icon

Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning

Add code
Bookmark button
Alert button
Apr 21, 2022
Gheorghe Comanici, Amelia Glaese, Anita Gergely, Daniel Toyama, Zafarali Ahmed, Tyler Jackson, Philippe Hamel, Doina Precup

Figure 1 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Figure 2 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Figure 3 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Figure 4 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Viaarxiv icon

Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention

Add code
Bookmark button
Alert button
Apr 19, 2022
Leo Schwinn, Doina Precup, Björn Eskofier, Dario Zanca

Figure 1 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Figure 2 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Figure 3 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Figure 4 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Viaarxiv icon

COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation

Add code
Bookmark button
Alert button
Apr 19, 2022
Jongmin Lee, Cosmin Paduraru, Daniel J. Mankowitz, Nicolas Heess, Doina Precup, Kee-Eung Kim, Arthur Guez

Figure 1 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 2 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 3 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 4 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Viaarxiv icon

Towards Painless Policy Optimization for Constrained MDPs

Add code
Bookmark button
Alert button
Apr 11, 2022
Arushi Jain, Sharan Vaswani, Reza Babanezhad, Csaba Szepesvari, Doina Precup

Figure 1 for Towards Painless Policy Optimization for Constrained MDPs
Figure 2 for Towards Painless Policy Optimization for Constrained MDPs
Figure 3 for Towards Painless Policy Optimization for Constrained MDPs
Figure 4 for Towards Painless Policy Optimization for Constrained MDPs
Viaarxiv icon