Picture for Doina Precup

Doina Precup

McGill University, Mila- Quebec Artificial Intelligence Institute

Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning

Add code
Jun 16, 2022
Figure 1 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Figure 2 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Figure 3 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Figure 4 for Understanding Decision-Time vs. Background Planning in Model-Based Reinforcement Learning
Viaarxiv icon

Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification

Add code
May 19, 2022
Figure 1 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Figure 2 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Figure 3 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Figure 4 for Improving Robustness against Real-World and Worst-Case Distribution Shifts through Decision Region Quantification
Viaarxiv icon

Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning

Add code
Apr 21, 2022
Figure 1 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Figure 2 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Figure 3 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Figure 4 for Learning how to Interact with a Complex Interface using Hierarchical Reinforcement Learning
Viaarxiv icon

Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention

Add code
Apr 19, 2022
Figure 1 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Figure 2 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Figure 3 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Figure 4 for Behind the Machine's Gaze: Biologically Constrained Neural Networks Exhibit Human-like Visual Attention
Viaarxiv icon

COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation

Add code
Apr 19, 2022
Figure 1 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 2 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 3 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Figure 4 for COptiDICE: Offline Constrained Reinforcement Learning via Stationary Distribution Correction Estimation
Viaarxiv icon

Towards Painless Policy Optimization for Constrained MDPs

Add code
Apr 11, 2022
Figure 1 for Towards Painless Policy Optimization for Constrained MDPs
Figure 2 for Towards Painless Policy Optimization for Constrained MDPs
Figure 3 for Towards Painless Policy Optimization for Constrained MDPs
Figure 4 for Towards Painless Policy Optimization for Constrained MDPs
Viaarxiv icon

Selective Credit Assignment

Add code
Feb 20, 2022
Figure 1 for Selective Credit Assignment
Figure 2 for Selective Credit Assignment
Figure 3 for Selective Credit Assignment
Figure 4 for Selective Credit Assignment
Viaarxiv icon

Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers

Add code
Feb 01, 2022
Figure 1 for Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers
Figure 2 for Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers
Figure 3 for Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers
Figure 4 for Improving Sample Efficiency of Value Based Models Using Attention and Vision Transformers
Viaarxiv icon

Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error

Add code
Jan 28, 2022
Figure 1 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 2 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 3 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Figure 4 for Why Should I Trust You, Bellman? The Bellman Error is a Poor Replacement for Value Error
Viaarxiv icon

The Paradox of Choice: Using Attention in Hierarchical Reinforcement Learning

Add code
Jan 24, 2022
Figure 1 for The Paradox of Choice: Using Attention in Hierarchical Reinforcement Learning
Figure 2 for The Paradox of Choice: Using Attention in Hierarchical Reinforcement Learning
Figure 3 for The Paradox of Choice: Using Attention in Hierarchical Reinforcement Learning
Figure 4 for The Paradox of Choice: Using Attention in Hierarchical Reinforcement Learning
Viaarxiv icon