Alert button
Picture for Brendan O'Donoghue

Brendan O'Donoghue

Alert button

Probabilistic Inference in Reinforcement Learning Done Right

Add code
Bookmark button
Alert button
Nov 22, 2023
Jean Tarbouriech, Tor Lattimore, Brendan O'Donoghue

Viaarxiv icon

Efficient exploration via epistemic-risk-seeking policy optimization

Add code
Bookmark button
Alert button
Feb 18, 2023
Brendan O'Donoghue

Figure 1 for Efficient exploration via epistemic-risk-seeking policy optimization
Figure 2 for Efficient exploration via epistemic-risk-seeking policy optimization
Figure 3 for Efficient exploration via epistemic-risk-seeking policy optimization
Figure 4 for Efficient exploration via epistemic-risk-seeking policy optimization
Viaarxiv icon

ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs

Add code
Bookmark button
Alert button
Feb 02, 2023
Ted Moskovitz, Brendan O'Donoghue, Vivek Veeriah, Sebastian Flennerhag, Satinder Singh, Tom Zahavy

Figure 1 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 2 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 3 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Figure 4 for ReLOAD: Reinforcement Learning with Optimistic Ascent-Descent for Last-Iterate Convergence in Constrained MDPs
Viaarxiv icon

Optimistic Meta-Gradients

Add code
Bookmark button
Alert button
Jan 09, 2023
Sebastian Flennerhag, Tom Zahavy, Brendan O'Donoghue, Hado van Hasselt, András György, Satinder Singh

Figure 1 for Optimistic Meta-Gradients
Figure 2 for Optimistic Meta-Gradients
Figure 3 for Optimistic Meta-Gradients
Figure 4 for Optimistic Meta-Gradients
Viaarxiv icon

POMRL: No-Regret Learning-to-Plan with Increasing Horizons

Add code
Bookmark button
Alert button
Dec 30, 2022
Khimya Khetarpal, Claire Vernade, Brendan O'Donoghue, Satinder Singh, Tom Zahavy

Figure 1 for POMRL: No-Regret Learning-to-Plan with Increasing Horizons
Figure 2 for POMRL: No-Regret Learning-to-Plan with Increasing Horizons
Figure 3 for POMRL: No-Regret Learning-to-Plan with Increasing Horizons
Figure 4 for POMRL: No-Regret Learning-to-Plan with Increasing Horizons
Viaarxiv icon

Variational Bayesian Optimistic Sampling

Add code
Bookmark button
Alert button
Oct 29, 2021
Brendan O'Donoghue, Tor Lattimore

Figure 1 for Variational Bayesian Optimistic Sampling
Figure 2 for Variational Bayesian Optimistic Sampling
Figure 3 for Variational Bayesian Optimistic Sampling
Figure 4 for Variational Bayesian Optimistic Sampling
Viaarxiv icon

Evaluating Predictive Distributions: Does Bayesian Deep Learning Work?

Add code
Bookmark button
Alert button
Oct 09, 2021
Ian Osband, Zheng Wen, Seyed Mohammad Asghari, Vikranth Dwaracherla, Botao Hao, Morteza Ibrahimi, Dieterich Lawson, Xiuyuan Lu, Brendan O'Donoghue, Benjamin Van Roy

Figure 1 for Evaluating Predictive Distributions: Does Bayesian Deep Learning Work?
Figure 2 for Evaluating Predictive Distributions: Does Bayesian Deep Learning Work?
Figure 3 for Evaluating Predictive Distributions: Does Bayesian Deep Learning Work?
Figure 4 for Evaluating Predictive Distributions: Does Bayesian Deep Learning Work?
Viaarxiv icon

Discovering Diverse Nearly Optimal Policies withSuccessor Features

Add code
Bookmark button
Alert button
Jun 01, 2021
Tom Zahavy, Brendan O'Donoghue, Andre Barreto, Volodymyr Mnih, Sebastian Flennerhag, Satinder Singh

Figure 1 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 2 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 3 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Figure 4 for Discovering Diverse Nearly Optimal Policies withSuccessor Features
Viaarxiv icon

Reward is enough for convex MDPs

Add code
Bookmark button
Alert button
Jun 01, 2021
Tom Zahavy, Brendan O'Donoghue, Guillaume Desjardins, Satinder Singh

Figure 1 for Reward is enough for convex MDPs
Figure 2 for Reward is enough for convex MDPs
Viaarxiv icon

Discovering a set of policies for the worst case reward

Add code
Bookmark button
Alert button
Feb 08, 2021
Tom Zahavy, Andre Barreto, Daniel J Mankowitz, Shaobo Hou, Brendan O'Donoghue, Iurii Kemaev, Satinder Baveja Singh

Figure 1 for Discovering a set of policies for the worst case reward
Figure 2 for Discovering a set of policies for the worst case reward
Figure 3 for Discovering a set of policies for the worst case reward
Figure 4 for Discovering a set of policies for the worst case reward
Viaarxiv icon