Alert button
Picture for Timothy Mann

Timothy Mann

Alert button

Non-Stationary Bandits with Intermediate Observations

Add code
Bookmark button
Alert button
Jun 03, 2020
Claire Vernade, Andras Gyorgy, Timothy Mann

Figure 1 for Non-Stationary Bandits with Intermediate Observations
Figure 2 for Non-Stationary Bandits with Intermediate Observations
Figure 3 for Non-Stationary Bandits with Intermediate Observations
Figure 4 for Non-Stationary Bandits with Intermediate Observations
Viaarxiv icon

Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations

Add code
Bookmark button
Alert button
Dec 06, 2019
Sven Gowal, Chongli Qin, Po-Sen Huang, Taylan Cemgil, Krishnamurthy Dvijotham, Timothy Mann, Pushmeet Kohli

Figure 1 for Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations
Figure 2 for Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations
Figure 3 for Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations
Figure 4 for Achieving Robustness in the Wild via Adversarial Mixing with Disentangled Representations
Viaarxiv icon

An Alternative Surrogate Loss for PGD-based Adversarial Testing

Add code
Bookmark button
Alert button
Oct 21, 2019
Sven Gowal, Jonathan Uesato, Chongli Qin, Po-Sen Huang, Timothy Mann, Pushmeet Kohli

Figure 1 for An Alternative Surrogate Loss for PGD-based Adversarial Testing
Figure 2 for An Alternative Surrogate Loss for PGD-based Adversarial Testing
Figure 3 for An Alternative Surrogate Loss for PGD-based Adversarial Testing
Figure 4 for An Alternative Surrogate Loss for PGD-based Adversarial Testing
Viaarxiv icon

Adaptive Temporal-Difference Learning for Policy Evaluation with Per-State Uncertainty Estimates

Add code
Bookmark button
Alert button
Jun 19, 2019
Hugo Penedones, Carlos Riquelme, Damien Vincent, Hartmut Maennel, Timothy Mann, Andre Barreto, Sylvain Gelly, Gergely Neu

Figure 1 for Adaptive Temporal-Difference Learning for Policy Evaluation with Per-State Uncertainty Estimates
Figure 2 for Adaptive Temporal-Difference Learning for Policy Evaluation with Per-State Uncertainty Estimates
Figure 3 for Adaptive Temporal-Difference Learning for Policy Evaluation with Per-State Uncertainty Estimates
Figure 4 for Adaptive Temporal-Difference Learning for Policy Evaluation with Per-State Uncertainty Estimates
Viaarxiv icon

Robust Reinforcement Learning for Continuous Control with Model Misspecification

Add code
Bookmark button
Alert button
Jun 18, 2019
Daniel J. Mankowitz, Nir Levine, Rae Jeong, Abbas Abdolmaleki, Jost Tobias Springenberg, Timothy Mann, Todd Hester, Martin Riedmiller

Figure 1 for Robust Reinforcement Learning for Continuous Control with Model Misspecification
Figure 2 for Robust Reinforcement Learning for Continuous Control with Model Misspecification
Figure 3 for Robust Reinforcement Learning for Continuous Control with Model Misspecification
Figure 4 for Robust Reinforcement Learning for Continuous Control with Model Misspecification
Viaarxiv icon

A Bayesian Approach to Robust Reinforcement Learning

Add code
Bookmark button
Alert button
May 20, 2019
Esther Derman, Daniel Mankowitz, Timothy Mann, Shie Mannor

Figure 1 for A Bayesian Approach to Robust Reinforcement Learning
Figure 2 for A Bayesian Approach to Robust Reinforcement Learning
Figure 3 for A Bayesian Approach to Robust Reinforcement Learning
Figure 4 for A Bayesian Approach to Robust Reinforcement Learning
Viaarxiv icon

On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models

Add code
Bookmark button
Alert button
Nov 05, 2018
Sven Gowal, Krishnamurthy Dvijotham, Robert Stanforth, Rudy Bunel, Chongli Qin, Jonathan Uesato, Relja Arandjelovic, Timothy Mann, Pushmeet Kohli

Figure 1 for On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models
Figure 2 for On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models
Figure 3 for On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models
Figure 4 for On the Effectiveness of Interval Bound Propagation for Training Verifiably Robust Models
Viaarxiv icon

A Dual Approach to Scalable Verification of Deep Networks

Add code
Bookmark button
Alert button
Aug 03, 2018
Krishnamurthy, Dvijotham, Robert Stanforth, Sven Gowal, Timothy Mann, Pushmeet Kohli

Figure 1 for A Dual Approach to Scalable Verification of Deep Networks
Viaarxiv icon

Temporal Difference Learning with Neural Networks - Study of the Leakage Propagation Problem

Add code
Bookmark button
Alert button
Jul 09, 2018
Hugo Penedones, Damien Vincent, Hartmut Maennel, Sylvain Gelly, Timothy Mann, Andre Barreto

Figure 1 for Temporal Difference Learning with Neural Networks - Study of the Leakage Propagation Problem
Figure 2 for Temporal Difference Learning with Neural Networks - Study of the Leakage Propagation Problem
Figure 3 for Temporal Difference Learning with Neural Networks - Study of the Leakage Propagation Problem
Figure 4 for Temporal Difference Learning with Neural Networks - Study of the Leakage Propagation Problem
Viaarxiv icon

Deep Reinforcement Learning in Large Discrete Action Spaces

Add code
Bookmark button
Alert button
Apr 04, 2016
Gabriel Dulac-Arnold, Richard Evans, Hado van Hasselt, Peter Sunehag, Timothy Lillicrap, Jonathan Hunt, Timothy Mann, Theophane Weber, Thomas Degris, Ben Coppin

Figure 1 for Deep Reinforcement Learning in Large Discrete Action Spaces
Figure 2 for Deep Reinforcement Learning in Large Discrete Action Spaces
Figure 3 for Deep Reinforcement Learning in Large Discrete Action Spaces
Figure 4 for Deep Reinforcement Learning in Large Discrete Action Spaces
Viaarxiv icon