Alert button
Picture for Shane Legg

Shane Legg

Alert button

Learning Human Objectives by Evaluating Hypothetical Behavior

Add code
Bookmark button
Alert button
Dec 05, 2019
Siddharth Reddy, Anca D. Dragan, Sergey Levine, Shane Legg, Jan Leike

Figure 1 for Learning Human Objectives by Evaluating Hypothetical Behavior
Figure 2 for Learning Human Objectives by Evaluating Hypothetical Behavior
Figure 3 for Learning Human Objectives by Evaluating Hypothetical Behavior
Figure 4 for Learning Human Objectives by Evaluating Hypothetical Behavior
Viaarxiv icon

Modeling AGI Safety Frameworks with Causal Influence Diagrams

Add code
Bookmark button
Alert button
Jun 20, 2019
Tom Everitt, Ramana Kumar, Victoria Krakovna, Shane Legg

Figure 1 for Modeling AGI Safety Frameworks with Causal Influence Diagrams
Figure 2 for Modeling AGI Safety Frameworks with Causal Influence Diagrams
Figure 3 for Modeling AGI Safety Frameworks with Causal Influence Diagrams
Figure 4 for Modeling AGI Safety Frameworks with Causal Influence Diagrams
Viaarxiv icon

Meta-learning of Sequential Strategies

Add code
Bookmark button
Alert button
May 08, 2019
Pedro A. Ortega, Jane X. Wang, Mark Rowland, Tim Genewein, Zeb Kurth-Nelson, Razvan Pascanu, Nicolas Heess, Joel Veness, Alex Pritzel, Pablo Sprechmann, Siddhant M. Jayakumar, Tom McGrath, Kevin Miller, Mohammad Azar, Ian Osband, Neil Rabinowitz, András György, Silvia Chiappa, Simon Osindero, Yee Whye Teh, Hado van Hasselt, Nando de Freitas, Matthew Botvinick, Shane Legg

Figure 1 for Meta-learning of Sequential Strategies
Figure 2 for Meta-learning of Sequential Strategies
Figure 3 for Meta-learning of Sequential Strategies
Figure 4 for Meta-learning of Sequential Strategies
Viaarxiv icon

Understanding Agent Incentives using Causal Influence Diagrams. Part I: Single Action Settings

Add code
Bookmark button
Alert button
Mar 12, 2019
Tom Everitt, Pedro A. Ortega, Elizabeth Barnes, Shane Legg

Figure 1 for Understanding Agent Incentives using Causal Influence Diagrams. Part I: Single Action Settings
Figure 2 for Understanding Agent Incentives using Causal Influence Diagrams. Part I: Single Action Settings
Figure 3 for Understanding Agent Incentives using Causal Influence Diagrams. Part I: Single Action Settings
Figure 4 for Understanding Agent Incentives using Causal Influence Diagrams. Part I: Single Action Settings
Viaarxiv icon

Soft-Bayes: Prod for Mixtures of Experts with Log-Loss

Add code
Bookmark button
Alert button
Jan 08, 2019
Laurent Orseau, Tor Lattimore, Shane Legg

Viaarxiv icon

Scaling shared model governance via model splitting

Add code
Bookmark button
Alert button
Dec 14, 2018
Miljan Martic, Jan Leike, Andrew Trask, Matteo Hessel, Shane Legg, Pushmeet Kohli

Figure 1 for Scaling shared model governance via model splitting
Figure 2 for Scaling shared model governance via model splitting
Figure 3 for Scaling shared model governance via model splitting
Figure 4 for Scaling shared model governance via model splitting
Viaarxiv icon

Scalable agent alignment via reward modeling: a research direction

Add code
Bookmark button
Alert button
Nov 19, 2018
Jan Leike, David Krueger, Tom Everitt, Miljan Martic, Vishal Maini, Shane Legg

Figure 1 for Scalable agent alignment via reward modeling: a research direction
Figure 2 for Scalable agent alignment via reward modeling: a research direction
Figure 3 for Scalable agent alignment via reward modeling: a research direction
Figure 4 for Scalable agent alignment via reward modeling: a research direction
Viaarxiv icon

Reward learning from human preferences and demonstrations in Atari

Add code
Bookmark button
Alert button
Nov 15, 2018
Borja Ibarz, Jan Leike, Tobias Pohlen, Geoffrey Irving, Shane Legg, Dario Amodei

Figure 1 for Reward learning from human preferences and demonstrations in Atari
Figure 2 for Reward learning from human preferences and demonstrations in Atari
Figure 3 for Reward learning from human preferences and demonstrations in Atari
Figure 4 for Reward learning from human preferences and demonstrations in Atari
Viaarxiv icon

Modeling Friends and Foes

Add code
Bookmark button
Alert button
Jun 30, 2018
Pedro A. Ortega, Shane Legg

Figure 1 for Modeling Friends and Foes
Figure 2 for Modeling Friends and Foes
Figure 3 for Modeling Friends and Foes
Figure 4 for Modeling Friends and Foes
Viaarxiv icon

IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures

Add code
Bookmark button
Alert button
Jun 28, 2018
Lasse Espeholt, Hubert Soyer, Remi Munos, Karen Simonyan, Volodymir Mnih, Tom Ward, Yotam Doron, Vlad Firoiu, Tim Harley, Iain Dunning, Shane Legg, Koray Kavukcuoglu

Figure 1 for IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
Figure 2 for IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
Figure 3 for IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
Figure 4 for IMPALA: Scalable Distributed Deep-RL with Importance Weighted Actor-Learner Architectures
Viaarxiv icon