Picture for Tom Everitt

Tom Everitt

DeepMind

Discovering Agents

Add code
Aug 24, 2022
Figure 1 for Discovering Agents
Figure 2 for Discovering Agents
Figure 3 for Discovering Agents
Figure 4 for Discovering Agents
Viaarxiv icon

Path-Specific Objectives for Safer Agent Incentives

Add code
Apr 21, 2022
Figure 1 for Path-Specific Objectives for Safer Agent Incentives
Figure 2 for Path-Specific Objectives for Safer Agent Incentives
Figure 3 for Path-Specific Objectives for Safer Agent Incentives
Figure 4 for Path-Specific Objectives for Safer Agent Incentives
Viaarxiv icon

A Complete Criterion for Value of Information in Soluble Influence Diagrams

Add code
Feb 23, 2022
Figure 1 for A Complete Criterion for Value of Information in Soluble Influence Diagrams
Figure 2 for A Complete Criterion for Value of Information in Soluble Influence Diagrams
Figure 3 for A Complete Criterion for Value of Information in Soluble Influence Diagrams
Figure 4 for A Complete Criterion for Value of Information in Soluble Influence Diagrams
Viaarxiv icon

Why Fair Labels Can Yield Unfair Predictions: Graphical Conditions for Introduced Unfairness

Add code
Feb 23, 2022
Figure 1 for Why Fair Labels Can Yield Unfair Predictions: Graphical Conditions for Introduced Unfairness
Figure 2 for Why Fair Labels Can Yield Unfair Predictions: Graphical Conditions for Introduced Unfairness
Figure 3 for Why Fair Labels Can Yield Unfair Predictions: Graphical Conditions for Introduced Unfairness
Figure 4 for Why Fair Labels Can Yield Unfair Predictions: Graphical Conditions for Introduced Unfairness
Viaarxiv icon

Shaking the foundations: delusions in sequence models for interaction and control

Add code
Oct 20, 2021
Figure 1 for Shaking the foundations: delusions in sequence models for interaction and control
Figure 2 for Shaking the foundations: delusions in sequence models for interaction and control
Figure 3 for Shaking the foundations: delusions in sequence models for interaction and control
Figure 4 for Shaking the foundations: delusions in sequence models for interaction and control
Viaarxiv icon

Alignment of Language Agents

Add code
Mar 26, 2021
Viaarxiv icon

How RL Agents Behave When Their Actions Are Modified

Add code
Feb 15, 2021
Figure 1 for How RL Agents Behave When Their Actions Are Modified
Figure 2 for How RL Agents Behave When Their Actions Are Modified
Figure 3 for How RL Agents Behave When Their Actions Are Modified
Figure 4 for How RL Agents Behave When Their Actions Are Modified
Viaarxiv icon

Equilibrium Refinements for Multi-Agent Influence Diagrams: Theory and Practice

Add code
Feb 09, 2021
Figure 1 for Equilibrium Refinements for Multi-Agent Influence Diagrams: Theory and Practice
Figure 2 for Equilibrium Refinements for Multi-Agent Influence Diagrams: Theory and Practice
Figure 3 for Equilibrium Refinements for Multi-Agent Influence Diagrams: Theory and Practice
Figure 4 for Equilibrium Refinements for Multi-Agent Influence Diagrams: Theory and Practice
Viaarxiv icon

Agent Incentives: A Causal Perspective

Add code
Feb 02, 2021
Figure 1 for Agent Incentives: A Causal Perspective
Figure 2 for Agent Incentives: A Causal Perspective
Figure 3 for Agent Incentives: A Causal Perspective
Figure 4 for Agent Incentives: A Causal Perspective
Viaarxiv icon

Avoiding Tampering Incentives in Deep RL via Decoupled Approval

Add code
Nov 17, 2020
Figure 1 for Avoiding Tampering Incentives in Deep RL via Decoupled Approval
Figure 2 for Avoiding Tampering Incentives in Deep RL via Decoupled Approval
Figure 3 for Avoiding Tampering Incentives in Deep RL via Decoupled Approval
Figure 4 for Avoiding Tampering Incentives in Deep RL via Decoupled Approval
Viaarxiv icon