Picture for Michael L. Littman

Michael L. Littman

Rutgers University

Stackelberg Punishment and Bully-Proofing Autonomous Vehicles

Add code
Aug 23, 2019
Figure 1 for Stackelberg Punishment and Bully-Proofing Autonomous Vehicles
Figure 2 for Stackelberg Punishment and Bully-Proofing Autonomous Vehicles
Viaarxiv icon

Interactive Learning of Environment Dynamics for Sequential Tasks

Add code
Jul 19, 2019
Figure 1 for Interactive Learning of Environment Dynamics for Sequential Tasks
Figure 2 for Interactive Learning of Environment Dynamics for Sequential Tasks
Figure 3 for Interactive Learning of Environment Dynamics for Sequential Tasks
Figure 4 for Interactive Learning of Environment Dynamics for Sequential Tasks
Viaarxiv icon

Deep Reinforcement Learning from Policy-Dependent Human Feedback

Add code
Feb 12, 2019
Figure 1 for Deep Reinforcement Learning from Policy-Dependent Human Feedback
Figure 2 for Deep Reinforcement Learning from Policy-Dependent Human Feedback
Figure 3 for Deep Reinforcement Learning from Policy-Dependent Human Feedback
Figure 4 for Deep Reinforcement Learning from Policy-Dependent Human Feedback
Viaarxiv icon

Successor Features Support Model-based and Model-free Reinforcement Learning

Add code
Jan 31, 2019
Figure 1 for Successor Features Support Model-based and Model-free Reinforcement Learning
Figure 2 for Successor Features Support Model-based and Model-free Reinforcement Learning
Figure 3 for Successor Features Support Model-based and Model-free Reinforcement Learning
Figure 4 for Successor Features Support Model-based and Model-free Reinforcement Learning
Viaarxiv icon

Theory of Minds: Understanding Behavior in Groups Through Inverse Planning

Add code
Jan 18, 2019
Figure 1 for Theory of Minds: Understanding Behavior in Groups Through Inverse Planning
Figure 2 for Theory of Minds: Understanding Behavior in Groups Through Inverse Planning
Figure 3 for Theory of Minds: Understanding Behavior in Groups Through Inverse Planning
Figure 4 for Theory of Minds: Understanding Behavior in Groups Through Inverse Planning
Viaarxiv icon

Mitigating Planner Overfitting in Model-Based Reinforcement Learning

Add code
Dec 03, 2018
Figure 1 for Mitigating Planner Overfitting in Model-Based Reinforcement Learning
Figure 2 for Mitigating Planner Overfitting in Model-Based Reinforcement Learning
Figure 3 for Mitigating Planner Overfitting in Model-Based Reinforcement Learning
Figure 4 for Mitigating Planner Overfitting in Model-Based Reinforcement Learning
Viaarxiv icon

Towards a Simple Approach to Multi-step Model-based Reinforcement Learning

Add code
Oct 31, 2018
Figure 1 for Towards a Simple Approach to Multi-step Model-based Reinforcement Learning
Figure 2 for Towards a Simple Approach to Multi-step Model-based Reinforcement Learning
Figure 3 for Towards a Simple Approach to Multi-step Model-based Reinforcement Learning
Figure 4 for Towards a Simple Approach to Multi-step Model-based Reinforcement Learning
Viaarxiv icon

Lipschitz Continuity in Model-based Reinforcement Learning

Add code
Jul 27, 2018
Figure 1 for Lipschitz Continuity in Model-based Reinforcement Learning
Figure 2 for Lipschitz Continuity in Model-based Reinforcement Learning
Figure 3 for Lipschitz Continuity in Model-based Reinforcement Learning
Figure 4 for Lipschitz Continuity in Model-based Reinforcement Learning
Viaarxiv icon

Equivalence Between Wasserstein and Value-Aware Loss for Model-based Reinforcement Learning

Add code
Jul 08, 2018
Viaarxiv icon

Transfer with Model Features in Reinforcement Learning

Add code
Jul 04, 2018
Figure 1 for Transfer with Model Features in Reinforcement Learning
Figure 2 for Transfer with Model Features in Reinforcement Learning
Figure 3 for Transfer with Model Features in Reinforcement Learning
Figure 4 for Transfer with Model Features in Reinforcement Learning
Viaarxiv icon