Picture for Frans A. Oliehoek

Frans A. Oliehoek

BADDr: Bayes-Adaptive Deep Dropout RL for POMDPs

Add code
Feb 17, 2022
Figure 1 for BADDr: Bayes-Adaptive Deep Dropout RL for POMDPs
Figure 2 for BADDr: Bayes-Adaptive Deep Dropout RL for POMDPs
Figure 3 for BADDr: Bayes-Adaptive Deep Dropout RL for POMDPs
Figure 4 for BADDr: Bayes-Adaptive Deep Dropout RL for POMDPs
Viaarxiv icon

Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems

Add code
Feb 03, 2022
Figure 1 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Figure 2 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Figure 3 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Figure 4 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Viaarxiv icon

Online Planning in POMDPs with Self-Improving Simulators

Add code
Jan 27, 2022
Figure 1 for Online Planning in POMDPs with Self-Improving Simulators
Figure 2 for Online Planning in POMDPs with Self-Improving Simulators
Figure 3 for Online Planning in POMDPs with Self-Improving Simulators
Figure 4 for Online Planning in POMDPs with Self-Improving Simulators
Viaarxiv icon

MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning

Add code
Dec 30, 2021
Figure 1 for MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning
Figure 2 for MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning
Figure 3 for MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning
Figure 4 for MORAL: Aligning AI with Human Norms through Multi-Objective Reinforced Active Learning
Viaarxiv icon

Multi-Agent MDP Homomorphic Networks

Add code
Oct 09, 2021
Figure 1 for Multi-Agent MDP Homomorphic Networks
Figure 2 for Multi-Agent MDP Homomorphic Networks
Figure 3 for Multi-Agent MDP Homomorphic Networks
Figure 4 for Multi-Agent MDP Homomorphic Networks
Viaarxiv icon

Difference Rewards Policy Gradients

Add code
Dec 21, 2020
Figure 1 for Difference Rewards Policy Gradients
Figure 2 for Difference Rewards Policy Gradients
Figure 3 for Difference Rewards Policy Gradients
Figure 4 for Difference Rewards Policy Gradients
Viaarxiv icon

Analog Circuit Design with Dyna-Style Reinforcement Learning

Add code
Nov 16, 2020
Figure 1 for Analog Circuit Design with Dyna-Style Reinforcement Learning
Figure 2 for Analog Circuit Design with Dyna-Style Reinforcement Learning
Figure 3 for Analog Circuit Design with Dyna-Style Reinforcement Learning
Figure 4 for Analog Circuit Design with Dyna-Style Reinforcement Learning
Viaarxiv icon

Loss Bounds for Approximate Influence-Based Abstraction

Add code
Nov 03, 2020
Figure 1 for Loss Bounds for Approximate Influence-Based Abstraction
Figure 2 for Loss Bounds for Approximate Influence-Based Abstraction
Figure 3 for Loss Bounds for Approximate Influence-Based Abstraction
Figure 4 for Loss Bounds for Approximate Influence-Based Abstraction
Viaarxiv icon

Multi-agent active perception with prediction rewards

Add code
Oct 22, 2020
Figure 1 for Multi-agent active perception with prediction rewards
Figure 2 for Multi-agent active perception with prediction rewards
Figure 3 for Multi-agent active perception with prediction rewards
Figure 4 for Multi-agent active perception with prediction rewards
Viaarxiv icon

Influence-Augmented Online Planning for Complex Environments

Add code
Oct 21, 2020
Figure 1 for Influence-Augmented Online Planning for Complex Environments
Figure 2 for Influence-Augmented Online Planning for Complex Environments
Figure 3 for Influence-Augmented Online Planning for Complex Environments
Figure 4 for Influence-Augmented Online Planning for Complex Environments
Viaarxiv icon