Alert button
Picture for Matthijs T. J. Spaan

Matthijs T. J. Spaan

Alert button

Exploring LLMs as a Source of Targeted Synthetic Textual Data to Minimize High Confidence Misclassifications

Add code
Bookmark button
Alert button
Apr 02, 2024
Philip Lippmann, Matthijs T. J. Spaan, Jie Yang

Viaarxiv icon

When Do Off-Policy and On-Policy Policy Gradient Methods Align?

Add code
Bookmark button
Alert button
Feb 19, 2024
Davide Mambelli, Stephan Bongers, Onno Zoeter, Matthijs T. J. Spaan, Frans A. Oliehoek

Viaarxiv icon

Reinforcement Learning by Guided Safe Exploration

Add code
Bookmark button
Alert button
Jul 26, 2023
Qisong Yang, Thiago D. Simão, Nils Jansen, Simon H. Tindemans, Matthijs T. J. Spaan

Figure 1 for Reinforcement Learning by Guided Safe Exploration
Figure 2 for Reinforcement Learning by Guided Safe Exploration
Figure 3 for Reinforcement Learning by Guided Safe Exploration
Figure 4 for Reinforcement Learning by Guided Safe Exploration
Viaarxiv icon

Diverse Projection Ensembles for Distributional Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 12, 2023
Moritz A. Zanger, Wendelin Böhmer, Matthijs T. J. Spaan

Figure 1 for Diverse Projection Ensembles for Distributional Reinforcement Learning
Figure 2 for Diverse Projection Ensembles for Distributional Reinforcement Learning
Figure 3 for Diverse Projection Ensembles for Distributional Reinforcement Learning
Figure 4 for Diverse Projection Ensembles for Distributional Reinforcement Learning
Viaarxiv icon

The Role of Diverse Replay for Generalisation in Reinforcement Learning

Add code
Bookmark button
Alert button
Jun 09, 2023
Max Weltevrede, Matthijs T. J. Spaan, Wendelin Böhmer

Figure 1 for The Role of Diverse Replay for Generalisation in Reinforcement Learning
Figure 2 for The Role of Diverse Replay for Generalisation in Reinforcement Learning
Figure 3 for The Role of Diverse Replay for Generalisation in Reinforcement Learning
Figure 4 for The Role of Diverse Replay for Generalisation in Reinforcement Learning
Viaarxiv icon

Bad Habits: Policy Confounding and Out-of-Trajectory Generalization in RL

Add code
Bookmark button
Alert button
Jun 04, 2023
Miguel Suau, Matthijs T. J. Spaan, Frans A. Oliehoek

Figure 1 for Bad Habits: Policy Confounding and Out-of-Trajectory Generalization in RL
Figure 2 for Bad Habits: Policy Confounding and Out-of-Trajectory Generalization in RL
Figure 3 for Bad Habits: Policy Confounding and Out-of-Trajectory Generalization in RL
Figure 4 for Bad Habits: Policy Confounding and Out-of-Trajectory Generalization in RL
Viaarxiv icon

Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems

Add code
Bookmark button
Alert button
Jul 01, 2022
Miguel Suau, Jinke He, Mustafa Mert Çelikok, Matthijs T. J. Spaan, Frans A. Oliehoek

Figure 1 for Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems
Figure 2 for Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems
Figure 3 for Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems
Figure 4 for Distributed Influence-Augmented Local Simulators for Parallel MARL in Large Networked Systems
Viaarxiv icon

Abstraction-Refinement for Hierarchical Probabilistic Models

Add code
Bookmark button
Alert button
Jun 06, 2022
Sebastian Junges, Matthijs T. J. Spaan

Figure 1 for Abstraction-Refinement for Hierarchical Probabilistic Models
Figure 2 for Abstraction-Refinement for Hierarchical Probabilistic Models
Figure 3 for Abstraction-Refinement for Hierarchical Probabilistic Models
Figure 4 for Abstraction-Refinement for Hierarchical Probabilistic Models
Viaarxiv icon

Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems

Add code
Bookmark button
Alert button
Feb 03, 2022
Miguel Suau, Jinke He, Matthijs T. J. Spaan, Frans A. Oliehoek

Figure 1 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Figure 2 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Figure 3 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Figure 4 for Influence-Augmented Local Simulators: A Scalable Solution for Fast Deep RL in Large Networked Systems
Viaarxiv icon

Exploiting Submodular Value Functions For Scaling Up Active Perception

Add code
Bookmark button
Alert button
Sep 21, 2020
Yash Satsangi, Shimon Whiteson, Frans A. Oliehoek, Matthijs T. J. Spaan

Figure 1 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Figure 2 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Figure 3 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Figure 4 for Exploiting Submodular Value Functions For Scaling Up Active Perception
Viaarxiv icon