Picture for Riccardo De Santi

Riccardo De Santi

Efficient Personalization of Generative Models via Optimal Experimental Design

Add code
Dec 22, 2025
Viaarxiv icon

Provable Maximum Entropy Manifold Exploration via Diffusion Models

Add code
Jun 18, 2025
Figure 1 for Provable Maximum Entropy Manifold Exploration via Diffusion Models
Figure 2 for Provable Maximum Entropy Manifold Exploration via Diffusion Models
Figure 3 for Provable Maximum Entropy Manifold Exploration via Diffusion Models
Figure 4 for Provable Maximum Entropy Manifold Exploration via Diffusion Models
Viaarxiv icon

Geometric Active Exploration in Markov Decision Processes: the Benefit of Abstraction

Add code
Jul 18, 2024
Figure 1 for Geometric Active Exploration in Markov Decision Processes: the Benefit of Abstraction
Figure 2 for Geometric Active Exploration in Markov Decision Processes: the Benefit of Abstraction
Viaarxiv icon

Global Reinforcement Learning: Beyond Linear and Convex Rewards via Submodular Semi-gradient Methods

Add code
Jul 13, 2024
Viaarxiv icon

Exploiting Causal Graph Priors with Posterior Sampling for Reinforcement Learning

Add code
Oct 11, 2023
Figure 1 for Exploiting Causal Graph Priors with Posterior Sampling for Reinforcement Learning
Figure 2 for Exploiting Causal Graph Priors with Posterior Sampling for Reinforcement Learning
Viaarxiv icon

Provably Efficient Causal Model-Based Reinforcement Learning for Systematic Generalization

Add code
Feb 14, 2022
Figure 1 for Provably Efficient Causal Model-Based Reinforcement Learning for Systematic Generalization
Figure 2 for Provably Efficient Causal Model-Based Reinforcement Learning for Systematic Generalization
Figure 3 for Provably Efficient Causal Model-Based Reinforcement Learning for Systematic Generalization
Figure 4 for Provably Efficient Causal Model-Based Reinforcement Learning for Systematic Generalization
Viaarxiv icon

The Importance of Non-Markovianity in Maximum State Entropy Exploration

Add code
Feb 07, 2022
Figure 1 for The Importance of Non-Markovianity in Maximum State Entropy Exploration
Figure 2 for The Importance of Non-Markovianity in Maximum State Entropy Exploration
Figure 3 for The Importance of Non-Markovianity in Maximum State Entropy Exploration
Viaarxiv icon

Challenging Common Assumptions in Convex Reinforcement Learning

Add code
Feb 03, 2022
Figure 1 for Challenging Common Assumptions in Convex Reinforcement Learning
Figure 2 for Challenging Common Assumptions in Convex Reinforcement Learning
Figure 3 for Challenging Common Assumptions in Convex Reinforcement Learning
Figure 4 for Challenging Common Assumptions in Convex Reinforcement Learning
Viaarxiv icon