Picture for Matthieu Geist

Matthieu Geist

INRIA Lorraine - LORIA

The Curious Price of Distributional Robustness in Reinforcement Learning with a Generative Model

Add code
May 26, 2023
Viaarxiv icon

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice

Add code
May 22, 2023
Viaarxiv icon

Get Back Here: Robust Imitation by Return-to-Distribution Planning

Add code
May 02, 2023
Viaarxiv icon

Twice Regularized Markov Decision Processes: The Equivalence between Robustness and Regularization

Add code
Mar 12, 2023
Figure 1 for Twice Regularized Markov Decision Processes: The Equivalence between Robustness and Regularization
Figure 2 for Twice Regularized Markov Decision Processes: The Equivalence between Robustness and Regularization
Figure 3 for Twice Regularized Markov Decision Processes: The Equivalence between Robustness and Regularization
Figure 4 for Twice Regularized Markov Decision Processes: The Equivalence between Robustness and Regularization
Viaarxiv icon

Towards Minimax Optimality of Model-based Robust Reinforcement Learning

Add code
Feb 10, 2023
Figure 1 for Towards Minimax Optimality of Model-based Robust Reinforcement Learning
Figure 2 for Towards Minimax Optimality of Model-based Robust Reinforcement Learning
Viaarxiv icon

Policy Gradient for s-Rectangular Robust Markov Decision Processes

Add code
Jan 31, 2023
Figure 1 for Policy Gradient for s-Rectangular Robust Markov Decision Processes
Figure 2 for Policy Gradient for s-Rectangular Robust Markov Decision Processes
Figure 3 for Policy Gradient for s-Rectangular Robust Markov Decision Processes
Figure 4 for Policy Gradient for s-Rectangular Robust Markov Decision Processes
Viaarxiv icon

Extreme Q-Learning: MaxEnt RL without Entropy

Add code
Jan 05, 2023
Viaarxiv icon

Policy Mirror Ascent for Efficient and Independent Learning in Mean Field Games

Add code
Dec 29, 2022
Viaarxiv icon

C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining

Add code
Nov 07, 2022
Figure 1 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Figure 2 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Figure 3 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Figure 4 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Viaarxiv icon

Learning Correlated Equilibria in Mean-Field Games

Add code
Aug 22, 2022
Figure 1 for Learning Correlated Equilibria in Mean-Field Games
Figure 2 for Learning Correlated Equilibria in Mean-Field Games
Figure 3 for Learning Correlated Equilibria in Mean-Field Games
Figure 4 for Learning Correlated Equilibria in Mean-Field Games
Viaarxiv icon