Alert button
Picture for Matthieu Geist

Matthieu Geist

Alert button

Policy Mirror Ascent for Efficient and Independent Learning in Mean Field Games

Add code
Bookmark button
Alert button
Dec 29, 2022
Batuhan Yardim, Semih Cayci, Matthieu Geist, Niao He

Figure 1 for Policy Mirror Ascent for Efficient and Independent Learning in Mean Field Games
Viaarxiv icon

C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining

Add code
Bookmark button
Alert button
Nov 07, 2022
Alexis Jacq, Manu Orsini, Gabriel Dulac-Arnold, Olivier Pietquin, Matthieu Geist, Olivier Bachem

Figure 1 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Figure 2 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Figure 3 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Figure 4 for C3PO: Learning to Achieve Arbitrary Goals via Massively Entropic Pretraining
Viaarxiv icon

Learning Correlated Equilibria in Mean-Field Games

Add code
Bookmark button
Alert button
Aug 22, 2022
Paul Muller, Romuald Elie, Mark Rowland, Mathieu Lauriere, Julien Perolat, Sarah Perrin, Matthieu Geist, Georgios Piliouras, Olivier Pietquin, Karl Tuyls

Figure 1 for Learning Correlated Equilibria in Mean-Field Games
Figure 2 for Learning Correlated Equilibria in Mean-Field Games
Figure 3 for Learning Correlated Equilibria in Mean-Field Games
Figure 4 for Learning Correlated Equilibria in Mean-Field Games
Viaarxiv icon

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
Bookmark button
Alert button
May 27, 2022
Tadashi Kozuno, Wenhao Yang, Nino Vieillard, Toshinori Kitamura, Yunhao Tang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Michal Valko, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári

Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

Learning Mean Field Games: A Survey

Add code
Bookmark button
Alert button
May 25, 2022
Mathieu Laurière, Sarah Perrin, Matthieu Geist, Olivier Pietquin

Figure 1 for Learning Mean Field Games: A Survey
Figure 2 for Learning Mean Field Games: A Survey
Figure 3 for Learning Mean Field Games: A Survey
Figure 4 for Learning Mean Field Games: A Survey
Viaarxiv icon

Learning Energy Networks with Generalized Fenchel-Young Losses

Add code
Bookmark button
Alert button
May 19, 2022
Mathieu Blondel, Felipe Llinares-López, Robert Dadashi, Léonard Hussenot, Matthieu Geist

Figure 1 for Learning Energy Networks with Generalized Fenchel-Young Losses
Figure 2 for Learning Energy Networks with Generalized Fenchel-Young Losses
Figure 3 for Learning Energy Networks with Generalized Fenchel-Young Losses
Figure 4 for Learning Energy Networks with Generalized Fenchel-Young Losses
Viaarxiv icon

Scalable Deep Reinforcement Learning Algorithms for Mean Field Games

Add code
Bookmark button
Alert button
Mar 22, 2022
Mathieu Laurière, Sarah Perrin, Sertan Girgin, Paul Muller, Ayush Jain, Theophile Cabannes, Georgios Piliouras, Julien Pérolat, Romuald Élie, Olivier Pietquin, Matthieu Geist

Figure 1 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Figure 2 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Figure 3 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Figure 4 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Viaarxiv icon

Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act

Add code
Bookmark button
Alert button
Mar 16, 2022
Alexis Jacq, Johan Ferret, Olivier Pietquin, Matthieu Geist

Figure 1 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Figure 2 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Figure 3 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Figure 4 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Viaarxiv icon

Continuous Control with Action Quantization from Demonstrations

Add code
Bookmark button
Alert button
Oct 19, 2021
Robert Dadashi, Léonard Hussenot, Damien Vincent, Sertan Girgin, Anton Raichuk, Matthieu Geist, Olivier Pietquin

Figure 1 for Continuous Control with Action Quantization from Demonstrations
Figure 2 for Continuous Control with Action Quantization from Demonstrations
Figure 3 for Continuous Control with Action Quantization from Demonstrations
Figure 4 for Continuous Control with Action Quantization from Demonstrations
Viaarxiv icon

Twice regularized MDPs and the equivalence between robustness and regularization

Add code
Bookmark button
Alert button
Oct 12, 2021
Esther Derman, Matthieu Geist, Shie Mannor

Figure 1 for Twice regularized MDPs and the equivalence between robustness and regularization
Figure 2 for Twice regularized MDPs and the equivalence between robustness and regularization
Figure 3 for Twice regularized MDPs and the equivalence between robustness and regularization
Figure 4 for Twice regularized MDPs and the equivalence between robustness and regularization
Viaarxiv icon