Picture for Matthieu Geist

Matthieu Geist

INRIA Lorraine - LORIA

KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal

Add code
May 27, 2022
Figure 1 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 2 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 3 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Figure 4 for KL-Entropy-Regularized RL with a Generative Model is Minimax Optimal
Viaarxiv icon

Learning Mean Field Games: A Survey

Add code
May 25, 2022
Figure 1 for Learning Mean Field Games: A Survey
Figure 2 for Learning Mean Field Games: A Survey
Figure 3 for Learning Mean Field Games: A Survey
Figure 4 for Learning Mean Field Games: A Survey
Viaarxiv icon

Learning Energy Networks with Generalized Fenchel-Young Losses

Add code
May 19, 2022
Figure 1 for Learning Energy Networks with Generalized Fenchel-Young Losses
Figure 2 for Learning Energy Networks with Generalized Fenchel-Young Losses
Figure 3 for Learning Energy Networks with Generalized Fenchel-Young Losses
Figure 4 for Learning Energy Networks with Generalized Fenchel-Young Losses
Viaarxiv icon

Scalable Deep Reinforcement Learning Algorithms for Mean Field Games

Add code
Mar 22, 2022
Figure 1 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Figure 2 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Figure 3 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Figure 4 for Scalable Deep Reinforcement Learning Algorithms for Mean Field Games
Viaarxiv icon

Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act

Add code
Mar 16, 2022
Figure 1 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Figure 2 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Figure 3 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Figure 4 for Lazy-MDPs: Towards Interpretable Reinforcement Learning by Learning When to Act
Viaarxiv icon

Continuous Control with Action Quantization from Demonstrations

Add code
Oct 19, 2021
Figure 1 for Continuous Control with Action Quantization from Demonstrations
Figure 2 for Continuous Control with Action Quantization from Demonstrations
Figure 3 for Continuous Control with Action Quantization from Demonstrations
Figure 4 for Continuous Control with Action Quantization from Demonstrations
Viaarxiv icon

Twice regularized MDPs and the equivalence between robustness and regularization

Add code
Oct 12, 2021
Figure 1 for Twice regularized MDPs and the equivalence between robustness and regularization
Figure 2 for Twice regularized MDPs and the equivalence between robustness and regularization
Figure 3 for Twice regularized MDPs and the equivalence between robustness and regularization
Figure 4 for Twice regularized MDPs and the equivalence between robustness and regularization
Viaarxiv icon

Large Batch Experience Replay

Add code
Oct 04, 2021
Figure 1 for Large Batch Experience Replay
Figure 2 for Large Batch Experience Replay
Figure 3 for Large Batch Experience Replay
Figure 4 for Large Batch Experience Replay
Viaarxiv icon

Generalization in Mean Field Games by Learning Master Policies

Add code
Sep 20, 2021
Figure 1 for Generalization in Mean Field Games by Learning Master Policies
Figure 2 for Generalization in Mean Field Games by Learning Master Policies
Figure 3 for Generalization in Mean Field Games by Learning Master Policies
Figure 4 for Generalization in Mean Field Games by Learning Master Policies
Viaarxiv icon

Implicitly Regularized RL with Implicit Q-Values

Add code
Aug 16, 2021
Figure 1 for Implicitly Regularized RL with Implicit Q-Values
Figure 2 for Implicitly Regularized RL with Implicit Q-Values
Figure 3 for Implicitly Regularized RL with Implicit Q-Values
Figure 4 for Implicitly Regularized RL with Implicit Q-Values
Viaarxiv icon