Alert button
Picture for Olivier Bachem

Olivier Bachem

Alert button

Google Research

Braxlines: Fast and Interactive Toolkit for RL-driven Behavior Engineering beyond Reward Maximization

Add code
Bookmark button
Alert button
Oct 10, 2021
Shixiang Shane Gu, Manfred Diaz, Daniel C. Freeman, Hiroki Furuta, Seyed Kamyar Seyed Ghasemipour, Anton Raichuk, Byron David, Erik Frey, Erwin Coumans, Olivier Bachem

Figure 1 for Braxlines: Fast and Interactive Toolkit for RL-driven Behavior Engineering beyond Reward Maximization
Figure 2 for Braxlines: Fast and Interactive Toolkit for RL-driven Behavior Engineering beyond Reward Maximization
Figure 3 for Braxlines: Fast and Interactive Toolkit for RL-driven Behavior Engineering beyond Reward Maximization
Figure 4 for Braxlines: Fast and Interactive Toolkit for RL-driven Behavior Engineering beyond Reward Maximization
Viaarxiv icon

A functional mirror ascent view of policy gradient methods with function approximation

Add code
Bookmark button
Alert button
Aug 12, 2021
Sharan Vaswani, Olivier Bachem, Simone Totaro, Robert Mueller, Matthieu Geist, Marlos C. Machado, Pablo Samuel Castro, Nicolas Le Roux

Figure 1 for A functional mirror ascent view of policy gradient methods with function approximation
Figure 2 for A functional mirror ascent view of policy gradient methods with function approximation
Figure 3 for A functional mirror ascent view of policy gradient methods with function approximation
Figure 4 for A functional mirror ascent view of policy gradient methods with function approximation
Viaarxiv icon

Representation Learning for Out-Of-Distribution Generalization in Reinforcement Learning

Add code
Bookmark button
Alert button
Jul 12, 2021
Andrea Dittadi, Frederik Träuble, Manuel Wüthrich, Felix Widmaier, Peter Gehler, Ole Winther, Francesco Locatello, Olivier Bachem, Bernhard Schölkopf, Stefan Bauer

Figure 1 for Representation Learning for Out-Of-Distribution Generalization in Reinforcement Learning
Figure 2 for Representation Learning for Out-Of-Distribution Generalization in Reinforcement Learning
Figure 3 for Representation Learning for Out-Of-Distribution Generalization in Reinforcement Learning
Figure 4 for Representation Learning for Out-Of-Distribution Generalization in Reinforcement Learning
Viaarxiv icon

Brax -- A Differentiable Physics Engine for Large Scale Rigid Body Simulation

Add code
Bookmark button
Alert button
Jun 24, 2021
C. Daniel Freeman, Erik Frey, Anton Raichuk, Sertan Girgin, Igor Mordatch, Olivier Bachem

Figure 1 for Brax -- A Differentiable Physics Engine for Large Scale Rigid Body Simulation
Figure 2 for Brax -- A Differentiable Physics Engine for Large Scale Rigid Body Simulation
Figure 3 for Brax -- A Differentiable Physics Engine for Large Scale Rigid Body Simulation
Figure 4 for Brax -- A Differentiable Physics Engine for Large Scale Rigid Body Simulation
Viaarxiv icon

Offline Reinforcement Learning as Anti-Exploration

Add code
Bookmark button
Alert button
Jun 11, 2021
Shideh Rezaeifar, Robert Dadashi, Nino Vieillard, Léonard Hussenot, Olivier Bachem, Olivier Pietquin, Matthieu Geist

Figure 1 for Offline Reinforcement Learning as Anti-Exploration
Figure 2 for Offline Reinforcement Learning as Anti-Exploration
Figure 3 for Offline Reinforcement Learning as Anti-Exploration
Figure 4 for Offline Reinforcement Learning as Anti-Exploration
Viaarxiv icon

Concave Utility Reinforcement Learning: the Mean-field Game viewpoint

Add code
Bookmark button
Alert button
Jun 09, 2021
Matthieu Geist, Julien Pérolat, Mathieu Laurière, Romuald Elie, Sarah Perrin, Olivier Bachem, Rémi Munos, Olivier Pietquin

Figure 1 for Concave Utility Reinforcement Learning: the Mean-field Game viewpoint
Figure 2 for Concave Utility Reinforcement Learning: the Mean-field Game viewpoint
Figure 3 for Concave Utility Reinforcement Learning: the Mean-field Game viewpoint
Figure 4 for Concave Utility Reinforcement Learning: the Mean-field Game viewpoint
Viaarxiv icon

What Matters for Adversarial Imitation Learning?

Add code
Bookmark button
Alert button
Jun 01, 2021
Manu Orsini, Anton Raichuk, Léonard Hussenot, Damien Vincent, Robert Dadashi, Sertan Girgin, Matthieu Geist, Olivier Bachem, Olivier Pietquin, Marcin Andrychowicz

Figure 1 for What Matters for Adversarial Imitation Learning?
Figure 2 for What Matters for Adversarial Imitation Learning?
Figure 3 for What Matters for Adversarial Imitation Learning?
Figure 4 for What Matters for Adversarial Imitation Learning?
Viaarxiv icon

Hyperparameter Selection for Imitation Learning

Add code
Bookmark button
Alert button
May 25, 2021
Leonard Hussenot, Marcin Andrychowicz, Damien Vincent, Robert Dadashi, Anton Raichuk, Lukasz Stafiniak, Sertan Girgin, Raphael Marinier, Nikola Momchev, Sabela Ramos, Manu Orsini, Olivier Bachem, Matthieu Geist, Olivier Pietquin

Figure 1 for Hyperparameter Selection for Imitation Learning
Figure 2 for Hyperparameter Selection for Imitation Learning
Figure 3 for Hyperparameter Selection for Imitation Learning
Figure 4 for Hyperparameter Selection for Imitation Learning
Viaarxiv icon

Scaling Hierarchical Agglomerative Clustering to Billion-sized Datasets

Add code
Bookmark button
Alert button
May 25, 2021
Baris Sumengen, Anand Rajagopalan, Gui Citovsky, David Simcha, Olivier Bachem, Pradipta Mitra, Sam Blasiak, Mason Liang, Sanjiv Kumar

Figure 1 for Scaling Hierarchical Agglomerative Clustering to Billion-sized Datasets
Figure 2 for Scaling Hierarchical Agglomerative Clustering to Billion-sized Datasets
Figure 3 for Scaling Hierarchical Agglomerative Clustering to Billion-sized Datasets
Figure 4 for Scaling Hierarchical Agglomerative Clustering to Billion-sized Datasets
Viaarxiv icon