Alert button
Picture for Olivier Pietquin

Olivier Pietquin

Alert button

Language Evolution with Deep Learning

Add code
Bookmark button
Alert button
Mar 18, 2024
Mathieu Rita, Paul Michel, Rahma Chaabouni, Olivier Pietquin, Emmanuel Dupoux, Florian Strub

Figure 1 for Language Evolution with Deep Learning
Figure 2 for Language Evolution with Deep Learning
Figure 3 for Language Evolution with Deep Learning
Figure 4 for Language Evolution with Deep Learning
Viaarxiv icon

Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning

Add code
Bookmark button
Alert button
Mar 06, 2024
Zida Wu, Mathieu Lauriere, Samuel Jia Cong Chua, Matthieu Geist, Olivier Pietquin, Ankur Mehta

Figure 1 for Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning
Figure 2 for Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning
Figure 3 for Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning
Figure 4 for Population-aware Online Mirror Descent for Mean-Field Games by Deep Reinforcement Learning
Viaarxiv icon

Back to Basics: Revisiting REINFORCE Style Optimization for Learning from Human Feedback in LLMs

Add code
Bookmark button
Alert button
Feb 26, 2024
Arash Ahmadian, Chris Cremer, Matthias Gallé, Marzieh Fadaee, Julia Kreutzer, Olivier Pietquin, Ahmet Üstün, Sara Hooker

Viaarxiv icon

MusicRL: Aligning Music Generation to Human Preferences

Add code
Bookmark button
Alert button
Feb 06, 2024
Geoffrey Cideron, Sertan Girgin, Mauro Verzetti, Damien Vincent, Matej Kastelic, Zalán Borsos, Brian McWilliams, Victor Ungureanu, Olivier Bachem, Olivier Pietquin, Matthieu Geist, Léonard Hussenot, Neil Zeghidour, Andrea Agostinelli

Viaarxiv icon

Learning Discrete-Time Major-Minor Mean Field Games

Add code
Bookmark button
Alert button
Dec 17, 2023
Kai Cui, Gökçe Dayanıklı, Mathieu Laurière, Matthieu Geist, Olivier Pietquin, Heinz Koeppl

Viaarxiv icon

On Imitation in Mean-field Games

Add code
Bookmark button
Alert button
Jun 26, 2023
Giorgia Ramponi, Pavel Kolev, Olivier Pietquin, Niao He, Mathieu Laurière, Matthieu Geist

Figure 1 for On Imitation in Mean-field Games
Figure 2 for On Imitation in Mean-field Games
Viaarxiv icon

Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback

Add code
Bookmark button
Alert button
May 31, 2023
Paul Roit, Johan Ferret, Lior Shani, Roee Aharoni, Geoffrey Cideron, Robert Dadashi, Matthieu Geist, Sertan Girgin, Léonard Hussenot, Orgad Keller, Nikola Momchev, Sabela Ramos, Piotr Stanczyk, Nino Vieillard, Olivier Bachem, Gal Elidan, Avinatan Hassidim, Olivier Pietquin, Idan Szpektor

Figure 1 for Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback
Figure 2 for Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback
Figure 3 for Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback
Figure 4 for Factually Consistent Summarization via Reinforcement Learning with Textual Entailment Feedback
Viaarxiv icon

Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice

Add code
Bookmark button
Alert button
May 22, 2023
Toshinori Kitamura, Tadashi Kozuno, Yunhao Tang, Nino Vieillard, Michal Valko, Wenhao Yang, Jincheng Mei, Pierre Ménard, Mohammad Gheshlaghi Azar, Rémi Munos, Olivier Pietquin, Matthieu Geist, Csaba Szepesvári, Wataru Kumagai, Yutaka Matsuo

Figure 1 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 2 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 3 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Figure 4 for Regularization and Variance-Weighted Regression Achieves Minimax Optimality in Linear MDPs: Theory and Practice
Viaarxiv icon

Get Back Here: Robust Imitation by Return-to-Distribution Planning

Add code
Bookmark button
Alert button
May 02, 2023
Geoffrey Cideron, Baruch Tabanpour, Sebastian Curi, Sertan Girgin, Leonard Hussenot, Gabriel Dulac-Arnold, Matthieu Geist, Olivier Pietquin, Robert Dadashi

Figure 1 for Get Back Here: Robust Imitation by Return-to-Distribution Planning
Figure 2 for Get Back Here: Robust Imitation by Return-to-Distribution Planning
Figure 3 for Get Back Here: Robust Imitation by Return-to-Distribution Planning
Figure 4 for Get Back Here: Robust Imitation by Return-to-Distribution Planning
Viaarxiv icon

Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision

Add code
Bookmark button
Alert button
Feb 07, 2023
Eugene Kharitonov, Damien Vincent, Zalán Borsos, Raphaël Marinier, Sertan Girgin, Olivier Pietquin, Matt Sharifi, Marco Tagliasacchi, Neil Zeghidour

Figure 1 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 2 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 3 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 4 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Viaarxiv icon