Alert button
Picture for Eugene Kharitonov

Eugene Kharitonov

Alert button

AudioPaLM: A Large Language Model That Can Speak and Listen

Add code
Bookmark button
Alert button
Jun 22, 2023
Paul K. Rubenstein, Chulayuth Asawaroengchai, Duc Dung Nguyen, Ankur Bapna, Zalán Borsos, Félix de Chaumont Quitry, Peter Chen, Dalia El Badawy, Wei Han, Eugene Kharitonov, Hannah Muckenhirn, Dirk Padfield, James Qin, Danny Rozenberg, Tara Sainath, Johan Schalkwyk, Matt Sharifi, Michelle Tadmor Ramanovich, Marco Tagliasacchi, Alexandru Tudor, Mihajlo Velimirović, Damien Vincent, Jiahui Yu, Yongqiang Wang, Vicky Zayats, Neil Zeghidour, Yu Zhang, Zhishuai Zhang, Lukas Zilka, Christian Frank

Figure 1 for AudioPaLM: A Large Language Model That Can Speak and Listen
Figure 2 for AudioPaLM: A Large Language Model That Can Speak and Listen
Figure 3 for AudioPaLM: A Large Language Model That Can Speak and Listen
Figure 4 for AudioPaLM: A Large Language Model That Can Speak and Listen
Viaarxiv icon

SoundStorm: Efficient Parallel Audio Generation

Add code
Bookmark button
Alert button
May 16, 2023
Zalán Borsos, Matt Sharifi, Damien Vincent, Eugene Kharitonov, Neil Zeghidour, Marco Tagliasacchi

Figure 1 for SoundStorm: Efficient Parallel Audio Generation
Figure 2 for SoundStorm: Efficient Parallel Audio Generation
Figure 3 for SoundStorm: Efficient Parallel Audio Generation
Figure 4 for SoundStorm: Efficient Parallel Audio Generation
Viaarxiv icon

Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision

Add code
Bookmark button
Alert button
Feb 07, 2023
Eugene Kharitonov, Damien Vincent, Zalán Borsos, Raphaël Marinier, Sertan Girgin, Olivier Pietquin, Matt Sharifi, Marco Tagliasacchi, Neil Zeghidour

Figure 1 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 2 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 3 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 4 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Viaarxiv icon

AudioLM: a Language Modeling Approach to Audio Generation

Add code
Bookmark button
Alert button
Sep 07, 2022
Zalán Borsos, Raphaël Marinier, Damien Vincent, Eugene Kharitonov, Olivier Pietquin, Matt Sharifi, Olivier Teboul, David Grangier, Marco Tagliasacchi, Neil Zeghidour

Figure 1 for AudioLM: a Language Modeling Approach to Audio Generation
Figure 2 for AudioLM: a Language Modeling Approach to Audio Generation
Figure 3 for AudioLM: a Language Modeling Approach to Audio Generation
Figure 4 for AudioLM: a Language Modeling Approach to Audio Generation
Viaarxiv icon

Generative Spoken Dialogue Language Modeling

Add code
Bookmark button
Alert button
Mar 30, 2022
Tu Anh Nguyen, Eugene Kharitonov, Jade Copet, Yossi Adi, Wei-Ning Hsu, Ali Elkahky, Paden Tomasello, Robin Algayres, Benoit Sagot, Abdelrahman Mohamed, Emmanuel Dupoux

Figure 1 for Generative Spoken Dialogue Language Modeling
Figure 2 for Generative Spoken Dialogue Language Modeling
Figure 3 for Generative Spoken Dialogue Language Modeling
Figure 4 for Generative Spoken Dialogue Language Modeling
Viaarxiv icon

textless-lib: a Library for Textless Spoken Language Processing

Add code
Bookmark button
Alert button
Feb 15, 2022
Eugene Kharitonov, Jade Copet, Kushal Lakhotia, Tu Anh Nguyen, Paden Tomasello, Ann Lee, Ali Elkahky, Wei-Ning Hsu, Abdelrahman Mohamed, Emmanuel Dupoux, Yossi Adi

Figure 1 for textless-lib: a Library for Textless Spoken Language Processing
Figure 2 for textless-lib: a Library for Textless Spoken Language Processing
Figure 3 for textless-lib: a Library for Textless Spoken Language Processing
Figure 4 for textless-lib: a Library for Textless Spoken Language Processing
Viaarxiv icon

Textless Speech Emotion Conversion using Decomposed and Discrete Representations

Add code
Bookmark button
Alert button
Nov 14, 2021
Felix Kreuk, Adam Polyak, Jade Copet, Eugene Kharitonov, Tu-Anh Nguyen, Morgane Rivière, Wei-Ning Hsu, Abdelrahman Mohamed, Emmanuel Dupoux, Yossi Adi

Figure 1 for Textless Speech Emotion Conversion using Decomposed and Discrete Representations
Figure 2 for Textless Speech Emotion Conversion using Decomposed and Discrete Representations
Figure 3 for Textless Speech Emotion Conversion using Decomposed and Discrete Representations
Figure 4 for Textless Speech Emotion Conversion using Decomposed and Discrete Representations
Viaarxiv icon

How BPE Affects Memorization in Transformers

Add code
Bookmark button
Alert button
Oct 06, 2021
Eugene Kharitonov, Marco Baroni, Dieuwke Hupkes

Figure 1 for How BPE Affects Memorization in Transformers
Figure 2 for How BPE Affects Memorization in Transformers
Figure 3 for How BPE Affects Memorization in Transformers
Figure 4 for How BPE Affects Memorization in Transformers
Viaarxiv icon

Text-Free Prosody-Aware Generative Spoken Language Modeling

Add code
Bookmark button
Alert button
Sep 07, 2021
Eugene Kharitonov, Ann Lee, Adam Polyak, Yossi Adi, Jade Copet, Kushal Lakhotia, Tu-Anh Nguyen, Morgane Rivière, Abdelrahman Mohamed, Emmanuel Dupoux, Wei-Ning Hsu

Figure 1 for Text-Free Prosody-Aware Generative Spoken Language Modeling
Figure 2 for Text-Free Prosody-Aware Generative Spoken Language Modeling
Figure 3 for Text-Free Prosody-Aware Generative Spoken Language Modeling
Figure 4 for Text-Free Prosody-Aware Generative Spoken Language Modeling
Viaarxiv icon

Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN

Add code
Bookmark button
Alert button
Jul 03, 2021
Rahma Chaabouni, Roberto Dessì, Eugene Kharitonov

Figure 1 for Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN
Figure 2 for Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN
Figure 3 for Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN
Figure 4 for Can Transformers Jump Around Right in Natural Language? Assessing Performance Transfer from SCAN
Viaarxiv icon