Alert button
Picture for Neil Zeghidour

Neil Zeghidour

Alert button

MAD Speech: Measures of Acoustic Diversity of Speech

Add code
Bookmark button
Alert button
Apr 16, 2024
Matthieu Futeral, Andrea Agostinelli, Marco Tagliasacchi, Neil Zeghidour, Eugene Kharitonov

Viaarxiv icon

MusicRL: Aligning Music Generation to Human Preferences

Add code
Bookmark button
Alert button
Feb 06, 2024
Geoffrey Cideron, Sertan Girgin, Mauro Verzetti, Damien Vincent, Matej Kastelic, Zalán Borsos, Brian McWilliams, Victor Ungureanu, Olivier Bachem, Olivier Pietquin, Matthieu Geist, Léonard Hussenot, Neil Zeghidour, Andrea Agostinelli

Viaarxiv icon

TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition

Add code
Bookmark button
Alert button
Aug 21, 2023
Hakan Erdogan, Scott Wisdom, Xuankai Chang, Zalán Borsos, Marco Tagliasacchi, Neil Zeghidour, John R. Hershey

Figure 1 for TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition
Figure 2 for TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition
Figure 3 for TokenSplit: Using Discrete Speech Representations for Direct, Refined, and Transcript-Conditioned Speech Separation and Recognition
Viaarxiv icon

AudioPaLM: A Large Language Model That Can Speak and Listen

Add code
Bookmark button
Alert button
Jun 22, 2023
Paul K. Rubenstein, Chulayuth Asawaroengchai, Duc Dung Nguyen, Ankur Bapna, Zalán Borsos, Félix de Chaumont Quitry, Peter Chen, Dalia El Badawy, Wei Han, Eugene Kharitonov, Hannah Muckenhirn, Dirk Padfield, James Qin, Danny Rozenberg, Tara Sainath, Johan Schalkwyk, Matt Sharifi, Michelle Tadmor Ramanovich, Marco Tagliasacchi, Alexandru Tudor, Mihajlo Velimirović, Damien Vincent, Jiahui Yu, Yongqiang Wang, Vicky Zayats, Neil Zeghidour, Yu Zhang, Zhishuai Zhang, Lukas Zilka, Christian Frank

Figure 1 for AudioPaLM: A Large Language Model That Can Speak and Listen
Figure 2 for AudioPaLM: A Large Language Model That Can Speak and Listen
Figure 3 for AudioPaLM: A Large Language Model That Can Speak and Listen
Figure 4 for AudioPaLM: A Large Language Model That Can Speak and Listen
Viaarxiv icon

SoundStorm: Efficient Parallel Audio Generation

Add code
Bookmark button
Alert button
May 16, 2023
Zalán Borsos, Matt Sharifi, Damien Vincent, Eugene Kharitonov, Neil Zeghidour, Marco Tagliasacchi

Figure 1 for SoundStorm: Efficient Parallel Audio Generation
Figure 2 for SoundStorm: Efficient Parallel Audio Generation
Figure 3 for SoundStorm: Efficient Parallel Audio Generation
Figure 4 for SoundStorm: Efficient Parallel Audio Generation
Viaarxiv icon

LMCodec: A Low Bitrate Speech Codec With Causal Transformer Models

Add code
Bookmark button
Alert button
Mar 23, 2023
Teerapat Jenrungrot, Michael Chinen, W. Bastiaan Kleijn, Jan Skoglund, Zalán Borsos, Neil Zeghidour, Marco Tagliasacchi

Figure 1 for LMCodec: A Low Bitrate Speech Codec With Causal Transformer Models
Figure 2 for LMCodec: A Low Bitrate Speech Codec With Causal Transformer Models
Figure 3 for LMCodec: A Low Bitrate Speech Codec With Causal Transformer Models
Figure 4 for LMCodec: A Low Bitrate Speech Codec With Causal Transformer Models
Viaarxiv icon

Speech Intelligibility Classifiers from 550k Disordered Speech Samples

Add code
Bookmark button
Alert button
Mar 15, 2023
Subhashini Venugopalan, Jimmy Tobin, Samuel J. Yang, Katie Seaver, Richard J. N. Cave, Pan-Pan Jiang, Neil Zeghidour, Rus Heywood, Jordan Green, Michael P. Brenner

Figure 1 for Speech Intelligibility Classifiers from 550k Disordered Speech Samples
Figure 2 for Speech Intelligibility Classifiers from 550k Disordered Speech Samples
Figure 3 for Speech Intelligibility Classifiers from 550k Disordered Speech Samples
Figure 4 for Speech Intelligibility Classifiers from 550k Disordered Speech Samples
Viaarxiv icon

DNArch: Learning Convolutional Neural Architectures by Backpropagation

Add code
Bookmark button
Alert button
Feb 10, 2023
David W. Romero, Neil Zeghidour

Figure 1 for DNArch: Learning Convolutional Neural Architectures by Backpropagation
Figure 2 for DNArch: Learning Convolutional Neural Architectures by Backpropagation
Figure 3 for DNArch: Learning Convolutional Neural Architectures by Backpropagation
Figure 4 for DNArch: Learning Convolutional Neural Architectures by Backpropagation
Viaarxiv icon

Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision

Add code
Bookmark button
Alert button
Feb 07, 2023
Eugene Kharitonov, Damien Vincent, Zalán Borsos, Raphaël Marinier, Sertan Girgin, Olivier Pietquin, Matt Sharifi, Marco Tagliasacchi, Neil Zeghidour

Figure 1 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 2 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 3 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Figure 4 for Speak, Read and Prompt: High-Fidelity Text-to-Speech with Minimal Supervision
Viaarxiv icon