Alert button

"speech": models, code, and papers
Alert button

FastPitch: Parallel Text-to-speech with Pitch Prediction

Add code
Bookmark button
Alert button
Jun 11, 2020
Adrian Łańcucki

Figure 1 for FastPitch: Parallel Text-to-speech with Pitch Prediction
Figure 2 for FastPitch: Parallel Text-to-speech with Pitch Prediction
Figure 3 for FastPitch: Parallel Text-to-speech with Pitch Prediction
Figure 4 for FastPitch: Parallel Text-to-speech with Pitch Prediction
Viaarxiv icon

A Spoken Drug Prescription Dataset in French for Spoken Language Understanding

Jul 17, 2022
Ali Can Kocabiyikoglu, François Portet, Prudence Gibert, Hervé Blanchon, Jean-Marc Babouchkine, Gaëtan Gavazzi

Figure 1 for A Spoken Drug Prescription Dataset in French for Spoken Language Understanding
Figure 2 for A Spoken Drug Prescription Dataset in French for Spoken Language Understanding
Figure 3 for A Spoken Drug Prescription Dataset in French for Spoken Language Understanding
Figure 4 for A Spoken Drug Prescription Dataset in French for Spoken Language Understanding
Viaarxiv icon

Assessing clinical utility of Machine Learning and Artificial Intelligence approaches to analyze speech recordings in Multiple Sclerosis: A Pilot Study

Sep 27, 2021
Emil Svoboda, Tomáš Bořil, Jan Rusz, Tereza Tykalová, Dana Horáková, Charles R. G. Guttman, Krastan B. Blagoev, Hiroto Hatabu, Vlad I. Valtchinov

Figure 1 for Assessing clinical utility of Machine Learning and Artificial Intelligence approaches to analyze speech recordings in Multiple Sclerosis: A Pilot Study
Figure 2 for Assessing clinical utility of Machine Learning and Artificial Intelligence approaches to analyze speech recordings in Multiple Sclerosis: A Pilot Study
Figure 3 for Assessing clinical utility of Machine Learning and Artificial Intelligence approaches to analyze speech recordings in Multiple Sclerosis: A Pilot Study
Figure 4 for Assessing clinical utility of Machine Learning and Artificial Intelligence approaches to analyze speech recordings in Multiple Sclerosis: A Pilot Study
Viaarxiv icon

Towards Unconstrained Audio Splicing Detection and Localization with Neural Networks

Jul 29, 2022
Denise Moussa, Germans Hirsch, Christian Riess

Figure 1 for Towards Unconstrained Audio Splicing Detection and Localization with Neural Networks
Figure 2 for Towards Unconstrained Audio Splicing Detection and Localization with Neural Networks
Figure 3 for Towards Unconstrained Audio Splicing Detection and Localization with Neural Networks
Figure 4 for Towards Unconstrained Audio Splicing Detection and Localization with Neural Networks
Viaarxiv icon

S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations

Add code
Bookmark button
Alert button
Oct 12, 2021
Wen-Chin Huang, Shu-Wen Yang, Tomoki Hayashi, Hung-Yi Lee, Shinji Watanabe, Tomoki Toda

Figure 1 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Figure 2 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Figure 3 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Figure 4 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Viaarxiv icon

Neural Speech Separation Using Spatially Distributed Microphones

Apr 28, 2020
Dongmei Wang, Zhuo Chen, Takuya Yoshioka

Figure 1 for Neural Speech Separation Using Spatially Distributed Microphones
Figure 2 for Neural Speech Separation Using Spatially Distributed Microphones
Figure 3 for Neural Speech Separation Using Spatially Distributed Microphones
Viaarxiv icon

Multi-channel Multi-frame ADL-MVDR for Target Speech Separation

Add code
Bookmark button
Alert button
Dec 24, 2020
Zhuohuang Zhang, Yong Xu, Meng Yu, Shi-Xiong Zhang, Lianwu Chen, Donald S. Williamson, Dong Yu

Figure 1 for Multi-channel Multi-frame ADL-MVDR for Target Speech Separation
Figure 2 for Multi-channel Multi-frame ADL-MVDR for Target Speech Separation
Figure 3 for Multi-channel Multi-frame ADL-MVDR for Target Speech Separation
Figure 4 for Multi-channel Multi-frame ADL-MVDR for Target Speech Separation
Viaarxiv icon

Nonlinear prediction with neural nets in ADPCM

Mar 22, 2022
Marcos Faundez-Zanuy, Francesc Vallverdu, Enric Monte

Figure 1 for Nonlinear prediction with neural nets in ADPCM
Figure 2 for Nonlinear prediction with neural nets in ADPCM
Figure 3 for Nonlinear prediction with neural nets in ADPCM
Figure 4 for Nonlinear prediction with neural nets in ADPCM
Viaarxiv icon

End-to-End Zero-Shot Voice Style Transfer with Location-Variable Convolutions

Add code
Bookmark button
Alert button
May 19, 2022
Wonjune Kang, Deb Roy

Figure 1 for End-to-End Zero-Shot Voice Style Transfer with Location-Variable Convolutions
Figure 2 for End-to-End Zero-Shot Voice Style Transfer with Location-Variable Convolutions
Figure 3 for End-to-End Zero-Shot Voice Style Transfer with Location-Variable Convolutions
Figure 4 for End-to-End Zero-Shot Voice Style Transfer with Location-Variable Convolutions
Viaarxiv icon

Continuous Silent Speech Recognition using EEG

Feb 29, 2020
Gautam Krishna, Co Tran, Mason Carnahan, Ahmed Tewfik

Figure 1 for Continuous Silent Speech Recognition using EEG
Figure 2 for Continuous Silent Speech Recognition using EEG
Figure 3 for Continuous Silent Speech Recognition using EEG
Figure 4 for Continuous Silent Speech Recognition using EEG
Viaarxiv icon