Alert button

"speech": models, code, and papers
Alert button

Learning Dependencies of Discrete Speech Representations with Neural Hidden Markov Models

Add code
Bookmark button
Alert button
Oct 29, 2022
Sung-Lin Yeh, Hao Tang

Figure 1 for Learning Dependencies of Discrete Speech Representations with Neural Hidden Markov Models
Figure 2 for Learning Dependencies of Discrete Speech Representations with Neural Hidden Markov Models
Figure 3 for Learning Dependencies of Discrete Speech Representations with Neural Hidden Markov Models
Figure 4 for Learning Dependencies of Discrete Speech Representations with Neural Hidden Markov Models
Viaarxiv icon

LibriS2S: A German-English Speech-to-Speech Translation Corpus

Add code
Bookmark button
Alert button
Apr 22, 2022
Pedro Jeuris, Jan Niehues

Figure 1 for LibriS2S: A German-English Speech-to-Speech Translation Corpus
Figure 2 for LibriS2S: A German-English Speech-to-Speech Translation Corpus
Figure 3 for LibriS2S: A German-English Speech-to-Speech Translation Corpus
Figure 4 for LibriS2S: A German-English Speech-to-Speech Translation Corpus
Viaarxiv icon

TorchDIVA: An Extensible Computational Model of Speech Production built on an Open-Source Machine Learning Library

Add code
Bookmark button
Alert button
Oct 17, 2022
Sean Kinahan, Julie Liss, Visar Berisha

Figure 1 for TorchDIVA: An Extensible Computational Model of Speech Production built on an Open-Source Machine Learning Library
Figure 2 for TorchDIVA: An Extensible Computational Model of Speech Production built on an Open-Source Machine Learning Library
Figure 3 for TorchDIVA: An Extensible Computational Model of Speech Production built on an Open-Source Machine Learning Library
Figure 4 for TorchDIVA: An Extensible Computational Model of Speech Production built on an Open-Source Machine Learning Library
Viaarxiv icon

Multilingual analysis of intelligibility classification using English, Korean, and Tamil dysarthric speech datasets

Sep 27, 2022
Eun Jung Yeo, Sunhee Kim, Minhwa Chung

Figure 1 for Multilingual analysis of intelligibility classification using English, Korean, and Tamil dysarthric speech datasets
Figure 2 for Multilingual analysis of intelligibility classification using English, Korean, and Tamil dysarthric speech datasets
Viaarxiv icon

Multi-Speaker Multi-Style Speech Synthesis with Timbre and Style Disentanglement

Add code
Bookmark button
Alert button
Nov 02, 2022
Wei Song, Yanghao Yue, Ya-jie Zhang, Zhengchen Zhang, Youzheng Wu, Xiaodong He

Figure 1 for Multi-Speaker Multi-Style Speech Synthesis with Timbre and Style Disentanglement
Figure 2 for Multi-Speaker Multi-Style Speech Synthesis with Timbre and Style Disentanglement
Figure 3 for Multi-Speaker Multi-Style Speech Synthesis with Timbre and Style Disentanglement
Figure 4 for Multi-Speaker Multi-Style Speech Synthesis with Timbre and Style Disentanglement
Viaarxiv icon

Advancing Stuttering Detection via Data Augmentation, Class-Balanced Loss and Multi-Contextual Deep Learning

Add code
Bookmark button
Alert button
Feb 21, 2023
Shakeel A. Sheikh, Md Sahidullah, Fabrice Hirsch, Slim Ouni

Figure 1 for Advancing Stuttering Detection via Data Augmentation, Class-Balanced Loss and Multi-Contextual Deep Learning
Figure 2 for Advancing Stuttering Detection via Data Augmentation, Class-Balanced Loss and Multi-Contextual Deep Learning
Figure 3 for Advancing Stuttering Detection via Data Augmentation, Class-Balanced Loss and Multi-Contextual Deep Learning
Figure 4 for Advancing Stuttering Detection via Data Augmentation, Class-Balanced Loss and Multi-Contextual Deep Learning
Viaarxiv icon

EBEN: Extreme bandwidth extension network applied to speech signals captured with noise-resilient microphones

Add code
Bookmark button
Alert button
Oct 25, 2022
Julien Hauret, Thomas Joubaud, Véronique Zimpfer, Éric Bavu

Figure 1 for EBEN: Extreme bandwidth extension network applied to speech signals captured with noise-resilient microphones
Figure 2 for EBEN: Extreme bandwidth extension network applied to speech signals captured with noise-resilient microphones
Figure 3 for EBEN: Extreme bandwidth extension network applied to speech signals captured with noise-resilient microphones
Figure 4 for EBEN: Extreme bandwidth extension network applied to speech signals captured with noise-resilient microphones
Viaarxiv icon

Decoding speech from non-invasive brain recordings

Add code
Bookmark button
Alert button
Aug 25, 2022
Alexandre Défossez, Charlotte Caucheteux, Jérémy Rapin, Ori Kabeli, Jean-Rémi King

Figure 1 for Decoding speech from non-invasive brain recordings
Figure 2 for Decoding speech from non-invasive brain recordings
Figure 3 for Decoding speech from non-invasive brain recordings
Figure 4 for Decoding speech from non-invasive brain recordings
Viaarxiv icon

An ASR-free Fluency Scoring Approach with Self-Supervised Learning

Feb 20, 2023
Wei Liu, Kaiqi Fu, Xiaohai Tian, Shuju Shi, Wei Li, Zejun Ma, Tan Lee

Figure 1 for An ASR-free Fluency Scoring Approach with Self-Supervised Learning
Figure 2 for An ASR-free Fluency Scoring Approach with Self-Supervised Learning
Figure 3 for An ASR-free Fluency Scoring Approach with Self-Supervised Learning
Figure 4 for An ASR-free Fluency Scoring Approach with Self-Supervised Learning
Viaarxiv icon

EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning

Add code
Bookmark button
Alert button
Jun 16, 2022
Li-Chin Chen, Po-Hsun Chen, Richard Tzong-Han Tsai, Yu Tsao

Figure 1 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Figure 2 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Figure 3 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Figure 4 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Viaarxiv icon