Alert button

"speech": models, code, and papers
Alert button

Hypergraph based semi-supervised learning algorithms applied to speech recognition problem: a novel approach

Oct 28, 2018
Loc Hoang Tran, Trang Hoang, Bui Hoang Nam Huynh

Figure 1 for Hypergraph based semi-supervised learning algorithms applied to speech recognition problem: a novel approach
Figure 2 for Hypergraph based semi-supervised learning algorithms applied to speech recognition problem: a novel approach
Figure 3 for Hypergraph based semi-supervised learning algorithms applied to speech recognition problem: a novel approach
Viaarxiv icon

A variance modeling framework based on variational autoencoders for speech enhancement

Add code
Bookmark button
Alert button
Feb 05, 2019
Simon Leglaive, Laurent Girin, Radu Horaud

Figure 1 for A variance modeling framework based on variational autoencoders for speech enhancement
Figure 2 for A variance modeling framework based on variational autoencoders for speech enhancement
Figure 3 for A variance modeling framework based on variational autoencoders for speech enhancement
Viaarxiv icon

UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions

Add code
Bookmark button
Alert button
Jul 01, 2019
Aciel Eshky, Manuel Sam Ribeiro, Joanne Cleland, Korin Richmond, Zoe Roxburgh, James Scobbie, Alan Wrench

Figure 1 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Figure 2 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Figure 3 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Figure 4 for UltraSuite: A Repository of Ultrasound and Acoustic Data from Child Speech Therapy Sessions
Viaarxiv icon

Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning

Add code
Bookmark button
Alert button
Jul 09, 2019
Yu Zhang, Ron J. Weiss, Heiga Zen, Yonghui Wu, Zhifeng Chen, RJ Skerry-Ryan, Ye Jia, Andrew Rosenberg, Bhuvana Ramabhadran

Figure 1 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 2 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 3 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 4 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Viaarxiv icon

SD-QA: Spoken Dialectal Question Answering for the Real World

Add code
Bookmark button
Alert button
Sep 24, 2021
Fahim Faisal, Sharlina Keshava, Md Mahfuz ibn Alam, Antonios Anastasopoulos

Figure 1 for SD-QA: Spoken Dialectal Question Answering for the Real World
Figure 2 for SD-QA: Spoken Dialectal Question Answering for the Real World
Figure 3 for SD-QA: Spoken Dialectal Question Answering for the Real World
Figure 4 for SD-QA: Spoken Dialectal Question Answering for the Real World
Viaarxiv icon

Multi-scale Octave Convolutions for Robust Speech Recognition

Oct 31, 2019
Joanna Rownicka, Peter Bell, Steve Renals

Figure 1 for Multi-scale Octave Convolutions for Robust Speech Recognition
Figure 2 for Multi-scale Octave Convolutions for Robust Speech Recognition
Figure 3 for Multi-scale Octave Convolutions for Robust Speech Recognition
Figure 4 for Multi-scale Octave Convolutions for Robust Speech Recognition
Viaarxiv icon

On the Contributions of Visual and Textual Supervision in Low-resource Semantic Speech Retrieval

Apr 24, 2019
Ankita Pasad, Bowen Shi, Herman Kamper, Karen Livescu

Figure 1 for On the Contributions of Visual and Textual Supervision in Low-resource Semantic Speech Retrieval
Figure 2 for On the Contributions of Visual and Textual Supervision in Low-resource Semantic Speech Retrieval
Figure 3 for On the Contributions of Visual and Textual Supervision in Low-resource Semantic Speech Retrieval
Viaarxiv icon

Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition

Nov 09, 2019
Alex Bie, Bharat Venkitesh, Joao Monteiro, Md. Akmal Haidar, Mehdi Rezagholizadeh

Figure 1 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Figure 2 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Figure 3 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Figure 4 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Viaarxiv icon

Constructing interval variables via faceted Rasch measurement and multitask deep learning: a hate speech application

Add code
Bookmark button
Alert button
Sep 22, 2020
Chris J. Kennedy, Geoff Bacon, Alexander Sahn, Claudia von Vacano

Figure 1 for Constructing interval variables via faceted Rasch measurement and multitask deep learning: a hate speech application
Figure 2 for Constructing interval variables via faceted Rasch measurement and multitask deep learning: a hate speech application
Figure 3 for Constructing interval variables via faceted Rasch measurement and multitask deep learning: a hate speech application
Figure 4 for Constructing interval variables via faceted Rasch measurement and multitask deep learning: a hate speech application
Viaarxiv icon

Prosody-Aware Neural Machine Translation for Dubbing

Dec 16, 2021
Derek Tam, Surafel M. Lakew, Yogesh Virkar, Prashant Mathur, Marcello Federico

Figure 1 for Prosody-Aware Neural Machine Translation for Dubbing
Figure 2 for Prosody-Aware Neural Machine Translation for Dubbing
Figure 3 for Prosody-Aware Neural Machine Translation for Dubbing
Figure 4 for Prosody-Aware Neural Machine Translation for Dubbing
Viaarxiv icon