Alert button
Picture for Bhuvana Ramabhadran

Bhuvana Ramabhadran

Alert button

Injecting Text in Self-Supervised Speech Pretraining

Add code
Bookmark button
Alert button
Aug 27, 2021
Zhehuai Chen, Yu Zhang, Andrew Rosenberg, Bhuvana Ramabhadran, Gary Wang, Pedro Moreno

Figure 1 for Injecting Text in Self-Supervised Speech Pretraining
Figure 2 for Injecting Text in Self-Supervised Speech Pretraining
Figure 3 for Injecting Text in Self-Supervised Speech Pretraining
Figure 4 for Injecting Text in Self-Supervised Speech Pretraining
Viaarxiv icon

LSTM Acoustic Models Learn to Align and Pronounce with Graphemes

Add code
Bookmark button
Alert button
Aug 13, 2020
Arindrima Datta, Guanlong Zhao, Bhuvana Ramabhadran, Eugene Weinstein

Figure 1 for LSTM Acoustic Models Learn to Align and Pronounce with Graphemes
Figure 2 for LSTM Acoustic Models Learn to Align and Pronounce with Graphemes
Figure 3 for LSTM Acoustic Models Learn to Align and Pronounce with Graphemes
Figure 4 for LSTM Acoustic Models Learn to Align and Pronounce with Graphemes
Viaarxiv icon

Language-agnostic Multilingual Modeling

Add code
Bookmark button
Alert button
Apr 20, 2020
Arindrima Datta, Bhuvana Ramabhadran, Jesse Emond, Anjuli Kannan, Brian Roark

Figure 1 for Language-agnostic Multilingual Modeling
Figure 2 for Language-agnostic Multilingual Modeling
Figure 3 for Language-agnostic Multilingual Modeling
Figure 4 for Language-agnostic Multilingual Modeling
Viaarxiv icon

Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and auto-regressive prosody prior

Add code
Bookmark button
Alert button
Feb 06, 2020
Guangzhi Sun, Yu Zhang, Ron J. Weiss, Yuan Cao, Heiga Zen, Andrew Rosenberg, Bhuvana Ramabhadran, Yonghui Wu

Figure 1 for Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and auto-regressive prosody prior
Figure 2 for Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and auto-regressive prosody prior
Figure 3 for Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and auto-regressive prosody prior
Figure 4 for Generating diverse and natural text-to-speech samples using a quantized fine-grained VAE and auto-regressive prosody prior
Viaarxiv icon

Speech Recognition with Augmented Synthesized Speech

Add code
Bookmark button
Alert button
Sep 25, 2019
Andrew Rosenberg, Yu Zhang, Bhuvana Ramabhadran, Ye Jia, Pedro Moreno, Yonghui Wu, Zelin Wu

Figure 1 for Speech Recognition with Augmented Synthesized Speech
Figure 2 for Speech Recognition with Augmented Synthesized Speech
Figure 3 for Speech Recognition with Augmented Synthesized Speech
Figure 4 for Speech Recognition with Augmented Synthesized Speech
Viaarxiv icon

Large-Scale Multilingual Speech Recognition with a Streaming End-to-End Model

Add code
Bookmark button
Alert button
Sep 11, 2019
Anjuli Kannan, Arindrima Datta, Tara N. Sainath, Eugene Weinstein, Bhuvana Ramabhadran, Yonghui Wu, Ankur Bapna, Zhifeng Chen, Seungji Lee

Figure 1 for Large-Scale Multilingual Speech Recognition with a Streaming End-to-End Model
Figure 2 for Large-Scale Multilingual Speech Recognition with a Streaming End-to-End Model
Figure 3 for Large-Scale Multilingual Speech Recognition with a Streaming End-to-End Model
Viaarxiv icon

Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning

Add code
Bookmark button
Alert button
Jul 24, 2019
Yu Zhang, Ron J. Weiss, Heiga Zen, Yonghui Wu, Zhifeng Chen, RJ Skerry-Ryan, Ye Jia, Andrew Rosenberg, Bhuvana Ramabhadran

Figure 1 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 2 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 3 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 4 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Viaarxiv icon

Joint Modeling of Accents and Acoustics for Multi-Accent Speech Recognition

Add code
Bookmark button
Alert button
Feb 07, 2018
Xuesong Yang, Kartik Audhkhasi, Andrew Rosenberg, Samuel Thomas, Bhuvana Ramabhadran, Mark Hasegawa-Johnson

Figure 1 for Joint Modeling of Accents and Acoustics for Multi-Accent Speech Recognition
Figure 2 for Joint Modeling of Accents and Acoustics for Multi-Accent Speech Recognition
Figure 3 for Joint Modeling of Accents and Acoustics for Multi-Accent Speech Recognition
Figure 4 for Joint Modeling of Accents and Acoustics for Multi-Accent Speech Recognition
Viaarxiv icon

Building competitive direct acoustics-to-word models for English conversational speech recognition

Add code
Bookmark button
Alert button
Dec 08, 2017
Kartik Audhkhasi, Brian Kingsbury, Bhuvana Ramabhadran, George Saon, Michael Picheny

Figure 1 for Building competitive direct acoustics-to-word models for English conversational speech recognition
Figure 2 for Building competitive direct acoustics-to-word models for English conversational speech recognition
Figure 3 for Building competitive direct acoustics-to-word models for English conversational speech recognition
Figure 4 for Building competitive direct acoustics-to-word models for English conversational speech recognition
Viaarxiv icon