Alert button
Picture for RJ Skerry-Ryan

RJ Skerry-Ryan

Alert button

Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning

Add code
Bookmark button
Alert button
Jul 09, 2019
Yu Zhang, Ron J. Weiss, Heiga Zen, Yonghui Wu, Zhifeng Chen, RJ Skerry-Ryan, Ye Jia, Andrew Rosenberg, Bhuvana Ramabhadran

Figure 1 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 2 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 3 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Figure 4 for Learning to Speak Fluently in a Foreign Language: Multilingual Speech Synthesis and Cross-Language Voice Cloning
Viaarxiv icon

Effective Use of Variational Embedding Capacity in Expressive End-to-End Speech Synthesis

Add code
Bookmark button
Alert button
Jul 09, 2019
Eric Battenberg, Soroosh Mariooryad, Daisy Stanton, RJ Skerry-Ryan, Matt Shannon, David Kao, Tom Bagby

Figure 1 for Effective Use of Variational Embedding Capacity in Expressive End-to-End Speech Synthesis
Figure 2 for Effective Use of Variational Embedding Capacity in Expressive End-to-End Speech Synthesis
Figure 3 for Effective Use of Variational Embedding Capacity in Expressive End-to-End Speech Synthesis
Figure 4 for Effective Use of Variational Embedding Capacity in Expressive End-to-End Speech Synthesis
Viaarxiv icon

Semi-Supervised Training for Improving Data Efficiency in End-to-End Speech Synthesis

Add code
Bookmark button
Alert button
Aug 30, 2018
Yu-An Chung, Yuxuan Wang, Wei-Ning Hsu, Yu Zhang, RJ Skerry-Ryan

Figure 1 for Semi-Supervised Training for Improving Data Efficiency in End-to-End Speech Synthesis
Figure 2 for Semi-Supervised Training for Improving Data Efficiency in End-to-End Speech Synthesis
Figure 3 for Semi-Supervised Training for Improving Data Efficiency in End-to-End Speech Synthesis
Figure 4 for Semi-Supervised Training for Improving Data Efficiency in End-to-End Speech Synthesis
Viaarxiv icon

Predicting Expressive Speaking Style From Text In End-To-End Speech Synthesis

Add code
Bookmark button
Alert button
Aug 04, 2018
Daisy Stanton, Yuxuan Wang, RJ Skerry-Ryan

Figure 1 for Predicting Expressive Speaking Style From Text In End-To-End Speech Synthesis
Figure 2 for Predicting Expressive Speaking Style From Text In End-To-End Speech Synthesis
Figure 3 for Predicting Expressive Speaking Style From Text In End-To-End Speech Synthesis
Figure 4 for Predicting Expressive Speaking Style From Text In End-To-End Speech Synthesis
Viaarxiv icon

Towards End-to-End Prosody Transfer for Expressive Speech Synthesis with Tacotron

Add code
Bookmark button
Alert button
Mar 24, 2018
RJ Skerry-Ryan, Eric Battenberg, Ying Xiao, Yuxuan Wang, Daisy Stanton, Joel Shor, Ron J. Weiss, Rob Clark, Rif A. Saurous

Figure 1 for Towards End-to-End Prosody Transfer for Expressive Speech Synthesis with Tacotron
Figure 2 for Towards End-to-End Prosody Transfer for Expressive Speech Synthesis with Tacotron
Figure 3 for Towards End-to-End Prosody Transfer for Expressive Speech Synthesis with Tacotron
Figure 4 for Towards End-to-End Prosody Transfer for Expressive Speech Synthesis with Tacotron
Viaarxiv icon

Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis

Add code
Bookmark button
Alert button
Mar 23, 2018
Yuxuan Wang, Daisy Stanton, Yu Zhang, RJ Skerry-Ryan, Eric Battenberg, Joel Shor, Ying Xiao, Fei Ren, Ye Jia, Rif A. Saurous

Figure 1 for Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis
Figure 2 for Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis
Figure 3 for Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis
Figure 4 for Style Tokens: Unsupervised Style Modeling, Control and Transfer in End-to-End Speech Synthesis
Viaarxiv icon

Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions

Add code
Bookmark button
Alert button
Feb 16, 2018
Jonathan Shen, Ruoming Pang, Ron J. Weiss, Mike Schuster, Navdeep Jaitly, Zongheng Yang, Zhifeng Chen, Yu Zhang, Yuxuan Wang, RJ Skerry-Ryan, Rif A. Saurous, Yannis Agiomyrgiannakis, Yonghui Wu

Figure 1 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 2 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 3 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 4 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Viaarxiv icon

Uncovering Latent Style Factors for Expressive Speech Synthesis

Add code
Bookmark button
Alert button
Nov 01, 2017
Yuxuan Wang, RJ Skerry-Ryan, Ying Xiao, Daisy Stanton, Joel Shor, Eric Battenberg, Rob Clark, Rif A. Saurous

Figure 1 for Uncovering Latent Style Factors for Expressive Speech Synthesis
Figure 2 for Uncovering Latent Style Factors for Expressive Speech Synthesis
Figure 3 for Uncovering Latent Style Factors for Expressive Speech Synthesis
Viaarxiv icon

Tacotron: Towards End-to-End Speech Synthesis

Add code
Bookmark button
Alert button
Apr 06, 2017
Yuxuan Wang, RJ Skerry-Ryan, Daisy Stanton, Yonghui Wu, Ron J. Weiss, Navdeep Jaitly, Zongheng Yang, Ying Xiao, Zhifeng Chen, Samy Bengio, Quoc Le, Yannis Agiomyrgiannakis, Rob Clark, Rif A. Saurous

Figure 1 for Tacotron: Towards End-to-End Speech Synthesis
Figure 2 for Tacotron: Towards End-to-End Speech Synthesis
Figure 3 for Tacotron: Towards End-to-End Speech Synthesis
Figure 4 for Tacotron: Towards End-to-End Speech Synthesis
Viaarxiv icon