Alert button
Picture for Zhifeng Chen

Zhifeng Chen

Alert button

Direct speech-to-speech translation with a sequence-to-sequence model

Add code
Bookmark button
Alert button
Apr 12, 2019
Ye Jia, Ron J. Weiss, Fadi Biadsy, Wolfgang Macherey, Melvin Johnson, Zhifeng Chen, Yonghui Wu

Figure 1 for Direct speech-to-speech translation with a sequence-to-sequence model
Figure 2 for Direct speech-to-speech translation with a sequence-to-sequence model
Figure 3 for Direct speech-to-speech translation with a sequence-to-sequence model
Figure 4 for Direct speech-to-speech translation with a sequence-to-sequence model
Viaarxiv icon

Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling

Add code
Bookmark button
Alert button
Feb 21, 2019
Jonathan Shen, Patrick Nguyen, Yonghui Wu, Zhifeng Chen, Mia X. Chen, Ye Jia, Anjuli Kannan, Tara Sainath, Yuan Cao, Chung-Cheng Chiu, Yanzhang He, Jan Chorowski, Smit Hinsu, Stella Laurenzo, James Qin, Orhan Firat, Wolfgang Macherey, Suyog Gupta, Ankur Bapna, Shuyuan Zhang, Ruoming Pang, Ron J. Weiss, Rohit Prabhavalkar, Qiao Liang, Benoit Jacob, Bowen Liang, HyoukJoong Lee, Ciprian Chelba, Sébastien Jean, Bo Li, Melvin Johnson, Rohan Anil, Rajat Tibrewal, Xiaobing Liu, Akiko Eriguchi, Navdeep Jaitly, Naveen Ari, Colin Cherry, Parisa Haghani, Otavio Good, Youlong Cheng, Raziel Alvarez, Isaac Caswell, Wei-Ning Hsu, Zongheng Yang, Kuan-Chieh Wang, Ekaterina Gonina, Katrin Tomanek, Ben Vanik, Zelin Wu, Llion Jones, Mike Schuster, Yanping Huang, Dehao Chen, Kazuki Irie, George Foster, John Richardson, Klaus Macherey, Antoine Bruguier, Heiga Zen, Colin Raffel, Shankar Kumar, Kanishka Rao, David Rybach, Matthew Murray, Vijayaditya Peddinti, Maxim Krikun, Michiel A. U. Bacchiani, Thomas B. Jablin, Rob Suderman, Ian Williams, Benjamin Lee, Deepti Bhatia, Justin Carlson, Semih Yavuz, Yu Zhang, Ian McGraw, Max Galkin, Qi Ge, Golan Pundak, Chad Whipkey, Todd Wang, Uri Alon, Dmitry Lepikhin, Ye Tian, Sara Sabour, William Chan, Shubham Toshniwal, Baohua Liao, Michael Nirschl, Pat Rondon

Figure 1 for Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling
Figure 2 for Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling
Figure 3 for Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling
Viaarxiv icon

GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

Add code
Bookmark button
Alert button
Dec 12, 2018
Yanping Huang, Yonglong Cheng, Dehao Chen, HyoukJoong Lee, Jiquan Ngiam, Quoc V. Le, Zhifeng Chen

Figure 1 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 2 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 3 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 4 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Viaarxiv icon

Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis

Add code
Bookmark button
Alert button
Nov 05, 2018
Ye Jia, Yu Zhang, Ron J. Weiss, Quan Wang, Jonathan Shen, Fei Ren, Zhifeng Chen, Patrick Nguyen, Ruoming Pang, Ignacio Lopez Moreno, Yonghui Wu

Figure 1 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Figure 2 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Figure 3 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Figure 4 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Viaarxiv icon

Hierarchical Generative Modeling for Controllable Speech Synthesis

Add code
Bookmark button
Alert button
Oct 16, 2018
Wei-Ning Hsu, Yu Zhang, Ron J. Weiss, Heiga Zen, Yonghui Wu, Yuxuan Wang, Yuan Cao, Ye Jia, Zhifeng Chen, Jonathan Shen, Patrick Nguyen, Ruoming Pang

Figure 1 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Figure 2 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Figure 3 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Figure 4 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Viaarxiv icon

The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation

Add code
Bookmark button
Alert button
Apr 27, 2018
Mia Xu Chen, Orhan Firat, Ankur Bapna, Melvin Johnson, Wolfgang Macherey, George Foster, Llion Jones, Niki Parmar, Mike Schuster, Zhifeng Chen, Yonghui Wu, Macduff Hughes

Figure 1 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Figure 2 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Figure 3 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Figure 4 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Viaarxiv icon

State-of-the-art Speech Recognition With Sequence-to-Sequence Models

Add code
Bookmark button
Alert button
Feb 23, 2018
Chung-Cheng Chiu, Tara N. Sainath, Yonghui Wu, Rohit Prabhavalkar, Patrick Nguyen, Zhifeng Chen, Anjuli Kannan, Ron J. Weiss, Kanishka Rao, Ekaterina Gonina, Navdeep Jaitly, Bo Li, Jan Chorowski, Michiel Bacchiani

Figure 1 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 2 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 3 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 4 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Viaarxiv icon

Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions

Add code
Bookmark button
Alert button
Feb 16, 2018
Jonathan Shen, Ruoming Pang, Ron J. Weiss, Mike Schuster, Navdeep Jaitly, Zongheng Yang, Zhifeng Chen, Yu Zhang, Yuxuan Wang, RJ Skerry-Ryan, Rif A. Saurous, Yannis Agiomyrgiannakis, Yonghui Wu

Figure 1 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 2 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 3 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 4 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Viaarxiv icon

An analysis of incorporating an external language model into a sequence-to-sequence model

Add code
Bookmark button
Alert button
Dec 06, 2017
Anjuli Kannan, Yonghui Wu, Patrick Nguyen, Tara N. Sainath, Zhifeng Chen, Rohit Prabhavalkar

Figure 1 for An analysis of incorporating an external language model into a sequence-to-sequence model
Figure 2 for An analysis of incorporating an external language model into a sequence-to-sequence model
Figure 3 for An analysis of incorporating an external language model into a sequence-to-sequence model
Figure 4 for An analysis of incorporating an external language model into a sequence-to-sequence model
Viaarxiv icon

No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models

Add code
Bookmark button
Alert button
Dec 05, 2017
Tara N. Sainath, Rohit Prabhavalkar, Shankar Kumar, Seungji Lee, Anjuli Kannan, David Rybach, Vlad Schogol, Patrick Nguyen, Bo Li, Yonghui Wu, Zhifeng Chen, Chung-Cheng Chiu

Figure 1 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Figure 2 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Figure 3 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Figure 4 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Viaarxiv icon