Picture for Zhifeng Chen

Zhifeng Chen

Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling

Add code
Feb 21, 2019
Figure 1 for Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling
Figure 2 for Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling
Figure 3 for Lingvo: a Modular and Scalable Framework for Sequence-to-Sequence Modeling
Viaarxiv icon

GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism

Add code
Dec 12, 2018
Figure 1 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 2 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 3 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Figure 4 for GPipe: Efficient Training of Giant Neural Networks using Pipeline Parallelism
Viaarxiv icon

Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis

Add code
Nov 05, 2018
Figure 1 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Figure 2 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Figure 3 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Figure 4 for Transfer Learning from Speaker Verification to Multispeaker Text-To-Speech Synthesis
Viaarxiv icon

Hierarchical Generative Modeling for Controllable Speech Synthesis

Add code
Oct 16, 2018
Figure 1 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Figure 2 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Figure 3 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Figure 4 for Hierarchical Generative Modeling for Controllable Speech Synthesis
Viaarxiv icon

The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation

Add code
Apr 27, 2018
Figure 1 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Figure 2 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Figure 3 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Figure 4 for The Best of Both Worlds: Combining Recent Advances in Neural Machine Translation
Viaarxiv icon

State-of-the-art Speech Recognition With Sequence-to-Sequence Models

Add code
Feb 23, 2018
Figure 1 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 2 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 3 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 4 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Viaarxiv icon

Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions

Add code
Feb 16, 2018
Figure 1 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 2 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 3 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 4 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Viaarxiv icon

An analysis of incorporating an external language model into a sequence-to-sequence model

Add code
Dec 06, 2017
Figure 1 for An analysis of incorporating an external language model into a sequence-to-sequence model
Figure 2 for An analysis of incorporating an external language model into a sequence-to-sequence model
Figure 3 for An analysis of incorporating an external language model into a sequence-to-sequence model
Figure 4 for An analysis of incorporating an external language model into a sequence-to-sequence model
Viaarxiv icon

No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models

Add code
Dec 05, 2017
Figure 1 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Figure 2 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Figure 3 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Figure 4 for No Need for a Lexicon? Evaluating the Value of the Pronunciation Lexica in End-to-End Models
Viaarxiv icon

Minimum Word Error Rate Training for Attention-based Sequence-to-Sequence Models

Add code
Dec 05, 2017
Figure 1 for Minimum Word Error Rate Training for Attention-based Sequence-to-Sequence Models
Figure 2 for Minimum Word Error Rate Training for Attention-based Sequence-to-Sequence Models
Figure 3 for Minimum Word Error Rate Training for Attention-based Sequence-to-Sequence Models
Figure 4 for Minimum Word Error Rate Training for Attention-based Sequence-to-Sequence Models
Viaarxiv icon