Picture for Ron J. Weiss

Ron J. Weiss

On Using Backpropagation for Speech Texture Generation and Voice Conversion

Add code
Mar 08, 2018
Figure 1 for On Using Backpropagation for Speech Texture Generation and Voice Conversion
Figure 2 for On Using Backpropagation for Speech Texture Generation and Voice Conversion
Figure 3 for On Using Backpropagation for Speech Texture Generation and Voice Conversion
Figure 4 for On Using Backpropagation for Speech Texture Generation and Voice Conversion
Viaarxiv icon

State-of-the-art Speech Recognition With Sequence-to-Sequence Models

Add code
Feb 23, 2018
Figure 1 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 2 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 3 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Figure 4 for State-of-the-art Speech Recognition With Sequence-to-Sequence Models
Viaarxiv icon

Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions

Add code
Feb 16, 2018
Figure 1 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 2 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 3 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Figure 4 for Natural TTS Synthesis by Conditioning WaveNet on Mel Spectrogram Predictions
Viaarxiv icon

Multilingual Speech Recognition With A Single End-To-End Model

Add code
Feb 15, 2018
Figure 1 for Multilingual Speech Recognition With A Single End-To-End Model
Figure 2 for Multilingual Speech Recognition With A Single End-To-End Model
Figure 3 for Multilingual Speech Recognition With A Single End-To-End Model
Figure 4 for Multilingual Speech Recognition With A Single End-To-End Model
Viaarxiv icon

Online and Linear-Time Attention by Enforcing Monotonic Alignments

Add code
Jun 29, 2017
Figure 1 for Online and Linear-Time Attention by Enforcing Monotonic Alignments
Figure 2 for Online and Linear-Time Attention by Enforcing Monotonic Alignments
Figure 3 for Online and Linear-Time Attention by Enforcing Monotonic Alignments
Figure 4 for Online and Linear-Time Attention by Enforcing Monotonic Alignments
Viaarxiv icon

Sequence-to-Sequence Models Can Directly Translate Foreign Speech

Add code
Jun 12, 2017
Figure 1 for Sequence-to-Sequence Models Can Directly Translate Foreign Speech
Figure 2 for Sequence-to-Sequence Models Can Directly Translate Foreign Speech
Figure 3 for Sequence-to-Sequence Models Can Directly Translate Foreign Speech
Figure 4 for Sequence-to-Sequence Models Can Directly Translate Foreign Speech
Viaarxiv icon

Tacotron: Towards End-to-End Speech Synthesis

Add code
Apr 06, 2017
Figure 1 for Tacotron: Towards End-to-End Speech Synthesis
Figure 2 for Tacotron: Towards End-to-End Speech Synthesis
Figure 3 for Tacotron: Towards End-to-End Speech Synthesis
Figure 4 for Tacotron: Towards End-to-End Speech Synthesis
Viaarxiv icon

CNN Architectures for Large-Scale Audio Classification

Add code
Jan 10, 2017
Figure 1 for CNN Architectures for Large-Scale Audio Classification
Figure 2 for CNN Architectures for Large-Scale Audio Classification
Figure 3 for CNN Architectures for Large-Scale Audio Classification
Figure 4 for CNN Architectures for Large-Scale Audio Classification
Viaarxiv icon