Picture for Jilong Wu

Jilong Wu

Self-Supervised Representations for Singing Voice Conversion

Add code
Mar 21, 2023
Figure 1 for Self-Supervised Representations for Singing Voice Conversion
Figure 2 for Self-Supervised Representations for Singing Voice Conversion
Figure 3 for Self-Supervised Representations for Singing Voice Conversion
Figure 4 for Self-Supervised Representations for Singing Voice Conversion
Viaarxiv icon

Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition

Add code
Mar 01, 2023
Figure 1 for Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition
Figure 2 for Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition
Figure 3 for Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition
Viaarxiv icon

Voice-preserving Zero-shot Multiple Accent Conversion

Add code
Nov 23, 2022
Figure 1 for Voice-preserving Zero-shot Multiple Accent Conversion
Figure 2 for Voice-preserving Zero-shot Multiple Accent Conversion
Figure 3 for Voice-preserving Zero-shot Multiple Accent Conversion
Figure 4 for Voice-preserving Zero-shot Multiple Accent Conversion
Viaarxiv icon

Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders

Add code
Oct 28, 2022
Figure 1 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Figure 2 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Figure 3 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Figure 4 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Viaarxiv icon

VocBench: A Neural Vocoder Benchmark for Speech Synthesis

Add code
Dec 06, 2021
Figure 1 for VocBench: A Neural Vocoder Benchmark for Speech Synthesis
Figure 2 for VocBench: A Neural Vocoder Benchmark for Speech Synthesis
Figure 3 for VocBench: A Neural Vocoder Benchmark for Speech Synthesis
Viaarxiv icon

Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling

Add code
Apr 01, 2021
Figure 1 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Figure 2 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Figure 3 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Figure 4 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Viaarxiv icon