Alert button
Picture for Jilong Wu

Jilong Wu

Alert button

Self-Supervised Representations for Singing Voice Conversion

Add code
Bookmark button
Alert button
Mar 21, 2023
Tejas Jayashankar, Jilong Wu, Leda Sari, David Kant, Vimal Manohar, Qing He

Figure 1 for Self-Supervised Representations for Singing Voice Conversion
Figure 2 for Self-Supervised Representations for Singing Voice Conversion
Figure 3 for Self-Supervised Representations for Singing Voice Conversion
Figure 4 for Self-Supervised Representations for Singing Voice Conversion
Viaarxiv icon

Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition

Add code
Bookmark button
Alert button
Mar 01, 2023
Philipp Klumpp, Pooja Chitkara, Leda Sarı, Prashant Serai, Jilong Wu, Irina-Elena Veliche, Rongqing Huang, Qing He

Figure 1 for Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition
Figure 2 for Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition
Figure 3 for Synthetic Cross-accent Data Augmentation for Automatic Speech Recognition
Viaarxiv icon

Voice-preserving Zero-shot Multiple Accent Conversion

Add code
Bookmark button
Alert button
Nov 23, 2022
Mumin Jin, Prashant Serai, Jilong Wu, Andros Tjandra, Vimal Manohar, Qing He

Figure 1 for Voice-preserving Zero-shot Multiple Accent Conversion
Figure 2 for Voice-preserving Zero-shot Multiple Accent Conversion
Figure 3 for Voice-preserving Zero-shot Multiple Accent Conversion
Figure 4 for Voice-preserving Zero-shot Multiple Accent Conversion
Viaarxiv icon

Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders

Add code
Bookmark button
Alert button
Oct 28, 2022
Jason Fong, Yun Wang, Prabhav Agrawal, Vimal Manohar, Jilong Wu, Thilo Köhler, Qing He

Figure 1 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Figure 2 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Figure 3 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Figure 4 for Towards zero-shot Text-based voice editing using acoustic context conditioning, utterance embeddings, and reference encoders
Viaarxiv icon

VocBench: A Neural Vocoder Benchmark for Speech Synthesis

Add code
Bookmark button
Alert button
Dec 06, 2021
Ehab A. AlBadawy, Andrew Gibiansky, Qing He, Jilong Wu, Ming-Ching Chang, Siwei Lyu

Figure 1 for VocBench: A Neural Vocoder Benchmark for Speech Synthesis
Figure 2 for VocBench: A Neural Vocoder Benchmark for Speech Synthesis
Figure 3 for VocBench: A Neural Vocoder Benchmark for Speech Synthesis
Viaarxiv icon

Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling

Add code
Bookmark button
Alert button
Apr 01, 2021
Qing He, Zhiping Xiu, Thilo Koehler, Jilong Wu

Figure 1 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Figure 2 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Figure 3 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Figure 4 for Multi-rate attention architecture for fast streamable Text-to-speech spectrum modeling
Viaarxiv icon