Alert button
Picture for Songxiang Liu

Songxiang Liu

Alert button

DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs

Add code
Bookmark button
Alert button
Jan 28, 2022
Songxiang Liu, Dan Su, Dong Yu

Figure 1 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Figure 2 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Figure 3 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Figure 4 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Viaarxiv icon

Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning

Add code
Bookmark button
Alert button
Nov 14, 2021
Songxiang Liu, Dan Su, Dong Yu

Figure 1 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Figure 2 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Figure 3 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Figure 4 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Viaarxiv icon

Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis

Add code
Bookmark button
Alert button
Sep 08, 2021
Songxiang Liu, Shan Yang, Dan Su, Dong Yu

Figure 1 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 2 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 3 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 4 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Viaarxiv icon

ASR-GLUE: A New Multi-task Benchmark for ASR-Robust Natural Language Understanding

Add code
Bookmark button
Alert button
Aug 30, 2021
Lingyun Feng, Jianwei Yu, Deng Cai, Songxiang Liu, Haitao Zheng, Yan Wang

Figure 1 for ASR-GLUE: A New Multi-task Benchmark for ASR-Robust Natural Language Understanding
Figure 2 for ASR-GLUE: A New Multi-task Benchmark for ASR-Robust Natural Language Understanding
Figure 3 for ASR-GLUE: A New Multi-task Benchmark for ASR-Robust Natural Language Understanding
Figure 4 for ASR-GLUE: A New Multi-task Benchmark for ASR-Robust Natural Language Understanding
Viaarxiv icon

DiffSVC: A Diffusion Probabilistic Model for Singing Voice Conversion

Add code
Bookmark button
Alert button
May 28, 2021
Songxiang Liu, Yuewen Cao, Dan Su, Helen Meng

Figure 1 for DiffSVC: A Diffusion Probabilistic Model for Singing Voice Conversion
Figure 2 for DiffSVC: A Diffusion Probabilistic Model for Singing Voice Conversion
Figure 3 for DiffSVC: A Diffusion Probabilistic Model for Singing Voice Conversion
Viaarxiv icon

VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention

Add code
Bookmark button
Alert button
Feb 12, 2021
Peng Liu, Yuewen Cao, Songxiang Liu, Na Hu, Guangzhi Li, Chao Weng, Dan Su

Figure 1 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Figure 2 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Figure 3 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Figure 4 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Viaarxiv icon

Any-to-Many Voice Conversion with Location-Relative Sequence-to-Sequence Modeling

Add code
Bookmark button
Alert button
Sep 06, 2020
Songxiang Liu, Yuewen Cao, Disong Wang, Xixin Wu, Xunying Liu, Helen Meng

Figure 1 for Any-to-Many Voice Conversion with Location-Relative Sequence-to-Sequence Modeling
Figure 2 for Any-to-Many Voice Conversion with Location-Relative Sequence-to-Sequence Modeling
Figure 3 for Any-to-Many Voice Conversion with Location-Relative Sequence-to-Sequence Modeling
Figure 4 for Any-to-Many Voice Conversion with Location-Relative Sequence-to-Sequence Modeling
Viaarxiv icon

Defense against adversarial attacks on spoofing countermeasures of ASV

Add code
Bookmark button
Alert button
Mar 06, 2020
Haibin Wu, Songxiang Liu, Helen Meng, Hung-yi Lee

Figure 1 for Defense against adversarial attacks on spoofing countermeasures of ASV
Figure 2 for Defense against adversarial attacks on spoofing countermeasures of ASV
Figure 3 for Defense against adversarial attacks on spoofing countermeasures of ASV
Viaarxiv icon

Multi-Layer Content Interaction Through Quaternion Product For Visual Question Answering

Add code
Bookmark button
Alert button
Feb 16, 2020
Lei Shi, Shijie Geng, Kai Shuang, Chiori Hori, Songxiang Liu, Peng Gao, Sen Su

Figure 1 for Multi-Layer Content Interaction Through Quaternion Product For Visual Question Answering
Figure 2 for Multi-Layer Content Interaction Through Quaternion Product For Visual Question Answering
Figure 3 for Multi-Layer Content Interaction Through Quaternion Product For Visual Question Answering
Figure 4 for Multi-Layer Content Interaction Through Quaternion Product For Visual Question Answering
Viaarxiv icon