Alert button
Picture for Songxiang Liu

Songxiang Liu

Alert button

UniAudio: An Audio Foundation Model Toward Universal Audio Generation

Add code
Bookmark button
Alert button
Oct 11, 2023
Dongchao Yang, Jinchuan Tian, Xu Tan, Rongjie Huang, Songxiang Liu, Xuankai Chang, Jiatong Shi, Sheng Zhao, Jiang Bian, Xixin Wu, Zhou Zhao, Shinji Watanabe, Helen Meng

Figure 1 for UniAudio: An Audio Foundation Model Toward Universal Audio Generation
Figure 2 for UniAudio: An Audio Foundation Model Toward Universal Audio Generation
Figure 3 for UniAudio: An Audio Foundation Model Toward Universal Audio Generation
Figure 4 for UniAudio: An Audio Foundation Model Toward Universal Audio Generation
Viaarxiv icon

SnakeGAN: A Universal Vocoder Leveraging DDSP Prior Knowledge and Periodic Inductive Bias

Add code
Bookmark button
Alert button
Sep 14, 2023
Sipan Li, Songxiang Liu, Luwen Zhang, Xiang Li, Yanyao Bian, Chao Weng, Zhiyong Wu, Helen Meng

Figure 1 for SnakeGAN: A Universal Vocoder Leveraging DDSP Prior Knowledge and Periodic Inductive Bias
Figure 2 for SnakeGAN: A Universal Vocoder Leveraging DDSP Prior Knowledge and Periodic Inductive Bias
Figure 3 for SnakeGAN: A Universal Vocoder Leveraging DDSP Prior Knowledge and Periodic Inductive Bias
Figure 4 for SnakeGAN: A Universal Vocoder Leveraging DDSP Prior Knowledge and Periodic Inductive Bias
Viaarxiv icon

The Singing Voice Conversion Challenge 2023

Add code
Bookmark button
Alert button
Jul 06, 2023
Wen-Chin Huang, Lester Phillip Violeta, Songxiang Liu, Jiatong Shi, Tomoki Toda

Figure 1 for The Singing Voice Conversion Challenge 2023
Figure 2 for The Singing Voice Conversion Challenge 2023
Figure 3 for The Singing Voice Conversion Challenge 2023
Figure 4 for The Singing Voice Conversion Challenge 2023
Viaarxiv icon

Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model

Add code
Bookmark button
Alert button
May 26, 2023
Xiang Li, Songxiang Liu, Max W. Y. Lam, Zhiyong Wu, Chao Weng, Helen Meng

Figure 1 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Figure 2 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Figure 3 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Figure 4 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Viaarxiv icon

HiFi-Codec: Group-residual Vector quantization for High Fidelity Audio Codec

Add code
Bookmark button
Alert button
May 07, 2023
Dongchao Yang, Songxiang Liu, Rongjie Huang, Jinchuan Tian, Chao Weng, Yuexian Zou

Figure 1 for HiFi-Codec: Group-residual Vector quantization for High Fidelity Audio Codec
Figure 2 for HiFi-Codec: Group-residual Vector quantization for High Fidelity Audio Codec
Viaarxiv icon

InstructTTS: Modelling Expressive TTS in Discrete Latent Space with Natural Language Style Prompt

Add code
Bookmark button
Alert button
Jan 31, 2023
Dongchao Yang, Songxiang Liu, Rongjie Huang, Guangzhi Lei, Chao Weng, Helen Meng, Dong Yu

Figure 1 for InstructTTS: Modelling Expressive TTS in Discrete Latent Space with Natural Language Style Prompt
Figure 2 for InstructTTS: Modelling Expressive TTS in Discrete Latent Space with Natural Language Style Prompt
Figure 3 for InstructTTS: Modelling Expressive TTS in Discrete Latent Space with Natural Language Style Prompt
Figure 4 for InstructTTS: Modelling Expressive TTS in Discrete Latent Space with Natural Language Style Prompt
Viaarxiv icon

NoreSpeech: Knowledge Distillation based Conditional Diffusion Model for Noise-robust Expressive TTS

Add code
Bookmark button
Alert button
Nov 04, 2022
Dongchao Yang, Songxiang Liu, Jianwei Yu, Helin Wang, Chao Weng, Yuexian Zou

Figure 1 for NoreSpeech: Knowledge Distillation based Conditional Diffusion Model for Noise-robust Expressive TTS
Figure 2 for NoreSpeech: Knowledge Distillation based Conditional Diffusion Model for Noise-robust Expressive TTS
Figure 3 for NoreSpeech: Knowledge Distillation based Conditional Diffusion Model for Noise-robust Expressive TTS
Viaarxiv icon

Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation

Add code
Bookmark button
Alert button
Feb 18, 2022
Disong Wang, Songxiang Liu, Xixin Wu, Hui Lu, Lifa Sun, Xunying Liu, Helen Meng

Figure 1 for Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation
Figure 2 for Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation
Figure 3 for Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation
Figure 4 for Speaker Identity Preservation in Dysarthric Speech Reconstruction by Adversarial Speaker Adaptation
Viaarxiv icon