Alert button
Picture for Dan Su

Dan Su

Alert button

AppQ: Warm-starting App Recommendation Based on View Graphs

Add code
Bookmark button
Alert button
Sep 08, 2021
Dan Su, Jiqiang Liu, Sencun Zhu, Xiaoyang Wang, Wei Wang, Xiangliang Zhang

Figure 1 for AppQ: Warm-starting App Recommendation Based on View Graphs
Figure 2 for AppQ: Warm-starting App Recommendation Based on View Graphs
Figure 3 for AppQ: Warm-starting App Recommendation Based on View Graphs
Figure 4 for AppQ: Warm-starting App Recommendation Based on View Graphs
Viaarxiv icon

Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis

Add code
Bookmark button
Alert button
Sep 08, 2021
Songxiang Liu, Shan Yang, Dan Su, Dong Yu

Figure 1 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 2 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 3 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 4 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Viaarxiv icon

Bilateral Denoising Diffusion Models

Add code
Bookmark button
Alert button
Aug 31, 2021
Max W. Y. Lam, Jun Wang, Rongjie Huang, Dan Su, Dong Yu

Figure 1 for Bilateral Denoising Diffusion Models
Figure 2 for Bilateral Denoising Diffusion Models
Figure 3 for Bilateral Denoising Diffusion Models
Figure 4 for Bilateral Denoising Diffusion Models
Viaarxiv icon

Glow-WaveGAN: Learning Speech Representations from GAN-based Variational Auto-Encoder For High Fidelity Flow-based Speech Synthesis

Add code
Bookmark button
Alert button
Jun 22, 2021
Jian Cong, Shan Yang, Lei Xie, Dan Su

Figure 1 for Glow-WaveGAN: Learning Speech Representations from GAN-based Variational Auto-Encoder For High Fidelity Flow-based Speech Synthesis
Figure 2 for Glow-WaveGAN: Learning Speech Representations from GAN-based Variational Auto-Encoder For High Fidelity Flow-based Speech Synthesis
Figure 3 for Glow-WaveGAN: Learning Speech Representations from GAN-based Variational Auto-Encoder For High Fidelity Flow-based Speech Synthesis
Figure 4 for Glow-WaveGAN: Learning Speech Representations from GAN-based Variational Auto-Encoder For High Fidelity Flow-based Speech Synthesis
Viaarxiv icon

Controllable Context-aware Conversational Speech Synthesis

Add code
Bookmark button
Alert button
Jun 21, 2021
Jian Cong, Shan Yang, Na Hu, Guangzhi Li, Lei Xie, Dan Su

Figure 1 for Controllable Context-aware Conversational Speech Synthesis
Figure 2 for Controllable Context-aware Conversational Speech Synthesis
Figure 3 for Controllable Context-aware Conversational Speech Synthesis
Figure 4 for Controllable Context-aware Conversational Speech Synthesis
Viaarxiv icon

GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio

Add code
Bookmark button
Alert button
Jun 13, 2021
Guoguo Chen, Shuzhou Chai, Guanbo Wang, Jiayu Du, Wei-Qiang Zhang, Chao Weng, Dan Su, Daniel Povey, Jan Trmal, Junbo Zhang, Mingjie Jin, Sanjeev Khudanpur, Shinji Watanabe, Shuaijiang Zhao, Wei Zou, Xiangang Li, Xuchen Yao, Yongqing Wang, Yujun Wang, Zhao You, Zhiyong Yan

Figure 1 for GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio
Figure 2 for GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio
Figure 3 for GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio
Figure 4 for GigaSpeech: An Evolving, Multi-domain ASR Corpus with 10,000 Hours of Transcribed Audio
Viaarxiv icon

Spoken Style Learning with Multi-modal Hierarchical Context Encoding for Conversational Text-to-Speech Synthesis

Add code
Bookmark button
Alert button
Jun 11, 2021
Jingbei Li, Yi Meng, Chenyi Li, Zhiyong Wu, Helen Meng, Chao Weng, Dan Su

Figure 1 for Spoken Style Learning with Multi-modal Hierarchical Context Encoding for Conversational Text-to-Speech Synthesis
Figure 2 for Spoken Style Learning with Multi-modal Hierarchical Context Encoding for Conversational Text-to-Speech Synthesis
Figure 3 for Spoken Style Learning with Multi-modal Hierarchical Context Encoding for Conversational Text-to-Speech Synthesis
Figure 4 for Spoken Style Learning with Multi-modal Hierarchical Context Encoding for Conversational Text-to-Speech Synthesis
Viaarxiv icon

Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Jun 08, 2021
Max W. Y. Lam, Jun Wang, Chao Weng, Dan Su, Dong Yu

Figure 1 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Figure 2 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Figure 3 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Figure 4 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Viaarxiv icon