Alert button
Picture for Max W. Y. Lam

Max W. Y. Lam

Alert button

Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model

Add code
Bookmark button
Alert button
May 26, 2023
Xiang Li, Songxiang Liu, Max W. Y. Lam, Zhiyong Wu, Chao Weng, Helen Meng

Figure 1 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Figure 2 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Figure 3 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Figure 4 for Diverse and Expressive Speech Prosody Prediction with Denoising Diffusion Probabilistic Model
Viaarxiv icon

Efficient Neural Music Generation

Add code
Bookmark button
Alert button
May 25, 2023
Max W. Y. Lam, Qiao Tian, Tang Li, Zongyu Yin, Siyuan Feng, Ming Tu, Yuliang Ji, Rui Xia, Mingbo Ma, Xuchen Song, Jitong Chen, Yuping Wang, Yuxuan Wang

Figure 1 for Efficient Neural Music Generation
Figure 2 for Efficient Neural Music Generation
Figure 3 for Efficient Neural Music Generation
Figure 4 for Efficient Neural Music Generation
Viaarxiv icon

FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis

Add code
Bookmark button
Alert button
Apr 21, 2022
Rongjie Huang, Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu, Yi Ren, Zhou Zhao

Figure 1 for FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis
Figure 2 for FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis
Figure 3 for FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis
Figure 4 for FastDiff: A Fast Conditional Diffusion Model for High-Quality Speech Synthesis
Viaarxiv icon

BDDM: Bilateral Denoising Diffusion Models for Fast and High-Quality Speech Synthesis

Add code
Bookmark button
Alert button
Mar 25, 2022
Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu

Figure 1 for BDDM: Bilateral Denoising Diffusion Models for Fast and High-Quality Speech Synthesis
Figure 2 for BDDM: Bilateral Denoising Diffusion Models for Fast and High-Quality Speech Synthesis
Figure 3 for BDDM: Bilateral Denoising Diffusion Models for Fast and High-Quality Speech Synthesis
Figure 4 for BDDM: Bilateral Denoising Diffusion Models for Fast and High-Quality Speech Synthesis
Viaarxiv icon

Bilateral Denoising Diffusion Models

Add code
Bookmark button
Alert button
Aug 31, 2021
Max W. Y. Lam, Jun Wang, Rongjie Huang, Dan Su, Dong Yu

Figure 1 for Bilateral Denoising Diffusion Models
Figure 2 for Bilateral Denoising Diffusion Models
Figure 3 for Bilateral Denoising Diffusion Models
Figure 4 for Bilateral Denoising Diffusion Models
Viaarxiv icon

Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Jun 08, 2021
Max W. Y. Lam, Jun Wang, Chao Weng, Dan Su, Dong Yu

Figure 1 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Figure 2 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Figure 3 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Figure 4 for Raw Waveform Encoder with Multi-Scale Globally Attentive Locally Recurrent Networks for End-to-End Speech Recognition
Viaarxiv icon

Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation

Add code
Bookmark button
Alert button
Mar 08, 2021
Max W. Y. Lam, Jun Wang, Dan Su, Dong Yu

Figure 1 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Figure 2 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Figure 3 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Figure 4 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Viaarxiv icon

Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect

Add code
Bookmark button
Alert button
Mar 02, 2021
Jun Wang, Max W. Y. Lam, Dan Su, Dong Yu

Figure 1 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Figure 2 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Figure 3 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Figure 4 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Viaarxiv icon