Alert button
Picture for Dianwen Ng

Dianwen Ng

Alert button

Noise robust distillation of self-supervised speech models via correlation metrics

Add code
Bookmark button
Alert button
Dec 19, 2023
Fabian Ritter-Gutierrez, Kuan-Po Huang, Dianwen Ng, Jeremy H. M. Wong, Hung-yi Lee, Eng Siong Chng, Nancy F. Chen

Viaarxiv icon

MossFormer2: Combining Transformer and RNN-Free Recurrent Network for Enhanced Time-Domain Monaural Speech Separation

Add code
Bookmark button
Alert button
Dec 19, 2023
Shengkui Zhao, Yukun Ma, Chongjia Ni, Chong Zhang, Hao Wang, Trung Hieu Nguyen, Kun Zhou, Jiaqi Yip, Dianwen Ng, Bin Ma

Viaarxiv icon

SPGM: Prioritizing Local Features for enhanced speech separation performance

Add code
Bookmark button
Alert button
Sep 22, 2023
Jia Qi Yip, Shengkui Zhao, Yukun Ma, Chongjia Ni, Chong Zhang, Hao Wang, Trung Hieu Nguyen, Kun Zhou, Dianwen Ng, Eng Siong Chng, Bin Ma

Figure 1 for SPGM: Prioritizing Local Features for enhanced speech separation performance
Figure 2 for SPGM: Prioritizing Local Features for enhanced speech separation performance
Figure 3 for SPGM: Prioritizing Local Features for enhanced speech separation performance
Figure 4 for SPGM: Prioritizing Local Features for enhanced speech separation performance
Viaarxiv icon

Are Soft Prompts Good Zero-shot Learners for Speech Recognition?

Add code
Bookmark button
Alert button
Sep 18, 2023
Dianwen Ng, Chong Zhang, Ruixi Zhang, Yukun Ma, Fabian Ritter-Gutierrez, Trung Hieu Nguyen, Chongjia Ni, Shengkui Zhao, Eng Siong Chng, Bin Ma

Figure 1 for Are Soft Prompts Good Zero-shot Learners for Speech Recognition?
Figure 2 for Are Soft Prompts Good Zero-shot Learners for Speech Recognition?
Figure 3 for Are Soft Prompts Good Zero-shot Learners for Speech Recognition?
Figure 4 for Are Soft Prompts Good Zero-shot Learners for Speech Recognition?
Viaarxiv icon

Analysis of Speech Separation Performance Degradation on Emotional Speech Mixtures

Add code
Bookmark button
Alert button
Sep 14, 2023
Jia Qi Yip, Dianwen Ng, Bin Ma, Chng Eng Siong

Figure 1 for Analysis of Speech Separation Performance Degradation on Emotional Speech Mixtures
Figure 2 for Analysis of Speech Separation Performance Degradation on Emotional Speech Mixtures
Figure 3 for Analysis of Speech Separation Performance Degradation on Emotional Speech Mixtures
Figure 4 for Analysis of Speech Separation Performance Degradation on Emotional Speech Mixtures
Viaarxiv icon

ACA-Net: Towards Lightweight Speaker Verification using Asymmetric Cross Attention

Add code
Bookmark button
Alert button
May 20, 2023
Jia Qi Yip, Tuan Truong, Dianwen Ng, Chong Zhang, Yukun Ma, Trung Hieu Nguyen, Chongjia Ni, Shengkui Zhao, Eng Siong Chng, Bin Ma

Figure 1 for ACA-Net: Towards Lightweight Speaker Verification using Asymmetric Cross Attention
Figure 2 for ACA-Net: Towards Lightweight Speaker Verification using Asymmetric Cross Attention
Figure 3 for ACA-Net: Towards Lightweight Speaker Verification using Asymmetric Cross Attention
Figure 4 for ACA-Net: Towards Lightweight Speaker Verification using Asymmetric Cross Attention
Viaarxiv icon

Contrastive Speech Mixup for Low-resource Keyword Spotting

Add code
Bookmark button
Alert button
May 02, 2023
Dianwen Ng, Ruixi Zhang, Jia Qi Yip, Chong Zhang, Yukun Ma, Trung Hieu Nguyen, Chongjia Ni, Eng Siong Chng, Bin Ma

Figure 1 for Contrastive Speech Mixup for Low-resource Keyword Spotting
Figure 2 for Contrastive Speech Mixup for Low-resource Keyword Spotting
Figure 3 for Contrastive Speech Mixup for Low-resource Keyword Spotting
Figure 4 for Contrastive Speech Mixup for Low-resource Keyword Spotting
Viaarxiv icon

Adaptive Knowledge Distillation between Text and Speech Pre-trained Models

Add code
Bookmark button
Alert button
Mar 07, 2023
Jinjie Ni, Yukun Ma, Wen Wang, Qian Chen, Dianwen Ng, Han Lei, Trung Hieu Nguyen, Chong Zhang, Bin Ma, Erik Cambria

Figure 1 for Adaptive Knowledge Distillation between Text and Speech Pre-trained Models
Figure 2 for Adaptive Knowledge Distillation between Text and Speech Pre-trained Models
Figure 3 for Adaptive Knowledge Distillation between Text and Speech Pre-trained Models
Figure 4 for Adaptive Knowledge Distillation between Text and Speech Pre-trained Models
Viaarxiv icon

deHuBERT: Disentangling Noise in a Self-supervised Model for Robust Speech Recognition

Add code
Bookmark button
Alert button
Feb 28, 2023
Dianwen Ng, Ruixi Zhang, Jia Qi Yip, Zhao Yang, Jinjie Ni, Chong Zhang, Yukun Ma, Chongjia Ni, Eng Siong Chng, Bin Ma

Figure 1 for deHuBERT: Disentangling Noise in a Self-supervised Model for Robust Speech Recognition
Figure 2 for deHuBERT: Disentangling Noise in a Self-supervised Model for Robust Speech Recognition
Figure 3 for deHuBERT: Disentangling Noise in a Self-supervised Model for Robust Speech Recognition
Figure 4 for deHuBERT: Disentangling Noise in a Self-supervised Model for Robust Speech Recognition
Viaarxiv icon

I2CR: Improving Noise Robustness on Keyword Spotting Using Inter-Intra Contrastive Regularization

Add code
Bookmark button
Alert button
Sep 14, 2022
Dianwen Ng, Jia Qi Yip, Tanmay Surana, Zhao Yang, Chong Zhang, Yukun Ma, Chongjia Ni, Eng Siong Chng, Bin Ma

Figure 1 for I2CR: Improving Noise Robustness on Keyword Spotting Using Inter-Intra Contrastive Regularization
Figure 2 for I2CR: Improving Noise Robustness on Keyword Spotting Using Inter-Intra Contrastive Regularization
Figure 3 for I2CR: Improving Noise Robustness on Keyword Spotting Using Inter-Intra Contrastive Regularization
Figure 4 for I2CR: Improving Noise Robustness on Keyword Spotting Using Inter-Intra Contrastive Regularization
Viaarxiv icon