Alert button
Picture for Helen Meng

Helen Meng

Alert button

Speech Enhancement with Fullband-Subband Cross-Attention Network

Add code
Bookmark button
Alert button
Nov 10, 2022
Jun Chen, Wei Rao, Zilin Wang, Zhiyong Wu, Yannan Wang, Tao Yu, Shidong Shang, Helen Meng

Figure 1 for Speech Enhancement with Fullband-Subband Cross-Attention Network
Figure 2 for Speech Enhancement with Fullband-Subband Cross-Attention Network
Figure 3 for Speech Enhancement with Fullband-Subband Cross-Attention Network
Viaarxiv icon

Exploiting prompt learning with pre-trained language models for Alzheimer's Disease detection

Add code
Bookmark button
Alert button
Oct 29, 2022
Yi Wang, Jiajun Deng, Tianzi Wang, Bo Zheng, Shoukang Hu, Xunying Liu, Helen Meng

Figure 1 for Exploiting prompt learning with pre-trained language models for Alzheimer's Disease detection
Figure 2 for Exploiting prompt learning with pre-trained language models for Alzheimer's Disease detection
Figure 3 for Exploiting prompt learning with pre-trained language models for Alzheimer's Disease detection
Viaarxiv icon

Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations

Add code
Bookmark button
Alert button
Oct 27, 2022
Haohan Guo, Fenglong Xie, Xixin Wu, Hui Lu, Helen Meng

Figure 1 for Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations
Figure 2 for Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations
Figure 3 for Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations
Figure 4 for Towards High-Quality Neural TTS for Low-Resource Languages by Learning Compact Speech Representations
Viaarxiv icon

Disentangled Speech Representation Learning for One-Shot Cross-lingual Voice Conversion Using $β$-VAE

Add code
Bookmark button
Alert button
Oct 25, 2022
Hui Lu, Disong Wang, Xixin Wu, Zhiyong Wu, Xunying Liu, Helen Meng

Figure 1 for Disentangled Speech Representation Learning for One-Shot Cross-lingual Voice Conversion Using $β$-VAE
Figure 2 for Disentangled Speech Representation Learning for One-Shot Cross-lingual Voice Conversion Using $β$-VAE
Figure 3 for Disentangled Speech Representation Learning for One-Shot Cross-lingual Voice Conversion Using $β$-VAE
Figure 4 for Disentangled Speech Representation Learning for One-Shot Cross-lingual Voice Conversion Using $β$-VAE
Viaarxiv icon

Robust Unsupervised Cross-Lingual Word Embedding using Domain Flow Interpolation

Add code
Bookmark button
Alert button
Oct 07, 2022
Liping Tang, Zhen Li, Zhiquan Luo, Helen Meng

Figure 1 for Robust Unsupervised Cross-Lingual Word Embedding using Domain Flow Interpolation
Figure 2 for Robust Unsupervised Cross-Lingual Word Embedding using Domain Flow Interpolation
Figure 3 for Robust Unsupervised Cross-Lingual Word Embedding using Domain Flow Interpolation
Figure 4 for Robust Unsupervised Cross-Lingual Word Embedding using Domain Flow Interpolation
Viaarxiv icon

Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection

Add code
Bookmark button
Alert button
Oct 03, 2022
Xuanjun Chen, Haibin Wu, Helen Meng, Hung-yi Lee, Jyh-Shing Roger Jang

Figure 1 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Figure 2 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Figure 3 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Figure 4 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Viaarxiv icon

A Multi-Stage Multi-Codebook VQ-VAE Approach to High-Performance Neural TTS

Add code
Bookmark button
Alert button
Sep 22, 2022
Haohan Guo, Fenglong Xie, Frank K. Soong, Xixin Wu, Helen Meng

Figure 1 for A Multi-Stage Multi-Codebook VQ-VAE Approach to High-Performance Neural TTS
Figure 2 for A Multi-Stage Multi-Codebook VQ-VAE Approach to High-Performance Neural TTS
Figure 3 for A Multi-Stage Multi-Codebook VQ-VAE Approach to High-Performance Neural TTS
Figure 4 for A Multi-Stage Multi-Codebook VQ-VAE Approach to High-Performance Neural TTS
Viaarxiv icon

Bayesian Neural Network Language Modeling for Speech Recognition

Add code
Bookmark button
Alert button
Aug 28, 2022
Boyang Xue, Shoukang Hu, Junhao Xu, Mengzhe Geng, Xunying Liu, Helen Meng

Figure 1 for Bayesian Neural Network Language Modeling for Speech Recognition
Figure 2 for Bayesian Neural Network Language Modeling for Speech Recognition
Figure 3 for Bayesian Neural Network Language Modeling for Speech Recognition
Figure 4 for Bayesian Neural Network Language Modeling for Speech Recognition
Viaarxiv icon

Towards Cross-speaker Reading Style Transfer on Audiobook Dataset

Add code
Bookmark button
Alert button
Aug 19, 2022
Xiang Li, Changhe Song, Xianhao Wei, Zhiyong Wu, Jia Jia, Helen Meng

Figure 1 for Towards Cross-speaker Reading Style Transfer on Audiobook Dataset
Figure 2 for Towards Cross-speaker Reading Style Transfer on Audiobook Dataset
Figure 3 for Towards Cross-speaker Reading Style Transfer on Audiobook Dataset
Figure 4 for Towards Cross-speaker Reading Style Transfer on Audiobook Dataset
Viaarxiv icon

Speech Representation Disentanglement with Adversarial Mutual Information Learning for One-shot Voice Conversion

Add code
Bookmark button
Alert button
Aug 18, 2022
SiCheng Yang, Methawee Tantrawenith, Haolin Zhuang, Zhiyong Wu, Aolan Sun, Jianzong Wang, Ning Cheng, Huaizhen Tang, Xintao Zhao, Jie Wang, Helen Meng

Figure 1 for Speech Representation Disentanglement with Adversarial Mutual Information Learning for One-shot Voice Conversion
Figure 2 for Speech Representation Disentanglement with Adversarial Mutual Information Learning for One-shot Voice Conversion
Figure 3 for Speech Representation Disentanglement with Adversarial Mutual Information Learning for One-shot Voice Conversion
Figure 4 for Speech Representation Disentanglement with Adversarial Mutual Information Learning for One-shot Voice Conversion
Viaarxiv icon