Alert button
Picture for Wen-Chin Huang

Wen-Chin Huang

Alert button

S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations

Add code
Bookmark button
Alert button
Oct 12, 2021
Wen-Chin Huang, Shu-Wen Yang, Tomoki Hayashi, Hung-Yi Lee, Shinji Watanabe, Tomoki Toda

Figure 1 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Figure 2 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Figure 3 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Figure 4 for S3PRL-VC: Open-source Voice Conversion Framework with Self-supervised Speech Representations
Viaarxiv icon

Generalization Ability of MOS Prediction Networks

Add code
Bookmark button
Alert button
Oct 06, 2021
Erica Cooper, Wen-Chin Huang, Tomoki Toda, Junichi Yamagishi

Figure 1 for Generalization Ability of MOS Prediction Networks
Figure 2 for Generalization Ability of MOS Prediction Networks
Figure 3 for Generalization Ability of MOS Prediction Networks
Figure 4 for Generalization Ability of MOS Prediction Networks
Viaarxiv icon

Noisy-to-Noisy Voice Conversion Framework with Denoising Model

Add code
Bookmark button
Alert button
Sep 22, 2021
Chao Xie, Yi-Chiao Wu, Patrick Lumban Tobing, Wen-Chin Huang, Tomoki Toda

Figure 1 for Noisy-to-Noisy Voice Conversion Framework with Denoising Model
Figure 2 for Noisy-to-Noisy Voice Conversion Framework with Denoising Model
Figure 3 for Noisy-to-Noisy Voice Conversion Framework with Denoising Model
Figure 4 for Noisy-to-Noisy Voice Conversion Framework with Denoising Model
Viaarxiv icon

Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion

Add code
Bookmark button
Alert button
Sep 08, 2021
Yi-Syuan Liou, Wen-Chin Huang, Ming-Chi Yen, Shu-Wei Tsai, Yu-Huai Peng, Tomoki Toda, Yu Tsao, Hsin-Min Wang

Figure 1 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 2 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 3 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 4 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Viaarxiv icon

On Prosody Modeling for ASR+TTS based Voice Conversion

Add code
Bookmark button
Alert button
Jul 20, 2021
Wen-Chin Huang, Tomoki Hayashi, Xinjian Li, Shinji Watanabe, Tomoki Toda

Figure 1 for On Prosody Modeling for ASR+TTS based Voice Conversion
Figure 2 for On Prosody Modeling for ASR+TTS based Voice Conversion
Figure 3 for On Prosody Modeling for ASR+TTS based Voice Conversion
Figure 4 for On Prosody Modeling for ASR+TTS based Voice Conversion
Viaarxiv icon

Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder

Add code
Bookmark button
Alert button
Jun 10, 2021
Yi-Chiao Wu, Cheng-Hung Hu, Hung-Shin Lee, Yu-Huai Peng, Wen-Chin Huang, Yu Tsao, Hsin-Min Wang, Tomoki Toda

Figure 1 for Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder
Figure 2 for Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder
Figure 3 for Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder
Viaarxiv icon

A Preliminary Study of a Two-Stage Paradigm for Preserving Speaker Identity in Dysarthric Voice Conversion

Add code
Bookmark button
Alert button
Jun 02, 2021
Wen-Chin Huang, Kazuhiro Kobayashi, Yu-Huai Peng, Ching-Feng Liu, Yu Tsao, Hsin-Min Wang, Tomoki Toda

Figure 1 for A Preliminary Study of a Two-Stage Paradigm for Preserving Speaker Identity in Dysarthric Voice Conversion
Figure 2 for A Preliminary Study of a Two-Stage Paradigm for Preserving Speaker Identity in Dysarthric Voice Conversion
Figure 3 for A Preliminary Study of a Two-Stage Paradigm for Preserving Speaker Identity in Dysarthric Voice Conversion
Figure 4 for A Preliminary Study of a Two-Stage Paradigm for Preserving Speaker Identity in Dysarthric Voice Conversion
Viaarxiv icon

Non-autoregressive sequence-to-sequence voice conversion

Add code
Bookmark button
Alert button
Apr 14, 2021
Tomoki Hayashi, Wen-Chin Huang, Kazuhiro Kobayashi, Tomoki Toda

Figure 1 for Non-autoregressive sequence-to-sequence voice conversion
Figure 2 for Non-autoregressive sequence-to-sequence voice conversion
Figure 3 for Non-autoregressive sequence-to-sequence voice conversion
Figure 4 for Non-autoregressive sequence-to-sequence voice conversion
Viaarxiv icon

The AS-NU System for the M2VoC Challenge

Add code
Bookmark button
Alert button
Apr 07, 2021
Cheng-Hung Hu, Yi-Chiao Wu, Wen-Chin Huang, Yu-Huai Peng, Yu-Wen Chen, Pin-Jui Ku, Tomoki Toda, Yu Tsao, Hsin-Min Wang

Figure 1 for The AS-NU System for the M2VoC Challenge
Figure 2 for The AS-NU System for the M2VoC Challenge
Figure 3 for The AS-NU System for the M2VoC Challenge
Figure 4 for The AS-NU System for the M2VoC Challenge
Viaarxiv icon

crank: An Open-Source Software for Nonparallel Voice Conversion Based on Vector-Quantized Variational Autoencoder

Add code
Bookmark button
Alert button
Mar 04, 2021
Kazuhiro Kobayashi, Wen-Chin Huang, Yi-Chiao Wu, Patrick Lumban Tobing, Tomoki Hayashi, Tomoki Toda

Figure 1 for crank: An Open-Source Software for Nonparallel Voice Conversion Based on Vector-Quantized Variational Autoencoder
Figure 2 for crank: An Open-Source Software for Nonparallel Voice Conversion Based on Vector-Quantized Variational Autoencoder
Viaarxiv icon