Alert button
Picture for Wen-Chin Huang

Wen-Chin Huang

Alert button

EMA2S: An End-to-End Multimodal Articulatory-to-Speech System

Add code
Bookmark button
Alert button
Feb 07, 2021
Yu-Wen Chen, Kuo-Hsuan Hung, Shang-Yi Chuang, Jonathan Sherman, Wen-Chin Huang, Xugang Lu, Yu Tsao

Figure 1 for EMA2S: An End-to-End Multimodal Articulatory-to-Speech System
Figure 2 for EMA2S: An End-to-End Multimodal Articulatory-to-Speech System
Figure 3 for EMA2S: An End-to-End Multimodal Articulatory-to-Speech System
Figure 4 for EMA2S: An End-to-End Multimodal Articulatory-to-Speech System
Viaarxiv icon

Speech Recognition by Simply Fine-tuning BERT

Add code
Bookmark button
Alert button
Jan 30, 2021
Wen-Chin Huang, Chia-Hua Wu, Shang-Bao Luo, Kuan-Yu Chen, Hsin-Min Wang, Tomoki Toda

Figure 1 for Speech Recognition by Simply Fine-tuning BERT
Figure 2 for Speech Recognition by Simply Fine-tuning BERT
Figure 3 for Speech Recognition by Simply Fine-tuning BERT
Figure 4 for Speech Recognition by Simply Fine-tuning BERT
Viaarxiv icon

The 2020 ESPnet update: new features, broadened applications, performance improvements, and future plans

Add code
Bookmark button
Alert button
Dec 23, 2020
Shinji Watanabe, Florian Boyer, Xuankai Chang, Pengcheng Guo, Tomoki Hayashi, Yosuke Higuchi, Takaaki Hori, Wen-Chin Huang, Hirofumi Inaguma, Naoyuki Kamo, Shigeki Karita, Chenda Li, Jing Shi, Aswin Shanmugam Subramanian, Wangyou Zhang

Figure 1 for The 2020 ESPnet update: new features, broadened applications, performance improvements, and future plans
Figure 2 for The 2020 ESPnet update: new features, broadened applications, performance improvements, and future plans
Viaarxiv icon

Any-to-One Sequence-to-Sequence Voice Conversion using Self-Supervised Discrete Speech Representations

Add code
Bookmark button
Alert button
Oct 23, 2020
Wen-Chin Huang, Yi-Chiao Wu, Tomoki Hayashi, Tomoki Toda

Figure 1 for Any-to-One Sequence-to-Sequence Voice Conversion using Self-Supervised Discrete Speech Representations
Figure 2 for Any-to-One Sequence-to-Sequence Voice Conversion using Self-Supervised Discrete Speech Representations
Figure 3 for Any-to-One Sequence-to-Sequence Voice Conversion using Self-Supervised Discrete Speech Representations
Figure 4 for Any-to-One Sequence-to-Sequence Voice Conversion using Self-Supervised Discrete Speech Representations
Viaarxiv icon

The NU Voice Conversion System for the Voice Conversion Challenge 2020: On the Effectiveness of Sequence-to-sequence Models and Autoregressive Neural Vocoders

Add code
Bookmark button
Alert button
Oct 09, 2020
Wen-Chin Huang, Patrick Lumban Tobing, Yi-Chiao Wu, Kazuhiro Kobayashi, Tomoki Toda

Figure 1 for The NU Voice Conversion System for the Voice Conversion Challenge 2020: On the Effectiveness of Sequence-to-sequence Models and Autoregressive Neural Vocoders
Figure 2 for The NU Voice Conversion System for the Voice Conversion Challenge 2020: On the Effectiveness of Sequence-to-sequence Models and Autoregressive Neural Vocoders
Figure 3 for The NU Voice Conversion System for the Voice Conversion Challenge 2020: On the Effectiveness of Sequence-to-sequence Models and Autoregressive Neural Vocoders
Figure 4 for The NU Voice Conversion System for the Voice Conversion Challenge 2020: On the Effectiveness of Sequence-to-sequence Models and Autoregressive Neural Vocoders
Viaarxiv icon

The Sequence-to-Sequence Baseline for the Voice Conversion Challenge 2020: Cascading ASR and TTS

Add code
Bookmark button
Alert button
Oct 06, 2020
Wen-Chin Huang, Tomoki Hayashi, Shinji Watanabe, Tomoki Toda

Figure 1 for The Sequence-to-Sequence Baseline for the Voice Conversion Challenge 2020: Cascading ASR and TTS
Figure 2 for The Sequence-to-Sequence Baseline for the Voice Conversion Challenge 2020: Cascading ASR and TTS
Figure 3 for The Sequence-to-Sequence Baseline for the Voice Conversion Challenge 2020: Cascading ASR and TTS
Figure 4 for The Sequence-to-Sequence Baseline for the Voice Conversion Challenge 2020: Cascading ASR and TTS
Viaarxiv icon

Pretraining Techniques for Sequence-to-Sequence Voice Conversion

Add code
Bookmark button
Alert button
Aug 07, 2020
Wen-Chin Huang, Tomoki Hayashi, Yi-Chiao Wu, Hirokazu Kameoka, Tomoki Toda

Figure 1 for Pretraining Techniques for Sequence-to-Sequence Voice Conversion
Figure 2 for Pretraining Techniques for Sequence-to-Sequence Voice Conversion
Figure 3 for Pretraining Techniques for Sequence-to-Sequence Voice Conversion
Figure 4 for Pretraining Techniques for Sequence-to-Sequence Voice Conversion
Viaarxiv icon

Many-to-Many Voice Transformer Network

Add code
Bookmark button
Alert button
Jun 07, 2020
Hirokazu Kameoka, Wen-Chin Huang, Kou Tanaka, Takuhiro Kaneko, Nobukatsu Hojo, Tomoki Toda

Figure 1 for Many-to-Many Voice Transformer Network
Figure 2 for Many-to-Many Voice Transformer Network
Figure 3 for Many-to-Many Voice Transformer Network
Figure 4 for Many-to-Many Voice Transformer Network
Viaarxiv icon

Unsupervised Representation Disentanglement using Cross Domain Features and Adversarial Learning in Variational Autoencoder based Voice Conversion

Add code
Bookmark button
Alert button
Feb 07, 2020
Wen-Chin Huang, Hao Luo, Hsin-Te Hwang, Chen-Chou Lo, Yu-Huai Peng, Yu Tsao, Hsin-Min Wang

Figure 1 for Unsupervised Representation Disentanglement using Cross Domain Features and Adversarial Learning in Variational Autoencoder based Voice Conversion
Figure 2 for Unsupervised Representation Disentanglement using Cross Domain Features and Adversarial Learning in Variational Autoencoder based Voice Conversion
Figure 3 for Unsupervised Representation Disentanglement using Cross Domain Features and Adversarial Learning in Variational Autoencoder based Voice Conversion
Figure 4 for Unsupervised Representation Disentanglement using Cross Domain Features and Adversarial Learning in Variational Autoencoder based Voice Conversion
Viaarxiv icon