Alert button
Picture for Yu Tsao

Yu Tsao

Alert button

Speech Enhancement Based on Cyclegan with Noise-informed Training

Oct 19, 2021
Wen-Yuan Ting, Syu-Siang Wang, Hsin-Li Chang, Borching Su, Yu Tsao

Figure 1 for Speech Enhancement Based on Cyclegan with Noise-informed Training
Figure 2 for Speech Enhancement Based on Cyclegan with Noise-informed Training
Figure 3 for Speech Enhancement Based on Cyclegan with Noise-informed Training
Figure 4 for Speech Enhancement Based on Cyclegan with Noise-informed Training
Viaarxiv icon

Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments

Oct 19, 2021
Yun-Ju Chan, Chiang-Jen Peng, Syu-Siang Wang, Hsin-Min Wang, Yu Tsao, Tai-Shih Chi

Figure 1 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Figure 2 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Figure 3 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Figure 4 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Viaarxiv icon

MetricGAN-U: Unsupervised speech enhancement/ dereverberation based only on noisy/ reverberated speech

Oct 12, 2021
Szu-Wei Fu, Cheng Yu, Kuo-Hsuan Hung, Mirco Ravanelli, Yu Tsao

Figure 1 for MetricGAN-U: Unsupervised speech enhancement/ dereverberation based only on noisy/ reverberated speech
Figure 2 for MetricGAN-U: Unsupervised speech enhancement/ dereverberation based only on noisy/ reverberated speech
Figure 3 for MetricGAN-U: Unsupervised speech enhancement/ dereverberation based only on noisy/ reverberated speech
Figure 4 for MetricGAN-U: Unsupervised speech enhancement/ dereverberation based only on noisy/ reverberated speech
Viaarxiv icon

Analyzing the Robustness of Unsupervised Speech Recognition

Oct 12, 2021
Guan-Ting Lin, Chan-Jan Hsu, Da-Rong Liu, Hung-Yi Lee, Yu Tsao

Figure 1 for Analyzing the Robustness of Unsupervised Speech Recognition
Figure 2 for Analyzing the Robustness of Unsupervised Speech Recognition
Figure 3 for Analyzing the Robustness of Unsupervised Speech Recognition
Figure 4 for Analyzing the Robustness of Unsupervised Speech Recognition
Viaarxiv icon

An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition

Oct 09, 2021
Xuankai Chang, Takashi Maekaku, Pengcheng Guo, Jing Shi, Yen-Ju Lu, Aswin Shanmugam Subramanian, Tianzi Wang, Shu-wen Yang, Yu Tsao, Hung-yi Lee, Shinji Watanabe

Figure 1 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Figure 2 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Figure 3 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Figure 4 for An Exploration of Self-Supervised Pretrained Representations for End-to-End Speech Recognition
Viaarxiv icon

A Study of Low-Resource Speech Commands Recognition based on Adversarial Reprogramming

Oct 08, 2021
Hao Yen, Pin-Jui Ku, Chao-Han Huck Yang, Hu Hu, Sabato Marco Siniscalchi, Pin-Yu Chen, Yu Tsao

Figure 1 for A Study of Low-Resource Speech Commands Recognition based on Adversarial Reprogramming
Figure 2 for A Study of Low-Resource Speech Commands Recognition based on Adversarial Reprogramming
Figure 3 for A Study of Low-Resource Speech Commands Recognition based on Adversarial Reprogramming
Figure 4 for A Study of Low-Resource Speech Commands Recognition based on Adversarial Reprogramming
Viaarxiv icon

Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion

Sep 08, 2021
Yi-Syuan Liou, Wen-Chin Huang, Ming-Chi Yen, Shu-Wei Tsai, Yu-Huai Peng, Tomoki Toda, Yu Tsao, Hsin-Min Wang

Figure 1 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 2 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 3 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 4 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Viaarxiv icon

A Study on Speech Enhancement Based on Diffusion Probabilistic Model

Jul 25, 2021
Yen-Ju Lu, Yu Tsao, Shinji Watanabe

Figure 1 for A Study on Speech Enhancement Based on Diffusion Probabilistic Model
Figure 2 for A Study on Speech Enhancement Based on Diffusion Probabilistic Model
Figure 3 for A Study on Speech Enhancement Based on Diffusion Probabilistic Model
Figure 4 for A Study on Speech Enhancement Based on Diffusion Probabilistic Model
Viaarxiv icon

SVSNet: An End-to-end Speaker Voice Similarity Assessment Model

Jul 20, 2021
Cheng-Hung Hu, Yu-Huai Peng, Junichi Yamagishi, Yu Tsao, Hsin-Min Wang

Figure 1 for SVSNet: An End-to-end Speaker Voice Similarity Assessment Model
Figure 2 for SVSNet: An End-to-end Speaker Voice Similarity Assessment Model
Figure 3 for SVSNet: An End-to-end Speaker Voice Similarity Assessment Model
Figure 4 for SVSNet: An End-to-end Speaker Voice Similarity Assessment Model
Viaarxiv icon

Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder

Jun 10, 2021
Yi-Chiao Wu, Cheng-Hung Hu, Hung-Shin Lee, Yu-Huai Peng, Wen-Chin Huang, Yu Tsao, Hsin-Min Wang, Tomoki Toda

Figure 1 for Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder
Figure 2 for Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder
Figure 3 for Relational Data Selection for Data Augmentation of Speaker-dependent Multi-band MelGAN Vocoder
Viaarxiv icon