Alert button
Picture for Po-chun Hsu

Po-chun Hsu

Alert button

Low-Resource Self-Supervised Learning with SSL-Enhanced TTS

Add code
Bookmark button
Alert button
Sep 29, 2023
Po-chun Hsu, Ali Elkahky, Wei-Ning Hsu, Yossi Adi, Tu Anh Nguyen, Jade Copet, Emmanuel Dupoux, Hung-yi Lee, Abdelrahman Mohamed

Figure 1 for Low-Resource Self-Supervised Learning with SSL-Enhanced TTS
Figure 2 for Low-Resource Self-Supervised Learning with SSL-Enhanced TTS
Figure 3 for Low-Resource Self-Supervised Learning with SSL-Enhanced TTS
Figure 4 for Low-Resource Self-Supervised Learning with SSL-Enhanced TTS
Viaarxiv icon

Learning Phone Recognition from Unpaired Audio and Phone Sequences Based on Generative Adversarial Network

Add code
Bookmark button
Alert button
Jul 29, 2022
Da-rong Liu, Po-chun Hsu, Yi-chen Chen, Sung-feng Huang, Shun-po Chuang, Da-yi Wu, Hung-yi Lee

Figure 1 for Learning Phone Recognition from Unpaired Audio and Phone Sequences Based on Generative Adversarial Network
Figure 2 for Learning Phone Recognition from Unpaired Audio and Phone Sequences Based on Generative Adversarial Network
Figure 3 for Learning Phone Recognition from Unpaired Audio and Phone Sequences Based on Generative Adversarial Network
Figure 4 for Learning Phone Recognition from Unpaired Audio and Phone Sequences Based on Generative Adversarial Network
Viaarxiv icon

Silence is Sweeter Than Speech: Self-Supervised Model Using Silence to Store Speaker Information

Add code
Bookmark button
Alert button
May 08, 2022
Chi-Luen Feng, Po-chun Hsu, Hung-yi Lee

Figure 1 for Silence is Sweeter Than Speech: Self-Supervised Model Using Silence to Store Speaker Information
Figure 2 for Silence is Sweeter Than Speech: Self-Supervised Model Using Silence to Store Speaker Information
Figure 3 for Silence is Sweeter Than Speech: Self-Supervised Model Using Silence to Store Speaker Information
Figure 4 for Silence is Sweeter Than Speech: Self-Supervised Model Using Silence to Store Speaker Information
Viaarxiv icon

Parallel Synthesis for Autoregressive Speech Generation

Add code
Bookmark button
Alert button
Apr 25, 2022
Po-chun Hsu, Da-rong Liu, Andy T. Liu, Hung-yi Lee

Figure 1 for Parallel Synthesis for Autoregressive Speech Generation
Figure 2 for Parallel Synthesis for Autoregressive Speech Generation
Figure 3 for Parallel Synthesis for Autoregressive Speech Generation
Figure 4 for Parallel Synthesis for Autoregressive Speech Generation
Viaarxiv icon

Universal Adaptor: Converting Mel-Spectrograms Between Different Configurations for Speech Synthesis

Add code
Bookmark button
Alert button
Apr 01, 2022
Fan-Lin Wang, Po-chun Hsu, Da-rong Liu, Hung-yi Lee

Figure 1 for Universal Adaptor: Converting Mel-Spectrograms Between Different Configurations for Speech Synthesis
Figure 2 for Universal Adaptor: Converting Mel-Spectrograms Between Different Configurations for Speech Synthesis
Figure 3 for Universal Adaptor: Converting Mel-Spectrograms Between Different Configurations for Speech Synthesis
Figure 4 for Universal Adaptor: Converting Mel-Spectrograms Between Different Configurations for Speech Synthesis
Viaarxiv icon

Spotting adversarial samples for speaker verification by neural vocoders

Add code
Bookmark button
Alert button
Jul 02, 2021
Haibin Wu, Po-chun Hsu, Ji Gao, Shanshan Zhang, Shen Huang, Jian Kang, Zhiyong Wu, Helen Meng, Hung-yi Lee

Figure 1 for Spotting adversarial samples for speaker verification by neural vocoders
Figure 2 for Spotting adversarial samples for speaker verification by neural vocoders
Figure 3 for Spotting adversarial samples for speaker verification by neural vocoders
Figure 4 for Spotting adversarial samples for speaker verification by neural vocoders
Viaarxiv icon

Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech

Add code
Bookmark button
Alert button
Mar 20, 2021
Chung-Ming Chien, Jheng-Hao Lin, Chien-yu Huang, Po-chun Hsu, Hung-yi Lee

Figure 1 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Figure 2 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Figure 3 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Figure 4 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Viaarxiv icon

Towards Robust Neural Vocoding for Speech Generation: A Survey

Add code
Bookmark button
Alert button
Dec 05, 2019
Po-chun Hsu, Chun-hsuan Wang, Andy T. Liu, Hung-yi Lee

Figure 1 for Towards Robust Neural Vocoding for Speech Generation: A Survey
Figure 2 for Towards Robust Neural Vocoding for Speech Generation: A Survey
Figure 3 for Towards Robust Neural Vocoding for Speech Generation: A Survey
Figure 4 for Towards Robust Neural Vocoding for Speech Generation: A Survey
Viaarxiv icon