Alert button
Picture for Hung-yi Lee

Hung-yi Lee

Alert button

Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention

Add code
Bookmark button
Alert button
Apr 23, 2021
Hsuan Su, Jiun-Hao Jhan, Fan-yun Sun, Saurav Sahay, Hung-yi Lee

Figure 1 for Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention
Figure 2 for Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention
Figure 3 for Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention
Figure 4 for Put Chatbot into Its Interlocutor's Shoes: New Framework to Learn Chatbot Responding with Intention
Viaarxiv icon

Utilizing Self-supervised Representations for MOS Prediction

Add code
Bookmark button
Alert button
Apr 21, 2021
Wei-Cheng Tseng, Chien-yu Huang, Wei-Tsung Kao, Yist Y. Lin, Hung-yi Lee

Figure 1 for Utilizing Self-supervised Representations for MOS Prediction
Figure 2 for Utilizing Self-supervised Representations for MOS Prediction
Figure 3 for Utilizing Self-supervised Representations for MOS Prediction
Figure 4 for Utilizing Self-supervised Representations for MOS Prediction
Viaarxiv icon

S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations

Add code
Bookmark button
Alert button
Apr 07, 2021
Jheng-hao Lin, Yist Y. Lin, Chung-Ming Chien, Hung-yi Lee

Figure 1 for S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations
Figure 2 for S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations
Figure 3 for S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations
Figure 4 for S2VC: A Framework for Any-to-Any Voice Conversion with Self-Supervised Pretrained Representations
Viaarxiv icon

Non-autoregressive Mandarin-English Code-switching Speech Recognition with Pinyin Mask-CTC and Word Embedding Regularization

Add code
Bookmark button
Alert button
Apr 06, 2021
Shun-Po Chuang, Heng-Jui Chang, Sung-Feng Huang, Hung-yi Lee

Figure 1 for Non-autoregressive Mandarin-English Code-switching Speech Recognition with Pinyin Mask-CTC and Word Embedding Regularization
Figure 2 for Non-autoregressive Mandarin-English Code-switching Speech Recognition with Pinyin Mask-CTC and Word Embedding Regularization
Figure 3 for Non-autoregressive Mandarin-English Code-switching Speech Recognition with Pinyin Mask-CTC and Word Embedding Regularization
Figure 4 for Non-autoregressive Mandarin-English Code-switching Speech Recognition with Pinyin Mask-CTC and Word Embedding Regularization
Viaarxiv icon

Towards Lifelong Learning of End-to-end ASR

Add code
Bookmark button
Alert button
Apr 04, 2021
Heng-Jui Chang, Hung-yi Lee, Lin-shan Lee

Figure 1 for Towards Lifelong Learning of End-to-end ASR
Figure 2 for Towards Lifelong Learning of End-to-end ASR
Figure 3 for Towards Lifelong Learning of End-to-end ASR
Figure 4 for Towards Lifelong Learning of End-to-end ASR
Viaarxiv icon

Auto-KWS 2021 Challenge: Task, Datasets, and Baselines

Add code
Bookmark button
Alert button
Mar 31, 2021
Jingsong Wang, Yuxuan He, Chunyu Zhao, Qijie Shao, Wei-Wei Tu, Tom Ko, Hung-yi Lee, Lei Xie

Figure 1 for Auto-KWS 2021 Challenge: Task, Datasets, and Baselines
Figure 2 for Auto-KWS 2021 Challenge: Task, Datasets, and Baselines
Figure 3 for Auto-KWS 2021 Challenge: Task, Datasets, and Baselines
Figure 4 for Auto-KWS 2021 Challenge: Task, Datasets, and Baselines
Viaarxiv icon

Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech

Add code
Bookmark button
Alert button
Mar 20, 2021
Chung-Ming Chien, Jheng-Hao Lin, Chien-yu Huang, Po-chun Hsu, Hung-yi Lee

Figure 1 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Figure 2 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Figure 3 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Figure 4 for Investigating on Incorporating Pretrained and Learnable Speaker Representations for Multi-Speaker Multi-Style Text-to-Speech
Viaarxiv icon