Alert button
Picture for Yi-Chen Chen

Yi-Chen Chen

Alert button

Speech Representation Learning Through Self-supervised Pretraining And Multi-task Finetuning

Oct 18, 2021
Yi-Chen Chen, Shu-wen Yang, Cheng-Kuang Lee, Simon See, Hung-yi Lee

Figure 1 for Speech Representation Learning Through Self-supervised Pretraining And Multi-task Finetuning
Figure 2 for Speech Representation Learning Through Self-supervised Pretraining And Multi-task Finetuning
Viaarxiv icon

SpeechNet: A Universal Modularized Model for Speech Processing Tasks

May 31, 2021
Yi-Chen Chen, Po-Han Chi, Shu-wen Yang, Kai-Wei Chang, Jheng-hao Lin, Sung-Feng Huang, Da-Rong Liu, Chi-Liang Liu, Cheng-Kuang Lee, Hung-yi Lee

Figure 1 for SpeechNet: A Universal Modularized Model for Speech Processing Tasks
Figure 2 for SpeechNet: A Universal Modularized Model for Speech Processing Tasks
Figure 3 for SpeechNet: A Universal Modularized Model for Speech Processing Tasks
Figure 4 for SpeechNet: A Universal Modularized Model for Speech Processing Tasks
Viaarxiv icon

Self-supervised Pre-training Reduces Label Permutation Instability of Speech Separation

Oct 29, 2020
Sung-Feng Huang, Shun-Po Chuang, Da-Rong Liu, Yi-Chen Chen, Gene-Ping Yang, Hung-yi Lee

Figure 1 for Self-supervised Pre-training Reduces Label Permutation Instability of Speech Separation
Figure 2 for Self-supervised Pre-training Reduces Label Permutation Instability of Speech Separation
Figure 3 for Self-supervised Pre-training Reduces Label Permutation Instability of Speech Separation
Figure 4 for Self-supervised Pre-training Reduces Label Permutation Instability of Speech Separation
Viaarxiv icon

DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation

May 13, 2020
Yi-Chen Chen, Jui-Yang Hsu, Cheng-Kuang Lee, Hung-yi Lee

Figure 1 for DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation
Figure 2 for DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation
Figure 3 for DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation
Figure 4 for DARTS-ASR: Differentiable Architecture Search for Multilingual Speech Recognition and Adaptation
Viaarxiv icon

AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition

Nov 27, 2019
Yi-Chen Chen, Zhaojun Yang, Ching-Feng Yeh, Mahaveer Jain, Michael L. Seltzer

Figure 1 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Figure 2 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Figure 3 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Figure 4 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Viaarxiv icon

From Semi-supervised to Almost-unsupervised Speech Recognition with Very-low Resource by Jointly Learning Phonetic Structures from Audio and Text Embeddings

Apr 10, 2019
Yi-Chen Chen, Sung-Feng Huang, Hung-yi Lee, Lin-shan Lee

Figure 1 for From Semi-supervised to Almost-unsupervised Speech Recognition with Very-low Resource by Jointly Learning Phonetic Structures from Audio and Text Embeddings
Figure 2 for From Semi-supervised to Almost-unsupervised Speech Recognition with Very-low Resource by Jointly Learning Phonetic Structures from Audio and Text Embeddings
Figure 3 for From Semi-supervised to Almost-unsupervised Speech Recognition with Very-low Resource by Jointly Learning Phonetic Structures from Audio and Text Embeddings
Figure 4 for From Semi-supervised to Almost-unsupervised Speech Recognition with Very-low Resource by Jointly Learning Phonetic Structures from Audio and Text Embeddings
Viaarxiv icon

Improved Audio Embeddings by Adjacency-Based Clustering with Applications in Spoken Term Detection

Nov 07, 2018
Sung-Feng Huang, Yi-Chen Chen, Hung-yi Lee, Lin-shan Lee

Figure 1 for Improved Audio Embeddings by Adjacency-Based Clustering with Applications in Spoken Term Detection
Figure 2 for Improved Audio Embeddings by Adjacency-Based Clustering with Applications in Spoken Term Detection
Figure 3 for Improved Audio Embeddings by Adjacency-Based Clustering with Applications in Spoken Term Detection
Figure 4 for Improved Audio Embeddings by Adjacency-Based Clustering with Applications in Spoken Term Detection
Viaarxiv icon

Almost-unsupervised Speech Recognition with Close-to-zero Resource Based on Phonetic Structures Learned from Very Small Unpaired Speech and Text Data

Oct 30, 2018
Yi-Chen Chen, Chia-Hao Shen, Sung-Feng Huang, Hung-yi Lee, Lin-shan Lee

Figure 1 for Almost-unsupervised Speech Recognition with Close-to-zero Resource Based on Phonetic Structures Learned from Very Small Unpaired Speech and Text Data
Figure 2 for Almost-unsupervised Speech Recognition with Close-to-zero Resource Based on Phonetic Structures Learned from Very Small Unpaired Speech and Text Data
Figure 3 for Almost-unsupervised Speech Recognition with Close-to-zero Resource Based on Phonetic Structures Learned from Very Small Unpaired Speech and Text Data
Figure 4 for Almost-unsupervised Speech Recognition with Close-to-zero Resource Based on Phonetic Structures Learned from Very Small Unpaired Speech and Text Data
Viaarxiv icon

Phonetic-and-Semantic Embedding of Spoken Words with Applications in Spoken Content Retrieval

Sep 03, 2018
Yi-Chen Chen, Sung-Feng Huang, Chia-Hao Shen, Hung-yi Lee, Lin-shan Lee

Figure 1 for Phonetic-and-Semantic Embedding of Spoken Words with Applications in Spoken Content Retrieval
Figure 2 for Phonetic-and-Semantic Embedding of Spoken Words with Applications in Spoken Content Retrieval
Figure 3 for Phonetic-and-Semantic Embedding of Spoken Words with Applications in Spoken Content Retrieval
Figure 4 for Phonetic-and-Semantic Embedding of Spoken Words with Applications in Spoken Content Retrieval
Viaarxiv icon