Alert button
Picture for Hung-yi Lee

Hung-yi Lee

Alert button

EURO: ESPnet Unsupervised ASR Open-source Toolkit

Add code
Bookmark button
Alert button
Dec 01, 2022
Dongji Gao, Jiatong Shi, Shun-Po Chuang, Leibny Paola Garcia, Hung-yi Lee, Shinji Watanabe, Sanjeev Khudanpur

Figure 1 for EURO: ESPnet Unsupervised ASR Open-source Toolkit
Figure 2 for EURO: ESPnet Unsupervised ASR Open-source Toolkit
Figure 3 for EURO: ESPnet Unsupervised ASR Open-source Toolkit
Figure 4 for EURO: ESPnet Unsupervised ASR Open-source Toolkit
Viaarxiv icon

CHAPTER: Exploiting Convolutional Neural Network Adapters for Self-supervised Speech Models

Add code
Bookmark button
Alert button
Dec 01, 2022
Zih-Ching Chen, Yu-Shun Sung, Hung-yi Lee

Figure 1 for CHAPTER: Exploiting Convolutional Neural Network Adapters for Self-supervised Speech Models
Figure 2 for CHAPTER: Exploiting Convolutional Neural Network Adapters for Self-supervised Speech Models
Figure 3 for CHAPTER: Exploiting Convolutional Neural Network Adapters for Self-supervised Speech Models
Figure 4 for CHAPTER: Exploiting Convolutional Neural Network Adapters for Self-supervised Speech Models
Viaarxiv icon

Model Extraction Attack against Self-supervised Speech Models

Add code
Bookmark button
Alert button
Nov 29, 2022
Tsu-Yuan Hsu, Chen-An Li, Tung-Yu Wu, Hung-yi Lee

Figure 1 for Model Extraction Attack against Self-supervised Speech Models
Figure 2 for Model Extraction Attack against Self-supervised Speech Models
Figure 3 for Model Extraction Attack against Self-supervised Speech Models
Figure 4 for Model Extraction Attack against Self-supervised Speech Models
Viaarxiv icon

Compressing Transformer-based self-supervised models for speech processing

Add code
Bookmark button
Alert button
Nov 17, 2022
Tzu-Quan Lin, Tsung-Huan Yang, Chun-Yao Chang, Kuang-Ming Chen, Tzu-hsun Feng, Hung-yi Lee, Hao Tang

Figure 1 for Compressing Transformer-based self-supervised models for speech processing
Figure 2 for Compressing Transformer-based self-supervised models for speech processing
Figure 3 for Compressing Transformer-based self-supervised models for speech processing
Figure 4 for Compressing Transformer-based self-supervised models for speech processing
Viaarxiv icon

MelHuBERT: A simplified HuBERT on Mel spectrogram

Add code
Bookmark button
Alert button
Nov 17, 2022
Tzu-Quan Lin, Hung-yi Lee, Hao Tang

Figure 1 for MelHuBERT: A simplified HuBERT on Mel spectrogram
Figure 2 for MelHuBERT: A simplified HuBERT on Mel spectrogram
Figure 3 for MelHuBERT: A simplified HuBERT on Mel spectrogram
Figure 4 for MelHuBERT: A simplified HuBERT on Mel spectrogram
Viaarxiv icon

Introducing Semantics into Speech Encoders

Add code
Bookmark button
Alert button
Nov 15, 2022
Derek Xu, Shuyan Dong, Changhan Wang, Suyoun Kim, Zhaojiang Lin, Akshat Shrivastava, Shang-Wen Li, Liang-Hsuan Tseng, Alexei Baevski, Guan-Ting Lin, Hung-yi Lee, Yizhou Sun, Wei Wang

Figure 1 for Introducing Semantics into Speech Encoders
Figure 2 for Introducing Semantics into Speech Encoders
Figure 3 for Introducing Semantics into Speech Encoders
Figure 4 for Introducing Semantics into Speech Encoders
Viaarxiv icon

Bridging Speech and Textual Pre-trained Models with Unsupervised ASR

Add code
Bookmark button
Alert button
Nov 06, 2022
Jiatong Shi, Chan-Jan Hsu, Holam Chung, Dongji Gao, Paola Garcia, Shinji Watanabe, Ann Lee, Hung-yi Lee

Figure 1 for Bridging Speech and Textual Pre-trained Models with Unsupervised ASR
Figure 2 for Bridging Speech and Textual Pre-trained Models with Unsupervised ASR
Figure 3 for Bridging Speech and Textual Pre-trained Models with Unsupervised ASR
Figure 4 for Bridging Speech and Textual Pre-trained Models with Unsupervised ASR
Viaarxiv icon

Once-for-All Sequence Compression for Self-Supervised Speech Models

Add code
Bookmark button
Alert button
Nov 04, 2022
Hsuan-Jui Chen, Yen Meng, Hung-yi Lee

Figure 1 for Once-for-All Sequence Compression for Self-Supervised Speech Models
Figure 2 for Once-for-All Sequence Compression for Self-Supervised Speech Models
Figure 3 for Once-for-All Sequence Compression for Self-Supervised Speech Models
Viaarxiv icon

M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval

Add code
Bookmark button
Alert button
Nov 02, 2022
Layne Berry, Yi-Jen Shih, Hsuan-Fu Wang, Heng-Jui Chang, Hung-yi Lee, David Harwath

Figure 1 for M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval
Figure 2 for M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval
Figure 3 for M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval
Figure 4 for M-SpeechCLIP: Leveraging Large-Scale, Pre-Trained Models for Multilingual Speech to Image Retrieval
Viaarxiv icon

T5lephone: Bridging Speech and Text Self-supervised Models for Spoken Language Understanding via Phoneme level T5

Add code
Bookmark button
Alert button
Nov 01, 2022
Chan-Jan Hsu, Ho-Lam Chung, Hung-yi Lee, Yu Tsao

Figure 1 for T5lephone: Bridging Speech and Text Self-supervised Models for Spoken Language Understanding via Phoneme level T5
Figure 2 for T5lephone: Bridging Speech and Text Self-supervised Models for Spoken Language Understanding via Phoneme level T5
Figure 3 for T5lephone: Bridging Speech and Text Self-supervised Models for Spoken Language Understanding via Phoneme level T5
Figure 4 for T5lephone: Bridging Speech and Text Self-supervised Models for Spoken Language Understanding via Phoneme level T5
Viaarxiv icon