Alert button
Picture for Kwangyoun Kim

Kwangyoun Kim

Alert button

Improving ASR Contextual Biasing with Guided Attention

Jan 16, 2024
Jiyang Tang, Kwangyoun Kim, Suwon Shon, Felix Wu, Prashant Sridhar, Shinji Watanabe

Viaarxiv icon

Generative Context-aware Fine-tuning of Self-supervised Speech Models

Dec 15, 2023
Suwon Shon, Kwangyoun Kim, Prashant Sridhar, Yi-Te Hsu, Shinji Watanabe, Karen Livescu

Viaarxiv icon

A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks

May 18, 2023
Yifan Peng, Kwangyoun Kim, Felix Wu, Brian Yan, Siddhant Arora, William Chen, Jiyang Tang, Suwon Shon, Prashant Sridhar, Shinji Watanabe

Figure 1 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Figure 2 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Figure 3 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Figure 4 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Viaarxiv icon

Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding

Feb 27, 2023
Yifan Peng, Kwangyoun Kim, Felix Wu, Prashant Sridhar, Shinji Watanabe

Figure 1 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 2 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 3 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 4 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Viaarxiv icon

Context-aware Fine-tuning of Self-supervised Speech Models

Dec 16, 2022
Suwon Shon, Felix Wu, Kwangyoun Kim, Prashant Sridhar, Karen Livescu, Shinji Watanabe

Figure 1 for Context-aware Fine-tuning of Self-supervised Speech Models
Figure 2 for Context-aware Fine-tuning of Self-supervised Speech Models
Figure 3 for Context-aware Fine-tuning of Self-supervised Speech Models
Figure 4 for Context-aware Fine-tuning of Self-supervised Speech Models
Viaarxiv icon

E-Branchformer: Branchformer with Enhanced merging for speech recognition

Sep 30, 2022
Kwangyoun Kim, Felix Wu, Yifan Peng, Jing Pan, Prashant Sridhar, Kyu J. Han, Shinji Watanabe

Figure 1 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 2 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 3 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 4 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Viaarxiv icon

Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages

May 02, 2022
Felix Wu, Kwangyoun Kim, Shinji Watanabe, Kyu Han, Ryan McDonald, Kilian Q. Weinberger, Yoav Artzi

Figure 1 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 2 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 3 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 4 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Viaarxiv icon

SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition

Oct 11, 2021
Jing Pan, Tao Lei, Kwangyoun Kim, Kyu Han, Shinji Watanabe

Figure 1 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 2 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 3 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 4 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Viaarxiv icon

Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition

Sep 14, 2021
Felix Wu, Kwangyoun Kim, Jing Pan, Kyu Han, Kilian Q. Weinberger, Yoav Artzi

Figure 1 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Figure 2 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Figure 3 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Figure 4 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Viaarxiv icon

Multi-mode Transformer Transducer with Stochastic Future Context

Jun 17, 2021
Kwangyoun Kim, Felix Wu, Prashant Sridhar, Kyu J. Han, Shinji Watanabe

Figure 1 for Multi-mode Transformer Transducer with Stochastic Future Context
Figure 2 for Multi-mode Transformer Transducer with Stochastic Future Context
Figure 3 for Multi-mode Transformer Transducer with Stochastic Future Context
Viaarxiv icon