Picture for Kwangyoun Kim

Kwangyoun Kim

DiscreteSLU: A Large Language Model with Self-Supervised Discrete Speech Units for Spoken Language Understanding

Add code
Jun 13, 2024
Viaarxiv icon

Improving ASR Contextual Biasing with Guided Attention

Add code
Jan 16, 2024
Figure 1 for Improving ASR Contextual Biasing with Guided Attention
Figure 2 for Improving ASR Contextual Biasing with Guided Attention
Figure 3 for Improving ASR Contextual Biasing with Guided Attention
Figure 4 for Improving ASR Contextual Biasing with Guided Attention
Viaarxiv icon

Generative Context-aware Fine-tuning of Self-supervised Speech Models

Add code
Dec 15, 2023
Viaarxiv icon

A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks

Add code
May 18, 2023
Figure 1 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Figure 2 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Figure 3 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Figure 4 for A Comparative Study on E-Branchformer vs Conformer in Speech Recognition, Translation, and Understanding Tasks
Viaarxiv icon

Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding

Add code
Feb 27, 2023
Figure 1 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 2 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 3 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Figure 4 for Structured Pruning of Self-Supervised Pre-trained Models for Speech Recognition and Understanding
Viaarxiv icon

Context-aware Fine-tuning of Self-supervised Speech Models

Add code
Dec 16, 2022
Figure 1 for Context-aware Fine-tuning of Self-supervised Speech Models
Figure 2 for Context-aware Fine-tuning of Self-supervised Speech Models
Figure 3 for Context-aware Fine-tuning of Self-supervised Speech Models
Figure 4 for Context-aware Fine-tuning of Self-supervised Speech Models
Viaarxiv icon

E-Branchformer: Branchformer with Enhanced merging for speech recognition

Add code
Sep 30, 2022
Figure 1 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 2 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 3 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Figure 4 for E-Branchformer: Branchformer with Enhanced merging for speech recognition
Viaarxiv icon

Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages

Add code
May 02, 2022
Figure 1 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 2 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 3 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Figure 4 for Wav2Seq: Pre-training Speech-to-Text Encoder-Decoder Models Using Pseudo Languages
Viaarxiv icon

SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition

Add code
Oct 11, 2021
Figure 1 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 2 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 3 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Figure 4 for SRU++: Pioneering Fast Recurrence with Attention for Speech Recognition
Viaarxiv icon

Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition

Add code
Sep 14, 2021
Figure 1 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Figure 2 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Figure 3 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Figure 4 for Performance-Efficiency Trade-offs in Unsupervised Pre-training for Speech Recognition
Viaarxiv icon