Alert button
Picture for Suyoun Kim

Suyoun Kim

Alert button

Cross-Attention End-to-End ASR for Two-Party Conversations

Add code
Bookmark button
Alert button
Jul 24, 2019
Suyoun Kim, Siddharth Dalmia, Florian Metze

Figure 1 for Cross-Attention End-to-End ASR for Two-Party Conversations
Figure 2 for Cross-Attention End-to-End ASR for Two-Party Conversations
Figure 3 for Cross-Attention End-to-End ASR for Two-Party Conversations
Viaarxiv icon

Gated Embeddings in End-to-End Speech Recognition for Conversational-Context Fusion

Add code
Bookmark button
Alert button
Jun 27, 2019
Suyoun Kim, Siddharth Dalmia, Florian Metze

Figure 1 for Gated Embeddings in End-to-End Speech Recognition for Conversational-Context Fusion
Figure 2 for Gated Embeddings in End-to-End Speech Recognition for Conversational-Context Fusion
Figure 3 for Gated Embeddings in End-to-End Speech Recognition for Conversational-Context Fusion
Figure 4 for Gated Embeddings in End-to-End Speech Recognition for Conversational-Context Fusion
Viaarxiv icon

Acoustic-to-Word Models with Conversational Context Information

Add code
Bookmark button
Alert button
May 21, 2019
Suyoun Kim, Florian Metze

Figure 1 for Acoustic-to-Word Models with Conversational Context Information
Figure 2 for Acoustic-to-Word Models with Conversational Context Information
Figure 3 for Acoustic-to-Word Models with Conversational Context Information
Figure 4 for Acoustic-to-Word Models with Conversational Context Information
Viaarxiv icon

Improved training for online end-to-end speech recognition systems

Add code
Bookmark button
Alert button
Aug 30, 2018
Suyoun Kim, Michael L. Seltzer, Jinyu Li, Rui Zhao

Figure 1 for Improved training for online end-to-end speech recognition systems
Figure 2 for Improved training for online end-to-end speech recognition systems
Figure 3 for Improved training for online end-to-end speech recognition systems
Viaarxiv icon

Dialog-context aware end-to-end speech recognition

Add code
Bookmark button
Alert button
Aug 07, 2018
Suyoun Kim, Florian Metze

Figure 1 for Dialog-context aware end-to-end speech recognition
Figure 2 for Dialog-context aware end-to-end speech recognition
Figure 3 for Dialog-context aware end-to-end speech recognition
Figure 4 for Dialog-context aware end-to-end speech recognition
Viaarxiv icon

Towards Language-Universal End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Nov 06, 2017
Suyoun Kim, Michael L. Seltzer

Figure 1 for Towards Language-Universal End-to-End Speech Recognition
Figure 2 for Towards Language-Universal End-to-End Speech Recognition
Figure 3 for Towards Language-Universal End-to-End Speech Recognition
Figure 4 for Towards Language-Universal End-to-End Speech Recognition
Viaarxiv icon

Joint CTC-Attention based End-to-End Speech Recognition using Multi-task Learning

Add code
Bookmark button
Alert button
Jan 31, 2017
Suyoun Kim, Takaaki Hori, Shinji Watanabe

Figure 1 for Joint CTC-Attention based End-to-End Speech Recognition using Multi-task Learning
Figure 2 for Joint CTC-Attention based End-to-End Speech Recognition using Multi-task Learning
Figure 3 for Joint CTC-Attention based End-to-End Speech Recognition using Multi-task Learning
Figure 4 for Joint CTC-Attention based End-to-End Speech Recognition using Multi-task Learning
Viaarxiv icon

Environmental Noise Embeddings for Robust Speech Recognition

Add code
Bookmark button
Alert button
Sep 29, 2016
Suyoun Kim, Bhiksha Raj, Ian Lane

Figure 1 for Environmental Noise Embeddings for Robust Speech Recognition
Figure 2 for Environmental Noise Embeddings for Robust Speech Recognition
Figure 3 for Environmental Noise Embeddings for Robust Speech Recognition
Figure 4 for Environmental Noise Embeddings for Robust Speech Recognition
Viaarxiv icon

Multimodal Transfer Deep Learning with Applications in Audio-Visual Recognition

Add code
Bookmark button
Alert button
Feb 18, 2016
Seungwhan Moon, Suyoun Kim, Haohan Wang

Figure 1 for Multimodal Transfer Deep Learning with Applications in Audio-Visual Recognition
Figure 2 for Multimodal Transfer Deep Learning with Applications in Audio-Visual Recognition
Figure 3 for Multimodal Transfer Deep Learning with Applications in Audio-Visual Recognition
Figure 4 for Multimodal Transfer Deep Learning with Applications in Audio-Visual Recognition
Viaarxiv icon

Recurrent Models for Auditory Attention in Multi-Microphone Distance Speech Recognition

Add code
Bookmark button
Alert button
Jan 07, 2016
Suyoun Kim, Ian Lane

Figure 1 for Recurrent Models for Auditory Attention in Multi-Microphone Distance Speech Recognition
Figure 2 for Recurrent Models for Auditory Attention in Multi-Microphone Distance Speech Recognition
Viaarxiv icon