Alert button
Picture for Ching-Feng Yeh

Ching-Feng Yeh

Alert button

Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Nov 03, 2020
Ching-Feng Yeh, Yongqiang Wang, Yangyang Shi, Chunyang Wu, Frank Zhang, Julian Chan, Michael L. Seltzer

Figure 1 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 2 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 3 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 4 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Viaarxiv icon

Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications

Add code
Bookmark button
Alert button
Oct 29, 2020
Yongqiang Wang, Yangyang Shi, Frank Zhang, Chunyang Wu, Julian Chan, Ching-Feng Yeh, Alex Xiao

Figure 1 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 2 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 3 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 4 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Viaarxiv icon

Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition

Add code
Bookmark button
Alert button
Oct 29, 2020
Yangyang Shi, Yongqiang Wang, Chunyang Wu, Ching-Feng Yeh, Julian Chan, Frank Zhang, Duc Le, Mike Seltzer

Figure 1 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 2 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 3 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 4 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Viaarxiv icon

Weak-Attention Suppression For Transformer Based Speech Recognition

Add code
Bookmark button
Alert button
May 18, 2020
Yangyang Shi, Yongqiang Wang, Chunyang Wu, Christian Fuegen, Frank Zhang, Duc Le, Ching-Feng Yeh, Michael L. Seltzer

Figure 1 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 2 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 3 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 4 for Weak-Attention Suppression For Transformer Based Speech Recognition
Viaarxiv icon

Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory

Add code
Bookmark button
Alert button
May 16, 2020
Chunyang Wu, Yongqiang Wang, Yangyang Shi, Ching-Feng Yeh, Frank Zhang

Figure 1 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 2 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 3 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 4 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Viaarxiv icon

AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition

Add code
Bookmark button
Alert button
Nov 27, 2019
Yi-Chen Chen, Zhaojun Yang, Ching-Feng Yeh, Mahaveer Jain, Michael L. Seltzer

Figure 1 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Figure 2 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Figure 3 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Figure 4 for AIPNet: Generative Adversarial Pre-training of Accent-invariant Networks for End-to-end Speech Recognition
Viaarxiv icon

RNN-T For Latency Controlled ASR With Improved Beam Search

Add code
Bookmark button
Alert button
Nov 05, 2019
Mahaveer Jain, Kjell Schubert, Jay Mahadeokar, Ching-Feng Yeh, Kaustubh Kalgaonkar, Anuroop Sriram, Christian Fuegen, Michael L. Seltzer

Figure 1 for RNN-T For Latency Controlled ASR With Improved Beam Search
Figure 2 for RNN-T For Latency Controlled ASR With Improved Beam Search
Figure 3 for RNN-T For Latency Controlled ASR With Improved Beam Search
Figure 4 for RNN-T For Latency Controlled ASR With Improved Beam Search
Viaarxiv icon

Transformer-Transducer: End-to-End Speech Recognition with Self-Attention

Add code
Bookmark button
Alert button
Oct 28, 2019
Ching-Feng Yeh, Jay Mahadeokar, Kaustubh Kalgaonkar, Yongqiang Wang, Duc Le, Mahaveer Jain, Kjell Schubert, Christian Fuegen, Michael L. Seltzer

Figure 1 for Transformer-Transducer: End-to-End Speech Recognition with Self-Attention
Figure 2 for Transformer-Transducer: End-to-End Speech Recognition with Self-Attention
Figure 3 for Transformer-Transducer: End-to-End Speech Recognition with Self-Attention
Figure 4 for Transformer-Transducer: End-to-End Speech Recognition with Self-Attention
Viaarxiv icon