Alert button
Picture for Frank Zhang

Frank Zhang

Alert button

Improving RNN Transducer Based ASR with Auxiliary Tasks

Add code
Bookmark button
Alert button
Nov 09, 2020
Chunxi Liu, Frank Zhang, Duc Le, Suyoun Kim, Yatharth Saraf, Geoffrey Zweig

Figure 1 for Improving RNN Transducer Based ASR with Auxiliary Tasks
Figure 2 for Improving RNN Transducer Based ASR with Auxiliary Tasks
Figure 3 for Improving RNN Transducer Based ASR with Auxiliary Tasks
Figure 4 for Improving RNN Transducer Based ASR with Auxiliary Tasks
Viaarxiv icon

Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Nov 03, 2020
Ching-Feng Yeh, Yongqiang Wang, Yangyang Shi, Chunyang Wu, Frank Zhang, Julian Chan, Michael L. Seltzer

Figure 1 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 2 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 3 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 4 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Viaarxiv icon

Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications

Add code
Bookmark button
Alert button
Oct 29, 2020
Yongqiang Wang, Yangyang Shi, Frank Zhang, Chunyang Wu, Julian Chan, Ching-Feng Yeh, Alex Xiao

Figure 1 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 2 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 3 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 4 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Viaarxiv icon

Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition

Add code
Bookmark button
Alert button
Oct 29, 2020
Yangyang Shi, Yongqiang Wang, Chunyang Wu, Ching-Feng Yeh, Julian Chan, Frank Zhang, Duc Le, Mike Seltzer

Figure 1 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 2 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 3 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 4 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Viaarxiv icon

Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces

Add code
Bookmark button
Alert button
May 19, 2020
Frank Zhang, Yongqiang Wang, Xiaohui Zhang, Chunxi Liu, Yatharth Saraf, Geoffrey Zweig

Figure 1 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Figure 2 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Figure 3 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Figure 4 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Viaarxiv icon

Weak-Attention Suppression For Transformer Based Speech Recognition

Add code
Bookmark button
Alert button
May 18, 2020
Yangyang Shi, Yongqiang Wang, Chunyang Wu, Christian Fuegen, Frank Zhang, Duc Le, Ching-Feng Yeh, Michael L. Seltzer

Figure 1 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 2 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 3 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 4 for Weak-Attention Suppression For Transformer Based Speech Recognition
Viaarxiv icon

Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory

Add code
Bookmark button
Alert button
May 16, 2020
Chunyang Wu, Yongqiang Wang, Yangyang Shi, Ching-Feng Yeh, Frank Zhang

Figure 1 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 2 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 3 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 4 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Viaarxiv icon

Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model

Add code
Bookmark button
Alert button
May 15, 2020
Da-Rong Liu, Chunxi Liu, Frank Zhang, Gabriel Synnaeve, Yatharth Saraf, Geoffrey Zweig

Figure 1 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Figure 2 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Figure 3 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Figure 4 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Viaarxiv icon