Picture for Frank Zhang

Frank Zhang

Jack

Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition

Add code
Nov 03, 2020
Figure 1 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 2 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 3 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Figure 4 for Streaming Attention-Based Models with Augmented Memory for End-to-End Speech Recognition
Viaarxiv icon

Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications

Add code
Oct 29, 2020
Figure 1 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 2 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 3 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Figure 4 for Transformer in action: a comparative study of transformer-based acoustic models for large scale speech recognition applications
Viaarxiv icon

Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition

Add code
Oct 29, 2020
Figure 1 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 2 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 3 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Figure 4 for Emformer: Efficient Memory Transformer Based Acoustic Model For Low Latency Streaming Speech Recognition
Viaarxiv icon

Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces

Add code
May 19, 2020
Figure 1 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Figure 2 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Figure 3 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Figure 4 for Fast, Simpler and More Accurate Hybrid ASR Systems Using Wordpieces
Viaarxiv icon

Weak-Attention Suppression For Transformer Based Speech Recognition

Add code
May 18, 2020
Figure 1 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 2 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 3 for Weak-Attention Suppression For Transformer Based Speech Recognition
Figure 4 for Weak-Attention Suppression For Transformer Based Speech Recognition
Viaarxiv icon

Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory

Add code
May 16, 2020
Figure 1 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 2 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 3 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Figure 4 for Streaming Transformer-based Acoustic Models Using Self-attention with Augmented Memory
Viaarxiv icon

Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model

Add code
May 15, 2020
Figure 1 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Figure 2 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Figure 3 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Figure 4 for Contextualizing ASR Lattice Rescoring with Hybrid Pointer Network Language Model
Viaarxiv icon

Training ASR models by Generation of Contextual Information

Add code
Oct 27, 2019
Figure 1 for Training ASR models by Generation of Contextual Information
Figure 2 for Training ASR models by Generation of Contextual Information
Figure 3 for Training ASR models by Generation of Contextual Information
Figure 4 for Training ASR models by Generation of Contextual Information
Viaarxiv icon

Deja-vu: Double Feature Presentation in Deep Transformer Networks

Add code
Oct 23, 2019
Figure 1 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Figure 2 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Figure 3 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Figure 4 for Deja-vu: Double Feature Presentation in Deep Transformer Networks
Viaarxiv icon

Transformer-based Acoustic Modeling for Hybrid Speech Recognition

Add code
Oct 22, 2019
Figure 1 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Figure 2 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Figure 3 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Figure 4 for Transformer-based Acoustic Modeling for Hybrid Speech Recognition
Viaarxiv icon