Picture for Ruoming Pang

Ruoming Pang

BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition

Add code
Oct 01, 2021
Figure 1 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Figure 2 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Figure 3 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Figure 4 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Viaarxiv icon

W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training

Add code
Aug 07, 2021
Figure 1 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Figure 2 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Figure 3 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Figure 4 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Viaarxiv icon

GSPMD: General and Scalable Parallelization for ML Computation Graphs

Add code
May 10, 2021
Figure 1 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Figure 2 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Figure 3 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Figure 4 for GSPMD: General and Scalable Parallelization for ML Computation Graphs
Viaarxiv icon

Scaling End-to-End Models for Large-Scale Multilingual ASR

Add code
Apr 30, 2021
Figure 1 for Scaling End-to-End Models for Large-Scale Multilingual ASR
Figure 2 for Scaling End-to-End Models for Large-Scale Multilingual ASR
Figure 3 for Scaling End-to-End Models for Large-Scale Multilingual ASR
Viaarxiv icon

Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models

Add code
Apr 25, 2021
Figure 1 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Figure 2 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Figure 3 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Figure 4 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Viaarxiv icon

Searching for Fast Model Families on Datacenter Accelerators

Add code
Feb 10, 2021
Figure 1 for Searching for Fast Model Families on Datacenter Accelerators
Figure 2 for Searching for Fast Model Families on Datacenter Accelerators
Figure 3 for Searching for Fast Model Families on Datacenter Accelerators
Figure 4 for Searching for Fast Model Families on Datacenter Accelerators
Viaarxiv icon

Transformer Based Deliberation for Two-Pass Speech Recognition

Add code
Jan 27, 2021
Figure 1 for Transformer Based Deliberation for Two-Pass Speech Recognition
Figure 2 for Transformer Based Deliberation for Two-Pass Speech Recognition
Figure 3 for Transformer Based Deliberation for Two-Pass Speech Recognition
Figure 4 for Transformer Based Deliberation for Two-Pass Speech Recognition
Viaarxiv icon

Cascaded encoders for unifying streaming and non-streaming ASR

Add code
Oct 27, 2020
Figure 1 for Cascaded encoders for unifying streaming and non-streaming ASR
Figure 2 for Cascaded encoders for unifying streaming and non-streaming ASR
Figure 3 for Cascaded encoders for unifying streaming and non-streaming ASR
Figure 4 for Cascaded encoders for unifying streaming and non-streaming ASR
Viaarxiv icon

Unsupervised Learning of Disentangled Speech Content and Style Representation

Add code
Oct 24, 2020
Figure 1 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Figure 2 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Figure 3 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Figure 4 for Unsupervised Learning of Disentangled Speech Content and Style Representation
Viaarxiv icon

Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data

Add code
Oct 22, 2020
Figure 1 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 2 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 3 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 4 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Viaarxiv icon