Picture for Chung-Cheng Chiu

Chung-Cheng Chiu

Self-supervised Learning with Random-projection Quantizer for Speech Recognition

Add code
Feb 03, 2022
Figure 1 for Self-supervised Learning with Random-projection Quantizer for Speech Recognition
Figure 2 for Self-supervised Learning with Random-projection Quantizer for Speech Recognition
Figure 3 for Self-supervised Learning with Random-projection Quantizer for Speech Recognition
Figure 4 for Self-supervised Learning with Random-projection Quantizer for Speech Recognition
Viaarxiv icon

Cross-attention conformer for context modeling in speech enhancement for ASR

Add code
Oct 30, 2021
Figure 1 for Cross-attention conformer for context modeling in speech enhancement for ASR
Figure 2 for Cross-attention conformer for context modeling in speech enhancement for ASR
Figure 3 for Cross-attention conformer for context modeling in speech enhancement for ASR
Figure 4 for Cross-attention conformer for context modeling in speech enhancement for ASR
Viaarxiv icon

BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition

Add code
Oct 01, 2021
Figure 1 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Figure 2 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Figure 3 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Figure 4 for BigSSL: Exploring the Frontier of Large-Scale Semi-Supervised Learning for Automatic Speech Recognition
Viaarxiv icon

W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training

Add code
Aug 07, 2021
Figure 1 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Figure 2 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Figure 3 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Figure 4 for W2v-BERT: Combining Contrastive Learning and Masked Language Modeling for Self-Supervised Speech Pre-Training
Viaarxiv icon

Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models

Add code
Apr 25, 2021
Figure 1 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Figure 2 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Figure 3 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Figure 4 for Bridging the gap between streaming and non-streaming ASR systems bydistilling ensembles of CTC and RNN-T models
Viaarxiv icon

Pushing the Limits of Non-Autoregressive Speech Recognition

Add code
Apr 12, 2021
Figure 1 for Pushing the Limits of Non-Autoregressive Speech Recognition
Figure 2 for Pushing the Limits of Non-Autoregressive Speech Recognition
Figure 3 for Pushing the Limits of Non-Autoregressive Speech Recognition
Figure 4 for Pushing the Limits of Non-Autoregressive Speech Recognition
Viaarxiv icon

Cascaded encoders for unifying streaming and non-streaming ASR

Add code
Oct 27, 2020
Figure 1 for Cascaded encoders for unifying streaming and non-streaming ASR
Figure 2 for Cascaded encoders for unifying streaming and non-streaming ASR
Figure 3 for Cascaded encoders for unifying streaming and non-streaming ASR
Figure 4 for Cascaded encoders for unifying streaming and non-streaming ASR
Viaarxiv icon

Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data

Add code
Oct 22, 2020
Figure 1 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 2 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 3 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 4 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Viaarxiv icon

FastEmit: Low-latency Streaming ASR with Sequence-level Emission Regularization

Add code
Oct 21, 2020
Figure 1 for FastEmit: Low-latency Streaming ASR with Sequence-level Emission Regularization
Figure 2 for FastEmit: Low-latency Streaming ASR with Sequence-level Emission Regularization
Figure 3 for FastEmit: Low-latency Streaming ASR with Sequence-level Emission Regularization
Figure 4 for FastEmit: Low-latency Streaming ASR with Sequence-level Emission Regularization
Viaarxiv icon

Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition

Add code
Oct 20, 2020
Figure 1 for Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition
Figure 2 for Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition
Figure 3 for Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition
Figure 4 for Pushing the Limits of Semi-Supervised Learning for Automatic Speech Recognition
Viaarxiv icon