Picture for Zhengkun Tian

Zhengkun Tian

Half-Truth: A Partially Fake Audio Detection Dataset

Add code
Apr 08, 2021
Figure 1 for Half-Truth: A Partially Fake Audio Detection Dataset
Figure 2 for Half-Truth: A Partially Fake Audio Detection Dataset
Figure 3 for Half-Truth: A Partially Fake Audio Detection Dataset
Figure 4 for Half-Truth: A Partially Fake Audio Detection Dataset
Viaarxiv icon

FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization

Add code
Apr 07, 2021
Figure 1 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Figure 2 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Figure 3 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Figure 4 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Viaarxiv icon

TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition

Add code
Apr 04, 2021
Figure 1 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Figure 2 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Figure 3 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Figure 4 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Viaarxiv icon

Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT

Add code
Feb 20, 2021
Figure 1 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Figure 2 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Figure 3 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Figure 4 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Viaarxiv icon

Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition

Add code
Nov 09, 2020
Figure 1 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Figure 2 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Figure 3 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Figure 4 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Viaarxiv icon

Decoupling Pronunciation and Language for End-to-end Code-switching Automatic Speech Recognition

Add code
Oct 28, 2020
Figure 1 for Decoupling Pronunciation and Language for End-to-end Code-switching Automatic Speech Recognition
Figure 2 for Decoupling Pronunciation and Language for End-to-end Code-switching Automatic Speech Recognition
Figure 3 for Decoupling Pronunciation and Language for End-to-end Code-switching Automatic Speech Recognition
Figure 4 for Decoupling Pronunciation and Language for End-to-end Code-switching Automatic Speech Recognition
Viaarxiv icon

Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition

Add code
May 30, 2020
Figure 1 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Figure 2 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Figure 3 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Figure 4 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Viaarxiv icon

Spike-Triggered Non-Autoregressive Transformer for End-to-End Speech Recognition

Add code
May 16, 2020
Figure 1 for Spike-Triggered Non-Autoregressive Transformer for End-to-End Speech Recognition
Figure 2 for Spike-Triggered Non-Autoregressive Transformer for End-to-End Speech Recognition
Figure 3 for Spike-Triggered Non-Autoregressive Transformer for End-to-End Speech Recognition
Figure 4 for Spike-Triggered Non-Autoregressive Transformer for End-to-End Speech Recognition
Viaarxiv icon

Adversarial Transfer Learning for Punctuation Restoration

Add code
Apr 01, 2020
Figure 1 for Adversarial Transfer Learning for Punctuation Restoration
Figure 2 for Adversarial Transfer Learning for Punctuation Restoration
Figure 3 for Adversarial Transfer Learning for Punctuation Restoration
Figure 4 for Adversarial Transfer Learning for Punctuation Restoration
Viaarxiv icon

Rnn-transducer with language bias for end-to-end Mandarin-English code-switching speech recognition

Add code
Feb 19, 2020
Figure 1 for Rnn-transducer with language bias for end-to-end Mandarin-English code-switching speech recognition
Figure 2 for Rnn-transducer with language bias for end-to-end Mandarin-English code-switching speech recognition
Figure 3 for Rnn-transducer with language bias for end-to-end Mandarin-English code-switching speech recognition
Figure 4 for Rnn-transducer with language bias for end-to-end Mandarin-English code-switching speech recognition
Viaarxiv icon