Alert button
Picture for Zhengqi Wen

Zhengqi Wen

Alert button

ADD 2022: the First Audio Deep Synthesis Detection Challenge

Feb 17, 2022
Jiangyan Yi, Ruibo Fu, Jianhua Tao, Shuai Nie, Haoxin Ma, Chenglong Wang, Tao Wang, Zhengkun Tian, Ye Bai, Cunhang Fan, Shan Liang, Shiming Wang, Shuai Zhang, Xinrui Yan, Le Xu, Zhengqi Wen, Haizhou Li

Figure 1 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Figure 2 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Figure 3 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Figure 4 for ADD 2022: the First Audio Deep Synthesis Detection Challenge
Viaarxiv icon

Singing-Tacotron: Global duration control attention and dynamic filter for End-to-end singing voice synthesis

Feb 16, 2022
Tao Wang, Ruibo Fu, Jiangyan Yi, Jianhua Tao, Zhengqi Wen

Figure 1 for Singing-Tacotron: Global duration control attention and dynamic filter for End-to-end singing voice synthesis
Figure 2 for Singing-Tacotron: Global duration control attention and dynamic filter for End-to-end singing voice synthesis
Figure 3 for Singing-Tacotron: Global duration control attention and dynamic filter for End-to-end singing voice synthesis
Figure 4 for Singing-Tacotron: Global duration control attention and dynamic filter for End-to-end singing voice synthesis
Viaarxiv icon

FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization

Apr 07, 2021
Zhengkun Tian, Jiangyan Yi, Ye Bai, Jianhua Tao, Shuai Zhang, Zhengqi Wen

Figure 1 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Figure 2 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Figure 3 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Figure 4 for FSR: Accelerating the Inference Process of Transducer-Based Models by Applying Fast-Skip Regularization
Viaarxiv icon

TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition

Apr 04, 2021
Zhengkun Tian, Jiangyan Yi, Jianhua Tao, Ye Bai, Shuai Zhang, Zhengqi Wen, Xuefei Liu

Figure 1 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Figure 2 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Figure 3 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Figure 4 for TSNAT: Two-Step Non-Autoregressvie Transformer Models for Speech Recognition
Viaarxiv icon

Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT

Feb 20, 2021
Ye Bai, Jiangyan Yi, Jianhua Tao, Zhengkun Tian, Zhengqi Wen, Shuai Zhang

Figure 1 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Figure 2 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Figure 3 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Figure 4 for Fast End-to-End Speech Recognition via a Non-Autoregressive Model and Cross-Modal Knowledge Transferring from BERT
Viaarxiv icon

Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT

Feb 15, 2021
Ye Bai, Jiangyan Yi, Jianhua Tao, Zhengkun Tian, Zhengqi Wen, Shuai Zhang

Figure 1 for Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT
Figure 2 for Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT
Figure 3 for Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT
Figure 4 for Fast End-to-End Speech Recognition via Non-Autoregressive Models and Cross-Modal Knowledge Transferring from BERT
Viaarxiv icon

Deep Time Delay Neural Network for Speech Enhancement with Full Data Learning

Nov 11, 2020
Cunhang Fan, Bin Liu, Jianhua Tao, Jiangyan Yi, Zhengqi Wen, Leichao Song

Figure 1 for Deep Time Delay Neural Network for Speech Enhancement with Full Data Learning
Figure 2 for Deep Time Delay Neural Network for Speech Enhancement with Full Data Learning
Figure 3 for Deep Time Delay Neural Network for Speech Enhancement with Full Data Learning
Figure 4 for Deep Time Delay Neural Network for Speech Enhancement with Full Data Learning
Viaarxiv icon

Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition

Nov 09, 2020
Cunhang Fan, Jiangyan Yi, Jianhua Tao, Zhengkun Tian, Bin Liu, Zhengqi Wen

Figure 1 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Figure 2 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Figure 3 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Figure 4 for Gated Recurrent Fusion with Joint Training Framework for Robust End-to-End Speech Recognition
Viaarxiv icon

Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition

May 30, 2020
Ye Bai, Jiangyan Yi, Jianhua Tao, Zhengkun Tian, Zhengqi Wen, Shuai Zhang

Figure 1 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Figure 2 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Figure 3 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Figure 4 for Listen Attentively, and Spell Once: Whole Sentence Generation via a Non-Autoregressive Architecture for Low-Latency Speech Recognition
Viaarxiv icon