Alert button
Picture for Zhuo Chen

Zhuo Chen

Alert button

Continuous Speech Separation with Recurrent Selective Attention Network

Add code
Bookmark button
Alert button
Oct 28, 2021
Yixuan Zhang, Zhuo Chen, Jian Wu, Takuya Yoshioka, Peidong Wang, Zhong Meng, Jinyu Li

Figure 1 for Continuous Speech Separation with Recurrent Selective Attention Network
Figure 2 for Continuous Speech Separation with Recurrent Selective Attention Network
Figure 3 for Continuous Speech Separation with Recurrent Selective Attention Network
Figure 4 for Continuous Speech Separation with Recurrent Selective Attention Network
Viaarxiv icon

Separating Long-Form Speech with Group-Wise Permutation Invariant Training

Add code
Bookmark button
Alert button
Oct 27, 2021
Wangyou Zhang, Zhuo Chen, Naoyuki Kanda, Shujie Liu, Jinyu Li, Sefik Emre Eskimez, Takuya Yoshioka, Xiong Xiao, Zhong Meng, Yanmin Qian, Furu Wei

Figure 1 for Separating Long-Form Speech with Group-Wise Permutation Invariant Training
Figure 2 for Separating Long-Form Speech with Group-Wise Permutation Invariant Training
Figure 3 for Separating Long-Form Speech with Group-Wise Permutation Invariant Training
Figure 4 for Separating Long-Form Speech with Group-Wise Permutation Invariant Training
Viaarxiv icon

VarArray: Array-Geometry-Agnostic Continuous Speech Separation

Add code
Bookmark button
Alert button
Oct 26, 2021
Takuya Yoshioka, Xiaofei Wang, Dongmei Wang, Min Tang, Zirun Zhu, Zhuo Chen, Naoyuki Kanda

Figure 1 for VarArray: Array-Geometry-Agnostic Continuous Speech Separation
Figure 2 for VarArray: Array-Geometry-Agnostic Continuous Speech Separation
Figure 3 for VarArray: Array-Geometry-Agnostic Continuous Speech Separation
Viaarxiv icon

One model to enhance them all: array geometry agnostic multi-channel personalized speech enhancement

Add code
Bookmark button
Alert button
Oct 20, 2021
Hassan Taherian, Sefik Emre Eskimez, Takuya Yoshioka, Huaming Wang, Zhuo Chen, Xuedong Huang

Figure 1 for One model to enhance them all: array geometry agnostic multi-channel personalized speech enhancement
Figure 2 for One model to enhance them all: array geometry agnostic multi-channel personalized speech enhancement
Figure 3 for One model to enhance them all: array geometry agnostic multi-channel personalized speech enhancement
Figure 4 for One model to enhance them all: array geometry agnostic multi-channel personalized speech enhancement
Viaarxiv icon

Personalized Speech Enhancement: New Models and Comprehensive Evaluation

Add code
Bookmark button
Alert button
Oct 18, 2021
Sefik Emre Eskimez, Takuya Yoshioka, Huaming Wang, Xiaofei Wang, Zhuo Chen, Xuedong Huang

Figure 1 for Personalized Speech Enhancement: New Models and Comprehensive Evaluation
Figure 2 for Personalized Speech Enhancement: New Models and Comprehensive Evaluation
Figure 3 for Personalized Speech Enhancement: New Models and Comprehensive Evaluation
Viaarxiv icon

All-neural beamformer for continuous speech separation

Add code
Bookmark button
Alert button
Oct 13, 2021
Zhuohuang Zhang, Takuya Yoshioka, Naoyuki Kanda, Zhuo Chen, Xiaofei Wang, Dongmei Wang, Sefik Emre Eskimez

Figure 1 for All-neural beamformer for continuous speech separation
Figure 2 for All-neural beamformer for continuous speech separation
Figure 3 for All-neural beamformer for continuous speech separation
Figure 4 for All-neural beamformer for continuous speech separation
Viaarxiv icon

UniSpeech-SAT: Universal Speech Representation Learning with Speaker Aware Pre-Training

Add code
Bookmark button
Alert button
Oct 12, 2021
Sanyuan Chen, Yu Wu, Chengyi Wang, Zhengyang Chen, Zhuo Chen, Shujie Liu, Jian Wu, Yao Qian, Furu Wei, Jinyu Li, Xiangzhan Yu

Figure 1 for UniSpeech-SAT: Universal Speech Representation Learning with Speaker Aware Pre-Training
Figure 2 for UniSpeech-SAT: Universal Speech Representation Learning with Speaker Aware Pre-Training
Figure 3 for UniSpeech-SAT: Universal Speech Representation Learning with Speaker Aware Pre-Training
Figure 4 for UniSpeech-SAT: Universal Speech Representation Learning with Speaker Aware Pre-Training
Viaarxiv icon

Transcribe-to-Diarize: Neural Speaker Diarization for Unlimited Number of Speakers using End-to-End Speaker-Attributed ASR

Add code
Bookmark button
Alert button
Oct 07, 2021
Naoyuki Kanda, Xiong Xiao, Yashesh Gaur, Xiaofei Wang, Zhong Meng, Zhuo Chen, Takuya Yoshioka

Figure 1 for Transcribe-to-Diarize: Neural Speaker Diarization for Unlimited Number of Speakers using End-to-End Speaker-Attributed ASR
Figure 2 for Transcribe-to-Diarize: Neural Speaker Diarization for Unlimited Number of Speakers using End-to-End Speaker-Attributed ASR
Figure 3 for Transcribe-to-Diarize: Neural Speaker Diarization for Unlimited Number of Speakers using End-to-End Speaker-Attributed ASR
Viaarxiv icon

Continuous Streaming Multi-Talker ASR with Dual-path Transducers

Add code
Bookmark button
Alert button
Sep 17, 2021
Desh Raj, Liang Lu, Zhuo Chen, Yashesh Gaur, Jinyu Li

Figure 1 for Continuous Streaming Multi-Talker ASR with Dual-path Transducers
Figure 2 for Continuous Streaming Multi-Talker ASR with Dual-path Transducers
Figure 3 for Continuous Streaming Multi-Talker ASR with Dual-path Transducers
Figure 4 for Continuous Streaming Multi-Talker ASR with Dual-path Transducers
Viaarxiv icon