Picture for Dan Su

Dan Su

Celine

TeCANet: Temporal-Contextual Attention Network for Environment-Aware Speech Dereverberation

Add code
Mar 31, 2021
Figure 1 for TeCANet: Temporal-Contextual Attention Network for Environment-Aware Speech Dereverberation
Figure 2 for TeCANet: Temporal-Contextual Attention Network for Environment-Aware Speech Dereverberation
Figure 3 for TeCANet: Temporal-Contextual Attention Network for Environment-Aware Speech Dereverberation
Figure 4 for TeCANet: Temporal-Contextual Attention Network for Environment-Aware Speech Dereverberation
Viaarxiv icon

Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation

Add code
Mar 08, 2021
Figure 1 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Figure 2 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Figure 3 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Figure 4 for Sandglasset: A Light Multi-Granularity Self-attentive Network For Time-Domain Speech Separation
Viaarxiv icon

Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect

Add code
Mar 02, 2021
Figure 1 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Figure 2 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Figure 3 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Figure 4 for Tune-In: Training Under Negative Environments with Interference for Attention Networks Simulating Cocktail Party Effect
Viaarxiv icon

Contrastive Separative Coding for Self-supervised Representation Learning

Add code
Mar 01, 2021
Figure 1 for Contrastive Separative Coding for Self-supervised Representation Learning
Figure 2 for Contrastive Separative Coding for Self-supervised Representation Learning
Viaarxiv icon

VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention

Add code
Feb 12, 2021
Figure 1 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Figure 2 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Figure 3 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Figure 4 for VARA-TTS: Non-Autoregressive Text-to-Speech Synthesis based on Very Deep VAE with Residual Attention
Viaarxiv icon

Effective Low-Cost Time-Domain Audio Separation Using Globally Attentive Locally Recurrent Networks

Add code
Jan 13, 2021
Figure 1 for Effective Low-Cost Time-Domain Audio Separation Using Globally Attentive Locally Recurrent Networks
Figure 2 for Effective Low-Cost Time-Domain Audio Separation Using Globally Attentive Locally Recurrent Networks
Figure 3 for Effective Low-Cost Time-Domain Audio Separation Using Globally Attentive Locally Recurrent Networks
Figure 4 for Effective Low-Cost Time-Domain Audio Separation Using Globally Attentive Locally Recurrent Networks
Viaarxiv icon

Phonetic Posteriorgrams based Many-to-Many Singing Voice Conversion via Adversarial Training

Add code
Dec 03, 2020
Figure 1 for Phonetic Posteriorgrams based Many-to-Many Singing Voice Conversion via Adversarial Training
Figure 2 for Phonetic Posteriorgrams based Many-to-Many Singing Voice Conversion via Adversarial Training
Figure 3 for Phonetic Posteriorgrams based Many-to-Many Singing Voice Conversion via Adversarial Training
Figure 4 for Phonetic Posteriorgrams based Many-to-Many Singing Voice Conversion via Adversarial Training
Viaarxiv icon

Non-Autoregressive Transformer ASR with CTC-Enhanced Decoder Input

Add code
Oct 28, 2020
Figure 1 for Non-Autoregressive Transformer ASR with CTC-Enhanced Decoder Input
Figure 2 for Non-Autoregressive Transformer ASR with CTC-Enhanced Decoder Input
Figure 3 for Non-Autoregressive Transformer ASR with CTC-Enhanced Decoder Input
Figure 4 for Non-Autoregressive Transformer ASR with CTC-Enhanced Decoder Input
Viaarxiv icon

Replay and Synthetic Speech Detection with Res2net Architecture

Add code
Oct 28, 2020
Figure 1 for Replay and Synthetic Speech Detection with Res2net Architecture
Figure 2 for Replay and Synthetic Speech Detection with Res2net Architecture
Figure 3 for Replay and Synthetic Speech Detection with Res2net Architecture
Figure 4 for Replay and Synthetic Speech Detection with Res2net Architecture
Viaarxiv icon

Dimsum @LaySumm 20: BART-based Approach for Scientific Document Summarization

Add code
Oct 19, 2020
Figure 1 for Dimsum @LaySumm 20: BART-based Approach for Scientific Document Summarization
Figure 2 for Dimsum @LaySumm 20: BART-based Approach for Scientific Document Summarization
Viaarxiv icon