Picture for Dan Su

Dan Su

Celine

Survey of Hallucination in Natural Language Generation

Add code
Feb 08, 2022
Figure 1 for Survey of Hallucination in Natural Language Generation
Figure 2 for Survey of Hallucination in Natural Language Generation
Figure 3 for Survey of Hallucination in Natural Language Generation
Figure 4 for Survey of Hallucination in Natural Language Generation
Viaarxiv icon

The CUHK-TENCENT speaker diarization system for the ICASSP 2022 multi-channel multi-party meeting transcription challenge

Add code
Feb 04, 2022
Figure 1 for The CUHK-TENCENT speaker diarization system for the ICASSP 2022 multi-channel multi-party meeting transcription challenge
Figure 2 for The CUHK-TENCENT speaker diarization system for the ICASSP 2022 multi-channel multi-party meeting transcription challenge
Figure 3 for The CUHK-TENCENT speaker diarization system for the ICASSP 2022 multi-channel multi-party meeting transcription challenge
Viaarxiv icon

DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs

Add code
Jan 28, 2022
Figure 1 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Figure 2 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Figure 3 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Figure 4 for DiffGAN-TTS: High-Fidelity and Efficient Text-to-Speech with Denoising Diffusion GANs
Viaarxiv icon

Consistent Training and Decoding For End-to-end Speech Recognition Using Lattice-free MMI

Add code
Dec 30, 2021
Figure 1 for Consistent Training and Decoding For End-to-end Speech Recognition Using Lattice-free MMI
Figure 2 for Consistent Training and Decoding For End-to-end Speech Recognition Using Lattice-free MMI
Figure 3 for Consistent Training and Decoding For End-to-end Speech Recognition Using Lattice-free MMI
Figure 4 for Consistent Training and Decoding For End-to-end Speech Recognition Using Lattice-free MMI
Viaarxiv icon

SpeechMoE2: Mixture-of-Experts Model with Improved Routing

Add code
Nov 23, 2021
Figure 1 for SpeechMoE2: Mixture-of-Experts Model with Improved Routing
Figure 2 for SpeechMoE2: Mixture-of-Experts Model with Improved Routing
Figure 3 for SpeechMoE2: Mixture-of-Experts Model with Improved Routing
Figure 4 for SpeechMoE2: Mixture-of-Experts Model with Improved Routing
Viaarxiv icon

Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning

Add code
Nov 14, 2021
Figure 1 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Figure 2 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Figure 3 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Figure 4 for Meta-Voice: Fast few-shot style transfer for expressive voice cloning using meta learning
Viaarxiv icon

Simple Attention Module based Speaker Verification with Iterative noisy label detection

Add code
Oct 13, 2021
Figure 1 for Simple Attention Module based Speaker Verification with Iterative noisy label detection
Figure 2 for Simple Attention Module based Speaker Verification with Iterative noisy label detection
Figure 3 for Simple Attention Module based Speaker Verification with Iterative noisy label detection
Figure 4 for Simple Attention Module based Speaker Verification with Iterative noisy label detection
Viaarxiv icon

AppQ: Warm-starting App Recommendation Based on View Graphs

Add code
Sep 08, 2021
Figure 1 for AppQ: Warm-starting App Recommendation Based on View Graphs
Figure 2 for AppQ: Warm-starting App Recommendation Based on View Graphs
Figure 3 for AppQ: Warm-starting App Recommendation Based on View Graphs
Figure 4 for AppQ: Warm-starting App Recommendation Based on View Graphs
Viaarxiv icon

Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis

Add code
Sep 08, 2021
Figure 1 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 2 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 3 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Figure 4 for Referee: Towards reference-free cross-speaker style transfer with low-quality data for expressive speech synthesis
Viaarxiv icon

Bilateral Denoising Diffusion Models

Add code
Aug 31, 2021
Figure 1 for Bilateral Denoising Diffusion Models
Figure 2 for Bilateral Denoising Diffusion Models
Figure 3 for Bilateral Denoising Diffusion Models
Figure 4 for Bilateral Denoising Diffusion Models
Viaarxiv icon