Picture for Hsin-Min Wang

Hsin-Min Wang

The VoiceMOS Challenge 2022

Add code
Mar 28, 2022
Figure 1 for The VoiceMOS Challenge 2022
Figure 2 for The VoiceMOS Challenge 2022
Figure 3 for The VoiceMOS Challenge 2022
Figure 4 for The VoiceMOS Challenge 2022
Viaarxiv icon

Subspace-based Representation and Learning for Phonotactic Spoken Language Recognition

Add code
Mar 28, 2022
Figure 1 for Subspace-based Representation and Learning for Phonotactic Spoken Language Recognition
Figure 2 for Subspace-based Representation and Learning for Phonotactic Spoken Language Recognition
Figure 3 for Subspace-based Representation and Learning for Phonotactic Spoken Language Recognition
Figure 4 for Subspace-based Representation and Learning for Phonotactic Spoken Language Recognition
Viaarxiv icon

Speech-enhanced and Noise-aware Networks for Robust Speech Recognition

Add code
Mar 25, 2022
Figure 1 for Speech-enhanced and Noise-aware Networks for Robust Speech Recognition
Figure 2 for Speech-enhanced and Noise-aware Networks for Robust Speech Recognition
Figure 3 for Speech-enhanced and Noise-aware Networks for Robust Speech Recognition
Figure 4 for Speech-enhanced and Noise-aware Networks for Robust Speech Recognition
Viaarxiv icon

Partially Fake Audio Detection by Self-attention-based Fake Span Discovery

Add code
Feb 15, 2022
Figure 1 for Partially Fake Audio Detection by Self-attention-based Fake Span Discovery
Figure 2 for Partially Fake Audio Detection by Self-attention-based Fake Span Discovery
Figure 3 for Partially Fake Audio Detection by Self-attention-based Fake Span Discovery
Figure 4 for Partially Fake Audio Detection by Self-attention-based Fake Span Discovery
Viaarxiv icon

EMGSE: Acoustic/EMG Fusion for Multimodal Speech Enhancement

Add code
Feb 14, 2022
Figure 1 for EMGSE: Acoustic/EMG Fusion for Multimodal Speech Enhancement
Figure 2 for EMGSE: Acoustic/EMG Fusion for Multimodal Speech Enhancement
Figure 3 for EMGSE: Acoustic/EMG Fusion for Multimodal Speech Enhancement
Figure 4 for EMGSE: Acoustic/EMG Fusion for Multimodal Speech Enhancement
Viaarxiv icon

Deep Learning-based Non-Intrusive Multi-Objective Speech Assessment Model with Cross-Domain Features

Add code
Dec 01, 2021
Figure 1 for Deep Learning-based Non-Intrusive Multi-Objective Speech Assessment Model with Cross-Domain Features
Figure 2 for Deep Learning-based Non-Intrusive Multi-Objective Speech Assessment Model with Cross-Domain Features
Figure 3 for Deep Learning-based Non-Intrusive Multi-Objective Speech Assessment Model with Cross-Domain Features
Figure 4 for Deep Learning-based Non-Intrusive Multi-Objective Speech Assessment Model with Cross-Domain Features
Viaarxiv icon

HASA-net: A non-intrusive hearing-aid speech assessment network

Add code
Nov 10, 2021
Figure 1 for HASA-net: A non-intrusive hearing-aid speech assessment network
Figure 2 for HASA-net: A non-intrusive hearing-aid speech assessment network
Figure 3 for HASA-net: A non-intrusive hearing-aid speech assessment network
Figure 4 for HASA-net: A non-intrusive hearing-aid speech assessment network
Viaarxiv icon

Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments

Add code
Oct 19, 2021
Figure 1 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Figure 2 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Figure 3 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Figure 4 for Speech Enhancement-assisted Stargan Voice Conversion in Noisy Environments
Viaarxiv icon

Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion

Add code
Sep 08, 2021
Figure 1 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 2 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 3 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Figure 4 for Time Alignment using Lip Images for Frame-based Electrolaryngeal Voice Conversion
Viaarxiv icon

SurpriseNet: Melody Harmonization Conditioning on User-controlled Surprise Contours

Add code
Aug 24, 2021
Figure 1 for SurpriseNet: Melody Harmonization Conditioning on User-controlled Surprise Contours
Figure 2 for SurpriseNet: Melody Harmonization Conditioning on User-controlled Surprise Contours
Figure 3 for SurpriseNet: Melody Harmonization Conditioning on User-controlled Surprise Contours
Figure 4 for SurpriseNet: Melody Harmonization Conditioning on User-controlled Surprise Contours
Viaarxiv icon