Alert button
Picture for Yusuke Yasuda

Yusuke Yasuda

Alert button

Automatic design optimization of preference-based subjective evaluation with online learning in crowdsourcing environment

Add code
Bookmark button
Alert button
Mar 10, 2024
Yusuke Yasuda, Tomoki Toda

Figure 1 for Automatic design optimization of preference-based subjective evaluation with online learning in crowdsourcing environment
Figure 2 for Automatic design optimization of preference-based subjective evaluation with online learning in crowdsourcing environment
Figure 3 for Automatic design optimization of preference-based subjective evaluation with online learning in crowdsourcing environment
Figure 4 for Automatic design optimization of preference-based subjective evaluation with online learning in crowdsourcing environment
Viaarxiv icon

Preference-based training framework for automatic speech quality assessment using deep neural network

Add code
Bookmark button
Alert button
Aug 29, 2023
Cheng-Hung Hu, Yusuke Yasuda, Tomoki Toda

Viaarxiv icon

The Singing Voice Conversion Challenge 2023

Add code
Bookmark button
Alert button
Jun 26, 2023
Wen-Chin Huang, Lester Phillip Violeta, Songxiang Liu, Jiatong Shi, Yusuke Yasuda, Tomoki Toda

Figure 1 for The Singing Voice Conversion Challenge 2023
Figure 2 for The Singing Voice Conversion Challenge 2023
Figure 3 for The Singing Voice Conversion Challenge 2023
Figure 4 for The Singing Voice Conversion Challenge 2023
Viaarxiv icon

Text-to-speech synthesis based on latent variable conversion using diffusion probabilistic model and variational autoencoder

Add code
Bookmark button
Alert button
Dec 16, 2022
Yusuke Yasuda, Tomoki Toda

Figure 1 for Text-to-speech synthesis based on latent variable conversion using diffusion probabilistic model and variational autoencoder
Figure 2 for Text-to-speech synthesis based on latent variable conversion using diffusion probabilistic model and variational autoencoder
Figure 3 for Text-to-speech synthesis based on latent variable conversion using diffusion probabilistic model and variational autoencoder
Viaarxiv icon

Investigation of Japanese PnG BERT language model in text-to-speech synthesis for pitch accent language

Add code
Bookmark button
Alert button
Dec 16, 2022
Yusuke Yasuda, Tomoki Toda

Figure 1 for Investigation of Japanese PnG BERT language model in text-to-speech synthesis for pitch accent language
Figure 2 for Investigation of Japanese PnG BERT language model in text-to-speech synthesis for pitch accent language
Figure 3 for Investigation of Japanese PnG BERT language model in text-to-speech synthesis for pitch accent language
Figure 4 for Investigation of Japanese PnG BERT language model in text-to-speech synthesis for pitch accent language
Viaarxiv icon

ESPnet2-TTS: Extending the Edge of TTS Research

Add code
Bookmark button
Alert button
Oct 15, 2021
Tomoki Hayashi, Ryuichi Yamamoto, Takenori Yoshimura, Peter Wu, Jiatong Shi, Takaaki Saeki, Yooncheol Ju, Yusuke Yasuda, Shinnosuke Takamichi, Shinji Watanabe

Figure 1 for ESPnet2-TTS: Extending the Edge of TTS Research
Figure 2 for ESPnet2-TTS: Extending the Edge of TTS Research
Figure 3 for ESPnet2-TTS: Extending the Edge of TTS Research
Figure 4 for ESPnet2-TTS: Extending the Edge of TTS Research
Viaarxiv icon

Pretraining Strategies, Waveform Model Choice, and Acoustic Configurations for Multi-Speaker End-to-End Speech Synthesis

Add code
Bookmark button
Alert button
Nov 10, 2020
Erica Cooper, Xin Wang, Yi Zhao, Yusuke Yasuda, Junichi Yamagishi

Figure 1 for Pretraining Strategies, Waveform Model Choice, and Acoustic Configurations for Multi-Speaker End-to-End Speech Synthesis
Figure 2 for Pretraining Strategies, Waveform Model Choice, and Acoustic Configurations for Multi-Speaker End-to-End Speech Synthesis
Figure 3 for Pretraining Strategies, Waveform Model Choice, and Acoustic Configurations for Multi-Speaker End-to-End Speech Synthesis
Figure 4 for Pretraining Strategies, Waveform Model Choice, and Acoustic Configurations for Multi-Speaker End-to-End Speech Synthesis
Viaarxiv icon

End-to-End Text-to-Speech using Latent Duration based on VQ-VAE

Add code
Bookmark button
Alert button
Oct 20, 2020
Yusuke Yasuda, Xin Wang, Junichi Yamagishi

Figure 1 for End-to-End Text-to-Speech using Latent Duration based on VQ-VAE
Figure 2 for End-to-End Text-to-Speech using Latent Duration based on VQ-VAE
Figure 3 for End-to-End Text-to-Speech using Latent Duration based on VQ-VAE
Figure 4 for End-to-End Text-to-Speech using Latent Duration based on VQ-VAE
Viaarxiv icon

Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis

Add code
Bookmark button
Alert button
May 20, 2020
Yusuke Yasuda, Xin Wang, Junichi Yamagishi

Figure 1 for Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis
Figure 2 for Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis
Figure 3 for Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis
Figure 4 for Investigation of learning abilities on linguistic features in sequence-to-sequence text-to-speech synthesis
Viaarxiv icon