Alert button
Picture for Tomoki Koriyama

Tomoki Koriyama

Alert button

Frame-Wise Breath Detection with Self-Training: An Exploration of Enhancing Breath Naturalness in Text-to-Speech

Add code
Bookmark button
Alert button
Feb 01, 2024
Dong Yang, Tomoki Koriyama, Yuki Saito

Viaarxiv icon

Duration-aware pause insertion using pre-trained language model for multi-speaker text-to-speech

Add code
Bookmark button
Alert button
Feb 27, 2023
Dong Yang, Tomoki Koriyama, Yuki Saito, Takaaki Saeki, Detai Xin, Hiroshi Saruwatari

Figure 1 for Duration-aware pause insertion using pre-trained language model for multi-speaker text-to-speech
Figure 2 for Duration-aware pause insertion using pre-trained language model for multi-speaker text-to-speech
Figure 3 for Duration-aware pause insertion using pre-trained language model for multi-speaker text-to-speech
Figure 4 for Duration-aware pause insertion using pre-trained language model for multi-speaker text-to-speech
Viaarxiv icon

Structured State Space Decoder for Speech Recognition and Synthesis

Add code
Bookmark button
Alert button
Oct 31, 2022
Koichi Miyazaki, Masato Murata, Tomoki Koriyama

Figure 1 for Structured State Space Decoder for Speech Recognition and Synthesis
Figure 2 for Structured State Space Decoder for Speech Recognition and Synthesis
Figure 3 for Structured State Space Decoder for Speech Recognition and Synthesis
Figure 4 for Structured State Space Decoder for Speech Recognition and Synthesis
Viaarxiv icon

UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022

Add code
Bookmark button
Alert button
Apr 05, 2022
Takaaki Saeki, Detai Xin, Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, Hiroshi Saruwatari

Figure 1 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Figure 2 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Figure 3 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Figure 4 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Viaarxiv icon

Multi-speaker Text-to-speech Synthesis Using Deep Gaussian Processes

Add code
Bookmark button
Alert button
Aug 07, 2020
Kentaro Mitsui, Tomoki Koriyama, Hiroshi Saruwatari

Viaarxiv icon

Utterance-level Sequential Modeling For Deep Gaussian Process Based Speech Synthesis Using Simple Recurrent Unit

Add code
Bookmark button
Alert button
Apr 22, 2020
Tomoki Koriyama, Hiroshi Saruwatari

Figure 1 for Utterance-level Sequential Modeling For Deep Gaussian Process Based Speech Synthesis Using Simple Recurrent Unit
Figure 2 for Utterance-level Sequential Modeling For Deep Gaussian Process Based Speech Synthesis Using Simple Recurrent Unit
Figure 3 for Utterance-level Sequential Modeling For Deep Gaussian Process Based Speech Synthesis Using Simple Recurrent Unit
Figure 4 for Utterance-level Sequential Modeling For Deep Gaussian Process Based Speech Synthesis Using Simple Recurrent Unit
Viaarxiv icon

Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking

Add code
Bookmark button
Alert button
Feb 09, 2019
Hiroki Tamaru, Yuki Saito, Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari

Figure 1 for Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking
Figure 2 for Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking
Figure 3 for Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking
Figure 4 for Generative Moment Matching Network-based Random Modulation Post-filter for DNN-based Singing Voice Synthesis and Neural Double-tracking
Viaarxiv icon

Sampling-based speech parameter generation using moment-matching networks

Add code
Bookmark button
Alert button
Apr 12, 2017
Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari

Figure 1 for Sampling-based speech parameter generation using moment-matching networks
Figure 2 for Sampling-based speech parameter generation using moment-matching networks
Figure 3 for Sampling-based speech parameter generation using moment-matching networks
Figure 4 for Sampling-based speech parameter generation using moment-matching networks
Viaarxiv icon