Alert button
Picture for Shinnosuke Takamichi

Shinnosuke Takamichi

Alert button

Mid-attribute speaker generation using optimal-transport-based interpolation of Gaussian mixture models

Add code
Bookmark button
Alert button
Oct 18, 2022
Aya Watanabe, Shinnosuke Takamichi, Yuki Saito, Detai Xin, Hiroshi Saruwatari

Figure 1 for Mid-attribute speaker generation using optimal-transport-based interpolation of Gaussian mixture models
Figure 2 for Mid-attribute speaker generation using optimal-transport-based interpolation of Gaussian mixture models
Figure 3 for Mid-attribute speaker generation using optimal-transport-based interpolation of Gaussian mixture models
Figure 4 for Mid-attribute speaker generation using optimal-transport-based interpolation of Gaussian mixture models
Viaarxiv icon

Spontaneous speech synthesis with linguistic-speech consistency training using pseudo-filled pauses

Add code
Bookmark button
Alert button
Oct 18, 2022
Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari

Figure 1 for Spontaneous speech synthesis with linguistic-speech consistency training using pseudo-filled pauses
Figure 2 for Spontaneous speech synthesis with linguistic-speech consistency training using pseudo-filled pauses
Figure 3 for Spontaneous speech synthesis with linguistic-speech consistency training using pseudo-filled pauses
Figure 4 for Spontaneous speech synthesis with linguistic-speech consistency training using pseudo-filled pauses
Viaarxiv icon

Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images

Add code
Bookmark button
Alert button
Oct 17, 2022
Hien Ohnaka, Shinnosuke Takamichi, Keisuke Imoto, Yuki Okamoto, Kazuki Fujii, Hiroshi Saruwatari

Figure 1 for Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images
Figure 2 for Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images
Figure 3 for Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images
Figure 4 for Visual onoma-to-wave: environmental sound synthesis from visual onomatopoeias and sound-source images
Viaarxiv icon

Empirical Study Incorporating Linguistic Knowledge on Filled Pauses for Personalized Spontaneous Speech Synthesis

Add code
Bookmark button
Alert button
Oct 14, 2022
Yuta Matsunaga, Takaaki Saeki, Shinnosuke Takamichi, Hiroshi Saruwatari

Figure 1 for Empirical Study Incorporating Linguistic Knowledge on Filled Pauses for Personalized Spontaneous Speech Synthesis
Figure 2 for Empirical Study Incorporating Linguistic Knowledge on Filled Pauses for Personalized Spontaneous Speech Synthesis
Figure 3 for Empirical Study Incorporating Linguistic Knowledge on Filled Pauses for Personalized Spontaneous Speech Synthesis
Figure 4 for Empirical Study Incorporating Linguistic Knowledge on Filled Pauses for Personalized Spontaneous Speech Synthesis
Viaarxiv icon

How Should We Evaluate Synthesized Environmental Sounds

Add code
Bookmark button
Alert button
Aug 16, 2022
Yuki Okamoto, Keisuke Imoto, Shinnosuke Takamichi, Takahiro Fukumori, Yoichi Yamashita

Figure 1 for How Should We Evaluate Synthesized Environmental Sounds
Figure 2 for How Should We Evaluate Synthesized Environmental Sounds
Figure 3 for How Should We Evaluate Synthesized Environmental Sounds
Figure 4 for How Should We Evaluate Synthesized Environmental Sounds
Viaarxiv icon

Exploring the Effectiveness of Self-supervised Learning and Classifier Chains in Emotion Recognition of Nonverbal Vocalizations

Add code
Bookmark button
Alert button
Jun 21, 2022
Detai Xin, Shinnosuke Takamichi, Hiroshi Saruwatari

Figure 1 for Exploring the Effectiveness of Self-supervised Learning and Classifier Chains in Emotion Recognition of Nonverbal Vocalizations
Figure 2 for Exploring the Effectiveness of Self-supervised Learning and Classifier Chains in Emotion Recognition of Nonverbal Vocalizations
Figure 3 for Exploring the Effectiveness of Self-supervised Learning and Classifier Chains in Emotion Recognition of Nonverbal Vocalizations
Figure 4 for Exploring the Effectiveness of Self-supervised Learning and Classifier Chains in Emotion Recognition of Nonverbal Vocalizations
Viaarxiv icon

Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History

Add code
Bookmark button
Alert button
Jun 16, 2022
Yuto Nishimura, Yuki Saito, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari

Figure 1 for Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History
Figure 2 for Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History
Figure 3 for Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History
Figure 4 for Acoustic Modeling for End-to-End Empathetic Dialogue Speech Synthesis Using Linguistic and Prosodic Contexts of Dialogue History
Viaarxiv icon

Speaking-Rate-Controllable HiFi-GAN Using Feature Interpolation

Add code
Bookmark button
Alert button
Apr 22, 2022
Detai Xin, Shinnosuke Takamichi, Takuma Okamoto, Hisashi Kawai, Hiroshi Saruwatari

Figure 1 for Speaking-Rate-Controllable HiFi-GAN Using Feature Interpolation
Figure 2 for Speaking-Rate-Controllable HiFi-GAN Using Feature Interpolation
Figure 3 for Speaking-Rate-Controllable HiFi-GAN Using Feature Interpolation
Figure 4 for Speaking-Rate-Controllable HiFi-GAN Using Feature Interpolation
Viaarxiv icon

UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022

Add code
Bookmark button
Alert button
Apr 05, 2022
Takaaki Saeki, Detai Xin, Wataru Nakata, Tomoki Koriyama, Shinnosuke Takamichi, Hiroshi Saruwatari

Figure 1 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Figure 2 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Figure 3 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Figure 4 for UTMOS: UTokyo-SaruLab System for VoiceMOS Challenge 2022
Viaarxiv icon

STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent

Add code
Bookmark button
Alert button
Mar 28, 2022
Yuki Saito, Yuto Nishimura, Shinnosuke Takamichi, Kentaro Tachibana, Hiroshi Saruwatari

Figure 1 for STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent
Figure 2 for STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent
Figure 3 for STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent
Figure 4 for STUDIES: Corpus of Japanese Empathetic Dialogue Speech Towards Friendly Voice Agent
Viaarxiv icon