Alert button
Picture for Yukiya Hono

Yukiya Hono

Alert button

Release of Pre-Trained Models for the Japanese Language

Add code
Bookmark button
Alert button
Apr 02, 2024
Kei Sawada, Tianyu Zhao, Makoto Shing, Kentaro Mitsui, Akio Kaga, Yukiya Hono, Toshiaki Wakatsuki, Koh Mitsuda

Viaarxiv icon

PeriodGrad: Towards Pitch-Controllable Neural Vocoder Based on a Diffusion Probabilistic Model

Add code
Bookmark button
Alert button
Feb 22, 2024
Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

Viaarxiv icon

An Integration of Pre-Trained Speech and Language Models for End-to-End Speech Recognition

Add code
Bookmark button
Alert button
Dec 06, 2023
Yukiya Hono, Koh Mitsuda, Tianyu Zhao, Kentaro Mitsui, Toshiaki Wakatsuki, Kei Sawada

Viaarxiv icon

Towards human-like spoken dialogue generation between AI agents from written dialogue

Add code
Bookmark button
Alert button
Oct 02, 2023
Kentaro Mitsui, Yukiya Hono, Kei Sawada

Viaarxiv icon

UniFLG: Unified Facial Landmark Generator from Text or Speech

Add code
Bookmark button
Alert button
Feb 28, 2023
Kentaro Mitsui, Yukiya Hono, Kei Sawada

Figure 1 for UniFLG: Unified Facial Landmark Generator from Text or Speech
Figure 2 for UniFLG: Unified Facial Landmark Generator from Text or Speech
Figure 3 for UniFLG: Unified Facial Landmark Generator from Text or Speech
Figure 4 for UniFLG: Unified Facial Landmark Generator from Text or Speech
Viaarxiv icon

Singing voice synthesis based on frame-level sequence-to-sequence models considering vocal timing deviation

Add code
Bookmark button
Alert button
Jan 05, 2023
Miku Nishihara, Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

Figure 1 for Singing voice synthesis based on frame-level sequence-to-sequence models considering vocal timing deviation
Figure 2 for Singing voice synthesis based on frame-level sequence-to-sequence models considering vocal timing deviation
Figure 3 for Singing voice synthesis based on frame-level sequence-to-sequence models considering vocal timing deviation
Figure 4 for Singing voice synthesis based on frame-level sequence-to-sequence models considering vocal timing deviation
Viaarxiv icon

Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism

Add code
Bookmark button
Alert button
Dec 28, 2022
Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

Figure 1 for Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism
Figure 2 for Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism
Figure 3 for Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism
Figure 4 for Singing Voice Synthesis Based on a Musical Note Position-Aware Attention Mechanism
Viaarxiv icon

Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System

Add code
Bookmark button
Alert button
Nov 21, 2022
Takenori Yoshimura, Shinji Takaki, Kazuhiro Nakamura, Keiichiro Oura, Yukiya Hono, Kei Hashimoto, Yoshihiko Nankaku, Keiichi Tokuda

Figure 1 for Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System
Figure 2 for Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System
Figure 3 for Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System
Figure 4 for Embedding a Differentiable Mel-cepstral Synthesis Filter to a Neural Speech Synthesis System
Viaarxiv icon

End-to-End Text-to-Speech Based on Latent Representation of Speaking Styles Using Spontaneous Dialogue

Add code
Bookmark button
Alert button
Jun 24, 2022
Kentaro Mitsui, Tianyu Zhao, Kei Sawada, Yukiya Hono, Yoshihiko Nankaku, Keiichi Tokuda

Figure 1 for End-to-End Text-to-Speech Based on Latent Representation of Speaking Styles Using Spontaneous Dialogue
Figure 2 for End-to-End Text-to-Speech Based on Latent Representation of Speaking Styles Using Spontaneous Dialogue
Figure 3 for End-to-End Text-to-Speech Based on Latent Representation of Speaking Styles Using Spontaneous Dialogue
Figure 4 for End-to-End Text-to-Speech Based on Latent Representation of Speaking Styles Using Spontaneous Dialogue
Viaarxiv icon

Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System

Add code
Bookmark button
Alert button
Aug 05, 2021
Yukiya Hono, Kei Hashimoto, Keiichiro Oura, Yoshihiko Nankaku, Keiichi Tokuda

Figure 1 for Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System
Figure 2 for Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System
Figure 3 for Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System
Figure 4 for Sinsy: A Deep Neural Network-Based Singing Voice Synthesis System
Viaarxiv icon