Alert button

"speech": models, code, and papers
Alert button

End-to-end LPCNet: A Neural Vocoder With Fully-Differentiable LPC Estimation

Add code
Bookmark button
Alert button
Mar 29, 2022
Krishna Subramani, Jean-Marc Valin, Umut Isik, Paris Smaragdis, Arvindh Krishnaswamy

Figure 1 for End-to-end LPCNet: A Neural Vocoder With Fully-Differentiable LPC Estimation
Figure 2 for End-to-end LPCNet: A Neural Vocoder With Fully-Differentiable LPC Estimation
Figure 3 for End-to-end LPCNet: A Neural Vocoder With Fully-Differentiable LPC Estimation
Viaarxiv icon

A Sparsity-promoting Dictionary Model for Variational Autoencoders

Add code
Bookmark button
Alert button
Mar 29, 2022
Mostafa Sadeghi, Paul Magron

Figure 1 for A Sparsity-promoting Dictionary Model for Variational Autoencoders
Figure 2 for A Sparsity-promoting Dictionary Model for Variational Autoencoders
Viaarxiv icon

Improving Performance of Seen and Unseen Speech Style Transfer in End-to-end Neural TTS

Add code
Bookmark button
Alert button
Jun 18, 2021
Xiaochun An, Frank K. Soong, Lei Xie

Figure 1 for Improving Performance of Seen and Unseen Speech Style Transfer in End-to-end Neural TTS
Figure 2 for Improving Performance of Seen and Unseen Speech Style Transfer in End-to-end Neural TTS
Figure 3 for Improving Performance of Seen and Unseen Speech Style Transfer in End-to-end Neural TTS
Figure 4 for Improving Performance of Seen and Unseen Speech Style Transfer in End-to-end Neural TTS
Viaarxiv icon

The USYD-JD Speech Translation System for IWSLT 2021

Add code
Bookmark button
Alert button
Jul 24, 2021
Liang Ding, Di Wu, Dacheng Tao

Figure 1 for The USYD-JD Speech Translation System for IWSLT 2021
Figure 2 for The USYD-JD Speech Translation System for IWSLT 2021
Figure 3 for The USYD-JD Speech Translation System for IWSLT 2021
Figure 4 for The USYD-JD Speech Translation System for IWSLT 2021
Viaarxiv icon

What Do We See in Them? Identifying Dimensions of Partner Models for Speech Interfaces Using a Psycholexical Approach

Feb 03, 2021
Philip R Doyle, Leigh Clark, Benjamin R Cowan

Figure 1 for What Do We See in Them? Identifying Dimensions of Partner Models for Speech Interfaces Using a Psycholexical Approach
Figure 2 for What Do We See in Them? Identifying Dimensions of Partner Models for Speech Interfaces Using a Psycholexical Approach
Figure 3 for What Do We See in Them? Identifying Dimensions of Partner Models for Speech Interfaces Using a Psycholexical Approach
Figure 4 for What Do We See in Them? Identifying Dimensions of Partner Models for Speech Interfaces Using a Psycholexical Approach
Viaarxiv icon

Multi-channel Speech Enhancement with 2-D Convolutional Time-frequency Domain Features and a Pre-trained Acoustic Model

Jul 26, 2021
Quandong Wang, Junnan Wu, Zhao Yan, Sichong Qian, Liyong Guo, Lichun Fan, Weiji Zhuang, Peng Gao, Yujun Wang

Figure 1 for Multi-channel Speech Enhancement with 2-D Convolutional Time-frequency Domain Features and a Pre-trained Acoustic Model
Figure 2 for Multi-channel Speech Enhancement with 2-D Convolutional Time-frequency Domain Features and a Pre-trained Acoustic Model
Figure 3 for Multi-channel Speech Enhancement with 2-D Convolutional Time-frequency Domain Features and a Pre-trained Acoustic Model
Figure 4 for Multi-channel Speech Enhancement with 2-D Convolutional Time-frequency Domain Features and a Pre-trained Acoustic Model
Viaarxiv icon

Internal Language Model Adaptation with Text-Only Data for End-to-End Speech Recognition

Oct 14, 2021
Zhong Meng, Yashesh Gaur, Naoyuki Kanda, Jinyu Li, Xie Chen, Yu Wu, Yifan Gong

Figure 1 for Internal Language Model Adaptation with Text-Only Data for End-to-End Speech Recognition
Figure 2 for Internal Language Model Adaptation with Text-Only Data for End-to-End Speech Recognition
Viaarxiv icon

Hierarchical Context-Aware Transformers for Non-Autoregressive Text to Speech

Add code
Bookmark button
Alert button
Jun 29, 2021
Jae-Sung Bae, Tae-Jun Bak, Young-Sun Joo, Hoon-Young Cho

Figure 1 for Hierarchical Context-Aware Transformers for Non-Autoregressive Text to Speech
Figure 2 for Hierarchical Context-Aware Transformers for Non-Autoregressive Text to Speech
Figure 3 for Hierarchical Context-Aware Transformers for Non-Autoregressive Text to Speech
Figure 4 for Hierarchical Context-Aware Transformers for Non-Autoregressive Text to Speech
Viaarxiv icon

Perception-Aware Attack: Creating Adversarial Music via Reverse-Engineering Human Perception

Jul 26, 2022
Rui Duan, Zhe Qu, Shangqing Zhao, Leah Ding, Yao Liu, Zhuo Lu

Figure 1 for Perception-Aware Attack: Creating Adversarial Music via Reverse-Engineering Human Perception
Figure 2 for Perception-Aware Attack: Creating Adversarial Music via Reverse-Engineering Human Perception
Figure 3 for Perception-Aware Attack: Creating Adversarial Music via Reverse-Engineering Human Perception
Figure 4 for Perception-Aware Attack: Creating Adversarial Music via Reverse-Engineering Human Perception
Viaarxiv icon

Self supervised learning for robust voice cloning

Add code
Bookmark button
Alert button
Apr 07, 2022
Konstantinos Klapsas, Nikolaos Ellinas, Karolos Nikitaras, Georgios Vamvoukakis, Panos Kakoulidis, Konstantinos Markopoulos, Spyros Raptis, June Sig Sung, Gunu Jho, Aimilios Chalamandaris, Pirros Tsiakoulis

Figure 1 for Self supervised learning for robust voice cloning
Figure 2 for Self supervised learning for robust voice cloning
Figure 3 for Self supervised learning for robust voice cloning
Viaarxiv icon