Alert button

"speech": models, code, and papers
Alert button

Wavebender GAN: An architecture for phonetically meaningful speech manipulation

Add code
Bookmark button
Alert button
Feb 22, 2022
Gustavo Teodoro Döhler Beck, Ulme Wennberg, Zofia Malisz, Gustav Eje Henter

Figure 1 for Wavebender GAN: An architecture for phonetically meaningful speech manipulation
Figure 2 for Wavebender GAN: An architecture for phonetically meaningful speech manipulation
Figure 3 for Wavebender GAN: An architecture for phonetically meaningful speech manipulation
Figure 4 for Wavebender GAN: An architecture for phonetically meaningful speech manipulation
Viaarxiv icon

A two-stage full-band speech enhancement model with effective spectral compression mapping

Jun 27, 2022
Zhongshu Hou, Qinwen Hu, Kai Chen, Jing Lu

Figure 1 for A two-stage full-band speech enhancement model with effective spectral compression mapping
Figure 2 for A two-stage full-band speech enhancement model with effective spectral compression mapping
Figure 3 for A two-stage full-band speech enhancement model with effective spectral compression mapping
Figure 4 for A two-stage full-band speech enhancement model with effective spectral compression mapping
Viaarxiv icon

Topic Modeling Based on Two-Step Flow Theory: Application to Tweets about Bitcoin

Mar 03, 2023
Aos Mulahuwaish, Matthew Loucks, Basheer Qolomany, Ala Al-Fuqaha

Figure 1 for Topic Modeling Based on Two-Step Flow Theory: Application to Tweets about Bitcoin
Figure 2 for Topic Modeling Based on Two-Step Flow Theory: Application to Tweets about Bitcoin
Figure 3 for Topic Modeling Based on Two-Step Flow Theory: Application to Tweets about Bitcoin
Figure 4 for Topic Modeling Based on Two-Step Flow Theory: Application to Tweets about Bitcoin
Viaarxiv icon

Tiny-Sepformer: A Tiny Time-Domain Transformer Network for Speech Separation

Jun 30, 2022
Jian Luo, Jianzong Wang, Ning Cheng, Edward Xiao, Xulong Zhang, Jing Xiao

Figure 1 for Tiny-Sepformer: A Tiny Time-Domain Transformer Network for Speech Separation
Figure 2 for Tiny-Sepformer: A Tiny Time-Domain Transformer Network for Speech Separation
Figure 3 for Tiny-Sepformer: A Tiny Time-Domain Transformer Network for Speech Separation
Figure 4 for Tiny-Sepformer: A Tiny Time-Domain Transformer Network for Speech Separation
Viaarxiv icon

Explainable Human-centered Traits from Head Motion and Facial Expression Dynamics

Feb 23, 2023
Surbhi Madan, Monika Gahalawat, Tanaya Guha, Roland Goecke, Ramanathan Subramanian

Figure 1 for Explainable Human-centered Traits from Head Motion and Facial Expression Dynamics
Figure 2 for Explainable Human-centered Traits from Head Motion and Facial Expression Dynamics
Figure 3 for Explainable Human-centered Traits from Head Motion and Facial Expression Dynamics
Figure 4 for Explainable Human-centered Traits from Head Motion and Facial Expression Dynamics
Viaarxiv icon

Generalization of Auto-Regressive Hidden Markov Models to Non-Linear Dynamics and Non-Euclidean Observation Space

Add code
Bookmark button
Alert button
Feb 23, 2023
Michele Ginesi, Paolo Fiorini

Figure 1 for Generalization of Auto-Regressive Hidden Markov Models to Non-Linear Dynamics and Non-Euclidean Observation Space
Figure 2 for Generalization of Auto-Regressive Hidden Markov Models to Non-Linear Dynamics and Non-Euclidean Observation Space
Figure 3 for Generalization of Auto-Regressive Hidden Markov Models to Non-Linear Dynamics and Non-Euclidean Observation Space
Figure 4 for Generalization of Auto-Regressive Hidden Markov Models to Non-Linear Dynamics and Non-Euclidean Observation Space
Viaarxiv icon

Incorporating Uncertainty from Speaker Embedding Estimation to Speaker Verification

Feb 23, 2023
Qiongqiong Wang, Kong Aik Lee, Tianchi Liu

Figure 1 for Incorporating Uncertainty from Speaker Embedding Estimation to Speaker Verification
Figure 2 for Incorporating Uncertainty from Speaker Embedding Estimation to Speaker Verification
Figure 3 for Incorporating Uncertainty from Speaker Embedding Estimation to Speaker Verification
Viaarxiv icon

Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages

Sep 08, 2022
Li Miao, Jian Wu, Piyush Behre, Shuangyu Chang, Sarangarajan Parthasarathy

Figure 1 for Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages
Figure 2 for Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages
Figure 3 for Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages
Figure 4 for Multilingual Transformer Language Model for Speech Recognition in Low-resource Languages
Viaarxiv icon

Lip to Speech Synthesis with Visual Context Attentional GAN

Apr 04, 2022
Minsu Kim, Joanna Hong, Yong Man Ro

Figure 1 for Lip to Speech Synthesis with Visual Context Attentional GAN
Figure 2 for Lip to Speech Synthesis with Visual Context Attentional GAN
Figure 3 for Lip to Speech Synthesis with Visual Context Attentional GAN
Figure 4 for Lip to Speech Synthesis with Visual Context Attentional GAN
Viaarxiv icon

Sentiment-Aware Automatic Speech Recognition pre-training for enhanced Speech Emotion Recognition

Jan 27, 2022
Ayoub Ghriss, Bo Yang, Viktor Rozgic, Elizabeth Shriberg, Chao Wang

Figure 1 for Sentiment-Aware Automatic Speech Recognition pre-training for enhanced Speech Emotion Recognition
Figure 2 for Sentiment-Aware Automatic Speech Recognition pre-training for enhanced Speech Emotion Recognition
Figure 3 for Sentiment-Aware Automatic Speech Recognition pre-training for enhanced Speech Emotion Recognition
Figure 4 for Sentiment-Aware Automatic Speech Recognition pre-training for enhanced Speech Emotion Recognition
Viaarxiv icon