Alert button

"speech": models, code, and papers
Alert button

EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning

Add code
Bookmark button
Alert button
Jun 16, 2022
Li-Chin Chen, Po-Hsun Chen, Richard Tzong-Han Tsai, Yu Tsao

Figure 1 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Figure 2 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Figure 3 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Figure 4 for EPG2S: Speech Generation and Speech Enhancement based on Electropalatography and Audio Signals using Multimodal Learning
Viaarxiv icon

Cross-modal Audio-visual Co-learning for Text-independent Speaker Verification

Add code
Bookmark button
Alert button
Feb 22, 2023
Meng Liu, Kong Aik Lee, Longbiao Wang, Hanyi Zhang, Chang Zeng, Jianwu Dang

Figure 1 for Cross-modal Audio-visual Co-learning for Text-independent Speaker Verification
Figure 2 for Cross-modal Audio-visual Co-learning for Text-independent Speaker Verification
Figure 3 for Cross-modal Audio-visual Co-learning for Text-independent Speaker Verification
Figure 4 for Cross-modal Audio-visual Co-learning for Text-independent Speaker Verification
Viaarxiv icon

AIRCADE: an Anechoic and IR Convolution-based Auralization Data-compilation Ensemble

Add code
Bookmark button
Alert button
Apr 24, 2023
Túlio Chiodi, Arthur dos Santos, Pedro Martins, Bruno Masiero

Figure 1 for AIRCADE: an Anechoic and IR Convolution-based Auralization Data-compilation Ensemble
Figure 2 for AIRCADE: an Anechoic and IR Convolution-based Auralization Data-compilation Ensemble
Figure 3 for AIRCADE: an Anechoic and IR Convolution-based Auralization Data-compilation Ensemble
Viaarxiv icon

Evaluating gesture-generation in a large-scale open challenge: The GENEA Challenge 2022

Add code
Bookmark button
Alert button
Mar 15, 2023
Taras Kucherenko, Pieter Wolfert, Youngwoo Yoon, Carla Viegas, Teodor Nikolov, Mihail Tsakov, Gustav Eje Henter

Figure 1 for Evaluating gesture-generation in a large-scale open challenge: The GENEA Challenge 2022
Figure 2 for Evaluating gesture-generation in a large-scale open challenge: The GENEA Challenge 2022
Figure 3 for Evaluating gesture-generation in a large-scale open challenge: The GENEA Challenge 2022
Figure 4 for Evaluating gesture-generation in a large-scale open challenge: The GENEA Challenge 2022
Viaarxiv icon

Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences

Add code
Bookmark button
Alert button
Mar 15, 2023
Yuan Tseng, Cheng-I Lai, Hung-yi Lee

Figure 1 for Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences
Figure 2 for Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences
Figure 3 for Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences
Figure 4 for Cascading and Direct Approaches to Unsupervised Constituency Parsing on Spoken Sentences
Viaarxiv icon

ASR2K: Speech Recognition for Around 2000 Languages without Audio

Add code
Bookmark button
Alert button
Sep 06, 2022
Xinjian Li, Florian Metze, David R Mortensen, Alan W Black, Shinji Watanabe

Figure 1 for ASR2K: Speech Recognition for Around 2000 Languages without Audio
Figure 2 for ASR2K: Speech Recognition for Around 2000 Languages without Audio
Figure 3 for ASR2K: Speech Recognition for Around 2000 Languages without Audio
Figure 4 for ASR2K: Speech Recognition for Around 2000 Languages without Audio
Viaarxiv icon

Extracting speaker and emotion information from self-supervised speech models via channel-wise correlations

Add code
Bookmark button
Alert button
Oct 15, 2022
Themos Stafylakis, Ladislav Mosner, Sofoklis Kakouros, Oldrich Plchot, Lukas Burget, Jan Cernocky

Figure 1 for Extracting speaker and emotion information from self-supervised speech models via channel-wise correlations
Figure 2 for Extracting speaker and emotion information from self-supervised speech models via channel-wise correlations
Figure 3 for Extracting speaker and emotion information from self-supervised speech models via channel-wise correlations
Figure 4 for Extracting speaker and emotion information from self-supervised speech models via channel-wise correlations
Viaarxiv icon

DualVoice: Speech Interaction that Discriminates between Normal and Whispered Voice Input

Aug 22, 2022
Jun Rekimoto

Figure 1 for DualVoice: Speech Interaction that Discriminates between Normal and Whispered Voice Input
Figure 2 for DualVoice: Speech Interaction that Discriminates between Normal and Whispered Voice Input
Figure 3 for DualVoice: Speech Interaction that Discriminates between Normal and Whispered Voice Input
Figure 4 for DualVoice: Speech Interaction that Discriminates between Normal and Whispered Voice Input
Viaarxiv icon

Memory Augmented Lookup Dictionary based Language Modeling for Automatic Speech Recognition

Dec 30, 2022
Yukun Feng, Ming Tu, Rui Xia, Chuanzeng Huang, Yuxuan Wang

Figure 1 for Memory Augmented Lookup Dictionary based Language Modeling for Automatic Speech Recognition
Figure 2 for Memory Augmented Lookup Dictionary based Language Modeling for Automatic Speech Recognition
Figure 3 for Memory Augmented Lookup Dictionary based Language Modeling for Automatic Speech Recognition
Figure 4 for Memory Augmented Lookup Dictionary based Language Modeling for Automatic Speech Recognition
Viaarxiv icon

Neural inhibition during speech planning contributes to contrastive hyperarticulation

Sep 25, 2022
Michael C. Stern, Jason A. Shaw

Figure 1 for Neural inhibition during speech planning contributes to contrastive hyperarticulation
Figure 2 for Neural inhibition during speech planning contributes to contrastive hyperarticulation
Figure 3 for Neural inhibition during speech planning contributes to contrastive hyperarticulation
Figure 4 for Neural inhibition during speech planning contributes to contrastive hyperarticulation
Viaarxiv icon