Alert button

"speech": models, code, and papers
Alert button

IANS: Intelligibility-aware Null-steering Beamforming for Dual-Microphone Arrays

Add code
Bookmark button
Alert button
Jul 09, 2023
Wen-Yuan Ting, Syu-Siang Wang, Yu Tsao, Borching Su

Figure 1 for IANS: Intelligibility-aware Null-steering Beamforming for Dual-Microphone Arrays
Figure 2 for IANS: Intelligibility-aware Null-steering Beamforming for Dual-Microphone Arrays
Viaarxiv icon

MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages

Add code
Bookmark button
Alert button
May 23, 2023
Cheikh M. Bamba Dione, David Adelani, Peter Nabende, Jesujoba Alabi, Thapelo Sindane, Happy Buzaaba, Shamsuddeen Hassan Muhammad, Chris Chinenye Emezue, Perez Ogayo, Anuoluwapo Aremu, Catherine Gitau, Derguene Mbaye, Jonathan Mukiibi, Blessing Sibanda, Bonaventure F. P. Dossou, Andiswa Bukula, Rooweither Mabuya, Allahsera Auguste Tapo, Edwin Munkoh-Buabeng, victoire Memdjokam Koagne, Fatoumata Ouoba Kabore, Amelia Taylor, Godson Kalipe, Tebogo Macucwa, Vukosi Marivate, Tajuddeen Gwadabe, Mboning Tchiaze Elvis, Ikechukwu Onyenwe, Gratien Atindogbe, Tolulope Adelani, Idris Akinade, Olanrewaju Samuel, Marien Nahimana, Théogène Musabeyezu, Emile Niyomutabazi, Ester Chimhenga, Kudzai Gotosa, Patrick Mizha, Apelete Agbolo, Seydou Traore, Chinedu Uchechukwu, Aliyu Yusuf, Muhammad Abdullahi, Dietrich Klakow

Figure 1 for MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages
Figure 2 for MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages
Figure 3 for MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages
Figure 4 for MasakhaPOS: Part-of-Speech Tagging for Typologically Diverse African Languages
Viaarxiv icon

Adaptation and Optimization of Automatic Speech Recognition (ASR) for the Maritime Domain in the Field of VHF Communication

Jun 01, 2023
Emin Cagatay Nakilcioglu, Maximilian Reimann, Ole John

Viaarxiv icon

Developmental Bootstrapping: From Simple Competences to Intelligent Human-Compatible AIs

Aug 23, 2023
Mark Stefik, Robert Price

Figure 1 for Developmental Bootstrapping: From Simple Competences to Intelligent Human-Compatible AIs
Figure 2 for Developmental Bootstrapping: From Simple Competences to Intelligent Human-Compatible AIs
Figure 3 for Developmental Bootstrapping: From Simple Competences to Intelligent Human-Compatible AIs
Figure 4 for Developmental Bootstrapping: From Simple Competences to Intelligent Human-Compatible AIs
Viaarxiv icon

Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions

Aug 10, 2023
Chinmay Rane, Kanishka Tyagi, Michael Manry

Figure 1 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Figure 2 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Figure 3 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Figure 4 for Optimizing Performance of Feedforward and Convolutional Neural Networks through Dynamic Activation Functions
Viaarxiv icon

Cross-speaker Emotion Transfer by Manipulating Speech Style Latents

Add code
Bookmark button
Alert button
Mar 15, 2023
Suhee Jo, Younggun Lee, Yookyung Shin, Yeongtae Hwang, Taesu Kim

Figure 1 for Cross-speaker Emotion Transfer by Manipulating Speech Style Latents
Figure 2 for Cross-speaker Emotion Transfer by Manipulating Speech Style Latents
Figure 3 for Cross-speaker Emotion Transfer by Manipulating Speech Style Latents
Figure 4 for Cross-speaker Emotion Transfer by Manipulating Speech Style Latents
Viaarxiv icon

Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding

Jul 22, 2023
Suyoun Kim, Akshat Shrivastava, Duc Le, Ju Lin, Ozlem Kalinli, Michael L. Seltzer

Figure 1 for Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
Figure 2 for Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
Figure 3 for Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
Figure 4 for Modality Confidence Aware Training for Robust End-to-End Spoken Language Understanding
Viaarxiv icon

Improving Continuous Sign Language Recognition with Cross-Lingual Signs

Add code
Bookmark button
Alert button
Aug 21, 2023
Fangyun Wei, Yutong Chen

Figure 1 for Improving Continuous Sign Language Recognition with Cross-Lingual Signs
Figure 2 for Improving Continuous Sign Language Recognition with Cross-Lingual Signs
Figure 3 for Improving Continuous Sign Language Recognition with Cross-Lingual Signs
Figure 4 for Improving Continuous Sign Language Recognition with Cross-Lingual Signs
Viaarxiv icon

EmoTalk: Speech-driven emotional disentanglement for 3D face animation

Add code
Bookmark button
Alert button
Mar 20, 2023
Ziqiao Peng, Haoyu Wu, Zhenbo Song, Hao Xu, Xiangyu Zhu, Hongyan Liu, Jun He, Zhaoxin Fan

Figure 1 for EmoTalk: Speech-driven emotional disentanglement for 3D face animation
Figure 2 for EmoTalk: Speech-driven emotional disentanglement for 3D face animation
Figure 3 for EmoTalk: Speech-driven emotional disentanglement for 3D face animation
Figure 4 for EmoTalk: Speech-driven emotional disentanglement for 3D face animation
Viaarxiv icon

BASEN: Time-Domain Brain-Assisted Speech Enhancement Network with Convolutional Cross Attention in Multi-talker Conditions

Add code
Bookmark button
Alert button
May 17, 2023
Jie Zhang, Qing-Tian Xu, Qiu-Shi Zhu, Zhen-Hua Ling

Figure 1 for BASEN: Time-Domain Brain-Assisted Speech Enhancement Network with Convolutional Cross Attention in Multi-talker Conditions
Figure 2 for BASEN: Time-Domain Brain-Assisted Speech Enhancement Network with Convolutional Cross Attention in Multi-talker Conditions
Figure 3 for BASEN: Time-Domain Brain-Assisted Speech Enhancement Network with Convolutional Cross Attention in Multi-talker Conditions
Viaarxiv icon