Alert button

"speech recognition": models, code, and papers
Alert button

Speech emotion recognition from voice messages recorded in the wild

Mar 04, 2024
Lucía Gómez-Zaragozá, Óscar Valls, Rocío del Amor, María José Castro-Bleda, Valery Naranjo, Mariano Alcañiz Raya, Javier Marín-Morales

Figure 1 for Speech emotion recognition from voice messages recorded in the wild
Figure 2 for Speech emotion recognition from voice messages recorded in the wild
Figure 3 for Speech emotion recognition from voice messages recorded in the wild
Figure 4 for Speech emotion recognition from voice messages recorded in the wild
Viaarxiv icon

Real-Time Multimodal Cognitive Assistant for Emergency Medical Services

Add code
Bookmark button
Alert button
Mar 11, 2024
Keshara Weerasinghe, Saahith Janapati, Xueren Ge, Sion Kim, Sneha Iyer, John A. Stankovic, Homa Alemzadeh

Figure 1 for Real-Time Multimodal Cognitive Assistant for Emergency Medical Services
Figure 2 for Real-Time Multimodal Cognitive Assistant for Emergency Medical Services
Figure 3 for Real-Time Multimodal Cognitive Assistant for Emergency Medical Services
Figure 4 for Real-Time Multimodal Cognitive Assistant for Emergency Medical Services
Viaarxiv icon

Advanced Artificial Intelligence Algorithms in Cochlear Implants: Review of Healthcare Strategies, Challenges, and Perspectives

Mar 17, 2024
Billel Essaid, Hamza Kheddar, Noureddine Batel, Abderrahmane Lakas, Muhammad E. H. Chowdhury

Viaarxiv icon

A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement

Mar 03, 2024
Ravi Shankar, Ke Tan, Buye Xu, Anurag Kumar

Figure 1 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Figure 2 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Figure 3 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Figure 4 for A Closer Look at Wav2Vec2 Embeddings for On-Device Single-Channel Speech Enhancement
Viaarxiv icon

Non-verbal information in spontaneous speech -- towards a new framework of analysis

Mar 13, 2024
Tirza Biron, Moshe Barboy, Eran Ben-Artzy, Alona Golubchik, Yanir Marmor, Smadar Szekely, Yaron Winter, David Harel

Figure 1 for Non-verbal information in spontaneous speech -- towards a new framework of analysis
Figure 2 for Non-verbal information in spontaneous speech -- towards a new framework of analysis
Figure 3 for Non-verbal information in spontaneous speech -- towards a new framework of analysis
Figure 4 for Non-verbal information in spontaneous speech -- towards a new framework of analysis
Viaarxiv icon

SpokeN-100: A Cross-Lingual Benchmarking Dataset for The Classification of Spoken Numbers in Different Languages

Mar 14, 2024
René Groh, Nina Goes, Andreas M. Kist

Figure 1 for SpokeN-100: A Cross-Lingual Benchmarking Dataset for The Classification of Spoken Numbers in Different Languages
Figure 2 for SpokeN-100: A Cross-Lingual Benchmarking Dataset for The Classification of Spoken Numbers in Different Languages
Figure 3 for SpokeN-100: A Cross-Lingual Benchmarking Dataset for The Classification of Spoken Numbers in Different Languages
Figure 4 for SpokeN-100: A Cross-Lingual Benchmarking Dataset for The Classification of Spoken Numbers in Different Languages
Viaarxiv icon

Neural Networks Hear You Loud And Clear: Hearing Loss Compensation Using Deep Neural Networks

Mar 15, 2024
Peter Leer, Jesper Jensen, Laurel Carney, Zheng-Hua Tan, Jan Østergaard, Lars Bramsløw

Figure 1 for Neural Networks Hear You Loud And Clear: Hearing Loss Compensation Using Deep Neural Networks
Figure 2 for Neural Networks Hear You Loud And Clear: Hearing Loss Compensation Using Deep Neural Networks
Figure 3 for Neural Networks Hear You Loud And Clear: Hearing Loss Compensation Using Deep Neural Networks
Figure 4 for Neural Networks Hear You Loud And Clear: Hearing Loss Compensation Using Deep Neural Networks
Viaarxiv icon

Fusion approaches for emotion recognition from speech using acoustic and text-based features

Mar 27, 2024
Leonardo Pepino, Pablo Riera, Luciana Ferrer, Agustin Gravano

Viaarxiv icon

The NeurIPS 2023 Machine Learning for Audio Workshop: Affective Audio Benchmarks and Novel Data

Mar 21, 2024
Alice Baird, Rachel Manzelli, Panagiotis Tzirakis, Chris Gagne, Haoqi Li, Sadie Allen, Sander Dieleman, Brian Kulis, Shrikanth S. Narayanan, Alan Cowen

Figure 1 for The NeurIPS 2023 Machine Learning for Audio Workshop: Affective Audio Benchmarks and Novel Data
Figure 2 for The NeurIPS 2023 Machine Learning for Audio Workshop: Affective Audio Benchmarks and Novel Data
Figure 3 for The NeurIPS 2023 Machine Learning for Audio Workshop: Affective Audio Benchmarks and Novel Data
Figure 4 for The NeurIPS 2023 Machine Learning for Audio Workshop: Affective Audio Benchmarks and Novel Data
Viaarxiv icon

Energy-Based Models with Applications to Speech and Language Processing

Add code
Bookmark button
Alert button
Mar 16, 2024
Zhijian Ou

Figure 1 for Energy-Based Models with Applications to Speech and Language Processing
Figure 2 for Energy-Based Models with Applications to Speech and Language Processing
Figure 3 for Energy-Based Models with Applications to Speech and Language Processing
Figure 4 for Energy-Based Models with Applications to Speech and Language Processing
Viaarxiv icon