Alert button

"speech recognition": models, code, and papers
Alert button

Semi-supervised cross-lingual speech emotion recognition

Jul 14, 2022
Mirko Agarla, Simone Bianco, Luigi Celona, Paolo Napoletano, Alexey Petrovsky, Flavio Piccoli, Raimondo Schettini, Ivan Shanin

Figure 1 for Semi-supervised cross-lingual speech emotion recognition
Figure 2 for Semi-supervised cross-lingual speech emotion recognition
Figure 3 for Semi-supervised cross-lingual speech emotion recognition
Figure 4 for Semi-supervised cross-lingual speech emotion recognition
Viaarxiv icon

A Comparison of Temporal Encoders for Neuromorphic Keyword Spotting with Few Neurons

Jan 24, 2023
Mattias Nilsson, Ton Juny Pina, Lyes Khacef, Foteini Liwicki, Elisabetta Chicca, Fredrik Sandin

Figure 1 for A Comparison of Temporal Encoders for Neuromorphic Keyword Spotting with Few Neurons
Figure 2 for A Comparison of Temporal Encoders for Neuromorphic Keyword Spotting with Few Neurons
Figure 3 for A Comparison of Temporal Encoders for Neuromorphic Keyword Spotting with Few Neurons
Figure 4 for A Comparison of Temporal Encoders for Neuromorphic Keyword Spotting with Few Neurons
Viaarxiv icon

Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition

Dec 10, 2021
Kenichi Kumatani, Robert Gmyr, Felipe Cruz Salinas, Linquan Liu, Wei Zuo, Devang Patel, Eric Sun, Yu Shi

Figure 1 for Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition
Figure 2 for Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition
Figure 3 for Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition
Figure 4 for Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition
Viaarxiv icon

Attention-based Neural Beamforming Layers for Multi-channel Speech Recognition

May 14, 2021
Bhargav Pulugundla, Yang Gao, Brian King, Gokce Keskin, Harish Mallidi, Minhua Wu, Jasha Droppo, Roland Maas

Figure 1 for Attention-based Neural Beamforming Layers for Multi-channel Speech Recognition
Figure 2 for Attention-based Neural Beamforming Layers for Multi-channel Speech Recognition
Figure 3 for Attention-based Neural Beamforming Layers for Multi-channel Speech Recognition
Figure 4 for Attention-based Neural Beamforming Layers for Multi-channel Speech Recognition
Viaarxiv icon

Exploring Effective Fusion Algorithms for Speech Based Self-Supervised Learning Models

Dec 20, 2022
Changli Tang, Yujin Wang, Xie Chen, Wei-Qiang Zhang

Figure 1 for Exploring Effective Fusion Algorithms for Speech Based Self-Supervised Learning Models
Figure 2 for Exploring Effective Fusion Algorithms for Speech Based Self-Supervised Learning Models
Figure 3 for Exploring Effective Fusion Algorithms for Speech Based Self-Supervised Learning Models
Figure 4 for Exploring Effective Fusion Algorithms for Speech Based Self-Supervised Learning Models
Viaarxiv icon

BART based semantic correction for Mandarin automatic speech recognition system

Mar 26, 2021
Yun Zhao, Xuerui Yang, Jinchao Wang, Yongyu Gao, Chao Yan, Yuanfu Zhou

Figure 1 for BART based semantic correction for Mandarin automatic speech recognition system
Figure 2 for BART based semantic correction for Mandarin automatic speech recognition system
Figure 3 for BART based semantic correction for Mandarin automatic speech recognition system
Figure 4 for BART based semantic correction for Mandarin automatic speech recognition system
Viaarxiv icon

Speech Emotion Recognition Using Quaternion Convolutional Neural Networks

Oct 31, 2021
Aneesh Muppidi, Martin Radfar

Figure 1 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Figure 2 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Figure 3 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Figure 4 for Speech Emotion Recognition Using Quaternion Convolutional Neural Networks
Viaarxiv icon

ElectrodeNet -- A Deep Learning Based Sound Coding Strategy for Cochlear Implants

May 26, 2023
Enoch Hsin-Ho Huang, Rong Chao, Yu Tsao, Chao-Min Wu

Figure 1 for ElectrodeNet -- A Deep Learning Based Sound Coding Strategy for Cochlear Implants
Figure 2 for ElectrodeNet -- A Deep Learning Based Sound Coding Strategy for Cochlear Implants
Figure 3 for ElectrodeNet -- A Deep Learning Based Sound Coding Strategy for Cochlear Implants
Figure 4 for ElectrodeNet -- A Deep Learning Based Sound Coding Strategy for Cochlear Implants
Viaarxiv icon

Data Augmentation with Locally-time Reversed Speech for Automatic Speech Recognition

Oct 09, 2021
Si-Ioi Ng, Tan Lee

Figure 1 for Data Augmentation with Locally-time Reversed Speech for Automatic Speech Recognition
Figure 2 for Data Augmentation with Locally-time Reversed Speech for Automatic Speech Recognition
Figure 3 for Data Augmentation with Locally-time Reversed Speech for Automatic Speech Recognition
Figure 4 for Data Augmentation with Locally-time Reversed Speech for Automatic Speech Recognition
Viaarxiv icon