Alert button

"speech recognition": models, code, and papers
Alert button

Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition

Add code
Bookmark button
Alert button
Jul 10, 2018
Chun-Fu Chen, Quanfu Fan, Neil Mallinar, Tom Sercu, Rogerio Feris

Figure 1 for Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition
Figure 2 for Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition
Figure 3 for Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition
Figure 4 for Big-Little Net: An Efficient Multi-Scale Feature Representation for Visual and Speech Recognition
Viaarxiv icon

Self Supervised Adversarial Domain Adaptation for Cross-Corpus and Cross-Language Speech Emotion Recognition

Apr 19, 2022
Siddique Latif, Rajib Rana, Sara Khalifa, Raja Jurdak, Björn Schuller

Figure 1 for Self Supervised Adversarial Domain Adaptation for Cross-Corpus and Cross-Language Speech Emotion Recognition
Figure 2 for Self Supervised Adversarial Domain Adaptation for Cross-Corpus and Cross-Language Speech Emotion Recognition
Figure 3 for Self Supervised Adversarial Domain Adaptation for Cross-Corpus and Cross-Language Speech Emotion Recognition
Figure 4 for Self Supervised Adversarial Domain Adaptation for Cross-Corpus and Cross-Language Speech Emotion Recognition
Viaarxiv icon

ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context

May 07, 2020
Wei Han, Zhengdong Zhang, Yu Zhang, Jiahui Yu, Chung-Cheng Chiu, James Qin, Anmol Gulati, Ruoming Pang, Yonghui Wu

Figure 1 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Figure 2 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Figure 3 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Figure 4 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Viaarxiv icon

An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions

Add code
Bookmark button
Alert button
Apr 28, 2019
Aswin Shanmugam Subramanian, Xiaofei Wang, Shinji Watanabe, Toru Taniguchi, Dung Tran, Yuya Fujita

Figure 1 for An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions
Figure 2 for An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions
Figure 3 for An Investigation of End-to-End Multichannel Speech Recognition for Reverberant and Mismatch Conditions
Viaarxiv icon

Data Augmentation for Training Dialog Models Robust to Speech Recognition Errors

Jun 10, 2020
Longshaokan Wang, Maryam Fazel-Zarandi, Aditya Tiwari, Spyros Matsoukas, Lazaros Polymenakos

Figure 1 for Data Augmentation for Training Dialog Models Robust to Speech Recognition Errors
Figure 2 for Data Augmentation for Training Dialog Models Robust to Speech Recognition Errors
Figure 3 for Data Augmentation for Training Dialog Models Robust to Speech Recognition Errors
Figure 4 for Data Augmentation for Training Dialog Models Robust to Speech Recognition Errors
Viaarxiv icon

Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer

Jul 29, 2022
Cong-Thanh Do, Mohan Li, Rama Doddipatla

Figure 1 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 2 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 3 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 4 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Viaarxiv icon

Comparison and Analysis of New Curriculum Criteria for End-to-End ASR

Add code
Bookmark button
Alert button
Aug 10, 2022
Georgios Karakasidis, Tamás Grósz, Mikko Kurimo

Figure 1 for Comparison and Analysis of New Curriculum Criteria for End-to-End ASR
Figure 2 for Comparison and Analysis of New Curriculum Criteria for End-to-End ASR
Figure 3 for Comparison and Analysis of New Curriculum Criteria for End-to-End ASR
Figure 4 for Comparison and Analysis of New Curriculum Criteria for End-to-End ASR
Viaarxiv icon

A Comparative Study of Data Augmentation Techniques for Deep Learning Based Emotion Recognition

Nov 09, 2022
Ravi Shankar, Abdouh Harouna Kenfack, Arjun Somayazulu, Archana Venkataraman

Figure 1 for A Comparative Study of Data Augmentation Techniques for Deep Learning Based Emotion Recognition
Figure 2 for A Comparative Study of Data Augmentation Techniques for Deep Learning Based Emotion Recognition
Figure 3 for A Comparative Study of Data Augmentation Techniques for Deep Learning Based Emotion Recognition
Figure 4 for A Comparative Study of Data Augmentation Techniques for Deep Learning Based Emotion Recognition
Viaarxiv icon

VAIS ASR: Building a conversational speech recognition system using language model combination

Add code
Bookmark button
Alert button
Oct 12, 2019
Quang Minh Nguyen, Thai Binh Nguyen, Ngoc Phuong Pham, The Loc Nguyen

Figure 1 for VAIS ASR: Building a conversational speech recognition system using language model combination
Figure 2 for VAIS ASR: Building a conversational speech recognition system using language model combination
Figure 3 for VAIS ASR: Building a conversational speech recognition system using language model combination
Figure 4 for VAIS ASR: Building a conversational speech recognition system using language model combination
Viaarxiv icon

Character-Aware Attention-Based End-to-End Speech Recognition

Jan 06, 2020
Zhong Meng, Yashesh Gaur, Jinyu Li, Yifan Gong

Figure 1 for Character-Aware Attention-Based End-to-End Speech Recognition
Figure 2 for Character-Aware Attention-Based End-to-End Speech Recognition
Figure 3 for Character-Aware Attention-Based End-to-End Speech Recognition
Figure 4 for Character-Aware Attention-Based End-to-End Speech Recognition
Viaarxiv icon