Alert button

"speech recognition": models, code, and papers
Alert button

CTC Variations Through New WFST Topologies

Add code
Bookmark button
Alert button
Oct 06, 2021
Aleksandr Laptev, Somshubra Majumdar, Boris Ginsburg

Figure 1 for CTC Variations Through New WFST Topologies
Figure 2 for CTC Variations Through New WFST Topologies
Figure 3 for CTC Variations Through New WFST Topologies
Figure 4 for CTC Variations Through New WFST Topologies
Viaarxiv icon

Structured Language Modeling for Speech Recognition

Jan 25, 2000
Ciprian Chelba, Frederick Jelinek

Figure 1 for Structured Language Modeling for Speech Recognition
Figure 2 for Structured Language Modeling for Speech Recognition
Figure 3 for Structured Language Modeling for Speech Recognition
Figure 4 for Structured Language Modeling for Speech Recognition
Viaarxiv icon

Exploiting Pre-Trained ASR Models for Alzheimer's Disease Recognition Through Spontaneous Speech

Add code
Bookmark button
Alert button
Oct 04, 2021
Ying Qin, Wei Liu, Zhiyuan Peng, Si-Ioi Ng, Jingyu Li, Haibo Hu, Tan Lee

Figure 1 for Exploiting Pre-Trained ASR Models for Alzheimer's Disease Recognition Through Spontaneous Speech
Figure 2 for Exploiting Pre-Trained ASR Models for Alzheimer's Disease Recognition Through Spontaneous Speech
Figure 3 for Exploiting Pre-Trained ASR Models for Alzheimer's Disease Recognition Through Spontaneous Speech
Figure 4 for Exploiting Pre-Trained ASR Models for Alzheimer's Disease Recognition Through Spontaneous Speech
Viaarxiv icon

End-to-end Speech Recognition with Word-based RNN Language Models

Aug 08, 2018
Takaaki Hori, Jaejin Cho, Shinji Watanabe

Figure 1 for End-to-end Speech Recognition with Word-based RNN Language Models
Figure 2 for End-to-end Speech Recognition with Word-based RNN Language Models
Figure 3 for End-to-end Speech Recognition with Word-based RNN Language Models
Figure 4 for End-to-end Speech Recognition with Word-based RNN Language Models
Viaarxiv icon

The Norwegian Parliamentary Speech Corpus

Add code
Bookmark button
Alert button
Jan 26, 2022
Per Erik Solberg, Pablo Ortiz

Figure 1 for The Norwegian Parliamentary Speech Corpus
Figure 2 for The Norwegian Parliamentary Speech Corpus
Figure 3 for The Norwegian Parliamentary Speech Corpus
Viaarxiv icon

HASP: A High-Performance Adaptive Mobile Security Enhancement Against Malicious Speech Recognition

Sep 04, 2018
Zirui Xu, Fuxun Yu, Chenchen Liu, Xiang Chen

Figure 1 for HASP: A High-Performance Adaptive Mobile Security Enhancement Against Malicious Speech Recognition
Figure 2 for HASP: A High-Performance Adaptive Mobile Security Enhancement Against Malicious Speech Recognition
Figure 3 for HASP: A High-Performance Adaptive Mobile Security Enhancement Against Malicious Speech Recognition
Figure 4 for HASP: A High-Performance Adaptive Mobile Security Enhancement Against Malicious Speech Recognition
Viaarxiv icon

Improving the fusion of acoustic and text representations in RNN-T

Jan 25, 2022
Chao Zhang, Bo Li, Zhiyun Lu, Tara N. Sainath, Shuo-yiin Chang

Figure 1 for Improving the fusion of acoustic and text representations in RNN-T
Figure 2 for Improving the fusion of acoustic and text representations in RNN-T
Figure 3 for Improving the fusion of acoustic and text representations in RNN-T
Figure 4 for Improving the fusion of acoustic and text representations in RNN-T
Viaarxiv icon

A Text-to-Speech Pipeline, Evaluation Methodology, and Initial Fine-Tuning Results for Child Speech Synthesis

Add code
Bookmark button
Alert button
Apr 04, 2022
Rishabh Jain, Mariam Yiwere, Dan Bigioi, Peter Corcoran, Horia Cucu

Figure 1 for A Text-to-Speech Pipeline, Evaluation Methodology, and Initial Fine-Tuning Results for Child Speech Synthesis
Figure 2 for A Text-to-Speech Pipeline, Evaluation Methodology, and Initial Fine-Tuning Results for Child Speech Synthesis
Figure 3 for A Text-to-Speech Pipeline, Evaluation Methodology, and Initial Fine-Tuning Results for Child Speech Synthesis
Figure 4 for A Text-to-Speech Pipeline, Evaluation Methodology, and Initial Fine-Tuning Results for Child Speech Synthesis
Viaarxiv icon

Hybrid Data Augmentation and Deep Attention-based Dilated Convolutional-Recurrent Neural Networks for Speech Emotion Recognition

Sep 18, 2021
Nhat Truong Pham, Duc Ngoc Minh Dang, Sy Dzung Nguyen

Figure 1 for Hybrid Data Augmentation and Deep Attention-based Dilated Convolutional-Recurrent Neural Networks for Speech Emotion Recognition
Figure 2 for Hybrid Data Augmentation and Deep Attention-based Dilated Convolutional-Recurrent Neural Networks for Speech Emotion Recognition
Figure 3 for Hybrid Data Augmentation and Deep Attention-based Dilated Convolutional-Recurrent Neural Networks for Speech Emotion Recognition
Figure 4 for Hybrid Data Augmentation and Deep Attention-based Dilated Convolutional-Recurrent Neural Networks for Speech Emotion Recognition
Viaarxiv icon

Residual Convolutional CTC Networks for Automatic Speech Recognition

Add code
Bookmark button
Alert button
Feb 24, 2017
Yisen Wang, Xuejiao Deng, Songbai Pu, Zhiheng Huang

Figure 1 for Residual Convolutional CTC Networks for Automatic Speech Recognition
Figure 2 for Residual Convolutional CTC Networks for Automatic Speech Recognition
Figure 3 for Residual Convolutional CTC Networks for Automatic Speech Recognition
Figure 4 for Residual Convolutional CTC Networks for Automatic Speech Recognition
Viaarxiv icon