Alert button

"speech recognition": models, code, and papers
Alert button

Improving Mandarin Speech Recogntion with Block-augmented Transformer

Add code
Bookmark button
Alert button
Jul 24, 2022
Xiaoming Ren, Huifeng Zhu, Liuwei Wei, Minghui Wu, Jie Hao

Figure 1 for Improving Mandarin Speech Recogntion with Block-augmented Transformer
Figure 2 for Improving Mandarin Speech Recogntion with Block-augmented Transformer
Figure 3 for Improving Mandarin Speech Recogntion with Block-augmented Transformer
Figure 4 for Improving Mandarin Speech Recogntion with Block-augmented Transformer
Viaarxiv icon

Speaker Identification using Speech Recognition

May 29, 2022
Syeda Rabia Arshad, Syed Mujtaba Haider, Abdul Basit Mughal

Figure 1 for Speaker Identification using Speech Recognition
Figure 2 for Speaker Identification using Speech Recognition
Viaarxiv icon

Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition

Add code
Bookmark button
Alert button
Nov 19, 2019
Jibin Wu, Emre Yilmaz, Malu Zhang, Haizhou Li, Kay Chen Tan

Figure 1 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Figure 2 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Figure 3 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Figure 4 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Viaarxiv icon

A baseline model for computationally inexpensive speech recognition for Kazakh using the Coqui STT framework

Add code
Bookmark button
Alert button
Jul 19, 2021
Ilnar Salimzianov

Figure 1 for A baseline model for computationally inexpensive speech recognition for Kazakh using the Coqui STT framework
Figure 2 for A baseline model for computationally inexpensive speech recognition for Kazakh using the Coqui STT framework
Viaarxiv icon

Blank Collapse: Compressing CTC emission for the faster decoding

Add code
Bookmark button
Alert button
Oct 31, 2022
Minkyu Jung, Ohhyeok Kwon, Seunghyun Seo, Soonshin Seo

Figure 1 for Blank Collapse: Compressing CTC emission for the faster decoding
Figure 2 for Blank Collapse: Compressing CTC emission for the faster decoding
Figure 3 for Blank Collapse: Compressing CTC emission for the faster decoding
Figure 4 for Blank Collapse: Compressing CTC emission for the faster decoding
Viaarxiv icon

Speech Recognition for Endangered and Extinct Samoyedic languages

Add code
Bookmark button
Alert button
Dec 09, 2020
Niko Partanen, Mika Hämäläinen, Tiina Klooster

Figure 1 for Speech Recognition for Endangered and Extinct Samoyedic languages
Figure 2 for Speech Recognition for Endangered and Extinct Samoyedic languages
Figure 3 for Speech Recognition for Endangered and Extinct Samoyedic languages
Figure 4 for Speech Recognition for Endangered and Extinct Samoyedic languages
Viaarxiv icon

Simplified Self-Attention for Transformer-based End-to-End Speech Recognition

May 21, 2020
Haoneng Luo, Shiliang Zhang, Ming Lei, Lei Xie

Figure 1 for Simplified Self-Attention for Transformer-based End-to-End Speech Recognition
Figure 2 for Simplified Self-Attention for Transformer-based End-to-End Speech Recognition
Figure 3 for Simplified Self-Attention for Transformer-based End-to-End Speech Recognition
Figure 4 for Simplified Self-Attention for Transformer-based End-to-End Speech Recognition
Viaarxiv icon

Cross-Modal Transformer-Based Neural Correction Models for Automatic Speech Recognition

Jul 04, 2021
Tomohiro Tanaka, Ryo Masumura, Mana Ihori, Akihiko Takashima, Takafumi Moriya, Takanori Ashihara, Shota Orihashi, Naoki Makishima

Figure 1 for Cross-Modal Transformer-Based Neural Correction Models for Automatic Speech Recognition
Figure 2 for Cross-Modal Transformer-Based Neural Correction Models for Automatic Speech Recognition
Figure 3 for Cross-Modal Transformer-Based Neural Correction Models for Automatic Speech Recognition
Figure 4 for Cross-Modal Transformer-Based Neural Correction Models for Automatic Speech Recognition
Viaarxiv icon

Modality Attention for End-to-End Audio-visual Speech Recognition

Nov 13, 2018
Pan Zhou, Wenwen Yang, Wei Chen, Yanfeng Wang, Jia Jia

Figure 1 for Modality Attention for End-to-End Audio-visual Speech Recognition
Figure 2 for Modality Attention for End-to-End Audio-visual Speech Recognition
Figure 3 for Modality Attention for End-to-End Audio-visual Speech Recognition
Viaarxiv icon

Hierarchical Multitask Learning for CTC-based Speech Recognition

Add code
Bookmark button
Alert button
Jul 17, 2018
Kalpesh Krishna, Shubham Toshniwal, Karen Livescu

Figure 1 for Hierarchical Multitask Learning for CTC-based Speech Recognition
Figure 2 for Hierarchical Multitask Learning for CTC-based Speech Recognition
Figure 3 for Hierarchical Multitask Learning for CTC-based Speech Recognition
Figure 4 for Hierarchical Multitask Learning for CTC-based Speech Recognition
Viaarxiv icon