Alert button

"speech recognition": models, code, and papers
Alert button

Extending Recurrent Neural Aligner for Streaming End-to-End Speech Recognition in Mandarin

Jun 17, 2018
Linhao Dong, Shiyu Zhou, Wei Chen, Bo Xu

Figure 1 for Extending Recurrent Neural Aligner for Streaming End-to-End Speech Recognition in Mandarin
Figure 2 for Extending Recurrent Neural Aligner for Streaming End-to-End Speech Recognition in Mandarin
Figure 3 for Extending Recurrent Neural Aligner for Streaming End-to-End Speech Recognition in Mandarin
Figure 4 for Extending Recurrent Neural Aligner for Streaming End-to-End Speech Recognition in Mandarin
Viaarxiv icon

Exploring End-to-End Techniques for Low-Resource Speech Recognition

Jul 02, 2018
Vladimir Bataev, Maxim Korenevsky, Ivan Medennikov, Alexander Zatvornitskiy

Figure 1 for Exploring End-to-End Techniques for Low-Resource Speech Recognition
Figure 2 for Exploring End-to-End Techniques for Low-Resource Speech Recognition
Figure 3 for Exploring End-to-End Techniques for Low-Resource Speech Recognition
Figure 4 for Exploring End-to-End Techniques for Low-Resource Speech Recognition
Viaarxiv icon

ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers

Add code
Bookmark button
Alert button
May 17, 2020
Jung-Woo Ha, Kihyun Nam, Jingu Kang, Sang-Woo Lee, Sohee Yang, Hyunhoon Jung, Eunmi Kim, Hyeji Kim, Soojin Kim, Hyun Ah Kim, Kyoungtae Doh, Chan Kyu Lee, Nako Sung, Sunghun Kim

Figure 1 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Figure 2 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Figure 3 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Figure 4 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Viaarxiv icon

Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition

Nov 09, 2019
Alex Bie, Bharat Venkitesh, Joao Monteiro, Md. Akmal Haidar, Mehdi Rezagholizadeh

Figure 1 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Figure 2 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Figure 3 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Figure 4 for Fully Quantizing a Simplified Transformer for End-to-end Speech Recognition
Viaarxiv icon

Semi-Supervised Model Training for Unbounded Conversational Speech Recognition

May 26, 2017
Shane Walker, Morten Pedersen, Iroro Orife, Jason Flaks

Figure 1 for Semi-Supervised Model Training for Unbounded Conversational Speech Recognition
Figure 2 for Semi-Supervised Model Training for Unbounded Conversational Speech Recognition
Figure 3 for Semi-Supervised Model Training for Unbounded Conversational Speech Recognition
Figure 4 for Semi-Supervised Model Training for Unbounded Conversational Speech Recognition
Viaarxiv icon

When Is TTS Augmentation Through a Pivot Language Useful?

Add code
Bookmark button
Alert button
Jul 20, 2022
Nathaniel Robinson, Perez Ogayo, Swetha Gangu, David R. Mortensen, Shinji Watanabe

Figure 1 for When Is TTS Augmentation Through a Pivot Language Useful?
Figure 2 for When Is TTS Augmentation Through a Pivot Language Useful?
Figure 3 for When Is TTS Augmentation Through a Pivot Language Useful?
Figure 4 for When Is TTS Augmentation Through a Pivot Language Useful?
Viaarxiv icon

A new Speech Feature Fusion method with cross gate parallel CNN for Speaker Recognition

Nov 24, 2022
Jiacheng Zhang, Wenyi Yan, Ye Zhang

Figure 1 for A new Speech Feature Fusion method with cross gate parallel CNN for Speaker Recognition
Figure 2 for A new Speech Feature Fusion method with cross gate parallel CNN for Speaker Recognition
Figure 3 for A new Speech Feature Fusion method with cross gate parallel CNN for Speaker Recognition
Figure 4 for A new Speech Feature Fusion method with cross gate parallel CNN for Speaker Recognition
Viaarxiv icon

Towards Transfer Learning of wav2vec 2.0 for Automatic Lyric Transcription

Add code
Bookmark button
Alert button
Jul 20, 2022
Longshen Ou, Xiangming Gu, Ye Wang

Figure 1 for Towards Transfer Learning of wav2vec 2.0 for Automatic Lyric Transcription
Figure 2 for Towards Transfer Learning of wav2vec 2.0 for Automatic Lyric Transcription
Figure 3 for Towards Transfer Learning of wav2vec 2.0 for Automatic Lyric Transcription
Figure 4 for Towards Transfer Learning of wav2vec 2.0 for Automatic Lyric Transcription
Viaarxiv icon

Phonological modeling for continuous speech recognition in Korean

Jul 18, 1996
WonIl Lee, Geunbae Lee, Jong-Hyeok Lee

Figure 1 for Phonological modeling for continuous speech recognition in Korean
Figure 2 for Phonological modeling for continuous speech recognition in Korean
Figure 3 for Phonological modeling for continuous speech recognition in Korean
Figure 4 for Phonological modeling for continuous speech recognition in Korean
Viaarxiv icon

ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context

May 16, 2020
Wei Han, Zhengdong Zhang, Yu Zhang, Jiahui Yu, Chung-Cheng Chiu, James Qin, Anmol Gulati, Ruoming Pang, Yonghui Wu

Figure 1 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Figure 2 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Figure 3 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Figure 4 for ContextNet: Improving Convolutional Neural Networks for Automatic Speech Recognition with Global Context
Viaarxiv icon