Alert button

"speech recognition": models, code, and papers
Alert button

Shallow Fusion of Weighted Finite-State Transducer and Language Model for Text Normalization

Add code
Bookmark button
Alert button
Mar 29, 2022
Evelina Bakhturina, Yang Zhang, Boris Ginsburg

Figure 1 for Shallow Fusion of Weighted Finite-State Transducer and Language Model for Text Normalization
Figure 2 for Shallow Fusion of Weighted Finite-State Transducer and Language Model for Text Normalization
Figure 3 for Shallow Fusion of Weighted Finite-State Transducer and Language Model for Text Normalization
Figure 4 for Shallow Fusion of Weighted Finite-State Transducer and Language Model for Text Normalization
Viaarxiv icon

Improved Regularization Techniques for End-to-End Speech Recognition

Dec 19, 2017
Yingbo Zhou, Caiming Xiong, Richard Socher

Figure 1 for Improved Regularization Techniques for End-to-End Speech Recognition
Figure 2 for Improved Regularization Techniques for End-to-End Speech Recognition
Figure 3 for Improved Regularization Techniques for End-to-End Speech Recognition
Figure 4 for Improved Regularization Techniques for End-to-End Speech Recognition
Viaarxiv icon

Is Attention always needed? A Case Study on Language Identification from Speech

Oct 05, 2021
Atanu Mandal, Santanu Pal, Indranil Dutta, Mahidas Bhattacharya, Sudip Kumar Naskar

Figure 1 for Is Attention always needed? A Case Study on Language Identification from Speech
Figure 2 for Is Attention always needed? A Case Study on Language Identification from Speech
Figure 3 for Is Attention always needed? A Case Study on Language Identification from Speech
Figure 4 for Is Attention always needed? A Case Study on Language Identification from Speech
Viaarxiv icon

M2MeT: The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Challenge

Add code
Bookmark button
Alert button
Oct 14, 2021
Fan Yu, Shiliang Zhang, Yihui Fu, Lei Xie, Siqi Zheng, Zhihao Du, Weilong Huang, Pengcheng Guo, Zhijie Yan, Bin Ma, Xin Xu, Hui Bu

Figure 1 for M2MeT: The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Challenge
Figure 2 for M2MeT: The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Challenge
Figure 3 for M2MeT: The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Challenge
Figure 4 for M2MeT: The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Challenge
Viaarxiv icon

Sequence-based Multi-lingual Low Resource Speech Recognition

Mar 06, 2018
Siddharth Dalmia, Ramon Sanabria, Florian Metze, Alan W. Black

Figure 1 for Sequence-based Multi-lingual Low Resource Speech Recognition
Viaarxiv icon

Representation learning through cross-modal conditional teacher-student training for speech emotion recognition

Nov 30, 2021
Sundararajan Srinivasan, Zhaocheng Huang, Katrin Kirchhoff

Figure 1 for Representation learning through cross-modal conditional teacher-student training for speech emotion recognition
Figure 2 for Representation learning through cross-modal conditional teacher-student training for speech emotion recognition
Figure 3 for Representation learning through cross-modal conditional teacher-student training for speech emotion recognition
Viaarxiv icon

High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model

Mar 17, 2020
Jinyu Li, Rui Zhao, Eric Sun, Jeremy H. M. Wong, Amit Das, Zhong Meng, Yifan Gong

Figure 1 for High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model
Figure 2 for High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model
Figure 3 for High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model
Figure 4 for High-Accuracy and Low-Latency Speech Recognition with Two-Head Contextual Layer Trajectory LSTM Model
Viaarxiv icon

Summary On The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Grand Challenge

Feb 08, 2022
Fan Yu, Shiliang Zhang, Pengcheng Guo, Yihui Fu, Zhihao Du, Siqi Zheng, Weilong Huang, Lei Xie, Zheng-Hua Tan, DeLiang Wang, Yanmin Qian, Kong Aik Lee, Zhijie Yan, Bin Ma, Xin Xu, Hui Bu

Figure 1 for Summary On The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Grand Challenge
Figure 2 for Summary On The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Grand Challenge
Figure 3 for Summary On The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Grand Challenge
Figure 4 for Summary On The ICASSP 2022 Multi-Channel Multi-Party Meeting Transcription Grand Challenge
Viaarxiv icon

Towards speech-to-text translation without speech recognition

Add code
Bookmark button
Alert button
Feb 13, 2017
Sameer Bansal, Herman Kamper, Adam Lopez, Sharon Goldwater

Figure 1 for Towards speech-to-text translation without speech recognition
Figure 2 for Towards speech-to-text translation without speech recognition
Figure 3 for Towards speech-to-text translation without speech recognition
Figure 4 for Towards speech-to-text translation without speech recognition
Viaarxiv icon

Integrating Source-channel and Attention-based Sequence-to-sequence Models for Speech Recognition

Sep 14, 2019
Qiujia Li, Chao Zhang, Philip C. Woodland

Figure 1 for Integrating Source-channel and Attention-based Sequence-to-sequence Models for Speech Recognition
Figure 2 for Integrating Source-channel and Attention-based Sequence-to-sequence Models for Speech Recognition
Figure 3 for Integrating Source-channel and Attention-based Sequence-to-sequence Models for Speech Recognition
Figure 4 for Integrating Source-channel and Attention-based Sequence-to-sequence Models for Speech Recognition
Viaarxiv icon