Alert button

"speech": models, code, and papers
Alert button

Evolution of Part-of-Speech in Classical Chinese

Sep 23, 2020
Bai Li

Figure 1 for Evolution of Part-of-Speech in Classical Chinese
Figure 2 for Evolution of Part-of-Speech in Classical Chinese
Figure 3 for Evolution of Part-of-Speech in Classical Chinese
Figure 4 for Evolution of Part-of-Speech in Classical Chinese
Viaarxiv icon

Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese

Add code
Bookmark button
Alert button
May 21, 2022
Kurt Micallef, Albert Gatt, Marc Tanti, Lonneke van der Plas, Claudia Borg

Figure 1 for Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese
Figure 2 for Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese
Figure 3 for Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese
Figure 4 for Pre-training Data Quality and Quantity for a Low-Resource Language: New Corpus and BERT Models for Maltese
Viaarxiv icon

A high quality and phonetic balanced speech corpus for Vietnamese

Apr 11, 2019
Pham Ngoc Phuong, Quoc Truong Do, Luong Chi Mai

Figure 1 for A high quality and phonetic balanced speech corpus for Vietnamese
Figure 2 for A high quality and phonetic balanced speech corpus for Vietnamese
Figure 3 for A high quality and phonetic balanced speech corpus for Vietnamese
Figure 4 for A high quality and phonetic balanced speech corpus for Vietnamese
Viaarxiv icon

Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data

Oct 22, 2020
Thibault Doutre, Wei Han, Min Ma, Zhiyun Lu, Chung-Cheng Chiu, Ruoming Pang, Arun Narayanan, Ananya Misra, Yu Zhang, Liangliang Cao

Figure 1 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 2 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 3 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Figure 4 for Improving Streaming Automatic Speech Recognition With Non-Streaming Model Distillation On Unsupervised Data
Viaarxiv icon

Don't shoot butterfly with rifles: Multi-channel Continuous Speech Separation with Early Exit Transformer

Add code
Bookmark button
Alert button
Oct 23, 2020
Sanyuan Chen, Yu Wu, Zhuo Chen, Takuya Yoshioka, Shujie Liu, Jinyu Li

Figure 1 for Don't shoot butterfly with rifles: Multi-channel Continuous Speech Separation with Early Exit Transformer
Figure 2 for Don't shoot butterfly with rifles: Multi-channel Continuous Speech Separation with Early Exit Transformer
Figure 3 for Don't shoot butterfly with rifles: Multi-channel Continuous Speech Separation with Early Exit Transformer
Viaarxiv icon

Proteno: Text Normalization with Limited Data for Fast Deployment in Text to Speech Systems

Add code
Bookmark button
Alert button
Apr 15, 2021
Shubhi Tyagi, Antonio Bonafonte, Jaime Lorenzo-Trueba, Javier Latorre

Figure 1 for Proteno: Text Normalization with Limited Data for Fast Deployment in Text to Speech Systems
Figure 2 for Proteno: Text Normalization with Limited Data for Fast Deployment in Text to Speech Systems
Figure 3 for Proteno: Text Normalization with Limited Data for Fast Deployment in Text to Speech Systems
Figure 4 for Proteno: Text Normalization with Limited Data for Fast Deployment in Text to Speech Systems
Viaarxiv icon

The DKU-Duke-Lenovo System Description for the Third DIHARD Speech Diarization Challenge

Feb 06, 2021
Weiqing Wang, Qingjian Lin, Danwei Cai, Lin Yang, Ming Li

Figure 1 for The DKU-Duke-Lenovo System Description for the Third DIHARD Speech Diarization Challenge
Figure 2 for The DKU-Duke-Lenovo System Description for the Third DIHARD Speech Diarization Challenge
Figure 3 for The DKU-Duke-Lenovo System Description for the Third DIHARD Speech Diarization Challenge
Figure 4 for The DKU-Duke-Lenovo System Description for the Third DIHARD Speech Diarization Challenge
Viaarxiv icon

VRAIN-UPV MLLP's system for the Blizzard Challenge 2021

Add code
Bookmark button
Alert button
Oct 29, 2021
Alejandro Pérez-González-de-Martos, Albert Sanchis, Alfons Juan

Figure 1 for VRAIN-UPV MLLP's system for the Blizzard Challenge 2021
Figure 2 for VRAIN-UPV MLLP's system for the Blizzard Challenge 2021
Figure 3 for VRAIN-UPV MLLP's system for the Blizzard Challenge 2021
Figure 4 for VRAIN-UPV MLLP's system for the Blizzard Challenge 2021
Viaarxiv icon

Customizable End-to-end Optimization of Online Neural Network-supported Dereverberation for Hearing Devices

Apr 06, 2022
Jean-Marie Lemercier, Joachim Thiemann, Raphael Koning, Timo Gerkmann

Figure 1 for Customizable End-to-end Optimization of Online Neural Network-supported Dereverberation for Hearing Devices
Figure 2 for Customizable End-to-end Optimization of Online Neural Network-supported Dereverberation for Hearing Devices
Figure 3 for Customizable End-to-end Optimization of Online Neural Network-supported Dereverberation for Hearing Devices
Figure 4 for Customizable End-to-end Optimization of Online Neural Network-supported Dereverberation for Hearing Devices
Viaarxiv icon

i-Code: An Integrative and Composable Multimodal Learning Framework

May 05, 2022
Ziyi Yang, Yuwei Fang, Chenguang Zhu, Reid Pryzant, Dongdong Chen, Yu Shi, Yichong Xu, Yao Qian, Mei Gao, Yi-Ling Chen, Liyang Lu, Yujia Xie, Robert Gmyr, Noel Codella, Naoyuki Kanda, Bin Xiao, Lu Yuan, Takuya Yoshioka, Michael Zeng, Xuedong Huang

Figure 1 for i-Code: An Integrative and Composable Multimodal Learning Framework
Figure 2 for i-Code: An Integrative and Composable Multimodal Learning Framework
Figure 3 for i-Code: An Integrative and Composable Multimodal Learning Framework
Figure 4 for i-Code: An Integrative and Composable Multimodal Learning Framework
Viaarxiv icon