Alert button

"speech": models, code, and papers
Alert button

TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech

Add code
Bookmark button
Alert button
Jul 12, 2020
Andy T. Liu, Shang-Wen Li, Hung-yi Lee

Figure 1 for TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech
Figure 2 for TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech
Figure 3 for TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech
Figure 4 for TERA: Self-Supervised Learning of Transformer Encoder Representation for Speech
Viaarxiv icon

The Phonetic Footprint of Parkinson's Disease

Dec 21, 2021
Philipp Klumpp, Tomás Arias-Vergara, Juan Camilo Vásquez-Correa, Paula Andrea Pérez-Toro, Juan Rafael Orozco-Arroyave, Anton Batliner, Elmar Nöth

Figure 1 for The Phonetic Footprint of Parkinson's Disease
Figure 2 for The Phonetic Footprint of Parkinson's Disease
Figure 3 for The Phonetic Footprint of Parkinson's Disease
Figure 4 for The Phonetic Footprint of Parkinson's Disease
Viaarxiv icon

Finnish Parliament ASR corpus - Analysis, benchmarks and statistics

Add code
Bookmark button
Alert button
Mar 28, 2022
Anja Virkkunen, Aku Rouhe, Nhan Phan, Mikko Kurimo

Figure 1 for Finnish Parliament ASR corpus - Analysis, benchmarks and statistics
Figure 2 for Finnish Parliament ASR corpus - Analysis, benchmarks and statistics
Figure 3 for Finnish Parliament ASR corpus - Analysis, benchmarks and statistics
Figure 4 for Finnish Parliament ASR corpus - Analysis, benchmarks and statistics
Viaarxiv icon

Improving Dysarthric Speech Intelligibility Using Cycle-consistent Adversarial Training

Add code
Bookmark button
Alert button
Jan 10, 2020
Seung Hee Yang, Minhwa Chung

Figure 1 for Improving Dysarthric Speech Intelligibility Using Cycle-consistent Adversarial Training
Figure 2 for Improving Dysarthric Speech Intelligibility Using Cycle-consistent Adversarial Training
Figure 3 for Improving Dysarthric Speech Intelligibility Using Cycle-consistent Adversarial Training
Viaarxiv icon

Streaming non-autoregressive model for any-to-many voice conversion

Add code
Bookmark button
Alert button
Jun 15, 2022
Ziyi Chen, Haoran Miao, Pengyuan Zhang

Figure 1 for Streaming non-autoregressive model for any-to-many voice conversion
Figure 2 for Streaming non-autoregressive model for any-to-many voice conversion
Figure 3 for Streaming non-autoregressive model for any-to-many voice conversion
Figure 4 for Streaming non-autoregressive model for any-to-many voice conversion
Viaarxiv icon

Speech Recognition using EEG signals recorded using dry electrodes

Aug 13, 2020
Gautam Krishna, Co Tran, Mason Carnahan, Morgan M Hagood, Ahmed H Tewfik

Figure 1 for Speech Recognition using EEG signals recorded using dry electrodes
Figure 2 for Speech Recognition using EEG signals recorded using dry electrodes
Figure 3 for Speech Recognition using EEG signals recorded using dry electrodes
Figure 4 for Speech Recognition using EEG signals recorded using dry electrodes
Viaarxiv icon

Integrating Text Inputs For Training and Adapting RNN Transducer ASR Models

Feb 26, 2022
Samuel Thomas, Brian Kingsbury, George Saon, Hong-Kwang J. Kuo

Figure 1 for Integrating Text Inputs For Training and Adapting RNN Transducer ASR Models
Figure 2 for Integrating Text Inputs For Training and Adapting RNN Transducer ASR Models
Figure 3 for Integrating Text Inputs For Training and Adapting RNN Transducer ASR Models
Figure 4 for Integrating Text Inputs For Training and Adapting RNN Transducer ASR Models
Viaarxiv icon

Lost in Interpreting: Speech Translation from Source or Interpreter?

Jun 17, 2021
Dominik Macháček, Matúš Žilinec, Ondřej Bojar

Figure 1 for Lost in Interpreting: Speech Translation from Source or Interpreter?
Figure 2 for Lost in Interpreting: Speech Translation from Source or Interpreter?
Figure 3 for Lost in Interpreting: Speech Translation from Source or Interpreter?
Figure 4 for Lost in Interpreting: Speech Translation from Source or Interpreter?
Viaarxiv icon

C3-DINO: Joint Contrastive and Non-contrastive Self-Supervised Learning for Speaker Verification

Aug 15, 2022
Chunlei Zhang, Dong Yu

Figure 1 for C3-DINO: Joint Contrastive and Non-contrastive Self-Supervised Learning for Speaker Verification
Figure 2 for C3-DINO: Joint Contrastive and Non-contrastive Self-Supervised Learning for Speaker Verification
Figure 3 for C3-DINO: Joint Contrastive and Non-contrastive Self-Supervised Learning for Speaker Verification
Figure 4 for C3-DINO: Joint Contrastive and Non-contrastive Self-Supervised Learning for Speaker Verification
Viaarxiv icon

Attention-based multi-task learning for speech-enhancement and speaker-identification in multi-speaker dialogue scenario

Jan 07, 2021
Chiang-Jen Peng, Yun-Ju Chan, Cheng Yu, Syu-Siang Wang, Yu Tsao, Tai-Shih Chi

Figure 1 for Attention-based multi-task learning for speech-enhancement and speaker-identification in multi-speaker dialogue scenario
Figure 2 for Attention-based multi-task learning for speech-enhancement and speaker-identification in multi-speaker dialogue scenario
Figure 3 for Attention-based multi-task learning for speech-enhancement and speaker-identification in multi-speaker dialogue scenario
Figure 4 for Attention-based multi-task learning for speech-enhancement and speaker-identification in multi-speaker dialogue scenario
Viaarxiv icon