Alert button

"speech": models, code, and papers
Alert button

Reducing Confusion in Active Learning for Part-Of-Speech Tagging

Nov 02, 2020
Aditi Chaudhary, Antonios Anastasopoulos, Zaid Sheikh, Graham Neubig

Figure 1 for Reducing Confusion in Active Learning for Part-Of-Speech Tagging
Figure 2 for Reducing Confusion in Active Learning for Part-Of-Speech Tagging
Figure 3 for Reducing Confusion in Active Learning for Part-Of-Speech Tagging
Figure 4 for Reducing Confusion in Active Learning for Part-Of-Speech Tagging
Viaarxiv icon

Machine Speech Chain with One-shot Speaker Adaptation

Mar 28, 2018
Andros Tjandra, Sakriani Sakti, Satoshi Nakamura

Figure 1 for Machine Speech Chain with One-shot Speaker Adaptation
Figure 2 for Machine Speech Chain with One-shot Speaker Adaptation
Figure 3 for Machine Speech Chain with One-shot Speaker Adaptation
Figure 4 for Machine Speech Chain with One-shot Speaker Adaptation
Viaarxiv icon

CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement

Sep 23, 2019
Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Amir Hussain

Figure 1 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Figure 2 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Figure 3 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Figure 4 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Viaarxiv icon

Empirical study on BlenderBot 2.0 Errors Analysis in terms of Model, Data and User-Centric Approach

Jan 10, 2022
Jungseob Lee, Midan Shim, Suhyune Son, Yujin Kim, Chanjun Park, Heuiseok Lim

Figure 1 for Empirical study on BlenderBot 2.0 Errors Analysis in terms of Model, Data and User-Centric Approach
Figure 2 for Empirical study on BlenderBot 2.0 Errors Analysis in terms of Model, Data and User-Centric Approach
Figure 3 for Empirical study on BlenderBot 2.0 Errors Analysis in terms of Model, Data and User-Centric Approach
Figure 4 for Empirical study on BlenderBot 2.0 Errors Analysis in terms of Model, Data and User-Centric Approach
Viaarxiv icon

NAS-Bench-Suite: NAS Evaluation is (Now) Surprisingly Easy

Feb 11, 2022
Yash Mehta, Colin White, Arber Zela, Arjun Krishnakumar, Guri Zabergja, Shakiba Moradian, Mahmoud Safari, Kaicheng Yu, Frank Hutter

Figure 1 for NAS-Bench-Suite: NAS Evaluation is (Now) Surprisingly Easy
Figure 2 for NAS-Bench-Suite: NAS Evaluation is (Now) Surprisingly Easy
Figure 3 for NAS-Bench-Suite: NAS Evaluation is (Now) Surprisingly Easy
Figure 4 for NAS-Bench-Suite: NAS Evaluation is (Now) Surprisingly Easy
Viaarxiv icon

Model-Based Approach for Measuring the Fairness in ASR

Sep 19, 2021
Zhe Liu, Irina-Elena Veliche, Fuchun Peng

Figure 1 for Model-Based Approach for Measuring the Fairness in ASR
Figure 2 for Model-Based Approach for Measuring the Fairness in ASR
Figure 3 for Model-Based Approach for Measuring the Fairness in ASR
Figure 4 for Model-Based Approach for Measuring the Fairness in ASR
Viaarxiv icon

ExcitNet vocoder: A neural excitation model for parametric speech synthesis systems

Nov 09, 2018
Eunwoo Song, Kyungguen Byun, Hong-Goo Kang

Figure 1 for ExcitNet vocoder: A neural excitation model for parametric speech synthesis systems
Figure 2 for ExcitNet vocoder: A neural excitation model for parametric speech synthesis systems
Figure 3 for ExcitNet vocoder: A neural excitation model for parametric speech synthesis systems
Figure 4 for ExcitNet vocoder: A neural excitation model for parametric speech synthesis systems
Viaarxiv icon

Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition

Nov 19, 2019
Jibin Wu, Emre Yilmaz, Malu Zhang, Haizhou Li, Kay Chen Tan

Figure 1 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Figure 2 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Figure 3 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Figure 4 for Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition
Viaarxiv icon

SingGAN: Generative Adversarial Network For High-Fidelity Singing Voice Generation

Oct 14, 2021
Feiyang Chen, Rongjie Huang, Chenye Cui, Yi Ren, Jinglin Liu, Zhou Zhao, Nicholas Yuan, Baoxing Huai

Figure 1 for SingGAN: Generative Adversarial Network For High-Fidelity Singing Voice Generation
Figure 2 for SingGAN: Generative Adversarial Network For High-Fidelity Singing Voice Generation
Figure 3 for SingGAN: Generative Adversarial Network For High-Fidelity Singing Voice Generation
Figure 4 for SingGAN: Generative Adversarial Network For High-Fidelity Singing Voice Generation
Viaarxiv icon

Beyond $L_p$ clipping: Equalization-based Psychoacoustic Attacks against ASRs

Oct 25, 2021
Hadi Abdullah, Muhammad Sajidur Rahman, Christian Peeters, Cassidy Gibson, Washington Garcia, Vincent Bindschaedler, Thomas Shrimpton, Patrick Traynor

Figure 1 for Beyond $L_p$ clipping: Equalization-based Psychoacoustic Attacks against ASRs
Figure 2 for Beyond $L_p$ clipping: Equalization-based Psychoacoustic Attacks against ASRs
Figure 3 for Beyond $L_p$ clipping: Equalization-based Psychoacoustic Attacks against ASRs
Figure 4 for Beyond $L_p$ clipping: Equalization-based Psychoacoustic Attacks against ASRs
Viaarxiv icon