Alert button

"speech": models, code, and papers
Alert button

Modeling Animal Vocalizations through Synthesizers

Add code
Bookmark button
Alert button
Oct 19, 2022
Masato Hagiwara, Maddie Cusimano, Jen-Yu Liu

Figure 1 for Modeling Animal Vocalizations through Synthesizers
Figure 2 for Modeling Animal Vocalizations through Synthesizers
Figure 3 for Modeling Animal Vocalizations through Synthesizers
Viaarxiv icon

G-Augment: Searching For The Meta-Structure Of Data Augmentation Policies For ASR

Oct 19, 2022
Gary Wang, Ekin D. Cubuk, Andrew Rosenberg, Shuyang Cheng, Ron J. Weiss, Bhuvana Ramabhadran, Pedro J. Moreno, Quoc V. Le, Daniel S. Park

Figure 1 for G-Augment: Searching For The Meta-Structure Of Data Augmentation Policies For ASR
Figure 2 for G-Augment: Searching For The Meta-Structure Of Data Augmentation Policies For ASR
Figure 3 for G-Augment: Searching For The Meta-Structure Of Data Augmentation Policies For ASR
Figure 4 for G-Augment: Searching For The Meta-Structure Of Data Augmentation Policies For ASR
Viaarxiv icon

LAE: Language-Aware Encoder for Monolingual and Multilingual ASR

Add code
Bookmark button
Alert button
Jun 05, 2022
Jinchuan Tian, Jianwei Yu, Chunlei Zhang, Chao Weng, Yuexian Zou, Dong Yu

Figure 1 for LAE: Language-Aware Encoder for Monolingual and Multilingual ASR
Figure 2 for LAE: Language-Aware Encoder for Monolingual and Multilingual ASR
Figure 3 for LAE: Language-Aware Encoder for Monolingual and Multilingual ASR
Figure 4 for LAE: Language-Aware Encoder for Monolingual and Multilingual ASR
Viaarxiv icon

Streaming Multi-talker Speech Recognition with Joint Speaker Identification

Add code
Bookmark button
Alert button
Apr 05, 2021
Liang Lu, Naoyuki Kanda, Jinyu Li, Yifan Gong

Figure 1 for Streaming Multi-talker Speech Recognition with Joint Speaker Identification
Figure 2 for Streaming Multi-talker Speech Recognition with Joint Speaker Identification
Figure 3 for Streaming Multi-talker Speech Recognition with Joint Speaker Identification
Figure 4 for Streaming Multi-talker Speech Recognition with Joint Speaker Identification
Viaarxiv icon

Disentanglement Learning for Variational Autoencoders Applied to Audio-Visual Speech Enhancement

May 19, 2021
Guillaume Carbajal, Julius Richter, Timo Gerkmann

Figure 1 for Disentanglement Learning for Variational Autoencoders Applied to Audio-Visual Speech Enhancement
Figure 2 for Disentanglement Learning for Variational Autoencoders Applied to Audio-Visual Speech Enhancement
Figure 3 for Disentanglement Learning for Variational Autoencoders Applied to Audio-Visual Speech Enhancement
Figure 4 for Disentanglement Learning for Variational Autoencoders Applied to Audio-Visual Speech Enhancement
Viaarxiv icon

MLP-ASR: Sequence-length agnostic all-MLP architectures for speech recognition

Feb 17, 2022
Jin Sakuma, Tatsuya Komatsu, Robin Scheibler

Figure 1 for MLP-ASR: Sequence-length agnostic all-MLP architectures for speech recognition
Figure 2 for MLP-ASR: Sequence-length agnostic all-MLP architectures for speech recognition
Figure 3 for MLP-ASR: Sequence-length agnostic all-MLP architectures for speech recognition
Figure 4 for MLP-ASR: Sequence-length agnostic all-MLP architectures for speech recognition
Viaarxiv icon

STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech

Add code
Bookmark button
Alert button
Mar 28, 2021
Keon Lee, Kyumin Park, Daeyoung Kim

Figure 1 for STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech
Figure 2 for STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech
Figure 3 for STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech
Figure 4 for STYLER: Style Factor Modeling with Rapidity and Robustness via Speech Decomposition for Expressive and Controllable Neural Text to Speech
Viaarxiv icon

Using Pre-Trained Language Models for Producing Counter Narratives Against Hate Speech: a Comparative Study

Apr 04, 2022
Serra Sinem Tekiroglu, Helena Bonaldi, Margherita Fanton, Marco Guerini

Figure 1 for Using Pre-Trained Language Models for Producing Counter Narratives Against Hate Speech: a Comparative Study
Figure 2 for Using Pre-Trained Language Models for Producing Counter Narratives Against Hate Speech: a Comparative Study
Figure 3 for Using Pre-Trained Language Models for Producing Counter Narratives Against Hate Speech: a Comparative Study
Figure 4 for Using Pre-Trained Language Models for Producing Counter Narratives Against Hate Speech: a Comparative Study
Viaarxiv icon

An Evaluation of Three-Stage Voice Conversion Framework for Noisy and Reverberant Conditions

Jun 30, 2022
Yeonjong Choi, Chao Xie, Tomoki Toda

Figure 1 for An Evaluation of Three-Stage Voice Conversion Framework for Noisy and Reverberant Conditions
Figure 2 for An Evaluation of Three-Stage Voice Conversion Framework for Noisy and Reverberant Conditions
Figure 3 for An Evaluation of Three-Stage Voice Conversion Framework for Noisy and Reverberant Conditions
Figure 4 for An Evaluation of Three-Stage Voice Conversion Framework for Noisy and Reverberant Conditions
Viaarxiv icon

Improving Automatic Hate Speech Detection with Multiword Expression Features

Add code
Bookmark button
Alert button
Jun 01, 2021
Nicolas Zampieri, Irina Illina, Dominique Fohr

Figure 1 for Improving Automatic Hate Speech Detection with Multiword Expression Features
Figure 2 for Improving Automatic Hate Speech Detection with Multiword Expression Features
Figure 3 for Improving Automatic Hate Speech Detection with Multiword Expression Features
Figure 4 for Improving Automatic Hate Speech Detection with Multiword Expression Features
Viaarxiv icon