Alert button

"speech": models, code, and papers
Alert button

Automatic Depression Detection: An Emotional Audio-Textual Corpus and a GRU/BiLSTM-based Model

Add code
Bookmark button
Alert button
Feb 15, 2022
Ying Shen, Huiyu Yang, Lin Lin

Figure 1 for Automatic Depression Detection: An Emotional Audio-Textual Corpus and a GRU/BiLSTM-based Model
Figure 2 for Automatic Depression Detection: An Emotional Audio-Textual Corpus and a GRU/BiLSTM-based Model
Figure 3 for Automatic Depression Detection: An Emotional Audio-Textual Corpus and a GRU/BiLSTM-based Model
Figure 4 for Automatic Depression Detection: An Emotional Audio-Textual Corpus and a GRU/BiLSTM-based Model
Viaarxiv icon

Generative Adversarial Networks for Unpaired Voice Transformation on Impaired Speech

Add code
Bookmark button
Alert button
Oct 30, 2018
Li-Wei Chen, Hung-Yi Lee, Yu Tsao

Figure 1 for Generative Adversarial Networks for Unpaired Voice Transformation on Impaired Speech
Figure 2 for Generative Adversarial Networks for Unpaired Voice Transformation on Impaired Speech
Figure 3 for Generative Adversarial Networks for Unpaired Voice Transformation on Impaired Speech
Figure 4 for Generative Adversarial Networks for Unpaired Voice Transformation on Impaired Speech
Viaarxiv icon

KBCNMUJAL@HASOC-Dravidian-CodeMix-FIRE2020: Using Machine Learning for Detection of Hate Speech and Offensive Code-Mixed Social Media text

Feb 19, 2021
Varsha Pathak, Manish Joshi, Prasad Joshi, Monica Mundada, Tanmay Joshi

Figure 1 for KBCNMUJAL@HASOC-Dravidian-CodeMix-FIRE2020: Using Machine Learning for Detection of Hate Speech and Offensive Code-Mixed Social Media text
Figure 2 for KBCNMUJAL@HASOC-Dravidian-CodeMix-FIRE2020: Using Machine Learning for Detection of Hate Speech and Offensive Code-Mixed Social Media text
Figure 3 for KBCNMUJAL@HASOC-Dravidian-CodeMix-FIRE2020: Using Machine Learning for Detection of Hate Speech and Offensive Code-Mixed Social Media text
Figure 4 for KBCNMUJAL@HASOC-Dravidian-CodeMix-FIRE2020: Using Machine Learning for Detection of Hate Speech and Offensive Code-Mixed Social Media text
Viaarxiv icon

Introducing the ICBe Dataset: Very High Recall and Precision Event Extraction from Narratives about International Crises

Add code
Bookmark button
Alert button
Feb 14, 2022
Rex W. Douglass, Thomas Leo Scherer, J. Andrés Gannon, Erik Gartzke, Jon Lindsay, Shannon Carcelli, Jonathan Wilkenfeld, David M. Quinn, Catherine Aiken, Jose Miguel Cabezas Navarro, Neil Lund, Egle Murauskaite, Diana Partridge

Figure 1 for Introducing the ICBe Dataset: Very High Recall and Precision Event Extraction from Narratives about International Crises
Figure 2 for Introducing the ICBe Dataset: Very High Recall and Precision Event Extraction from Narratives about International Crises
Figure 3 for Introducing the ICBe Dataset: Very High Recall and Precision Event Extraction from Narratives about International Crises
Figure 4 for Introducing the ICBe Dataset: Very High Recall and Precision Event Extraction from Narratives about International Crises
Viaarxiv icon

Visual Acoustic Matching

Feb 14, 2022
Changan Chen, Ruohan Gao, Paul Calamia, Kristen Grauman

Figure 1 for Visual Acoustic Matching
Figure 2 for Visual Acoustic Matching
Figure 3 for Visual Acoustic Matching
Figure 4 for Visual Acoustic Matching
Viaarxiv icon

Dialog-context aware end-to-end speech recognition

Aug 07, 2018
Suyoun Kim, Florian Metze

Figure 1 for Dialog-context aware end-to-end speech recognition
Figure 2 for Dialog-context aware end-to-end speech recognition
Figure 3 for Dialog-context aware end-to-end speech recognition
Figure 4 for Dialog-context aware end-to-end speech recognition
Viaarxiv icon

Visualization and Interpretation of Latent Spaces for Controlling Expressive Speech Synthesis through Audio Analysis

Add code
Bookmark button
Alert button
Mar 27, 2019
Noé Tits, Fengna Wang, Kevin El Haddad, Vincent Pagel, Thierry Dutoit

Figure 1 for Visualization and Interpretation of Latent Spaces for Controlling Expressive Speech Synthesis through Audio Analysis
Figure 2 for Visualization and Interpretation of Latent Spaces for Controlling Expressive Speech Synthesis through Audio Analysis
Figure 3 for Visualization and Interpretation of Latent Spaces for Controlling Expressive Speech Synthesis through Audio Analysis
Figure 4 for Visualization and Interpretation of Latent Spaces for Controlling Expressive Speech Synthesis through Audio Analysis
Viaarxiv icon

A Comparison of Label-Synchronous and Frame-Synchronous End-to-End Models for Speech Recognition

May 25, 2020
Linhao Dong, Cheng Yi, Jianzong Wang, Shiyu Zhou, Shuang Xu, Xueli Jia, Bo Xu

Figure 1 for A Comparison of Label-Synchronous and Frame-Synchronous End-to-End Models for Speech Recognition
Figure 2 for A Comparison of Label-Synchronous and Frame-Synchronous End-to-End Models for Speech Recognition
Figure 3 for A Comparison of Label-Synchronous and Frame-Synchronous End-to-End Models for Speech Recognition
Figure 4 for A Comparison of Label-Synchronous and Frame-Synchronous End-to-End Models for Speech Recognition
Viaarxiv icon

On using 2D sequence-to-sequence models for speech recognition

Add code
Bookmark button
Alert button
Nov 20, 2019
Parnia Bahar, Albert Zeyer, Ralf Schlüter, Hermann Ney

Figure 1 for On using 2D sequence-to-sequence models for speech recognition
Figure 2 for On using 2D sequence-to-sequence models for speech recognition
Figure 3 for On using 2D sequence-to-sequence models for speech recognition
Figure 4 for On using 2D sequence-to-sequence models for speech recognition
Viaarxiv icon

Channel-Attention Dense U-Net for Multichannel Speech Enhancement

Jan 30, 2020
Bahareh Tolooshams, Ritwik Giri, Andrew H. Song, Umut Isik, Arvindh Krishnaswamy

Figure 1 for Channel-Attention Dense U-Net for Multichannel Speech Enhancement
Figure 2 for Channel-Attention Dense U-Net for Multichannel Speech Enhancement
Figure 3 for Channel-Attention Dense U-Net for Multichannel Speech Enhancement
Figure 4 for Channel-Attention Dense U-Net for Multichannel Speech Enhancement
Viaarxiv icon