Alert button

"speech": models, code, and papers
Alert button

Gaze-enhanced Crossmodal Embeddings for Emotion Recognition

Apr 30, 2022
Ahmed Abdou, Ekta Sood, Philipp Müller, Andreas Bulling

Figure 1 for Gaze-enhanced Crossmodal Embeddings for Emotion Recognition
Figure 2 for Gaze-enhanced Crossmodal Embeddings for Emotion Recognition
Figure 3 for Gaze-enhanced Crossmodal Embeddings for Emotion Recognition
Figure 4 for Gaze-enhanced Crossmodal Embeddings for Emotion Recognition
Viaarxiv icon

End-to-End Spectro-Temporal Graph Attention Networks for Speaker Verification Anti-Spoofing and Speech Deepfake Detection

Add code
Bookmark button
Alert button
Jul 27, 2021
Hemlata Tak, Jee-weon Jung, Jose Patino, Madhu Kamble, Massimiliano Todisco, Nicholas Evans

Figure 1 for End-to-End Spectro-Temporal Graph Attention Networks for Speaker Verification Anti-Spoofing and Speech Deepfake Detection
Figure 2 for End-to-End Spectro-Temporal Graph Attention Networks for Speaker Verification Anti-Spoofing and Speech Deepfake Detection
Figure 3 for End-to-End Spectro-Temporal Graph Attention Networks for Speaker Verification Anti-Spoofing and Speech Deepfake Detection
Figure 4 for End-to-End Spectro-Temporal Graph Attention Networks for Speaker Verification Anti-Spoofing and Speech Deepfake Detection
Viaarxiv icon

Masked Autoencoders that Listen

Add code
Bookmark button
Alert button
Jul 13, 2022
Po-Yao, Huang, Hu Xu, Juncheng Li, Alexei Baevski, Michael Auli, Wojciech Galuba, Florian Metze, Christoph Feichtenhofer

Figure 1 for Masked Autoencoders that Listen
Figure 2 for Masked Autoencoders that Listen
Figure 3 for Masked Autoencoders that Listen
Figure 4 for Masked Autoencoders that Listen
Viaarxiv icon

Transfer learning from High-Resource to Low-Resource Language Improves Speech Affect Recognition Classification Accuracy

Mar 04, 2021
Sara Durrani, Umair Arshad

Figure 1 for Transfer learning from High-Resource to Low-Resource Language Improves Speech Affect Recognition Classification Accuracy
Figure 2 for Transfer learning from High-Resource to Low-Resource Language Improves Speech Affect Recognition Classification Accuracy
Figure 3 for Transfer learning from High-Resource to Low-Resource Language Improves Speech Affect Recognition Classification Accuracy
Figure 4 for Transfer learning from High-Resource to Low-Resource Language Improves Speech Affect Recognition Classification Accuracy
Viaarxiv icon

TinySpeech: Attention Condensers for Deep Speech Recognition Neural Networks on Edge Devices

Aug 23, 2020
Alexander Wong, Mahmoud Famouri, Maya Pavlova, Siddharth Surana

Figure 1 for TinySpeech: Attention Condensers for Deep Speech Recognition Neural Networks on Edge Devices
Figure 2 for TinySpeech: Attention Condensers for Deep Speech Recognition Neural Networks on Edge Devices
Figure 3 for TinySpeech: Attention Condensers for Deep Speech Recognition Neural Networks on Edge Devices
Viaarxiv icon

Adaptation Algorithms for Speech Recognition: An Overview

Add code
Bookmark button
Alert button
Aug 14, 2020
Peter Bell, Joachim Fainberg, Ondrej Klejch, Jinyu Li, Steve Renals, Pawel Swietojanski

Figure 1 for Adaptation Algorithms for Speech Recognition: An Overview
Figure 2 for Adaptation Algorithms for Speech Recognition: An Overview
Figure 3 for Adaptation Algorithms for Speech Recognition: An Overview
Figure 4 for Adaptation Algorithms for Speech Recognition: An Overview
Viaarxiv icon

Vietnamese Capitalization and Punctuation Recovery Models

Add code
Bookmark button
Alert button
Jul 04, 2022
Hoang Thi Thu Uyen, Nguyen Anh Tu, Ta Duc Huy

Figure 1 for Vietnamese Capitalization and Punctuation Recovery Models
Figure 2 for Vietnamese Capitalization and Punctuation Recovery Models
Figure 3 for Vietnamese Capitalization and Punctuation Recovery Models
Viaarxiv icon

ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers

Add code
Bookmark button
Alert button
Apr 20, 2020
Jung-Woo Ha, Kihyun Nam, Jin Gu Kang, Sang-Woo Lee, Sohee Yang, Hyunhoon Jung, Eunmi Kim, Hyeji Kim, Soojin Kim, Hyun Ah Kim, Kyoungtae Doh, Chan Kyu Lee, Sunghun Kim

Figure 1 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Figure 2 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Figure 3 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Figure 4 for ClovaCall: Korean Goal-Oriented Dialog Speech Corpus for Automatic Speech Recognition of Contact Centers
Viaarxiv icon

MASRI-HEADSET: A Maltese Corpus for Speech Recognition

Add code
Bookmark button
Alert button
Aug 13, 2020
Carlos Mena, Albert Gatt, Andrea DeMarco, Claudia Borg, Lonneke van der Plas, Amanda Muscat, Ian Padovani

Figure 1 for MASRI-HEADSET: A Maltese Corpus for Speech Recognition
Figure 2 for MASRI-HEADSET: A Maltese Corpus for Speech Recognition
Figure 3 for MASRI-HEADSET: A Maltese Corpus for Speech Recognition
Figure 4 for MASRI-HEADSET: A Maltese Corpus for Speech Recognition
Viaarxiv icon

Consistent Transcription and Translation of Speech

Add code
Bookmark button
Alert button
Jul 24, 2020
Matthias Sperber, Hendra Setiawan, Christian Gollan, Udhyakumar Nallasamy, Matthias Paulik

Figure 1 for Consistent Transcription and Translation of Speech
Figure 2 for Consistent Transcription and Translation of Speech
Figure 3 for Consistent Transcription and Translation of Speech
Figure 4 for Consistent Transcription and Translation of Speech
Viaarxiv icon