Alert button

"speech": models, code, and papers
Alert button

A Joint Model for Dropped Pronoun Recovery and Conversational Discourse Parsing in Chinese Conversational Speech

Add code
Bookmark button
Alert button
Jun 07, 2021
Jingxuan Yang, Kerui Xu, Jun Xu, Si Li, Sheng Gao, Jun Guo, Nianwen Xue, Ji-Rong Wen

Figure 1 for A Joint Model for Dropped Pronoun Recovery and Conversational Discourse Parsing in Chinese Conversational Speech
Figure 2 for A Joint Model for Dropped Pronoun Recovery and Conversational Discourse Parsing in Chinese Conversational Speech
Figure 3 for A Joint Model for Dropped Pronoun Recovery and Conversational Discourse Parsing in Chinese Conversational Speech
Figure 4 for A Joint Model for Dropped Pronoun Recovery and Conversational Discourse Parsing in Chinese Conversational Speech
Viaarxiv icon

Speech Prediction in Silent Videos using Variational Autoencoders

Nov 14, 2020
Ravindra Yadav, Ashish Sardana, Vinay P Namboodiri, Rajesh M Hegde

Figure 1 for Speech Prediction in Silent Videos using Variational Autoencoders
Figure 2 for Speech Prediction in Silent Videos using Variational Autoencoders
Figure 3 for Speech Prediction in Silent Videos using Variational Autoencoders
Figure 4 for Speech Prediction in Silent Videos using Variational Autoencoders
Viaarxiv icon

SyncNet: Using Causal Convolutions and Correlating Objective for Time Delay Estimation in Audio Signals

Add code
Bookmark button
Alert button
Mar 28, 2022
Akshay Raina, Vipul Arora

Figure 1 for SyncNet: Using Causal Convolutions and Correlating Objective for Time Delay Estimation in Audio Signals
Figure 2 for SyncNet: Using Causal Convolutions and Correlating Objective for Time Delay Estimation in Audio Signals
Figure 3 for SyncNet: Using Causal Convolutions and Correlating Objective for Time Delay Estimation in Audio Signals
Figure 4 for SyncNet: Using Causal Convolutions and Correlating Objective for Time Delay Estimation in Audio Signals
Viaarxiv icon

Breaking the Data Barrier: Towards Robust Speech Translation via Adversarial Stability Training

Sep 25, 2019
Qiao Cheng, Meiyuan Fang, Yaqian Han, Jin Huang, Yitao Duan

Figure 1 for Breaking the Data Barrier: Towards Robust Speech Translation via Adversarial Stability Training
Figure 2 for Breaking the Data Barrier: Towards Robust Speech Translation via Adversarial Stability Training
Figure 3 for Breaking the Data Barrier: Towards Robust Speech Translation via Adversarial Stability Training
Figure 4 for Breaking the Data Barrier: Towards Robust Speech Translation via Adversarial Stability Training
Viaarxiv icon

ETHOS: an Online Hate Speech Detection Dataset

Add code
Bookmark button
Alert button
Jun 11, 2020
Ioannis Mollas, Zoe Chrysopoulou, Stamatis Karlos, Grigorios Tsoumakas

Figure 1 for ETHOS: an Online Hate Speech Detection Dataset
Figure 2 for ETHOS: an Online Hate Speech Detection Dataset
Figure 3 for ETHOS: an Online Hate Speech Detection Dataset
Figure 4 for ETHOS: an Online Hate Speech Detection Dataset
Viaarxiv icon

Decoupling Speaker-Independent Emotions for Voice Conversion Via Source-Filter Networks

Oct 04, 2021
Zhaojie Luo, Shoufeng Lin, Rui Liu, Jun Baba, Yuichiro Yoshikawa, Ishiguro Hiroshi

Figure 1 for Decoupling Speaker-Independent Emotions for Voice Conversion Via Source-Filter Networks
Figure 2 for Decoupling Speaker-Independent Emotions for Voice Conversion Via Source-Filter Networks
Figure 3 for Decoupling Speaker-Independent Emotions for Voice Conversion Via Source-Filter Networks
Figure 4 for Decoupling Speaker-Independent Emotions for Voice Conversion Via Source-Filter Networks
Viaarxiv icon

Deep Annotation of Therapeutic Working Alliance in Psychotherapy

Apr 12, 2022
Baihan Lin, Guillermo Cecchi, Djallel Bouneffouf

Figure 1 for Deep Annotation of Therapeutic Working Alliance in Psychotherapy
Figure 2 for Deep Annotation of Therapeutic Working Alliance in Psychotherapy
Figure 3 for Deep Annotation of Therapeutic Working Alliance in Psychotherapy
Figure 4 for Deep Annotation of Therapeutic Working Alliance in Psychotherapy
Viaarxiv icon

Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners

Add code
Bookmark button
Alert button
May 24, 2022
Zhenhailong Wang, Manling Li, Ruochen Xu, Luowei Zhou, Jie Lei, Xudong Lin, Shuohang Wang, Ziyi Yang, Chenguang Zhu, Derek Hoiem, Shih-Fu Chang, Mohit Bansal, Heng Ji

Figure 1 for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners
Figure 2 for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners
Figure 3 for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners
Figure 4 for Language Models with Image Descriptors are Strong Few-Shot Video-Language Learners
Viaarxiv icon

Low-Latency Sequence-to-Sequence Speech Recognition and Translation by Partial Hypothesis Selection

Add code
Bookmark button
Alert button
May 22, 2020
Danni Liu, Gerasimos Spanakis, Jan Niehues

Figure 1 for Low-Latency Sequence-to-Sequence Speech Recognition and Translation by Partial Hypothesis Selection
Figure 2 for Low-Latency Sequence-to-Sequence Speech Recognition and Translation by Partial Hypothesis Selection
Figure 3 for Low-Latency Sequence-to-Sequence Speech Recognition and Translation by Partial Hypothesis Selection
Figure 4 for Low-Latency Sequence-to-Sequence Speech Recognition and Translation by Partial Hypothesis Selection
Viaarxiv icon

Synthesising 3D Facial Motion from "In-the-Wild" Speech

Apr 15, 2019
Panagiotis Tzirakis, Athanasios Papaioannou, Alexander Lattas, Michail Tarasiou, Björn Schuller, Stefanos Zafeiriou

Figure 1 for Synthesising 3D Facial Motion from "In-the-Wild" Speech
Figure 2 for Synthesising 3D Facial Motion from "In-the-Wild" Speech
Figure 3 for Synthesising 3D Facial Motion from "In-the-Wild" Speech
Figure 4 for Synthesising 3D Facial Motion from "In-the-Wild" Speech
Viaarxiv icon