Alert button

"speech": models, code, and papers
Alert button

Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors

Mar 01, 2022
Yang Wu, Yanyan Zhao, Hao Yang, Song Chen, Bing Qin, Xiaohuan Cao, Wenting Zhao

Figure 1 for Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors
Figure 2 for Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors
Figure 3 for Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors
Figure 4 for Sentiment Word Aware Multimodal Refinement for Multimodal Sentiment Analysis with ASR Errors
Viaarxiv icon

Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference

Feb 14, 2022
Gianna Paulin, Francesco Conti, Lukas Cavigelli, Luca Benini

Figure 1 for Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
Figure 2 for Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
Figure 3 for Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
Figure 4 for Vau da muntanialas: Energy-efficient multi-die scalable acceleration of RNN inference
Viaarxiv icon

Towards Learning Fine-Grained Disentangled Representations from Speech

Aug 08, 2018
Yuan Gong, Christian Poellabauer

Figure 1 for Towards Learning Fine-Grained Disentangled Representations from Speech
Figure 2 for Towards Learning Fine-Grained Disentangled Representations from Speech
Viaarxiv icon

End-to-End Feedback Loss in Speech Chain Framework via Straight-Through Estimator

Oct 31, 2018
Andros Tjandra, Sakriani Sakti, Satoshi Nakamura

Figure 1 for End-to-End Feedback Loss in Speech Chain Framework via Straight-Through Estimator
Figure 2 for End-to-End Feedback Loss in Speech Chain Framework via Straight-Through Estimator
Figure 3 for End-to-End Feedback Loss in Speech Chain Framework via Straight-Through Estimator
Figure 4 for End-to-End Feedback Loss in Speech Chain Framework via Straight-Through Estimator
Viaarxiv icon

Unified Streaming and Non-streaming Two-pass End-to-end Model for Speech Recognition

Dec 10, 2020
Binbin Zhang, Di Wu, Zhuoyuan Yao, Xiong Wang, Fan Yu, Chao Yang, Liyong Guo, Yaguang Hu, Lei Xie, Xin Lei

Figure 1 for Unified Streaming and Non-streaming Two-pass End-to-end Model for Speech Recognition
Figure 2 for Unified Streaming and Non-streaming Two-pass End-to-end Model for Speech Recognition
Figure 3 for Unified Streaming and Non-streaming Two-pass End-to-end Model for Speech Recognition
Figure 4 for Unified Streaming and Non-streaming Two-pass End-to-end Model for Speech Recognition
Viaarxiv icon

Learning Spontaneity to Improve Emotion Recognition In Speech

Jun 13, 2018
Karttikeya Mangalam, Tanaya Guha

Figure 1 for Learning Spontaneity to Improve Emotion Recognition In Speech
Figure 2 for Learning Spontaneity to Improve Emotion Recognition In Speech
Figure 3 for Learning Spontaneity to Improve Emotion Recognition In Speech
Figure 4 for Learning Spontaneity to Improve Emotion Recognition In Speech
Viaarxiv icon

Unsupervised Cross-Modal Alignment of Speech and Text Embedding Spaces

Sep 20, 2018
Yu-An Chung, Wei-Hung Weng, Schrasing Tong, James Glass

Figure 1 for Unsupervised Cross-Modal Alignment of Speech and Text Embedding Spaces
Figure 2 for Unsupervised Cross-Modal Alignment of Speech and Text Embedding Spaces
Figure 3 for Unsupervised Cross-Modal Alignment of Speech and Text Embedding Spaces
Figure 4 for Unsupervised Cross-Modal Alignment of Speech and Text Embedding Spaces
Viaarxiv icon

Speech Sentiment Analysis via Pre-trained Features from End-to-end ASR Models

Nov 21, 2019
Zhiyun Lu, Liangliang Cao, Yu Zhang, Chung-Cheng Chiu, James Fan

Figure 1 for Speech Sentiment Analysis via Pre-trained Features from End-to-end ASR Models
Figure 2 for Speech Sentiment Analysis via Pre-trained Features from End-to-end ASR Models
Figure 3 for Speech Sentiment Analysis via Pre-trained Features from End-to-end ASR Models
Figure 4 for Speech Sentiment Analysis via Pre-trained Features from End-to-end ASR Models
Viaarxiv icon

Sampling-based speech parameter generation using moment-matching networks

Apr 12, 2017
Shinnosuke Takamichi, Tomoki Koriyama, Hiroshi Saruwatari

Figure 1 for Sampling-based speech parameter generation using moment-matching networks
Figure 2 for Sampling-based speech parameter generation using moment-matching networks
Figure 3 for Sampling-based speech parameter generation using moment-matching networks
Figure 4 for Sampling-based speech parameter generation using moment-matching networks
Viaarxiv icon

Speech, Head, and Eye-based Cues for Continuous Affect Prediction

Jul 23, 2019
Jonny O'Dwyer

Figure 1 for Speech, Head, and Eye-based Cues for Continuous Affect Prediction
Figure 2 for Speech, Head, and Eye-based Cues for Continuous Affect Prediction
Figure 3 for Speech, Head, and Eye-based Cues for Continuous Affect Prediction
Viaarxiv icon