Alert button

"speech": models, code, and papers
Alert button

Multi-Speaker Expressive Speech Synthesis via Multiple Factors Decoupling

Nov 19, 2022
Xinfa Zhu, Yi Lei, Kun Song, Yongmao Zhang, Tao Li, Lei Xie

Figure 1 for Multi-Speaker Expressive Speech Synthesis via Multiple Factors Decoupling
Figure 2 for Multi-Speaker Expressive Speech Synthesis via Multiple Factors Decoupling
Figure 3 for Multi-Speaker Expressive Speech Synthesis via Multiple Factors Decoupling
Figure 4 for Multi-Speaker Expressive Speech Synthesis via Multiple Factors Decoupling
Viaarxiv icon

Down the Rabbit Hole: Detecting Online Extremism, Radicalisation, and Politicised Hate Speech

Jan 27, 2023
Jarod Govers, Philip Feldman, Aaron Dant, Panos Patros

Figure 1 for Down the Rabbit Hole: Detecting Online Extremism, Radicalisation, and Politicised Hate Speech
Figure 2 for Down the Rabbit Hole: Detecting Online Extremism, Radicalisation, and Politicised Hate Speech
Figure 3 for Down the Rabbit Hole: Detecting Online Extremism, Radicalisation, and Politicised Hate Speech
Figure 4 for Down the Rabbit Hole: Detecting Online Extremism, Radicalisation, and Politicised Hate Speech
Viaarxiv icon

Text-to-speech synthesis from dark data with evaluation-in-the-loop data selection

Oct 26, 2022
Kentaro Seki, Shinnosuke Takamichi, Takaaki Saeki, Hiroshi Saruwatari

Viaarxiv icon

Reverberation as Supervision for Speech Separation

Nov 15, 2022
Rohith Aralikatti, Christoph Boeddeker, Gordon Wichern, Aswin Shanmugam Subramanian, Jonathan Le Roux

Figure 1 for Reverberation as Supervision for Speech Separation
Figure 2 for Reverberation as Supervision for Speech Separation
Figure 3 for Reverberation as Supervision for Speech Separation
Figure 4 for Reverberation as Supervision for Speech Separation
Viaarxiv icon

Speech Enhancement with Perceptually-motivated Optimization and Dual Transformations

Sep 24, 2022
Xucheng Wan, Kai Liu, Ziqing Du, Huan Zhou

Figure 1 for Speech Enhancement with Perceptually-motivated Optimization and Dual Transformations
Figure 2 for Speech Enhancement with Perceptually-motivated Optimization and Dual Transformations
Figure 3 for Speech Enhancement with Perceptually-motivated Optimization and Dual Transformations
Figure 4 for Speech Enhancement with Perceptually-motivated Optimization and Dual Transformations
Viaarxiv icon

A Teacher-student Framework for Unsupervised Speech Enhancement Using Noise Remixing Training and Two-stage Inference

Oct 27, 2022
Li-Wei Chen, Yao-Fei Cheng, Hung-Shin Lee, Yu Tsao, Hsin-Min Wang

Figure 1 for A Teacher-student Framework for Unsupervised Speech Enhancement Using Noise Remixing Training and Two-stage Inference
Figure 2 for A Teacher-student Framework for Unsupervised Speech Enhancement Using Noise Remixing Training and Two-stage Inference
Figure 3 for A Teacher-student Framework for Unsupervised Speech Enhancement Using Noise Remixing Training and Two-stage Inference
Viaarxiv icon

Automatic Severity Assessment of Dysarthric speech by using Self-supervised Model with Multi-task Learning

Oct 27, 2022
Eun Jung Yeo, Kwanghee Choi, Sunhee Kim, Minhwa Chung

Figure 1 for Automatic Severity Assessment of Dysarthric speech by using Self-supervised Model with Multi-task Learning
Figure 2 for Automatic Severity Assessment of Dysarthric speech by using Self-supervised Model with Multi-task Learning
Figure 3 for Automatic Severity Assessment of Dysarthric speech by using Self-supervised Model with Multi-task Learning
Figure 4 for Automatic Severity Assessment of Dysarthric speech by using Self-supervised Model with Multi-task Learning
Viaarxiv icon

Neural Fourier Shift for Binaural Speech Rendering

Nov 02, 2022
Jin Woo Lee, Kyogu Lee

Figure 1 for Neural Fourier Shift for Binaural Speech Rendering
Figure 2 for Neural Fourier Shift for Binaural Speech Rendering
Figure 3 for Neural Fourier Shift for Binaural Speech Rendering
Figure 4 for Neural Fourier Shift for Binaural Speech Rendering
Viaarxiv icon

Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning

Dec 10, 2022
Chen Chen, Yuchen Hu, Qiang Zhang, Heqing Zou, Beier Zhu, Eng Siong Chng

Figure 1 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Figure 2 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Figure 3 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Figure 4 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Viaarxiv icon

Learning to Jointly Transcribe and Subtitle for End-to-End Spontaneous Speech Recognition

Oct 14, 2022
Jakob Poncelet, Hugo Van hamme

Figure 1 for Learning to Jointly Transcribe and Subtitle for End-to-End Spontaneous Speech Recognition
Figure 2 for Learning to Jointly Transcribe and Subtitle for End-to-End Spontaneous Speech Recognition
Figure 3 for Learning to Jointly Transcribe and Subtitle for End-to-End Spontaneous Speech Recognition
Figure 4 for Learning to Jointly Transcribe and Subtitle for End-to-End Spontaneous Speech Recognition
Viaarxiv icon