Alert button

"speech": models, code, and papers
Alert button

DeLoRes: Decorrelating Latent Spaces for Low-Resource Audio Representation Learning

Add code
Bookmark button
Alert button
Apr 13, 2022
Sreyan Ghosh, Ashish Seth, and Deepak Mittal, Maneesh Singh, S Umesh

Figure 1 for DeLoRes: Decorrelating Latent Spaces for Low-Resource Audio Representation Learning
Figure 2 for DeLoRes: Decorrelating Latent Spaces for Low-Resource Audio Representation Learning
Figure 3 for DeLoRes: Decorrelating Latent Spaces for Low-Resource Audio Representation Learning
Figure 4 for DeLoRes: Decorrelating Latent Spaces for Low-Resource Audio Representation Learning
Viaarxiv icon

WaveFuzz: A Clean-Label Poisoning Attack to Protect Your Voice

Mar 25, 2022
Yunjie Ge, Qian Wang, Jingfeng Zhang, Juntao Zhou, Yunzhu Zhang, Chao Shen

Figure 1 for WaveFuzz: A Clean-Label Poisoning Attack to Protect Your Voice
Figure 2 for WaveFuzz: A Clean-Label Poisoning Attack to Protect Your Voice
Figure 3 for WaveFuzz: A Clean-Label Poisoning Attack to Protect Your Voice
Figure 4 for WaveFuzz: A Clean-Label Poisoning Attack to Protect Your Voice
Viaarxiv icon

End-to-end Microphone Permutation and Number Invariant Multi-channel Speech Separation

Add code
Bookmark button
Alert button
Oct 30, 2019
Yi Luo, Zhuo Chen, Nima Mesgarani, Takuya Yoshioka

Figure 1 for End-to-end Microphone Permutation and Number Invariant Multi-channel Speech Separation
Figure 2 for End-to-end Microphone Permutation and Number Invariant Multi-channel Speech Separation
Figure 3 for End-to-end Microphone Permutation and Number Invariant Multi-channel Speech Separation
Viaarxiv icon

TransCouplet:Transformer based Chinese Couplet Generation

Add code
Bookmark button
Alert button
Dec 03, 2021
Kuan-Yu Chiang, Shihao Lin, Joe Chen, Qian Yin, Qizhen Jin

Figure 1 for TransCouplet:Transformer based Chinese Couplet Generation
Figure 2 for TransCouplet:Transformer based Chinese Couplet Generation
Figure 3 for TransCouplet:Transformer based Chinese Couplet Generation
Figure 4 for TransCouplet:Transformer based Chinese Couplet Generation
Viaarxiv icon

Towards Visually Grounded Sub-Word Speech Unit Discovery

Feb 21, 2019
David Harwath, James Glass

Figure 1 for Towards Visually Grounded Sub-Word Speech Unit Discovery
Figure 2 for Towards Visually Grounded Sub-Word Speech Unit Discovery
Figure 3 for Towards Visually Grounded Sub-Word Speech Unit Discovery
Figure 4 for Towards Visually Grounded Sub-Word Speech Unit Discovery
Viaarxiv icon

On Laughter and Speech-Laugh, Based on Observations of Child-Robot Interaction

Aug 30, 2019
Anton Batliner, Stefan Steidl, Florian Eyben, Björn Schuller

Figure 1 for On Laughter and Speech-Laugh, Based on Observations of Child-Robot Interaction
Figure 2 for On Laughter and Speech-Laugh, Based on Observations of Child-Robot Interaction
Figure 3 for On Laughter and Speech-Laugh, Based on Observations of Child-Robot Interaction
Figure 4 for On Laughter and Speech-Laugh, Based on Observations of Child-Robot Interaction
Viaarxiv icon

LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example

Add code
Bookmark button
Alert button
Oct 11, 2021
Hieu-Thi Luong, Junichi Yamagishi

Figure 1 for LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example
Figure 2 for LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example
Figure 3 for LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example
Figure 4 for LaughNet: synthesizing laughter utterances from waveform silhouettes and a single laughter example
Viaarxiv icon

Low-Resource Speech-to-Text Translation

Add code
Bookmark button
Alert button
Jun 18, 2018
Sameer Bansal, Herman Kamper, Karen Livescu, Adam Lopez, Sharon Goldwater

Figure 1 for Low-Resource Speech-to-Text Translation
Figure 2 for Low-Resource Speech-to-Text Translation
Figure 3 for Low-Resource Speech-to-Text Translation
Figure 4 for Low-Resource Speech-to-Text Translation
Viaarxiv icon

CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement

Add code
Bookmark button
Alert button
Sep 23, 2019
Mandar Gogate, Kia Dashtipour, Ahsan Adeel, Amir Hussain

Figure 1 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Figure 2 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Figure 3 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Figure 4 for CochleaNet: A Robust Language-independent Audio-Visual Model for Speech Enhancement
Viaarxiv icon

Top-Down Influence? Predicting CEO Personality and Risk Impact from Speech Transcripts

Jan 19, 2022
Kilian Theil, Dirk Hovy, Heiner Stuckenschmidt

Figure 1 for Top-Down Influence? Predicting CEO Personality and Risk Impact from Speech Transcripts
Figure 2 for Top-Down Influence? Predicting CEO Personality and Risk Impact from Speech Transcripts
Figure 3 for Top-Down Influence? Predicting CEO Personality and Risk Impact from Speech Transcripts
Figure 4 for Top-Down Influence? Predicting CEO Personality and Risk Impact from Speech Transcripts
Viaarxiv icon