Alert button

"speech": models, code, and papers
Alert button

Fully Unsupervised Training of Few-shot Keyword Spotting

Oct 07, 2022
Dongjune Lee, Minchan Kim, Sung Hwan Mun, Min Hyun Han, Nam Soo Kim

Figure 1 for Fully Unsupervised Training of Few-shot Keyword Spotting
Figure 2 for Fully Unsupervised Training of Few-shot Keyword Spotting
Figure 3 for Fully Unsupervised Training of Few-shot Keyword Spotting
Viaarxiv icon

Residual Adapters for Parameter-Efficient ASR Adaptation to Atypical and Accented Speech

Sep 14, 2021
Katrin Tomanek, Vicky Zayats, Dirk Padfield, Kara Vaillancourt, Fadi Biadsy

Figure 1 for Residual Adapters for Parameter-Efficient ASR Adaptation to Atypical and Accented Speech
Figure 2 for Residual Adapters for Parameter-Efficient ASR Adaptation to Atypical and Accented Speech
Figure 3 for Residual Adapters for Parameter-Efficient ASR Adaptation to Atypical and Accented Speech
Figure 4 for Residual Adapters for Parameter-Efficient ASR Adaptation to Atypical and Accented Speech
Viaarxiv icon

Investigation of Deep Neural Network Acoustic Modelling Approaches for Low Resource Accented Mandarin Speech Recognition

Jan 24, 2022
Xurong Xie, Xiang Sui, Xunying Liu, Lan Wang

Viaarxiv icon

Attribute Inference Attack of Speech Emotion Recognition in Federated Learning Settings

Add code
Bookmark button
Alert button
Dec 26, 2021
Tiantian Feng, Hanieh Hashemi, Rajat Hebbar, Murali Annavaram, Shrikanth S. Narayanan

Figure 1 for Attribute Inference Attack of Speech Emotion Recognition in Federated Learning Settings
Figure 2 for Attribute Inference Attack of Speech Emotion Recognition in Federated Learning Settings
Figure 3 for Attribute Inference Attack of Speech Emotion Recognition in Federated Learning Settings
Figure 4 for Attribute Inference Attack of Speech Emotion Recognition in Federated Learning Settings
Viaarxiv icon

Learning a Neural Diff for Speech Models

Aug 17, 2021
Jonathan Macoskey, Grant P. Strimel, Ariya Rastrow

Figure 1 for Learning a Neural Diff for Speech Models
Figure 2 for Learning a Neural Diff for Speech Models
Viaarxiv icon

Task-aware Warping Factors in Mask-based Speech Enhancement

Aug 27, 2021
Qiongqiong Wang, Kong Aik Lee, Takafumi Koshinaka, Koji Okabe, Hitoshi Yamamoto

Figure 1 for Task-aware Warping Factors in Mask-based Speech Enhancement
Figure 2 for Task-aware Warping Factors in Mask-based Speech Enhancement
Figure 3 for Task-aware Warping Factors in Mask-based Speech Enhancement
Figure 4 for Task-aware Warping Factors in Mask-based Speech Enhancement
Viaarxiv icon

Mono vs Multilingual BERT for Hate Speech Detection and Text Classification: A Case Study in Marathi

Apr 19, 2022
Abhishek Velankar, Hrushikesh Patil, Raviraj Joshi

Figure 1 for Mono vs Multilingual BERT for Hate Speech Detection and Text Classification: A Case Study in Marathi
Viaarxiv icon

Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection

Add code
Bookmark button
Alert button
Oct 03, 2022
Xuanjun Chen, Haibin Wu, Helen Meng, Hung-yi Lee, Jyh-Shing Roger Jang

Figure 1 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Figure 2 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Figure 3 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Figure 4 for Push-Pull: Characterizing the Adversarial Robustness for Audio-Visual Active Speaker Detection
Viaarxiv icon

FaceFormer: Speech-Driven 3D Facial Animation with Transformers

Add code
Bookmark button
Alert button
Dec 28, 2021
Yingruo Fan, Zhaojiang Lin, Jun Saito, Wenping Wang, Taku Komura

Figure 1 for FaceFormer: Speech-Driven 3D Facial Animation with Transformers
Figure 2 for FaceFormer: Speech-Driven 3D Facial Animation with Transformers
Figure 3 for FaceFormer: Speech-Driven 3D Facial Animation with Transformers
Figure 4 for FaceFormer: Speech-Driven 3D Facial Animation with Transformers
Viaarxiv icon

PercepNet+: A Phase and SNR Aware PercepNet for Real-Time Speech Enhancement

Mar 04, 2022
Xiaofeng Ge, Jiangyu Han, Yanhua Long, Haixin Guan

Figure 1 for PercepNet+: A Phase and SNR Aware PercepNet for Real-Time Speech Enhancement
Figure 2 for PercepNet+: A Phase and SNR Aware PercepNet for Real-Time Speech Enhancement
Figure 3 for PercepNet+: A Phase and SNR Aware PercepNet for Real-Time Speech Enhancement
Figure 4 for PercepNet+: A Phase and SNR Aware PercepNet for Real-Time Speech Enhancement
Viaarxiv icon