Alert button

"speech": models, code, and papers
Alert button

IIITT@LT-EDI-EACL2021-Hope Speech Detection: There is always Hope in Transformers

Add code
Bookmark button
Alert button
Apr 19, 2021
Karthik Puranik, Adeep Hande, Ruba Priyadharshini, Sajeetha Thavareesan, Bharathi Raja Chakravarthi

Figure 1 for IIITT@LT-EDI-EACL2021-Hope Speech Detection: There is always Hope in Transformers
Figure 2 for IIITT@LT-EDI-EACL2021-Hope Speech Detection: There is always Hope in Transformers
Figure 3 for IIITT@LT-EDI-EACL2021-Hope Speech Detection: There is always Hope in Transformers
Figure 4 for IIITT@LT-EDI-EACL2021-Hope Speech Detection: There is always Hope in Transformers
Viaarxiv icon

Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators

Nov 09, 2021
Marko Stamenovic, Nils L. Westhausen, Li-Chia Yang, Carl Jensen, Alex Pawlicki

Figure 1 for Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators
Figure 2 for Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators
Figure 3 for Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators
Figure 4 for Weight, Block or Unit? Exploring Sparsity Tradeoffs for Speech Enhancement on Tiny Neural Accelerators
Viaarxiv icon

A Time-domain Generalized Wiener Filter for Multi-channel Speech Separation

Add code
Bookmark button
Alert button
Dec 07, 2021
Yi Luo

Figure 1 for A Time-domain Generalized Wiener Filter for Multi-channel Speech Separation
Figure 2 for A Time-domain Generalized Wiener Filter for Multi-channel Speech Separation
Figure 3 for A Time-domain Generalized Wiener Filter for Multi-channel Speech Separation
Figure 4 for A Time-domain Generalized Wiener Filter for Multi-channel Speech Separation
Viaarxiv icon

Adversarial synthesis based data-augmentation for code-switched spoken language identification

May 30, 2022
Parth Shastri, Chirag Patil, Poorval Wanere, Dr. Shrinivas Mahajan, Dr. Abhishek Bhatt, Dr. Hardik Sailor

Figure 1 for Adversarial synthesis based data-augmentation for code-switched spoken language identification
Figure 2 for Adversarial synthesis based data-augmentation for code-switched spoken language identification
Figure 3 for Adversarial synthesis based data-augmentation for code-switched spoken language identification
Figure 4 for Adversarial synthesis based data-augmentation for code-switched spoken language identification
Viaarxiv icon

Unsupervised Mismatch Localization in Cross-Modal Sequential Data

Add code
Bookmark button
Alert button
May 05, 2022
Wei Wei, Huang Hengguan, Gu Xiangming, Wang Hao, Wang Ye

Figure 1 for Unsupervised Mismatch Localization in Cross-Modal Sequential Data
Figure 2 for Unsupervised Mismatch Localization in Cross-Modal Sequential Data
Figure 3 for Unsupervised Mismatch Localization in Cross-Modal Sequential Data
Figure 4 for Unsupervised Mismatch Localization in Cross-Modal Sequential Data
Viaarxiv icon

CAB: Comprehensive Attention Benchmarking on Long Sequence Modeling

Add code
Bookmark button
Alert button
Oct 19, 2022
Jun Zhang, Shuyang Jiang, Jiangtao Feng, Lin Zheng, Lingpeng Kong

Figure 1 for CAB: Comprehensive Attention Benchmarking on Long Sequence Modeling
Figure 2 for CAB: Comprehensive Attention Benchmarking on Long Sequence Modeling
Figure 3 for CAB: Comprehensive Attention Benchmarking on Long Sequence Modeling
Figure 4 for CAB: Comprehensive Attention Benchmarking on Long Sequence Modeling
Viaarxiv icon

Reducing language context confusion for end-to-end code-switching automatic speech recognition

Jan 28, 2022
Shuai Zhang, Jiangyan Yi, Zhengkun Tian, Jianhua Tao, Yu Ting Yeung, Liqun Deng

Figure 1 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Figure 2 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Figure 3 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Figure 4 for Reducing language context confusion for end-to-end code-switching automatic speech recognition
Viaarxiv icon

Advancing Speech Recognition With No Speech Or With Noisy Speech

Jul 17, 2019
Gautam Krishna, Co Tran, Mason Carnahan, Ahmed H Tewfik

Figure 1 for Advancing Speech Recognition With No Speech Or With Noisy Speech
Figure 2 for Advancing Speech Recognition With No Speech Or With Noisy Speech
Figure 3 for Advancing Speech Recognition With No Speech Or With Noisy Speech
Figure 4 for Advancing Speech Recognition With No Speech Or With Noisy Speech
Viaarxiv icon

Meta-Learning for Adaptive Filters with Higher-Order Frequency Dependencies

Add code
Bookmark button
Alert button
Sep 20, 2022
Junkai Wu, Jonah Casebeer, Nicholas J. Bryan, Paris Smaragdis

Figure 1 for Meta-Learning for Adaptive Filters with Higher-Order Frequency Dependencies
Figure 2 for Meta-Learning for Adaptive Filters with Higher-Order Frequency Dependencies
Figure 3 for Meta-Learning for Adaptive Filters with Higher-Order Frequency Dependencies
Figure 4 for Meta-Learning for Adaptive Filters with Higher-Order Frequency Dependencies
Viaarxiv icon

Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer

Jul 29, 2022
Cong-Thanh Do, Mohan Li, Rama Doddipatla

Figure 1 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 2 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 3 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Figure 4 for Multiple-hypothesis RNN-T Loss for Unsupervised Fine-tuning and Self-training of Neural Transducer
Viaarxiv icon