Alert button
Picture for Heqing Zou

Heqing Zou

Alert button

MIR-GAN: Refining Frame-Level Modality-Invariant Representations with Adversarial Network for Audio-Visual Speech Recognition

Add code
Bookmark button
Alert button
Jun 18, 2023
Yuchen Hu, Chen Chen, Ruizhe Li, Heqing Zou, Eng Siong Chng

Figure 1 for MIR-GAN: Refining Frame-Level Modality-Invariant Representations with Adversarial Network for Audio-Visual Speech Recognition
Figure 2 for MIR-GAN: Refining Frame-Level Modality-Invariant Representations with Adversarial Network for Audio-Visual Speech Recognition
Figure 3 for MIR-GAN: Refining Frame-Level Modality-Invariant Representations with Adversarial Network for Audio-Visual Speech Recognition
Figure 4 for MIR-GAN: Refining Frame-Level Modality-Invariant Representations with Adversarial Network for Audio-Visual Speech Recognition
Viaarxiv icon

Towards Balanced Active Learning for Multimodal Classification

Add code
Bookmark button
Alert button
Jun 14, 2023
Meng Shen, Yizheng Huang, Jianxiong Yin, Heqing Zou, Deepu Rajan, Simon See

Figure 1 for Towards Balanced Active Learning for Multimodal Classification
Figure 2 for Towards Balanced Active Learning for Multimodal Classification
Figure 3 for Towards Balanced Active Learning for Multimodal Classification
Figure 4 for Towards Balanced Active Learning for Multimodal Classification
Viaarxiv icon

UniS-MMC: Multimodal Classification via Unimodality-supervised Multimodal Contrastive Learning

Add code
Bookmark button
Alert button
May 16, 2023
Heqing Zou, Meng Shen, Chen Chen, Yuchen Hu, Deepu Rajan, Eng Siong Chng

Figure 1 for UniS-MMC: Multimodal Classification via Unimodality-supervised Multimodal Contrastive Learning
Figure 2 for UniS-MMC: Multimodal Classification via Unimodality-supervised Multimodal Contrastive Learning
Figure 3 for UniS-MMC: Multimodal Classification via Unimodality-supervised Multimodal Contrastive Learning
Figure 4 for UniS-MMC: Multimodal Classification via Unimodality-supervised Multimodal Contrastive Learning
Viaarxiv icon

Cross-Modal Global Interaction and Local Alignment for Audio-Visual Speech Recognition

Add code
Bookmark button
Alert button
May 16, 2023
Yuchen Hu, Ruizhe Li, Chen Chen, Heqing Zou, Qiushi Zhu, Eng Siong Chng

Figure 1 for Cross-Modal Global Interaction and Local Alignment for Audio-Visual Speech Recognition
Figure 2 for Cross-Modal Global Interaction and Local Alignment for Audio-Visual Speech Recognition
Figure 3 for Cross-Modal Global Interaction and Local Alignment for Audio-Visual Speech Recognition
Figure 4 for Cross-Modal Global Interaction and Local Alignment for Audio-Visual Speech Recognition
Viaarxiv icon

Unsupervised Noise adaptation using Data Simulation

Add code
Bookmark button
Alert button
Feb 23, 2023
Chen Chen, Yuchen Hu, Heqing Zou, Linhui Sun, Eng Siong Chng

Figure 1 for Unsupervised Noise adaptation using Data Simulation
Figure 2 for Unsupervised Noise adaptation using Data Simulation
Figure 3 for Unsupervised Noise adaptation using Data Simulation
Figure 4 for Unsupervised Noise adaptation using Data Simulation
Viaarxiv icon

Unifying Speech Enhancement and Separation with Gradient Modulation for End-to-End Noise-Robust Speech Separation

Add code
Bookmark button
Alert button
Feb 22, 2023
Yuchen Hu, Chen Chen, Heqing Zou, Xionghu Zhong, Eng Siong Chng

Figure 1 for Unifying Speech Enhancement and Separation with Gradient Modulation for End-to-End Noise-Robust Speech Separation
Figure 2 for Unifying Speech Enhancement and Separation with Gradient Modulation for End-to-End Noise-Robust Speech Separation
Figure 3 for Unifying Speech Enhancement and Separation with Gradient Modulation for End-to-End Noise-Robust Speech Separation
Figure 4 for Unifying Speech Enhancement and Separation with Gradient Modulation for End-to-End Noise-Robust Speech Separation
Viaarxiv icon

Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning

Add code
Bookmark button
Alert button
Dec 10, 2022
Chen Chen, Yuchen Hu, Qiang Zhang, Heqing Zou, Beier Zhu, Eng Siong Chng

Figure 1 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Figure 2 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Figure 3 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Figure 4 for Leveraging Modality-specific Representations for Audio-visual Speech Recognition via Reinforcement Learning
Viaarxiv icon

Self-critical Sequence Training for Automatic Speech Recognition

Add code
Bookmark button
Alert button
Apr 13, 2022
Chen Chen, Yuchen Hu, Nana Hou, Xiaofeng Qi, Heqing Zou, Eng Siong Chng

Figure 1 for Self-critical Sequence Training for Automatic Speech Recognition
Figure 2 for Self-critical Sequence Training for Automatic Speech Recognition
Figure 3 for Self-critical Sequence Training for Automatic Speech Recognition
Figure 4 for Self-critical Sequence Training for Automatic Speech Recognition
Viaarxiv icon

Interactive Audio-text Representation for Automated Audio Captioning with Contrastive Learning

Add code
Bookmark button
Alert button
Apr 12, 2022
Chen Chen, Nana Hou, Yuchen Hu, Heqing Zou, Xiaofeng Qi, Eng Siong Chng

Figure 1 for Interactive Audio-text Representation for Automated Audio Captioning with Contrastive Learning
Figure 2 for Interactive Audio-text Representation for Automated Audio Captioning with Contrastive Learning
Figure 3 for Interactive Audio-text Representation for Automated Audio Captioning with Contrastive Learning
Figure 4 for Interactive Audio-text Representation for Automated Audio Captioning with Contrastive Learning
Viaarxiv icon

Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information

Add code
Bookmark button
Alert button
Mar 29, 2022
Heqing Zou, Yuke Si, Chen Chen, Deepu Rajan, Eng Siong Chng

Figure 1 for Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information
Figure 2 for Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information
Figure 3 for Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information
Figure 4 for Speech Emotion Recognition with Co-Attention based Multi-level Acoustic Information
Viaarxiv icon