Alert button
Picture for Xin Fang

Xin Fang

Alert button

Multitask frame-level learning for few-shot sound event detection

Add code
Bookmark button
Alert button
Mar 17, 2024
Liang Zou, Genwei Yan, Ruoyu Wang, Jun Du, Meng Lei, Tian Gao, Xin Fang

Figure 1 for Multitask frame-level learning for few-shot sound event detection
Figure 2 for Multitask frame-level learning for few-shot sound event detection
Figure 3 for Multitask frame-level learning for few-shot sound event detection
Figure 4 for Multitask frame-level learning for few-shot sound event detection
Viaarxiv icon

SegRap2023: A Benchmark of Organs-at-Risk and Gross Tumor Volume Segmentation for Radiotherapy Planning of Nasopharyngeal Carcinoma

Add code
Bookmark button
Alert button
Dec 15, 2023
Xiangde Luo, Jia Fu, Yunxin Zhong, Shuolin Liu, Bing Han, Mehdi Astaraki, Simone Bendazzoli, Iuliana Toma-Dasu, Yiwen Ye, Ziyang Chen, Yong Xia, Yanzhou Su, Jin Ye, Junjun He, Zhaohu Xing, Hongqiu Wang, Lei Zhu, Kaixiang Yang, Xin Fang, Zhiwei Wang, Chan Woong Lee, Sang Joon Park, Jaehee Chun, Constantin Ulrich, Klaus H. Maier-Hein, Nchongmaje Ndipenoch, Alina Miron, Yongmin Li, Yimeng Zhang, Yu Chen, Lu Bai, Jinlong Huang, Chengyang An, Lisheng Wang, Kaiwen Huang, Yunqi Gu, Tao Zhou, Mu Zhou, Shichuan Zhang, Wenjun Liao, Guotai Wang, Shaoting Zhang

Figure 1 for SegRap2023: A Benchmark of Organs-at-Risk and Gross Tumor Volume Segmentation for Radiotherapy Planning of Nasopharyngeal Carcinoma
Figure 2 for SegRap2023: A Benchmark of Organs-at-Risk and Gross Tumor Volume Segmentation for Radiotherapy Planning of Nasopharyngeal Carcinoma
Figure 3 for SegRap2023: A Benchmark of Organs-at-Risk and Gross Tumor Volume Segmentation for Radiotherapy Planning of Nasopharyngeal Carcinoma
Figure 4 for SegRap2023: A Benchmark of Organs-at-Risk and Gross Tumor Volume Segmentation for Radiotherapy Planning of Nasopharyngeal Carcinoma
Viaarxiv icon

AST-SED: An Effective Sound Event Detection Method Based on Audio Spectrogram Transformer

Add code
Bookmark button
Alert button
Mar 07, 2023
Kang Li, Yan Song, Li-Rong Dai, Ian McLoughlin, Xin Fang, Lin Liu

Figure 1 for AST-SED: An Effective Sound Event Detection Method Based on Audio Spectrogram Transformer
Figure 2 for AST-SED: An Effective Sound Event Detection Method Based on Audio Spectrogram Transformer
Figure 3 for AST-SED: An Effective Sound Event Detection Method Based on Audio Spectrogram Transformer
Figure 4 for AST-SED: An Effective Sound Event Detection Method Based on Audio Spectrogram Transformer
Viaarxiv icon

Deep Virtual-to-Real Distillation for Pedestrian Crossing Prediction

Add code
Bookmark button
Alert button
Nov 02, 2022
Jie Bai, Xin Fang, Jianwu Fang, Jianru Xue, Changwei Yuan

Figure 1 for Deep Virtual-to-Real Distillation for Pedestrian Crossing Prediction
Figure 2 for Deep Virtual-to-Real Distillation for Pedestrian Crossing Prediction
Figure 3 for Deep Virtual-to-Real Distillation for Pedestrian Crossing Prediction
Figure 4 for Deep Virtual-to-Real Distillation for Pedestrian Crossing Prediction
Viaarxiv icon

A Complementary Joint Training Approach Using Unpaired Speech and Text for Low-Resource Automatic Speech Recognition

Add code
Bookmark button
Alert button
Apr 05, 2022
Ye-Qian Du, Jie Zhang, Qiu-Shi Zhu, Li-Rong Dai, Ming-Hui Wu, Xin Fang, Zhou-Wang Yang

Figure 1 for A Complementary Joint Training Approach Using Unpaired Speech and Text for Low-Resource Automatic Speech Recognition
Figure 2 for A Complementary Joint Training Approach Using Unpaired Speech and Text for Low-Resource Automatic Speech Recognition
Figure 3 for A Complementary Joint Training Approach Using Unpaired Speech and Text for Low-Resource Automatic Speech Recognition
Figure 4 for A Complementary Joint Training Approach Using Unpaired Speech and Text for Low-Resource Automatic Speech Recognition
Viaarxiv icon

Learning Contextually Fused Audio-visual Representations for Audio-visual Speech Recognition

Add code
Bookmark button
Alert button
Feb 15, 2022
Zi-Qiang Zhang, Jie Zhang, Jian-Shu Zhang, Ming-Hui Wu, Xin Fang, Li-Rong Dai

Figure 1 for Learning Contextually Fused Audio-visual Representations for Audio-visual Speech Recognition
Figure 2 for Learning Contextually Fused Audio-visual Representations for Audio-visual Speech Recognition
Figure 3 for Learning Contextually Fused Audio-visual Representations for Audio-visual Speech Recognition
Figure 4 for Learning Contextually Fused Audio-visual Representations for Audio-visual Speech Recognition
Viaarxiv icon

A Noise-Robust Self-supervised Pre-training Model Based Speech Representation Learning for Automatic Speech Recognition

Add code
Bookmark button
Alert button
Jan 22, 2022
Qiu-Shi Zhu, Jie Zhang, Zi-Qiang Zhang, Ming-Hui Wu, Xin Fang, Li-Rong Dai

Figure 1 for A Noise-Robust Self-supervised Pre-training Model Based Speech Representation Learning for Automatic Speech Recognition
Figure 2 for A Noise-Robust Self-supervised Pre-training Model Based Speech Representation Learning for Automatic Speech Recognition
Figure 3 for A Noise-Robust Self-supervised Pre-training Model Based Speech Representation Learning for Automatic Speech Recognition
Figure 4 for A Noise-Robust Self-supervised Pre-training Model Based Speech Representation Learning for Automatic Speech Recognition
Viaarxiv icon

USTC-NELSLIP System Description for DIHARD-III Challenge

Add code
Bookmark button
Alert button
Mar 19, 2021
Yuxuan Wang, Maokui He, Shutong Niu, Lei Sun, Tian Gao, Xin Fang, Jia Pan, Jun Du, Chin-Hui Lee

Figure 1 for USTC-NELSLIP System Description for DIHARD-III Challenge
Figure 2 for USTC-NELSLIP System Description for DIHARD-III Challenge
Figure 3 for USTC-NELSLIP System Description for DIHARD-III Challenge
Figure 4 for USTC-NELSLIP System Description for DIHARD-III Challenge
Viaarxiv icon

XLST: Cross-lingual Self-training to Learn Multilingual Representation for Low Resource Speech Recognition

Add code
Bookmark button
Alert button
Mar 15, 2021
Zi-Qiang Zhang, Yan Song, Ming-Hui Wu, Xin Fang, Li-Rong Dai

Figure 1 for XLST: Cross-lingual Self-training to Learn Multilingual Representation for Low Resource Speech Recognition
Figure 2 for XLST: Cross-lingual Self-training to Learn Multilingual Representation for Low Resource Speech Recognition
Figure 3 for XLST: Cross-lingual Self-training to Learn Multilingual Representation for Low Resource Speech Recognition
Figure 4 for XLST: Cross-lingual Self-training to Learn Multilingual Representation for Low Resource Speech Recognition
Viaarxiv icon