Alert button
Picture for Chengfeng Zhou

Chengfeng Zhou

Alert button

Towards Discriminative Representation with Meta-learning for Colonoscopic Polyp Re-Identification

Aug 02, 2023
Suncheng Xiang, Qingzhong Chen, Shilun Cai, Chengfeng Zhou, Crystal Cai, Sijia Du, Dahong Qian

Figure 1 for Towards Discriminative Representation with Meta-learning for Colonoscopic Polyp Re-Identification
Figure 2 for Towards Discriminative Representation with Meta-learning for Colonoscopic Polyp Re-Identification
Figure 3 for Towards Discriminative Representation with Meta-learning for Colonoscopic Polyp Re-Identification
Figure 4 for Towards Discriminative Representation with Meta-learning for Colonoscopic Polyp Re-Identification

Colonoscopic Polyp Re-Identification aims to match the same polyp from a large gallery with images from different views taken using different cameras and plays an important role in the prevention and treatment of colorectal cancer in computer-aided diagnosis. However, traditional methods for object ReID directly adopting CNN models trained on the ImageNet dataset usually produce unsatisfactory retrieval performance on colonoscopic datasets due to the large domain gap. Additionally, these methods neglect to explore the potential of self-discrepancy among intra-class relations in the colonoscopic polyp dataset, which remains an open research problem in the medical community. To solve this dilemma, we propose a simple but effective training method named Colo-ReID, which can help our model to learn more general and discriminative knowledge based on the meta-learning strategy in scenarios with fewer samples. Based on this, a dynamic Meta-Learning Regulation mechanism called MLR is introduced to further boost the performance of polyp re-identification. To the best of our knowledge, this is the first attempt to leverage the meta-learning paradigm instead of traditional machine learning to effectively train deep models in the task of colonoscopic polyp re-identification. Empirical results show that our method significantly outperforms current state-of-the-art methods by a clear margin.

* arXiv admin note: text overlap with arXiv:2307.10625 
Viaarxiv icon

Learning Robust Visual-Semantic Embedding for Generalizable Person Re-identification

Apr 19, 2023
Suncheng Xiang, Jingsheng Gao, Mengyuan Guan, Jiacheng Ruan, Chengfeng Zhou, Ting Liu, Dahong Qian, Yuzhuo Fu

Figure 1 for Learning Robust Visual-Semantic Embedding for Generalizable Person Re-identification
Figure 2 for Learning Robust Visual-Semantic Embedding for Generalizable Person Re-identification
Figure 3 for Learning Robust Visual-Semantic Embedding for Generalizable Person Re-identification
Figure 4 for Learning Robust Visual-Semantic Embedding for Generalizable Person Re-identification

Generalizable person re-identification (Re-ID) is a very hot research topic in machine learning and computer vision, which plays a significant role in realistic scenarios due to its various applications in public security and video surveillance. However, previous methods mainly focus on the visual representation learning, while neglect to explore the potential of semantic features during training, which easily leads to poor generalization capability when adapted to the new domain. In this paper, we propose a Multi-Modal Equivalent Transformer called MMET for more robust visual-semantic embedding learning on visual, textual and visual-textual tasks respectively. To further enhance the robust feature learning in the context of transformer, a dynamic masking mechanism called Masked Multimodal Modeling strategy (MMM) is introduced to mask both the image patches and the text tokens, which can jointly works on multimodal or unimodal data and significantly boost the performance of generalizable person Re-ID. Extensive experiments on benchmark datasets demonstrate the competitive performance of our method over previous approaches. We hope this method could advance the research towards visual-semantic representation learning. Our source code is also publicly available at https://github.com/JeremyXSC/MMET.

Viaarxiv icon

A simple normalization technique using window statistics to improve the out-of-distribution generalization on medical images

Jul 14, 2022
Chengfeng Zhou, Songchang Chen, Chenming Xu, Jun Wang, Feng Liu, Chun Zhang, Juan Ye, Hefeng Huang, Dahong Qian

Figure 1 for A simple normalization technique using window statistics to improve the out-of-distribution generalization on medical images
Figure 2 for A simple normalization technique using window statistics to improve the out-of-distribution generalization on medical images
Figure 3 for A simple normalization technique using window statistics to improve the out-of-distribution generalization on medical images
Figure 4 for A simple normalization technique using window statistics to improve the out-of-distribution generalization on medical images

Since data scarcity and data heterogeneity are prevailing for medical images, well-trained Convolutional Neural Networks (CNNs) using previous normalization methods may perform poorly when deployed to a new site. However, a reliable model for real-world clinical applications should be able to generalize well both on in-distribution (IND) and out-of-distribution (OOD) data (e.g., the new site data). In this study, we present a novel normalization technique called window normalization (WIN) to improve the model generalization on heterogeneous medical images, which is a simple yet effective alternative to existing normalization methods. Specifically, WIN perturbs the normalizing statistics with the local statistics computed on the window of features. This feature-level augmentation technique regularizes the models well and improves their OOD generalization significantly. Taking its advantage, we propose a novel self-distillation method called WIN-WIN for classification tasks. WIN-WIN is easily implemented with twice forward passes and a consistency constraint, which can be a simple extension for existing methods. Extensive experimental results on various tasks (6 tasks) and datasets (24 datasets) demonstrate the generality and effectiveness of our methods.

Viaarxiv icon

A simple normalization technique using window statistics to improve the out-of-distribution generalization in medical images

Jul 07, 2022
Chengfeng Zhou, Songchang Chen, Chenming Xu, Jun Wang, Chun Zhang, Juan Ye, Hefeng Huang, Dahong Qian

Figure 1 for A simple normalization technique using window statistics to improve the out-of-distribution generalization in medical images
Figure 2 for A simple normalization technique using window statistics to improve the out-of-distribution generalization in medical images
Figure 3 for A simple normalization technique using window statistics to improve the out-of-distribution generalization in medical images
Figure 4 for A simple normalization technique using window statistics to improve the out-of-distribution generalization in medical images

Since data scarcity and data heterogeneity are prevailing for medical images, well-trained Convolutional Neural Networks (CNNs) using previous normalization methods may perform poorly when deployed to a new site. However, a reliable model for real-world applications should be able to generalize well both on in-distribution (IND) and out-of-distribution (OOD) data (e.g., the new site data). In this study, we present a novel normalization technique called window normalization (WIN), which is a simple yet effective alternative to existing normalization methods. Specifically, WIN perturbs the normalizing statistics with the local statistics computed on a window of features. This feature-level augmentation technique regularizes the models well and improves their OOD generalization significantly. Taking its advantage, we propose a novel self-distillation method called WIN-WIN to further improve the OOD generalization in classification. WIN-WIN is easily implemented with twice forward passes and a consistency constraint, which can be a simple extension for existing methods. Extensive experimental results on various tasks (such as glaucoma detection, breast cancer detection, chromosome classification, optic disc and cup segmentation, etc.) and datasets (26 datasets) demonstrate the generality and effectiveness of our methods. The code is available at https://github.com/joe1chief/windowNormalizaion.

Viaarxiv icon