Alert button
Picture for Mingye Xie

Mingye Xie

Alert button

EGE-UNet: an Efficient Group Enhanced UNet for skin lesion segmentation

Jul 17, 2023
Jiacheng Ruan, Mingye Xie, Jingsheng Gao, Ting Liu, Yuzhuo Fu

Figure 1 for EGE-UNet: an Efficient Group Enhanced UNet for skin lesion segmentation
Figure 2 for EGE-UNet: an Efficient Group Enhanced UNet for skin lesion segmentation
Figure 3 for EGE-UNet: an Efficient Group Enhanced UNet for skin lesion segmentation
Figure 4 for EGE-UNet: an Efficient Group Enhanced UNet for skin lesion segmentation

Transformer and its variants have been widely used for medical image segmentation. However, the large number of parameter and computational load of these models make them unsuitable for mobile health applications. To address this issue, we propose a more efficient approach, the Efficient Group Enhanced UNet (EGE-UNet). We incorporate a Group multi-axis Hadamard Product Attention module (GHPA) and a Group Aggregation Bridge module (GAB) in a lightweight manner. The GHPA groups input features and performs Hadamard Product Attention mechanism (HPA) on different axes to extract pathological information from diverse perspectives. The GAB effectively fuses multi-scale information by grouping low-level features, high-level features, and a mask generated by the decoder at each stage. Comprehensive experiments on the ISIC2017 and ISIC2018 datasets demonstrate that EGE-UNet outperforms existing state-of-the-art methods. In short, compared to the TransFuse, our model achieves superior segmentation performance while reducing parameter and computation costs by 494x and 160x, respectively. Moreover, to our best knowledge, this is the first model with a parameter count limited to just 50KB. Our code is available at https://github.com/JCruan519/EGE-UNet.

* 10 pages, 4 figures, 2 tables. This paper has been early accepted by MICCAI 2023 and has received the MICCAI Student-Author Registration (STAR) Award 
Viaarxiv icon

MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation

Nov 03, 2022
Jiacheng Ruan, Suncheng Xiang, Mingye Xie, Ting Liu, Yuzhuo Fu

Figure 1 for MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation
Figure 2 for MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation
Figure 3 for MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation
Figure 4 for MALUNet: A Multi-Attention and Light-weight UNet for Skin Lesion Segmentation

Recently, some pioneering works have preferred applying more complex modules to improve segmentation performances. However, it is not friendly for actual clinical environments due to limited computing resources. To address this challenge, we propose a light-weight model to achieve competitive performances for skin lesion segmentation at the lowest cost of parameters and computational complexity so far. Briefly, we propose four modules: (1) DGA consists of dilated convolution and gated attention mechanisms to extract global and local feature information; (2) IEA, which is based on external attention to characterize the overall datasets and enhance the connection between samples; (3) CAB is composed of 1D convolution and fully connected layers to perform a global and local fusion of multi-stage features to generate attention maps at channel axis; (4) SAB, which operates on multi-stage features by a shared 2D convolution to generate attention maps at spatial axis. We combine four modules with our U-shape architecture and obtain a light-weight medical image segmentation model dubbed as MALUNet. Compared with UNet, our model improves the mIoU and DSC metrics by 2.39% and 1.49%, respectively, with a 44x and 166x reduction in the number of parameters and computational complexity. In addition, we conduct comparison experiments on two skin lesion segmentation datasets (ISIC2017 and ISIC2018). Experimental results show that our model achieves state-of-the-art in balancing the number of parameters, computational complexity and segmentation performances. Code is available at https://github.com/JCruan519/MALUNet.

* 7 pages, 7 figures, 5 tables; This work has been accepted as a regular paper in IEEE BIBM 2022 
Viaarxiv icon

MEW-UNet: Multi-axis representation learning in frequency domain for medical image segmentation

Oct 25, 2022
Jiacheng Ruan, Mingye Xie, Suncheng Xiang, Ting Liu, Yuzhuo Fu

Figure 1 for MEW-UNet: Multi-axis representation learning in frequency domain for medical image segmentation
Figure 2 for MEW-UNet: Multi-axis representation learning in frequency domain for medical image segmentation
Figure 3 for MEW-UNet: Multi-axis representation learning in frequency domain for medical image segmentation
Figure 4 for MEW-UNet: Multi-axis representation learning in frequency domain for medical image segmentation

Recently, Visual Transformer (ViT) has been widely used in various fields of computer vision due to applying self-attention mechanism in the spatial domain to modeling global knowledge. Especially in medical image segmentation (MIS), many works are devoted to combining ViT and CNN, and even some works directly utilize pure ViT-based models. However, recent works improved models in the aspect of spatial domain while ignoring the importance of frequency domain information. Therefore, we propose Multi-axis External Weights UNet (MEW-UNet) for MIS based on the U-shape architecture by replacing self-attention in ViT with our Multi-axis External Weights block. Specifically, our block performs a Fourier transform on the three axes of the input feature and assigns the external weight in the frequency domain, which is generated by our Weights Generator. Then, an inverse Fourier transform is performed to change the features back to the spatial domain. We evaluate our model on four datasets and achieve state-of-the-art performances. In particular, on the Synapse dataset, our method outperforms MT-UNet by 10.15mm in terms of HD95. Code is available at https://github.com/JCruan519/MEW-UNet.

* 5 pages, 3 figures, 4 tables 
Viaarxiv icon