Alert button
Picture for Yuyao Zhang

Yuyao Zhang

Alert button

Self-Prompting Large Vision Models for Few-Shot Medical Image Segmentation

Aug 15, 2023
Qi Wu, Yuyao Zhang, Marawan Elbatel

Figure 1 for Self-Prompting Large Vision Models for Few-Shot Medical Image Segmentation
Figure 2 for Self-Prompting Large Vision Models for Few-Shot Medical Image Segmentation
Figure 3 for Self-Prompting Large Vision Models for Few-Shot Medical Image Segmentation
Figure 4 for Self-Prompting Large Vision Models for Few-Shot Medical Image Segmentation

Recent advancements in large foundation models have shown promising potential in the medical industry due to their flexible prompting capability. One such model, the Segment Anything Model (SAM), a prompt-driven segmentation model, has shown remarkable performance improvements, surpassing state-of-the-art approaches in medical image segmentation. However, existing methods primarily rely on tuning strategies that require extensive data or prior prompts tailored to the specific task, making it particularly challenging when only a limited number of data samples are available. In this paper, we propose a novel perspective on self-prompting in medical vision applications. Specifically, we harness the embedding space of SAM to prompt itself through a simple yet effective linear pixel-wise classifier. By preserving the encoding capabilities of the large model, the contextual information from its decoder, and leveraging its interactive promptability, we achieve competitive results on multiple datasets (i.e. improvement of more than 15% compared to fine-tuning the mask decoder using a few images).

* 8.5 pages + 2 pages of supplementary materials + 2 pages of references, 3 figures, submitted to 5th MICCAI Workshop on Domain Adaptation and Representation Transfer (DART) 
Viaarxiv icon

Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction

Jun 27, 2023
Qing Wu, Lixuan Chen, Ce Wang, Hongjiang Wei, S. Kevin Zhou, Jingyi Yu, Yuyao Zhang

Figure 1 for Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction
Figure 2 for Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction
Figure 3 for Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction
Figure 4 for Unsupervised Polychromatic Neural Representation for CT Metal Artifact Reduction

Emerging neural reconstruction techniques based on tomography (e.g., NeRF, NeAT, and NeRP) have started showing unique capabilities in medical imaging. In this work, we present a novel Polychromatic neural representation (Polyner) to tackle the challenging problem of CT imaging when metallic implants exist within the human body. The artifacts arise from the drastic variation of metal's attenuation coefficients at various energy levels of the X-ray spectrum, leading to a nonlinear metal effect in CT measurements. Reconstructing CT images from metal-affected measurements hence poses a complicated nonlinear inverse problem where empirical models adopted in previous metal artifact reduction (MAR) approaches lead to signal loss and strongly aliased reconstructions. Polyner instead models the MAR problem from a nonlinear inverse problem perspective. Specifically, we first derive a polychromatic forward model to accurately simulate the nonlinear CT acquisition process. Then, we incorporate our forward model into the implicit neural representation to accomplish reconstruction. Lastly, we adopt a regularizer to preserve the physical properties of the CT images across different energy levels while effectively constraining the solution space. Our Polyner is an unsupervised method and does not require any external training data. Experimenting with multiple datasets shows that our Polyner achieves comparable or better performance than supervised methods on in-domain datasets while demonstrating significant performance improvements on out-of-domain datasets. To the best of our knowledge, our Polyner is the first unsupervised MAR method that outperforms its supervised counterparts.

* 19 pages 
Viaarxiv icon

Mining fMRI Dynamics with Parcellation Prior for Brain Disease Diagnosis

May 04, 2023
Xiaozhao Liu, Mianxin Liu, Lang Mei, Yuyao Zhang, Feng Shi, Han Zhang, Dinggang Shen

Figure 1 for Mining fMRI Dynamics with Parcellation Prior for Brain Disease Diagnosis
Figure 2 for Mining fMRI Dynamics with Parcellation Prior for Brain Disease Diagnosis
Figure 3 for Mining fMRI Dynamics with Parcellation Prior for Brain Disease Diagnosis
Figure 4 for Mining fMRI Dynamics with Parcellation Prior for Brain Disease Diagnosis

To characterize atypical brain dynamics under diseases, prevalent studies investigate functional magnetic resonance imaging (fMRI). However, most of the existing analyses compress rich spatial-temporal information as the brain functional networks (BFNs) and directly investigate the whole-brain network without neurological priors about functional subnetworks. We thus propose a novel graph learning framework to mine fMRI signals with topological priors from brain parcellation for disease diagnosis. Specifically, we 1) detect diagnosis-related temporal features using a "Transformer" for a higher-level BFN construction, and process it with a following graph convolutional network, and 2) apply an attention-based multiple instance learning strategy to emphasize the disease-affected subnetworks to further enhance the diagnosis performance and interpretability. Experiments demonstrate higher effectiveness of our method than compared methods in the diagnosis of early mild cognitive impairment. More importantly, our method is capable of localizing crucial brain subnetworks during the diagnosis, providing insights into the pathogenic source of mild cognitive impairment.

* 5 pages, 2 figures, conference paper, accepted by IEEE International Symposium on Biomedical Imaging (ISBI) 2023 
Viaarxiv icon

Self-supervised arbitrary scale super-resolution framework for anisotropic MRI

May 02, 2023
Haonan Zhang, Yuhan Zhang, Qing Wu, Jiangjie Wu, Zhiming Zhen, Feng Shi, Jianmin Yuan, Hongjiang Wei, Chen Liu, Yuyao Zhang

Figure 1 for Self-supervised arbitrary scale super-resolution framework for anisotropic MRI
Figure 2 for Self-supervised arbitrary scale super-resolution framework for anisotropic MRI
Figure 3 for Self-supervised arbitrary scale super-resolution framework for anisotropic MRI
Figure 4 for Self-supervised arbitrary scale super-resolution framework for anisotropic MRI

In this paper, we propose an efficient self-supervised arbitrary-scale super-resolution (SR) framework to reconstruct isotropic magnetic resonance (MR) images from anisotropic MRI inputs without involving external training data. The proposed framework builds a training dataset using in-the-wild anisotropic MR volumes with arbitrary image resolution. We then formulate the 3D volume SR task as a SR problem for 2D image slices. The anisotropic volume's high-resolution (HR) plane is used to build the HR-LR image pairs for model training. We further adapt the implicit neural representation (INR) network to implement the 2D arbitrary-scale image SR model. Finally, we leverage the well-trained proposed model to up-sample the 2D LR plane extracted from the anisotropic MR volumes to their HR views. The isotropic MR volumes thus can be reconstructed by stacking and averaging the generated HR slices. Our proposed framework has two major advantages: (1) It only involves the arbitrary-resolution anisotropic MR volumes, which greatly improves the model practicality in real MR imaging scenarios (e.g., clinical brain image acquisition); (2) The INR-based SR model enables arbitrary-scale image SR from the arbitrary-resolution input image, which significantly improves model training efficiency. We perform experiments on a simulated public adult brain dataset and a real collected 7T brain dataset. The results indicate that our current framework greatly outperforms two well-known self-supervised models for anisotropic MR image SR tasks.

* 10 pages, 5 figures 
Viaarxiv icon

Spatiotemporal implicit neural representation for unsupervised dynamic MRI reconstruction

Dec 31, 2022
Jie Feng, Ruimin Feng, Qing Wu, Zhiyong Zhang, Yuyao Zhang, Hongjiang Wei

Figure 1 for Spatiotemporal implicit neural representation for unsupervised dynamic MRI reconstruction
Figure 2 for Spatiotemporal implicit neural representation for unsupervised dynamic MRI reconstruction
Figure 3 for Spatiotemporal implicit neural representation for unsupervised dynamic MRI reconstruction
Figure 4 for Spatiotemporal implicit neural representation for unsupervised dynamic MRI reconstruction

Supervised Deep-Learning (DL)-based reconstruction algorithms have shown state-of-the-art results for highly-undersampled dynamic Magnetic Resonance Imaging (MRI) reconstruction. However, the requirement of excessive high-quality ground-truth data hinders their applications due to the generalization problem. Recently, Implicit Neural Representation (INR) has appeared as a powerful DL-based tool for solving the inverse problem by characterizing the attributes of a signal as a continuous function of corresponding coordinates in an unsupervised manner. In this work, we proposed an INR-based method to improve dynamic MRI reconstruction from highly undersampled k-space data, which only takes spatiotemporal coordinates as inputs. Specifically, the proposed INR represents the dynamic MRI images as an implicit function and encodes them into neural networks. The weights of the network are learned from sparsely-acquired (k, t)-space data itself only, without external training datasets or prior images. Benefiting from the strong implicit continuity regularization of INR together with explicit regularization for low-rankness and sparsity, our proposed method outperforms the compared scan-specific methods at various acceleration factors. E.g., experiments on retrospective cardiac cine datasets show an improvement of 5.5 ~ 7.1 dB in PSNR for extremely high accelerations (up to 41.6-fold). The high-quality and inner continuity of the images provided by INR has great potential to further improve the spatiotemporal resolution of dynamic MRI, without the need of any training data.

* 9 pages, 5 figures 
Viaarxiv icon

Joint Rigid Motion Correction and Sparse-View CT via Self-Calibrating Neural Field

Nov 06, 2022
Qing Wu, Xin Li, Hongjiang Wei, Jingyi Yu, Yuyao Zhang

Figure 1 for Joint Rigid Motion Correction and Sparse-View CT via Self-Calibrating Neural Field
Figure 2 for Joint Rigid Motion Correction and Sparse-View CT via Self-Calibrating Neural Field
Figure 3 for Joint Rigid Motion Correction and Sparse-View CT via Self-Calibrating Neural Field
Figure 4 for Joint Rigid Motion Correction and Sparse-View CT via Self-Calibrating Neural Field

Neural Radiance Field (NeRF) has widely received attention in Sparse-View Computed Tomography (SVCT) reconstruction tasks as a self-supervised deep learning framework. NeRF-based SVCT methods represent the desired CT image as a continuous function of spatial coordinates and train a Multi-Layer Perceptron (MLP) to learn the function by minimizing loss on the SV sinogram. Benefiting from the continuous representation provided by NeRF, the high-quality CT image can be reconstructed. However, existing NeRF-based SVCT methods strictly suppose there is completely no relative motion during the CT acquisition because they require \textit{accurate} projection poses to model the X-rays that scan the SV sinogram. Therefore, these methods suffer from severe performance drops for real SVCT imaging with motion. In this work, we propose a self-calibrating neural field to recover the artifacts-free image from the rigid motion-corrupted SV sinogram without using any external data. Specifically, we parametrize the inaccurate projection poses caused by rigid motion as trainable variables and then jointly optimize these pose variables and the MLP. We conduct numerical experiments on a public CT image dataset. The results indicate our model significantly outperforms two representative NeRF-based methods for SVCT reconstruction tasks with four different levels of rigid motion.

* 5 pages 
Viaarxiv icon

A scan-specific unsupervised method for parallel MRI reconstruction via implicit neural representation

Oct 19, 2022
Ruimin Feng, Qing Wu, Yuyao Zhang, Hongjiang Wei

Figure 1 for A scan-specific unsupervised method for parallel MRI reconstruction via implicit neural representation
Figure 2 for A scan-specific unsupervised method for parallel MRI reconstruction via implicit neural representation
Figure 3 for A scan-specific unsupervised method for parallel MRI reconstruction via implicit neural representation
Figure 4 for A scan-specific unsupervised method for parallel MRI reconstruction via implicit neural representation

Parallel imaging is a widely-used technique to accelerate magnetic resonance imaging (MRI). However, current methods still perform poorly in reconstructing artifact-free MRI images from highly undersampled k-space data. Recently, implicit neural representation (INR) has emerged as a new deep learning paradigm for learning the internal continuity of an object. In this study, we adopted INR to parallel MRI reconstruction. The MRI image was modeled as a continuous function of spatial coordinates. This function was parameterized by a neural network and learned directly from the measured k-space itself without additional fully sampled high-quality training data. Benefitting from the powerful continuous representations provided by INR, the proposed method outperforms existing methods by suppressing the aliasing artifacts and noise, especially at higher acceleration rates and smaller sizes of the auto-calibration signals. The high-quality results and scanning specificity make the proposed method hold the potential for further accelerating the data acquisition of parallel MRI.

* conference 
Viaarxiv icon

SCULPTOR: Skeleton-Consistent Face Creation Using a Learned Parametric Generator

Sep 14, 2022
Zesong Qiu, Yuwei Li, Dongming He, Qixuan Zhang, Longwen Zhang, Yinghao Zhang, Jingya Wang, Lan Xu, Xudong Wang, Yuyao Zhang, Jingyi Yu

Figure 1 for SCULPTOR: Skeleton-Consistent Face Creation Using a Learned Parametric Generator
Figure 2 for SCULPTOR: Skeleton-Consistent Face Creation Using a Learned Parametric Generator
Figure 3 for SCULPTOR: Skeleton-Consistent Face Creation Using a Learned Parametric Generator
Figure 4 for SCULPTOR: Skeleton-Consistent Face Creation Using a Learned Parametric Generator

Recent years have seen growing interest in 3D human faces modelling due to its wide applications in digital human, character generation and animation. Existing approaches overwhelmingly emphasized on modeling the exterior shapes, textures and skin properties of faces, ignoring the inherent correlation between inner skeletal structures and appearance. In this paper, we present SCULPTOR, 3D face creations with Skeleton Consistency Using a Learned Parametric facial generaTOR, aiming to facilitate easy creation of both anatomically correct and visually convincing face models via a hybrid parametric-physical representation. At the core of SCULPTOR is LUCY, the first large-scale shape-skeleton face dataset in collaboration with plastic surgeons. Named after the fossils of one of the oldest known human ancestors, our LUCY dataset contains high-quality Computed Tomography (CT) scans of the complete human head before and after orthognathic surgeries, critical for evaluating surgery results. LUCY consists of 144 scans of 72 subjects (31 male and 41 female) where each subject has two CT scans taken pre- and post-orthognathic operations. Based on our LUCY dataset, we learn a novel skeleton consistent parametric facial generator, SCULPTOR, which can create the unique and nuanced facial features that help define a character and at the same time maintain physiological soundness. Our SCULPTOR jointly models the skull, face geometry and face appearance under a unified data-driven framework, by separating the depiction of a 3D face into shape blend shape, pose blend shape and facial expression blend shape. SCULPTOR preserves both anatomic correctness and visual realism in facial generation tasks compared with existing methods. Finally, we showcase the robustness and effectiveness of SCULPTOR in various fancy applications unseen before.

* 16 page, 13 figs 
Viaarxiv icon

Continuous longitudinal fetus brain atlas construction via implicit neural representation

Sep 14, 2022
Lixuan Chen, Jiangjie Wu, Qing Wu, Hongjiang Wei, Yuyao Zhang

Figure 1 for Continuous longitudinal fetus brain atlas construction via implicit neural representation
Figure 2 for Continuous longitudinal fetus brain atlas construction via implicit neural representation
Figure 3 for Continuous longitudinal fetus brain atlas construction via implicit neural representation
Figure 4 for Continuous longitudinal fetus brain atlas construction via implicit neural representation

Longitudinal fetal brain atlas is a powerful tool for understanding and characterizing the complex process of fetus brain development. Existing fetus brain atlases are typically constructed by averaged brain images on discrete time points independently over time. Due to the differences in onto-genetic trends among samples at different time points, the resulting atlases suffer from temporal inconsistency, which may lead to estimating error of the brain developmental characteristic parameters along the timeline. To this end, we proposed a multi-stage deep-learning framework to tackle the time inconsistency issue as a 4D (3D brain volume + 1D age) image data denoising task. Using implicit neural representation, we construct a continuous and noise-free longitudinal fetus brain atlas as a function of the 4D spatial-temporal coordinate. Experimental results on two public fetal brain atlases (CRL and FBA-Chinese atlases) show that the proposed method can significantly improve the atlas temporal consistency while maintaining good fetus brain structure representation. In addition, the continuous longitudinal fetus brain atlases can also be extensively applied to generate finer 4D atlases in both spatial and temporal resolution.

* 11 pages, 4 figures 
Viaarxiv icon

Noise2SR: Learning to Denoise from Super-Resolved Single Noisy Fluorescence Image

Sep 14, 2022
Xuanyu Tian, Qing Wu, Hongjiang Wei, Yuyao Zhang

Figure 1 for Noise2SR: Learning to Denoise from Super-Resolved Single Noisy Fluorescence Image
Figure 2 for Noise2SR: Learning to Denoise from Super-Resolved Single Noisy Fluorescence Image
Figure 3 for Noise2SR: Learning to Denoise from Super-Resolved Single Noisy Fluorescence Image
Figure 4 for Noise2SR: Learning to Denoise from Super-Resolved Single Noisy Fluorescence Image

Fluorescence microscopy is a key driver to promote discoveries of biomedical research. However, with the limitation of microscope hardware and characteristics of the observed samples, the fluorescence microscopy images are susceptible to noise. Recently, a few self-supervised deep learning (DL) denoising methods have been proposed. However, the training efficiency and denoising performance of existing methods are relatively low in real scene noise removal. To address this issue, this paper proposed self-supervised image denoising method Noise2SR (N2SR) to train a simple and effective image denoising model based on single noisy observation. Our Noise2SR denoising model is designed for training with paired noisy images of different dimensions. Benefiting from this training strategy, Noise2SR is more efficiently self-supervised and able to restore more image details from a single noisy observation. Experimental results of simulated noise and real microscopy noise removal show that Noise2SR outperforms two blind-spot based self-supervised deep learning image denoising methods. We envision that Noise2SR has the potential to improve more other kind of scientific imaging quality.

* MICCAI 2022  
* 12 pages, 6 figures 
Viaarxiv icon