Alert button
Picture for Shandong Wu

Shandong Wu

Alert button

FedPerfix: Towards Partial Model Personalization of Vision Transformers in Federated Learning

Aug 17, 2023
Guangyu Sun, Matias Mendieta, Jun Luo, Shandong Wu, Chen Chen

Figure 1 for FedPerfix: Towards Partial Model Personalization of Vision Transformers in Federated Learning
Figure 2 for FedPerfix: Towards Partial Model Personalization of Vision Transformers in Federated Learning
Figure 3 for FedPerfix: Towards Partial Model Personalization of Vision Transformers in Federated Learning
Figure 4 for FedPerfix: Towards Partial Model Personalization of Vision Transformers in Federated Learning

Personalized Federated Learning (PFL) represents a promising solution for decentralized learning in heterogeneous data environments. Partial model personalization has been proposed to improve the efficiency of PFL by selectively updating local model parameters instead of aggregating all of them. However, previous work on partial model personalization has mainly focused on Convolutional Neural Networks (CNNs), leaving a gap in understanding how it can be applied to other popular models such as Vision Transformers (ViTs). In this work, we investigate where and how to partially personalize a ViT model. Specifically, we empirically evaluate the sensitivity to data distribution of each type of layer. Based on the insights that the self-attention layer and the classification head are the most sensitive parts of a ViT, we propose a novel approach called FedPerfix, which leverages plugins to transfer information from the aggregated model to the local client as a personalization. Finally, we evaluate the proposed approach on CIFAR-100, OrganAMNIST, and Office-Home datasets and demonstrate its effectiveness in improving the model's performance compared to several advanced PFL methods.

* 2023 IEEE/CVF International Conference on Computer Vision (ICCV) 
Viaarxiv icon

Human not in the loop: objective sample difficulty measures for Curriculum Learning

Feb 02, 2023
Zhengbo Zhou, Jun Luo, Gene Kitamura, Shandong Wu

Figure 1 for Human not in the loop: objective sample difficulty measures for Curriculum Learning
Figure 2 for Human not in the loop: objective sample difficulty measures for Curriculum Learning
Figure 3 for Human not in the loop: objective sample difficulty measures for Curriculum Learning
Figure 4 for Human not in the loop: objective sample difficulty measures for Curriculum Learning

Curriculum learning is a learning method that trains models in a meaningful order from easier to harder samples. A key here is to devise automatic and objective difficulty measures of samples. In the medical domain, previous work applied domain knowledge from human experts to qualitatively assess classification difficulty of medical images to guide curriculum learning, which requires extra annotation efforts, relies on subjective human experience, and may introduce bias. In this work, we propose a new automated curriculum learning technique using the variance of gradients (VoG) to compute an objective difficulty measure of samples and evaluated its effects on elbow fracture classification from X-ray images. Specifically, we used VoG as a metric to rank each sample in terms of the classification difficulty, where high VoG scores indicate more difficult cases for classification, to guide the curriculum training process We compared the proposed technique to a baseline (without curriculum learning), a previous method that used human annotations on classification difficulty, and anti-curriculum learning. Our experiment results showed comparable and higher performance for the binary and multi-class bone fracture classification tasks.

* ISBI 2023 
Viaarxiv icon

PGFed: Personalize Each Client's Global Objective for Federated Learning

Dec 02, 2022
Jun Luo, Matias Mendieta, Chen Chen, Shandong Wu

Figure 1 for PGFed: Personalize Each Client's Global Objective for Federated Learning
Figure 2 for PGFed: Personalize Each Client's Global Objective for Federated Learning
Figure 3 for PGFed: Personalize Each Client's Global Objective for Federated Learning
Figure 4 for PGFed: Personalize Each Client's Global Objective for Federated Learning

The mediocre performance of conventional federated learning (FL) over heterogeneous data has been facilitating personalized FL solutions, where, unlike conventional FL which trains a single global consensus model, different models are allowed for different clients. However, in most existing personalized FL algorithms, the collaborative knowledge across the federation was only implicitly passed to the clients in ways such as model aggregation or regularization. We observed that this implicit knowledge transfer fails to maximize the potential value of each client's empirical risk toward other clients. Based on our observation, in this work, we propose Personalized Global Federated Learning (PGFed), a novel personalized FL framework that enables each client to personalize its own global objective by explicitly and adaptively aggregating the empirical risks of itself and other clients. To avoid massive ($O(N^2)$) communication overhead and potential privacy leakage, each client's risk is estimated through a first-order approximation for other clients' adaptive risk aggregation. On top of PGFed, we develop a momentum upgrade, dubbed PGFedMo, to more efficiently utilize clients' empirical risks. Our extensive experiments under different federated settings with benchmark datasets show consistent improvements of PGFed over the compared state-of-the-art alternatives.

* 14 pages, 6 figures 
Viaarxiv icon

Cross-Domain Self-Supervised Deep Learning for Robust Alzheimer's Disease Progression Modeling

Nov 15, 2022
Saba Dadsetan, Mohsen Hejrati, Shandong Wu, Somaye Hashemifar

Figure 1 for Cross-Domain Self-Supervised Deep Learning for Robust Alzheimer's Disease Progression Modeling
Figure 2 for Cross-Domain Self-Supervised Deep Learning for Robust Alzheimer's Disease Progression Modeling
Figure 3 for Cross-Domain Self-Supervised Deep Learning for Robust Alzheimer's Disease Progression Modeling
Figure 4 for Cross-Domain Self-Supervised Deep Learning for Robust Alzheimer's Disease Progression Modeling

Developing successful artificial intelligence systems in practice depends both on robust deep learning models as well as large high quality data. Acquiring and labeling data can become prohibitively expensive and time-consuming in many real-world applications such as clinical disease models. Self-supervised learning has demonstrated great potential in increasing model accuracy and robustness in small data regimes. In addition, many clinical imaging and disease modeling applications rely heavily on regression of continuous quantities. However, the applicability of self-supervised learning for these medical-imaging regression tasks has not been extensively studied. In this study, we develop a cross-domain self-supervised learning approach for disease prognostic modeling as a regression problem using 3D images as input. We demonstrate that self-supervised pre-training can improve the prediction of Alzheimer's Disease progression from brain MRI. We also show that pre-training on extended (but not labeled) brain MRI data outperforms pre-training on natural images. We further observe that the highest performance is achieved when both natural images and extended brain-MRI data are used for pre-training.

Viaarxiv icon

Medical Knowledge-Guided Deep Learning for Imbalanced Medical Image Classification

Nov 20, 2021
Long Gao, Chang Liu, Dooman Arefan, Ashok Panigrahy, Margarita L. Zuley, Shandong Wu

Figure 1 for Medical Knowledge-Guided Deep Learning for Imbalanced Medical Image Classification
Figure 2 for Medical Knowledge-Guided Deep Learning for Imbalanced Medical Image Classification
Figure 3 for Medical Knowledge-Guided Deep Learning for Imbalanced Medical Image Classification
Figure 4 for Medical Knowledge-Guided Deep Learning for Imbalanced Medical Image Classification

Deep learning models have gained remarkable performance on a variety of image classification tasks. However, many models suffer from limited performance in clinical or medical settings when data are imbalanced. To address this challenge, we propose a medical-knowledge-guided one-class classification approach that leverages domain-specific knowledge of classification tasks to boost the model's performance. The rationale behind our approach is that some existing prior medical knowledge can be incorporated into data-driven deep learning to facilitate model learning. We design a deep learning-based one-class classification pipeline for imbalanced image classification, and demonstrate in three use cases how we take advantage of medical knowledge of each specific classification task by generating additional middle classes to achieve higher classification performances. We evaluate our approach on three different clinical image classification tasks (a total of 8459 images) and show superior model performance when compared to six state-of-the-art methods. All codes of this work will be publicly available upon acceptance of the paper.

* 11 pages 
Viaarxiv icon

Constrained Deep One-Class Feature Learning For Classifying Imbalanced Medical Images

Nov 20, 2021
Long Gao, Chang Liu, Dooman Arefan, Ashok Panigrahy, Shandong Wu

Figure 1 for Constrained Deep One-Class Feature Learning For Classifying Imbalanced Medical Images
Figure 2 for Constrained Deep One-Class Feature Learning For Classifying Imbalanced Medical Images
Figure 3 for Constrained Deep One-Class Feature Learning For Classifying Imbalanced Medical Images
Figure 4 for Constrained Deep One-Class Feature Learning For Classifying Imbalanced Medical Images

Medical image data are usually imbalanced across different classes. One-class classification has attracted increasing attention to address the data imbalance problem by distinguishing the samples of the minority class from the majority class. Previous methods generally aim to either learn a new feature space to map training samples together or to fit training samples by autoencoder-like models. These methods mainly focus on capturing either compact or descriptive features, where the information of the samples of a given one class is not sufficiently utilized. In this paper, we propose a novel deep learning-based method to learn compact features by adding constraints on the bottleneck features, and to preserve descriptive features by training an autoencoder at the same time. Through jointly optimizing the constraining loss and the autoencoder's reconstruction loss, our method can learn more relevant features associated with the given class, making the majority and minority samples more distinguishable. Experimental results on three clinical datasets (including the MRI breast images, FFDM breast images and chest X-ray images) obtains state-of-art performance compared to previous methods.

* 10 pages 
Viaarxiv icon

Deep Curriculum Learning in Task Space for Multi-Class Based Mammography Diagnosis

Oct 21, 2021
Jun Luo, Dooman Arefan, Margarita Zuley, Jules Sumkin, Shandong Wu

Figure 1 for Deep Curriculum Learning in Task Space for Multi-Class Based Mammography Diagnosis
Figure 2 for Deep Curriculum Learning in Task Space for Multi-Class Based Mammography Diagnosis
Figure 3 for Deep Curriculum Learning in Task Space for Multi-Class Based Mammography Diagnosis
Figure 4 for Deep Curriculum Learning in Task Space for Multi-Class Based Mammography Diagnosis

Mammography is used as a standard screening procedure for the potential patients of breast cancer. Over the past decade, it has been shown that deep learning techniques have succeeded in reaching near-human performance in a number of tasks, and its application in mammography is one of the topics that medical researchers most concentrate on. In this work, we propose an end-to-end Curriculum Learning (CL) strategy in task space for classifying the three categories of Full-Field Digital Mammography (FFDM), namely Malignant, Negative, and False recall. Specifically, our method treats this three-class classification as a "harder" task in terms of CL, and create an "easier" sub-task of classifying False recall against the combined group of Negative and Malignant. We introduce a loss scheduler to dynamically weight the contribution of the losses from the two tasks throughout the entire training process. We conduct experiments on an FFDM datasets of 1,709 images using 5-fold cross validation. The results show that our curriculum learning strategy can boost the performance for classifying the three categories of FFDM compared to the baseline strategies for model training.

* 4-page abstract. Full paper to appear at SPIE Medical Imaging 2022 
Viaarxiv icon

Knowledge-Guided Multiview Deep Curriculum Learning for Elbow Fracture Classification

Oct 20, 2021
Jun Luo, Gene Kitamura, Dooman Arefan, Emine Doganay, Ashok Panigrahy, Shandong Wu

Figure 1 for Knowledge-Guided Multiview Deep Curriculum Learning for Elbow Fracture Classification
Figure 2 for Knowledge-Guided Multiview Deep Curriculum Learning for Elbow Fracture Classification
Figure 3 for Knowledge-Guided Multiview Deep Curriculum Learning for Elbow Fracture Classification
Figure 4 for Knowledge-Guided Multiview Deep Curriculum Learning for Elbow Fracture Classification

Elbow fracture diagnosis often requires patients to take both frontal and lateral views of elbow X-ray radiographs. In this paper, we propose a multiview deep learning method for an elbow fracture subtype classification task. Our strategy leverages transfer learning by first training two single-view models, one for frontal view and the other for lateral view, and then transferring the weights to the corresponding layers in the proposed multiview network architecture. Meanwhile, quantitative medical knowledge was integrated into the training process through a curriculum learning framework, which enables the model to first learn from "easier" samples and then transition to "harder" samples to reach better performance. In addition, our multiview network can work both in a dual-view setting and with a single view as input. We evaluate our method through extensive experiments on a classification task of elbow fracture with a dataset of 1,964 images. Results show that our method outperforms two related methods on bone fracture study in multiple settings, and our technique is able to boost the performance of the compared methods. The code is available at https://github.com/ljaiverson/multiview-curriculum.

* In International Workshop on Machine Learning in Medical Imaging (pp. 555-564) at MICCAI 2021. Springer, Cham  
* MICCAI 2021 workshop. DOI: https://doi.org/10.1007/978-3-030-87589-3_57. URL: https://link.springer.com/chapter/10.1007/978-3-030-87589-3_57 
Viaarxiv icon

Medical Knowledge-Guided Deep Curriculum Learning for Elbow Fracture Diagnosis from X-Ray Images

Oct 20, 2021
Jun Luo, Gene Kitamura, Emine Doganay, Dooman Arefan, Shandong Wu

Elbow fractures are one of the most common fracture types. Diagnoses on elbow fractures often need the help of radiographic imaging to be read and analyzed by a specialized radiologist with years of training. Thanks to the recent advances of deep learning, a model that can classify and detect different types of bone fractures needs only hours of training and has shown promising results. However, most existing deep learning models are purely data-driven, lacking incorporation of known domain knowledge from human experts. In this work, we propose a novel deep learning method to diagnose elbow fracture from elbow X-ray images by integrating domain-specific medical knowledge into a curriculum learning framework. In our method, the training data are permutated by sampling without replacement at the beginning of each training epoch. The sampling probability of each training sample is guided by a scoring criterion constructed based on clinically known knowledge from human experts, where the scoring indicates the diagnosis difficultness of different elbow fracture subtypes. We also propose an algorithm that updates the sampling probabilities at each epoch, which is applicable to other sampling-based curriculum learning frameworks. We design an experiment with 1865 elbow X-ray images for a fracture/normal binary classification task and compare our proposed method to a baseline method and a previous method using multiple metrics. Our results show that the proposed method achieves the highest classification performance. Also, our proposed probability update algorithm boosts the performance of the previous method.

* SPIE Medical Imaging 2021: Computer-Aided Diagnosis  
* SPIE Medical Imaging 2021. DOI: https://doi.org/10.1117/12.2582184. URL: https://www.spiedigitallibrary.org/conference-proceedings-of-spie/11597/1159712/Medical-knowledge-guided-deep-curriculum-learning-for-elbow-fracture-diagnosis/10.1117/12.2582184.short?SSO=1 
Viaarxiv icon

Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning

Oct 15, 2021
Jun Luo, Shandong Wu

Figure 1 for Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning
Figure 2 for Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning
Figure 3 for Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning
Figure 4 for Adapt to Adaptation: Learning Personalization for Cross-Silo Federated Learning

The goal of conventional federated learning (FL) is to train a global model for a federation of clients with decentralized data, reducing the systemic privacy risk of centralized training. The distribution shift across non-IID datasets, also known as the data heterogeneity, often poses a challenge for this one-global-model-fits-all solution. In this work, we propose APPLE, a personalized cross-silo FL framework that adaptively learns how much each client can benefit from other clients' models. We also introduce a method to flexibly control the focus of training APPLE between global and local objectives. We empirically evaluate our method's convergence and generalization behavior and performed extensive experiments on two benchmark datasets and two medical imaging datasets under two non-IID settings. The results show that the proposed personalized FL framework, APPLE, achieves state-of-the-art performance compared to several other personalized FL approaches in the literature.

* 15 pages 
Viaarxiv icon