Alert button
Picture for Ruixuan Wang

Ruixuan Wang

Alert button

Class Attention to Regions of Lesion for Imbalanced Medical Image Recognition

Jul 20, 2023
Jia-Xin Zhuang, Jiabin Cai, Jianguo Zhang, Wei-shi Zheng, Ruixuan Wang

Figure 1 for Class Attention to Regions of Lesion for Imbalanced Medical Image Recognition
Figure 2 for Class Attention to Regions of Lesion for Imbalanced Medical Image Recognition
Figure 3 for Class Attention to Regions of Lesion for Imbalanced Medical Image Recognition
Figure 4 for Class Attention to Regions of Lesion for Imbalanced Medical Image Recognition

Automated medical image classification is the key component in intelligent diagnosis systems. However, most medical image datasets contain plenty of samples of common diseases and just a handful of rare ones, leading to major class imbalances. Currently, it is an open problem in intelligent diagnosis to effectively learn from imbalanced training data. In this paper, we propose a simple yet effective framework, named \textbf{C}lass \textbf{A}ttention to \textbf{RE}gions of the lesion (CARE), to handle data imbalance issues by embedding attention into the training process of \textbf{C}onvolutional \textbf{N}eural \textbf{N}etworks (CNNs). The proposed attention module helps CNNs attend to lesion regions of rare diseases, therefore helping CNNs to learn their characteristics more effectively. In addition, this attention module works only during the training phase and does not change the architecture of the original network, so it can be directly combined with any existing CNN architecture. The CARE framework needs bounding boxes to represent the lesion regions of rare diseases. To alleviate the need for manual annotation, we further developed variants of CARE by leveraging the traditional saliency methods or a pretrained segmentation model for bounding box generation. Results show that the CARE variants with automated bounding box generation are comparable to the original CARE framework with \textit{manual} bounding box annotations. A series of experiments on an imbalanced skin image dataset and a pneumonia dataset indicates that our method can effectively help the network focus on the lesion regions of rare diseases and remarkably improves the classification performance of rare diseases.

* Accepted by Neurocomputing on July 2023. 37 pages 
Viaarxiv icon

Adapter Learning in Pretrained Feature Extractor for Continual Learning of Diseases

Apr 18, 2023
Wentao Zhang, Yujun Huang, Tong Zhang, Qingsong Zou, Wei-Shi Zheng, Ruixuan Wang

Figure 1 for Adapter Learning in Pretrained Feature Extractor for Continual Learning of Diseases
Figure 2 for Adapter Learning in Pretrained Feature Extractor for Continual Learning of Diseases
Figure 3 for Adapter Learning in Pretrained Feature Extractor for Continual Learning of Diseases
Figure 4 for Adapter Learning in Pretrained Feature Extractor for Continual Learning of Diseases

Currently intelligent diagnosis systems lack the ability of continually learning to diagnose new diseases once deployed, under the condition of preserving old disease knowledge. In particular, updating an intelligent diagnosis system with training data of new diseases would cause catastrophic forgetting of old disease knowledge. To address the catastrophic forgetting issue, a novel adapter-based strategy is proposed to help effectively learn a set of new diseases at each round (or task) of continual learning, without changing the shared feature extractor. The learnable lightweight task-specific adapter(s) can be flexibly designed (e.g., two convolutional layers) and then added to the pretrained and fixed feature extractor. Together with a specially designed task-specific head which absorbs all previously learned old diseases as a single 'out-of-distribution' category, task-specific adapter(s) can help the pretrained feature extractor more effectively extract discriminative features between diseases. In addition, a simple yet effective fine-tuning is applied to collaboratively fine-tune multiple task-specific heads such that outputs from different heads are comparable and consequently the appropriate classifier head can be more accurately selected during model inference. Extensive empirical evaluations on three image datasets demonstrate the superior performance of the proposed method in continual learning of new diseases. The source code will be released publicly.

* 10 pages 
Viaarxiv icon

PAMI: partition input and aggregate outputs for model interpretation

Feb 08, 2023
Wei Shi, Wentao Zhang, Weishi Zheng, Ruixuan Wang

Figure 1 for PAMI: partition input and aggregate outputs for model interpretation
Figure 2 for PAMI: partition input and aggregate outputs for model interpretation
Figure 3 for PAMI: partition input and aggregate outputs for model interpretation
Figure 4 for PAMI: partition input and aggregate outputs for model interpretation

There is an increasing demand for interpretation of model predictions especially in high-risk applications. Various visualization approaches have been proposed to estimate the part of input which is relevant to a specific model prediction. However, most approaches require model structure and parameter details in order to obtain the visualization results, and in general much effort is required to adapt each approach to multiple types of tasks particularly when model backbone and input format change over tasks. In this study, a simple yet effective visualization framework called PAMI is proposed based on the observation that deep learning models often aggregate features from local regions for model predictions. The basic idea is to mask majority of the input and use the corresponding model output as the relative contribution of the preserved input part to the original model prediction. For each input, since only a set of model outputs are collected and aggregated, PAMI does not require any model detail and can be applied to various prediction tasks with different model backbones and input formats. Extensive experiments on multiple tasks confirm the proposed method performs better than existing visualization approaches in more precisely finding class-specific input regions, and when applied to different model backbones and input formats. The source code will be released publicly.

* 19 pages 
Viaarxiv icon

Adaptively Integrated Knowledge Distillation and Prediction Uncertainty for Continual Learning

Jan 18, 2023
Kanghao Chen, Sijia Liu, Ruixuan Wang, Wei-Shi Zheng

Figure 1 for Adaptively Integrated Knowledge Distillation and Prediction Uncertainty for Continual Learning
Figure 2 for Adaptively Integrated Knowledge Distillation and Prediction Uncertainty for Continual Learning
Figure 3 for Adaptively Integrated Knowledge Distillation and Prediction Uncertainty for Continual Learning
Figure 4 for Adaptively Integrated Knowledge Distillation and Prediction Uncertainty for Continual Learning

Current deep learning models often suffer from catastrophic forgetting of old knowledge when continually learning new knowledge. Existing strategies to alleviate this issue often fix the trade-off between keeping old knowledge (stability) and learning new knowledge (plasticity). However, the stability-plasticity trade-off during continual learning may need to be dynamically changed for better model performance. In this paper, we propose two novel ways to adaptively balance model stability and plasticity. The first one is to adaptively integrate multiple levels of old knowledge and transfer it to each block level in the new model. The second one uses prediction uncertainty of old knowledge to naturally tune the importance of learning new knowledge during model training. To our best knowledge, this is the first time to connect model prediction uncertainty and knowledge distillation for continual learning. In addition, this paper applies a modified CutMix particularly to augment the data for old knowledge, further alleviating the catastrophic forgetting issue. Extensive evaluations on the CIFAR100 and the ImageNet datasets confirmed the effectiveness of the proposed method for continual learning.

Viaarxiv icon

Assessing and Analyzing the Resilience of Graph Neural Networks Against Hardware Faults

Dec 07, 2022
Xun Jiao, Ruixuan Wang, Fred Lin, Daniel Moore, Sriram Sankar

Figure 1 for Assessing and Analyzing the Resilience of Graph Neural Networks Against Hardware Faults
Figure 2 for Assessing and Analyzing the Resilience of Graph Neural Networks Against Hardware Faults
Figure 3 for Assessing and Analyzing the Resilience of Graph Neural Networks Against Hardware Faults
Figure 4 for Assessing and Analyzing the Resilience of Graph Neural Networks Against Hardware Faults

Graph neural networks (GNNs) have recently emerged as a promising learning paradigm in learning graph-structured data and have demonstrated wide success across various domains such as recommendation systems, social networks, and electronic design automation (EDA). Like other deep learning (DL) methods, GNNs are being deployed in sophisticated modern hardware systems, as well as dedicated accelerators. However, despite the popularity of GNNs and the recent efforts of bringing GNNs to hardware, the fault tolerance and resilience of GNNs has generally been overlooked. Inspired by the inherent algorithmic resilience of DL methods, this paper conducts, for the first time, a large-scale and empirical study of GNN resilience, aiming to understand the relationship between hardware faults and GNN accuracy. By developing a customized fault injection tool on top of PyTorch, we perform extensive fault injection experiments to various GNN models and application datasets. We observe that the error resilience of GNN models varies by orders of magnitude with respect to different models and application datasets. Further, we explore a low-cost error mitigation mechanism for GNN to enhance its resilience. This GNN resilience study aims to open up new directions and opportunities for future GNN accelerator design and architectural optimization.

Viaarxiv icon

Learning Discriminative Representation via Metric Learning for Imbalanced Medical Image Classification

Jul 14, 2022
Chenghua Zeng, Huijuan Lu, Kanghao Chen, Ruixuan Wang, Wei-Shi Zheng

Figure 1 for Learning Discriminative Representation via Metric Learning for Imbalanced Medical Image Classification
Figure 2 for Learning Discriminative Representation via Metric Learning for Imbalanced Medical Image Classification
Figure 3 for Learning Discriminative Representation via Metric Learning for Imbalanced Medical Image Classification
Figure 4 for Learning Discriminative Representation via Metric Learning for Imbalanced Medical Image Classification

Data imbalance between common and rare diseases during model training often causes intelligent diagnosis systems to have biased predictions towards common diseases. The state-of-the-art approaches apply a two-stage learning framework to alleviate the class-imbalance issue, where the first stage focuses on training of a general feature extractor and the second stage focuses on fine-tuning the classifier head for class rebalancing. However, existing two-stage approaches do not consider the fine-grained property between different diseases, often causing the first stage less effective for medical image classification than for natural image classification tasks. In this study, we propose embedding metric learning into the first stage of the two-stage framework specially to help the feature extractor learn to extract more discriminative feature representations. Extensive experiments mainly on three medical image datasets show that the proposed approach consistently outperforms existing onestage and two-stage approaches, suggesting that metric learning can be used as an effective plug-in component in the two-stage framework for fine-grained class-imbalanced image classification tasks.

Viaarxiv icon

PCCT: Progressive Class-Center Triplet Loss for Imbalanced Medical Image Classification

Jul 11, 2022
Kanghao Chen, Weixian Lei, Rong Zhang, Shen Zhao, Wei-shi Zheng, Ruixuan Wang

Figure 1 for PCCT: Progressive Class-Center Triplet Loss for Imbalanced Medical Image Classification
Figure 2 for PCCT: Progressive Class-Center Triplet Loss for Imbalanced Medical Image Classification
Figure 3 for PCCT: Progressive Class-Center Triplet Loss for Imbalanced Medical Image Classification
Figure 4 for PCCT: Progressive Class-Center Triplet Loss for Imbalanced Medical Image Classification

Imbalanced training data is a significant challenge for medical image classification. In this study, we propose a novel Progressive Class-Center Triplet (PCCT) framework to alleviate the class imbalance issue particularly for diagnosis of rare diseases, mainly by carefully designing the triplet sampling strategy and the triplet loss formation. Specifically, the PCCT framework includes two successive stages. In the first stage, PCCT trains the diagnosis system via a class-balanced triplet loss to coarsely separate distributions of different classes. In the second stage, the PCCT framework further improves the diagnosis system via a class-center involved triplet loss to cause a more compact distribution for each class. For the class-balanced triplet loss, triplets are sampled equally for each class at each training iteration, thus alleviating the imbalanced data issue. For the class-center involved triplet loss, the positive and negative samples in each triplet are replaced by their corresponding class centers, which enforces data representations of the same class closer to the class center. Furthermore, the class-center involved triplet loss is extended to the pair-wise ranking loss and the quadruplet loss, which demonstrates the generalization of the proposed framework. Extensive experiments support that the PCCT framework works effectively for medical image classification with imbalanced training images. On two skin image datasets and one chest X-ray dataset, the proposed approach respectively obtains the mean F1 score 86.2, 65.2, and 90.66 over all classes and 81.4, 63.87, and 81.92 for rare classes, achieving state-of-the-art performance and outperforming the widely used methods for the class imbalance issue.

Viaarxiv icon

Task-oriented Self-supervised Learning for Anomaly Detection in Electroencephalography

Jul 04, 2022
Yaojia Zheng, Zhouwu Liu, Rong Mo, Ziyi Chen, Wei-shi Zheng, Ruixuan Wang

Figure 1 for Task-oriented Self-supervised Learning for Anomaly Detection in Electroencephalography
Figure 2 for Task-oriented Self-supervised Learning for Anomaly Detection in Electroencephalography
Figure 3 for Task-oriented Self-supervised Learning for Anomaly Detection in Electroencephalography
Figure 4 for Task-oriented Self-supervised Learning for Anomaly Detection in Electroencephalography

Accurate automated analysis of electroencephalography (EEG) would largely help clinicians effectively monitor and diagnose patients with various brain diseases. Compared to supervised learning with labelled disease EEG data which can train a model to analyze specific diseases but would fail to monitor previously unseen statuses, anomaly detection based on only normal EEGs can detect any potential anomaly in new EEGs. Different from existing anomaly detection strategies which do not consider any property of unavailable abnormal data during model development, a task-oriented self-supervised learning approach is proposed here which makes use of available normal EEGs and expert knowledge about abnormal EEGs to train a more effective feature extractor for the subsequent development of anomaly detector. In addition, a specific two branch convolutional neural network with larger kernels is designed as the feature extractor such that it can more easily extract both larger scale and small-scale features which often appear in unavailable abnormal EEGs. The effectively designed and trained feature extractor has shown to be able to extract better feature representations from EEGs for development of anomaly detector based on normal data and future anomaly detection for new EEGs, as demonstrated on three EEG datasets. The code is available at https://github.com/ironing/EEG-AD.

Viaarxiv icon

Continual Learning with Bayesian Model based on a Fixed Pre-trained Feature Extractor

Apr 28, 2022
Yang Yang, Zhiying Cui, Junjie Xu, Changhong Zhong, Wei-Shi Zheng, Ruixuan Wang

Figure 1 for Continual Learning with Bayesian Model based on a Fixed Pre-trained Feature Extractor
Figure 2 for Continual Learning with Bayesian Model based on a Fixed Pre-trained Feature Extractor
Figure 3 for Continual Learning with Bayesian Model based on a Fixed Pre-trained Feature Extractor
Figure 4 for Continual Learning with Bayesian Model based on a Fixed Pre-trained Feature Extractor

Deep learning has shown its human-level performance in various applications. However, current deep learning models are characterised by catastrophic forgetting of old knowledge when learning new classes. This poses a challenge particularly in intelligent diagnosis systems where initially only training data of a limited number of diseases are available. In this case, updating the intelligent system with data of new diseases would inevitably downgrade its performance on previously learned diseases. Inspired by the process of learning new knowledge in human brains, we propose a Bayesian generative model for continual learning built on a fixed pre-trained feature extractor. In this model, knowledge of each old class can be compactly represented by a collection of statistical distributions, e.g. with Gaussian mixture models, and naturally kept from forgetting in continual learning over time. Unlike existing class-incremental learning methods, the proposed approach is not sensitive to the continual learning process and can be additionally well applied to the data-incremental learning scenario. Experiments on multiple medical and natural image classification tasks showed that the proposed approach outperforms state-of-the-art approaches which even keep some images of old classes during continual learning of new classes.

Viaarxiv icon

EnHDC: Ensemble Learning for Brain-Inspired Hyperdimensional Computing

Mar 25, 2022
Ruixuan Wang, Dongning Ma, Xun Jiao

Figure 1 for EnHDC: Ensemble Learning for Brain-Inspired Hyperdimensional Computing
Figure 2 for EnHDC: Ensemble Learning for Brain-Inspired Hyperdimensional Computing

Ensemble learning is a classical learning method utilizing a group of weak learners to form a strong learner, which aims to increase the accuracy of the model. Recently, brain-inspired hyperdimensional computing (HDC) becomes an emerging computational paradigm that has achieved success in various domains such as human activity recognition, voice recognition, and bio-medical signal classification. HDC mimics the brain cognition and leverages high-dimensional vectors (e.g., 10000 dimensions) with fully distributed holographic representation and (pseudo-)randomness. This paper presents the first effort in exploring ensemble learning in the context of HDC and proposes the first ensemble HDC model referred to as EnHDC. EnHDC uses a majority voting-based mechanism to synergistically integrate the prediction outcomes of multiple base HDC classifiers. To enhance the diversity of base classifiers, we vary the encoding mechanisms, dimensions, and data width settings among base classifiers. By applying EnHDC on a wide range of applications, results show that the EnHDC can achieve on average 3.2\% accuracy improvement over a single HDC classifier. Further, we show that EnHDC with reduced dimensionality, e.g., 1000 dimensions, can achieve similar or even surpass the accuracy of baseline HDC with higher dimensionality, e.g., 10000 dimensions. This leads to a 20\% reduction of storage requirement of HDC model, which is key to enabling HDC on low-power computing platforms.

Viaarxiv icon