Alert button
Picture for Chenyu Wang

Chenyu Wang

Alert button

Improving Multiple Sclerosis Lesion Segmentation Across Clinical Sites: A Federated Learning Approach with Noise-Resilient Training

Aug 31, 2023
Lei Bai, Dongang Wang, Michael Barnett, Mariano Cabezas, Weidong Cai, Fernando Calamante, Kain Kyle, Dongnan Liu, Linda Ly, Aria Nguyen, Chun-Chien Shieh, Ryan Sullivan, Hengrui Wang, Geng Zhan, Wanli Ouyang, Chenyu Wang

Accurately measuring the evolution of Multiple Sclerosis (MS) with magnetic resonance imaging (MRI) critically informs understanding of disease progression and helps to direct therapeutic strategy. Deep learning models have shown promise for automatically segmenting MS lesions, but the scarcity of accurately annotated data hinders progress in this area. Obtaining sufficient data from a single clinical site is challenging and does not address the heterogeneous need for model robustness. Conversely, the collection of data from multiple sites introduces data privacy concerns and potential label noise due to varying annotation standards. To address this dilemma, we explore the use of the federated learning framework while considering label noise. Our approach enables collaboration among multiple clinical sites without compromising data privacy under a federated learning paradigm that incorporates a noise-robust training strategy based on label correction. Specifically, we introduce a Decoupled Hard Label Correction (DHLC) strategy that considers the imbalanced distribution and fuzzy boundaries of MS lesions, enabling the correction of false annotations based on prediction confidence. We also introduce a Centrally Enhanced Label Correction (CELC) strategy, which leverages the aggregated central model as a correction teacher for all sites, enhancing the reliability of the correction process. Extensive experiments conducted on two multi-site datasets demonstrate the effectiveness and robustness of our proposed methods, indicating their potential for clinical applications in multi-site collaborations.

* 11 pages, 4 figures, journal submission 
Viaarxiv icon

A hybrid Decoder-DeepONet operator regression framework for unaligned observation data

Aug 18, 2023
Bo Chen, Chenyu Wang, Weipeng Li, Haiyang Fu

Deep neural operators (DNOs) have been utilized to approximate nonlinear mappings between function spaces. However, DNOs face the challenge of increased dimensionality and computational cost associated with unaligned observation data. In this study, we propose a hybrid Decoder-DeepONet operator regression framework to handle unaligned data effectively. Additionally, we introduce a Multi-Decoder-DeepONet, which utilizes an average field of training data as input augmentation. The consistencies of the frameworks with the operator approximation theory are provided, on the basis of the universal approximation theorem. Two numerical experiments, Darcy problem and flow-field around an airfoil, are conducted to validate the efficiency and accuracy of the proposed methods. Results illustrate the advantages of Decoder-DeepONet and Multi-Decoder-DeepONet in handling unaligned observation data and showcase their potentials in improving prediction accuracy.

* 35 pages, 10 figures, 11 tables 
Viaarxiv icon

Precise Few-shot Fat-free Thigh Muscle Segmentation in T1-weighted MRI

Apr 27, 2023
Sheng Chen, Zihao Tang, Dongnan Liu, Ché Fornusek, Michael Barnett, Chenyu Wang, Mariano Cabezas, Weidong Cai

Figure 1 for Precise Few-shot Fat-free Thigh Muscle Segmentation in T1-weighted MRI
Figure 2 for Precise Few-shot Fat-free Thigh Muscle Segmentation in T1-weighted MRI
Figure 3 for Precise Few-shot Fat-free Thigh Muscle Segmentation in T1-weighted MRI
Figure 4 for Precise Few-shot Fat-free Thigh Muscle Segmentation in T1-weighted MRI

Precise thigh muscle volumes are crucial to monitor the motor functionality of patients with diseases that may result in various degrees of thigh muscle loss. T1-weighted MRI is the default surrogate to obtain thigh muscle masks due to its contrast between muscle and fat signals. Deep learning approaches have recently been widely used to obtain these masks through segmentation. However, due to the insufficient amount of precise annotations, thigh muscle masks generated by deep learning approaches tend to misclassify intra-muscular fat (IMF) as muscle impacting the analysis of muscle volumetrics. As IMF is infiltrated inside the muscle, human annotations require expertise and time. Thus, precise muscle masks where IMF is excluded are limited in practice. To alleviate this, we propose a few-shot segmentation framework to generate thigh muscle masks excluding IMF. In our framework, we design a novel pseudo-label correction and evaluation scheme, together with a new noise robust loss for exploiting high certainty areas. The proposed framework only takes $1\%$ of the fine-annotated training dataset, and achieves comparable performance with fully supervised methods according to the experimental results.

* ISBI2023, Few-shot, Intra-muscular fat, Thigh muscle segmentation, Pseudo-label denoising, MRI 
Viaarxiv icon

VECtor: A Versatile Event-Centric Benchmark for Multi-Sensor SLAM

Jul 04, 2022
Ling Gao, Yuxuan Liang, Jiaqi Yang, Shaoxun Wu, Chenyu Wang, Jiaben Chen, Laurent Kneip

Figure 1 for VECtor: A Versatile Event-Centric Benchmark for Multi-Sensor SLAM
Figure 2 for VECtor: A Versatile Event-Centric Benchmark for Multi-Sensor SLAM
Figure 3 for VECtor: A Versatile Event-Centric Benchmark for Multi-Sensor SLAM
Figure 4 for VECtor: A Versatile Event-Centric Benchmark for Multi-Sensor SLAM

Event cameras have recently gained in popularity as they hold strong potential to complement regular cameras in situations of high dynamics or challenging illumination. An important problem that may benefit from the addition of an event camera is given by Simultaneous Localization And Mapping (SLAM). However, in order to ensure progress on event-inclusive multi-sensor SLAM, novel benchmark sequences are needed. Our contribution is the first complete set of benchmark datasets captured with a multi-sensor setup containing an event-based stereo camera, a regular stereo camera, multiple depth sensors, and an inertial measurement unit. The setup is fully hardware-synchronized and underwent accurate extrinsic calibration. All sequences come with ground truth data captured by highly accurate external reference devices such as a motion capture system. Individual sequences include both small and large-scale environments, and cover the specific challenges targeted by dynamic vision sensors.

* IEEE Robotics and Automation Letters, 2022  
Viaarxiv icon

Accurate Instance-Level CAD Model Retrieval in a Large-Scale Database

Jul 04, 2022
Jiaxin Wei, Lan Hu, Chenyu Wang, Laurent Kneip

Figure 1 for Accurate Instance-Level CAD Model Retrieval in a Large-Scale Database
Figure 2 for Accurate Instance-Level CAD Model Retrieval in a Large-Scale Database
Figure 3 for Accurate Instance-Level CAD Model Retrieval in a Large-Scale Database
Figure 4 for Accurate Instance-Level CAD Model Retrieval in a Large-Scale Database

We present a new solution to the fine-grained retrieval of clean CAD models from a large-scale database in order to recover detailed object shape geometries for RGBD scans. Unlike previous work simply indexing into a moderately small database using an object shape descriptor and accepting the top retrieval result, we argue that in the case of a large-scale database a more accurate model may be found within a neighborhood of the descriptor. More importantly, we propose that the distinctiveness deficiency of shape descriptors at the instance level can be compensated by a geometry-based re-ranking of its neighborhood. Our approach first leverages the discriminative power of learned representations to distinguish between different categories of models and then uses a novel robust point set distance metric to re-rank the CAD neighborhood, enabling fine-grained retrieval in a large shape database. Evaluation on a real-world dataset shows that our geometry-based re-ranking is a conceptually simple but highly effective method that can lead to a significant improvement in retrieval accuracy compared to the state-of-the-art.

* Accepted by IROS 2022 
Viaarxiv icon

MS Lesion Segmentation: Revisiting Weighting Mechanisms for Federated Learning

May 03, 2022
Dongnan Liu, Mariano Cabezas, Dongang Wang, Zihao Tang, Lei Bai, Geng Zhan, Yuling Luo, Kain Kyle, Linda Ly, James Yu, Chun-Chien Shieh, Aria Nguyen, Ettikan Kandasamy Karuppiah, Ryan Sullivan, Fernando Calamante, Michael Barnett, Wanli Ouyang, Weidong Cai, Chenyu Wang

Figure 1 for MS Lesion Segmentation: Revisiting Weighting Mechanisms for Federated Learning
Figure 2 for MS Lesion Segmentation: Revisiting Weighting Mechanisms for Federated Learning
Figure 3 for MS Lesion Segmentation: Revisiting Weighting Mechanisms for Federated Learning
Figure 4 for MS Lesion Segmentation: Revisiting Weighting Mechanisms for Federated Learning

Federated learning (FL) has been widely employed for medical image analysis to facilitate multi-client collaborative learning without sharing raw data. Despite great success, FL's performance is limited for multiple sclerosis (MS) lesion segmentation tasks, due to variance in lesion characteristics imparted by different scanners and acquisition parameters. In this work, we propose the first FL MS lesion segmentation framework via two effective re-weighting mechanisms. Specifically, a learnable weight is assigned to each local node during the aggregation process, based on its segmentation performance. In addition, the segmentation loss function in each client is also re-weighted according to the lesion volume for the data during training. Comparison experiments on two FL MS segmentation scenarios using public and clinical datasets have demonstrated the effectiveness of the proposed method by outperforming other FL methods significantly. Furthermore, the segmentation performance of FL incorporating our proposed aggregation mechanism can exceed centralised training with all the raw data. The extensive evaluation also indicated the superiority of our method when estimating brain volume differences estimation after lesion inpainting.

* 10 pages, 3 figures, and 7 tables 
Viaarxiv icon

Precise Indoor Positioning Based on UWB and Deep Learning

Apr 17, 2022
Chenyu Wang, Zihuai Lin

Figure 1 for Precise Indoor Positioning Based on UWB and Deep Learning
Figure 2 for Precise Indoor Positioning Based on UWB and Deep Learning
Figure 3 for Precise Indoor Positioning Based on UWB and Deep Learning
Figure 4 for Precise Indoor Positioning Based on UWB and Deep Learning

We examined UWB-based indoor location in conjunction with a fingerprint technique in this work. We built a connection between the measured and real distances for the UWB indoor positioning system. This connection is used to produce a distance database that may be used to generate fringerprints. We created a BP neural network to classify the target node to the relevant fringerpint using the distance database. Our suggested deep learning technology considerably enhances location accuracy when compared to existing trilateration systems.

Viaarxiv icon

DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation

Mar 30, 2022
Yu Tang, Chenyu Wang, Yufan Zhang, Yuliang Liu, Xingcheng Zhang, Linbo Qiao, Zhiquan Lai, Dongsheng Li

Figure 1 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 2 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 3 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation
Figure 4 for DELTA: Dynamically Optimizing GPU Memory beyond Tensor Recomputation

The further development of deep neural networks is hampered by the limited GPU memory resource. Therefore, the optimization of GPU memory resources is highly demanded. Swapping and recomputation are commonly applied to make better use of GPU memory in deep learning. However, as an emerging domain, several challenges remain:1)The efficiency of recomputation is limited for both static and dynamic methods. 2)Swapping requires offloading parameters manually, which incurs a great time cost. 3) There is no such dynamic and fine-grained method that involves tensor swapping together with tensor recomputation nowadays. To remedy the above issues, we propose a novel scheduler manager named DELTA(Dynamic tEnsor offLoad and recompuTAtion). To the best of our knowledge, we are the first to make a reasonable dynamic runtime scheduler on the combination of tensor swapping and tensor recomputation without user oversight. In DELTA, we propose a filter algorithm to select the optimal tensors to be released out of GPU memory and present a director algorithm to select a proper action for each of these tensors. Furthermore, prefetching and overlapping are deliberately considered to overcome the time cost caused by swapping and recomputing tensors. Experimental results show that DELTA not only saves 40%-70% of GPU memory, surpassing the state-of-the-art method to a great extent but also gets comparable convergence results as the baseline with acceptable time delay. Also, DELTA gains 2.04$\times$ maximum batchsize when training ResNet-50 and 2.25$\times$ when training ResNet-101 compared with the baseline. Besides, comparisons between the swapping cost and recomputation cost in our experiments demonstrate the importance of making a reasonable dynamic scheduler on tensor swapping and tensor recomputation, which refutes the arguments in some related work that swapping should be the first and best choice.

* 12 pages, 6 figures 
Viaarxiv icon

Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement

Jan 06, 2022
Dongnan Liu, Chaoyi Zhang, Yang Song, Heng Huang, Chenyu Wang, Michael Barnett, Weidong Cai

Figure 1 for Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement
Figure 2 for Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement
Figure 3 for Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement
Figure 4 for Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement

Recent advances in unsupervised domain adaptation (UDA) techniques have witnessed great success in cross-domain computer vision tasks, enhancing the generalization ability of data-driven deep learning architectures by bridging the domain distribution gaps. For the UDA-based cross-domain object detection methods, the majority of them alleviate the domain bias by inducing the domain-invariant feature generation via adversarial learning strategy. However, their domain discriminators have limited classification ability due to the unstable adversarial training process. Therefore, the extracted features induced by them cannot be perfectly domain-invariant and still contain domain-private factors, bringing obstacles to further alleviate the cross-domain discrepancy. To tackle this issue, we design a Domain Disentanglement Faster-RCNN (DDF) to eliminate the source-specific information in the features for detection task learning. Our DDF method facilitates the feature disentanglement at the global and local stages, with a Global Triplet Disentanglement (GTD) module and an Instance Similarity Disentanglement (ISD) module, respectively. By outperforming state-of-the-art methods on four benchmark UDA object detection tasks, our DDF method is demonstrated to be effective with wide applicability.

* Accepted to appear in IEEE Transactions on Multimedia; source code: https://github.com/dliu5812/DDF 
Viaarxiv icon

HAGEN: Homophily-Aware Graph Convolutional Recurrent Network for Crime Forecasting

Sep 27, 2021
Chenyu Wang, Zongyu Lin, Xiaochen Yang, Jiao Sun, Mingxuan Yue, Cyrus Shahabi

Figure 1 for HAGEN: Homophily-Aware Graph Convolutional Recurrent Network for Crime Forecasting
Figure 2 for HAGEN: Homophily-Aware Graph Convolutional Recurrent Network for Crime Forecasting
Figure 3 for HAGEN: Homophily-Aware Graph Convolutional Recurrent Network for Crime Forecasting
Figure 4 for HAGEN: Homophily-Aware Graph Convolutional Recurrent Network for Crime Forecasting

The crime forecasting is an important problem as it greatly contributes to urban safety. Typically, the goal of the problem is to predict different types of crimes for each geographical region (like a neighborhood or censor tract) in the near future. Since nearby regions usually have similar socioeconomic characteristics which indicate similar crime patterns, recent state-of-the-art solutions constructed a distance-based region graph and utilized Graph Neural Network (GNN) techniques for crime forecasting, because the GNN techniques could effectively exploit the latent relationships between neighboring region nodes in the graph. However, this distance-based pre-defined graph cannot fully capture crime correlation between regions that are far from each other but share similar crime patterns. Hence, to make an accurate crime prediction, the main challenge is to learn a better graph that reveals the dependencies between regions in crime occurrences and meanwhile captures the temporal patterns from historical crime records. To address these challenges, we propose an end-to-end graph convolutional recurrent network called HAGEN with several novel designs for crime prediction. Specifically, our framework could jointly capture the crime correlation between regions and the temporal crime dynamics by combining an adaptive region graph learning module with the Diffusion Convolution Gated Recurrent Unit (DCGRU). Based on the homophily assumption of GNN, we propose a homophily-aware constraint to regularize the optimization of the region graph so that neighboring region nodes on the learned graph share similar crime patterns, thus fitting the mechanism of diffusion convolution. It also incorporates crime embedding to model the interdependencies between regions and crime categories. Empirical experiments and comprehensive analysis on two real-world datasets showcase the effectiveness of HAGEN.

Viaarxiv icon