Alert button
Picture for Rajarsi Gupta

Rajarsi Gupta

Alert button

Open and reusable deep learning for pathology with WSInfer and QuPath

Sep 08, 2023
Jakub R. Kaczmarzyk, Alan O'Callaghan, Fiona Inglis, Tahsin Kurc, Rajarsi Gupta, Erich Bremer, Peter Bankhead, Joel H. Saltz

Figure 1 for Open and reusable deep learning for pathology with WSInfer and QuPath

The field of digital pathology has seen a proliferation of deep learning models in recent years. Despite substantial progress, it remains rare for other researchers and pathologists to be able to access models published in the literature and apply them to their own images. This is due to difficulties in both sharing and running models. To address these concerns, we introduce WSInfer: a new, open-source software ecosystem designed to make deep learning for pathology more streamlined and accessible. WSInfer comprises three main elements: 1) a Python package and command line tool to efficiently apply patch-based deep learning inference to whole slide images; 2) a QuPath extension that provides an alternative inference engine through user-friendly and interactive software, and 3) a model zoo, which enables pathology models and metadata to be easily shared in a standardized form. Together, these contributions aim to encourage wider reuse, exploration, and interrogation of deep learning models for research purposes, by putting them into the hands of pathologists and eliminating a need for coding experience when accessed through QuPath. The WSInfer source code is hosted on GitHub and documentation is available at https://wsinfer.readthedocs.io.

Viaarxiv icon

Halcyon -- A Pathology Imaging and Feature analysis and Management System

Apr 07, 2023
Erich Bremer, Tammy DiPrima, Joseph Balsamo, Jonas Almeida, Rajarsi Gupta, Joel Saltz

Figure 1 for Halcyon -- A Pathology Imaging and Feature analysis and Management System
Figure 2 for Halcyon -- A Pathology Imaging and Feature analysis and Management System
Figure 3 for Halcyon -- A Pathology Imaging and Feature analysis and Management System
Figure 4 for Halcyon -- A Pathology Imaging and Feature analysis and Management System

Halcyon is a new pathology imaging analysis and feature management system based on W3C linked-data open standards and is designed to scale to support the needs for the voluminous production of features from deep-learning feature pipelines. Halcyon can support multiple users with a web-based UX with access to all user data over a standards-based web API allowing for integration with other processes and software systems. Identity management and data security is also provided.

* 15 pages, 11 figures. arXiv admin note: text overlap with arXiv:2005.06469 
Viaarxiv icon

Topology-Guided Multi-Class Cell Context Generation for Digital Pathology

Apr 05, 2023
Shahira Abousamra, Rajarsi Gupta, Tahsin Kurc, Dimitris Samaras, Joel Saltz, Chao Chen

Figure 1 for Topology-Guided Multi-Class Cell Context Generation for Digital Pathology
Figure 2 for Topology-Guided Multi-Class Cell Context Generation for Digital Pathology
Figure 3 for Topology-Guided Multi-Class Cell Context Generation for Digital Pathology
Figure 4 for Topology-Guided Multi-Class Cell Context Generation for Digital Pathology

In digital pathology, the spatial context of cells is important for cell classification, cancer diagnosis and prognosis. To model such complex cell context, however, is challenging. Cells form different mixtures, lineages, clusters and holes. To model such structural patterns in a learnable fashion, we introduce several mathematical tools from spatial statistics and topological data analysis. We incorporate such structural descriptors into a deep generative model as both conditional inputs and a differentiable loss. This way, we are able to generate high quality multi-class cell layouts for the first time. We show that the topology-rich cell layouts can be used for data augmentation and improve the performance of downstream tasks such as cell classification.

* To be published in proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2023 
Viaarxiv icon

ViT-DAE: Transformer-driven Diffusion Autoencoder for Histopathology Image Analysis

Apr 03, 2023
Xuan Xu, Saarthak Kapse, Rajarsi Gupta, Prateek Prasanna

Figure 1 for ViT-DAE: Transformer-driven Diffusion Autoencoder for Histopathology Image Analysis
Figure 2 for ViT-DAE: Transformer-driven Diffusion Autoencoder for Histopathology Image Analysis
Figure 3 for ViT-DAE: Transformer-driven Diffusion Autoencoder for Histopathology Image Analysis
Figure 4 for ViT-DAE: Transformer-driven Diffusion Autoencoder for Histopathology Image Analysis

Generative AI has received substantial attention in recent years due to its ability to synthesize data that closely resembles the original data source. While Generative Adversarial Networks (GANs) have provided innovative approaches for histopathological image analysis, they suffer from limitations such as mode collapse and overfitting in discriminator. Recently, Denoising Diffusion models have demonstrated promising results in computer vision. These models exhibit superior stability during training, better distribution coverage, and produce high-quality diverse images. Additionally, they display a high degree of resilience to noise and perturbations, making them well-suited for use in digital pathology, where images commonly contain artifacts and exhibit significant variations in staining. In this paper, we present a novel approach, namely ViT-DAE, which integrates vision transformers (ViT) and diffusion autoencoders for high-quality histopathology image synthesis. This marks the first time that ViT has been introduced to diffusion autoencoders in computational pathology, allowing the model to better capture the complex and intricate details of histopathology images. We demonstrate the effectiveness of ViT-DAE on three publicly available datasets. Our approach outperforms recent GAN-based and vanilla DAE methods in generating realistic images.

* Submitted to MICCAI 2023 
Viaarxiv icon

Gigapixel Whole-Slide Images Classification using Locally Supervised Learning

Jul 17, 2022
Jingwei Zhang, Xin Zhang, Ke Ma, Rajarsi Gupta, Joel Saltz, Maria Vakalopoulou, Dimitris Samaras

Figure 1 for Gigapixel Whole-Slide Images Classification using Locally Supervised Learning
Figure 2 for Gigapixel Whole-Slide Images Classification using Locally Supervised Learning
Figure 3 for Gigapixel Whole-Slide Images Classification using Locally Supervised Learning
Figure 4 for Gigapixel Whole-Slide Images Classification using Locally Supervised Learning

Histopathology whole slide images (WSIs) play a very important role in clinical studies and serve as the gold standard for many cancer diagnoses. However, generating automatic tools for processing WSIs is challenging due to their enormous sizes. Currently, to deal with this issue, conventional methods rely on a multiple instance learning (MIL) strategy to process a WSI at patch level. Although effective, such methods are computationally expensive, because tiling a WSI into patches takes time and does not explore the spatial relations between these tiles. To tackle these limitations, we propose a locally supervised learning framework which processes the entire slide by exploring the entire local and global information that it contains. This framework divides a pre-trained network into several modules and optimizes each module locally using an auxiliary model. We also introduce a random feature reconstruction unit (RFR) to preserve distinguishing features during training and improve the performance of our method by 1% to 3%. Extensive experiments on three publicly available WSI datasets: TCGA-NSCLC, TCGA-RCC and LKS, highlight the superiority of our method on different classification tasks. Our method outperforms the state-of-the-art MIL methods by 2% to 5% in accuracy, while being 7 to 10 times faster. Additionally, when dividing it into eight modules, our method requires as little as 20% of the total gpu memory required by end-to-end training. Our code is available at https://github.com/cvlab-stonybrook/local_learning_wsi.

* Accepted to MICCAI 2022 
Viaarxiv icon

AI and Pathology: Steering Treatment and Predicting Outcomes

Jun 15, 2022
Rajarsi Gupta, Jakub Kaczmarzyk, Soma Kobayashi, Tahsin Kurc, Joel Saltz

Figure 1 for AI and Pathology: Steering Treatment and Predicting Outcomes
Figure 2 for AI and Pathology: Steering Treatment and Predicting Outcomes
Figure 3 for AI and Pathology: Steering Treatment and Predicting Outcomes

The combination of data analysis methods, increasing computing capacity, and improved sensors enable quantitative granular, multi-scale, cell-based analyses. We describe the rich set of application challenges related to tissue interpretation and survey AI methods currently used to address these challenges. We focus on a particular class of targeted human tissue analysis - histopathology - aimed at quantitative characterization of disease state, patient outcome prediction and treatment steering.

Viaarxiv icon

Evaluating histopathology transfer learning with ChampKit

Jun 14, 2022
Jakub R. Kaczmarzyk, Tahsin M. Kurc, Shahira Abousamra, Rajarsi Gupta, Joel H. Saltz, Peter K. Koo

Figure 1 for Evaluating histopathology transfer learning with ChampKit
Figure 2 for Evaluating histopathology transfer learning with ChampKit
Figure 3 for Evaluating histopathology transfer learning with ChampKit
Figure 4 for Evaluating histopathology transfer learning with ChampKit

Histopathology remains the gold standard for diagnosis of various cancers. Recent advances in computer vision, specifically deep learning, have facilitated the analysis of histopathology images for various tasks, including immune cell detection and microsatellite instability classification. The state-of-the-art for each task often employs base architectures that have been pretrained for image classification on ImageNet. The standard approach to develop classifiers in histopathology tends to focus narrowly on optimizing models for a single task, not considering the aspects of modeling innovations that improve generalization across tasks. Here we present ChampKit (Comprehensive Histopathology Assessment of Model Predictions toolKit): an extensible, fully reproducible benchmarking toolkit that consists of a broad collection of patch-level image classification tasks across different cancers. ChampKit enables a way to systematically document the performance impact of proposed improvements in models and methodology. ChampKit source code and data are freely accessible at https://github.com/kaczmarj/champkit .

* Submitted to NeurIPS 2022 Track on Datasets and Benchmarks. Source code available at https://github.com/kaczmarj/champkit 
Viaarxiv icon

Federated Learning for the Classification of Tumor Infiltrating Lymphocytes

Apr 01, 2022
Ujjwal Baid, Sarthak Pati, Tahsin M. Kurc, Rajarsi Gupta, Erich Bremer, Shahira Abousamra, Siddhesh P. Thakur, Joel H. Saltz, Spyridon Bakas

Figure 1 for Federated Learning for the Classification of Tumor Infiltrating Lymphocytes
Figure 2 for Federated Learning for the Classification of Tumor Infiltrating Lymphocytes
Figure 3 for Federated Learning for the Classification of Tumor Infiltrating Lymphocytes
Figure 4 for Federated Learning for the Classification of Tumor Infiltrating Lymphocytes

We evaluate the performance of federated learning (FL) in developing deep learning models for analysis of digitized tissue sections. A classification application was considered as the example use case, on quantifiying the distribution of tumor infiltrating lymphocytes within whole slide images (WSIs). A deep learning classification model was trained using 50*50 square micron patches extracted from the WSIs. We simulated a FL environment in which a dataset, generated from WSIs of cancer from numerous anatomical sites available by The Cancer Genome Atlas repository, is partitioned in 8 different nodes. Our results show that the model trained with the federated training approach achieves similar performance, both quantitatively and qualitatively, to that of a model trained with all the training data pooled at a centralized location. Our study shows that FL has tremendous potential for enabling development of more robust and accurate models for histopathology image analysis without having to collect large and diverse training data at a single location.

Viaarxiv icon

Visual attention analysis of pathologists examining whole slide images of Prostate cancer

Feb 17, 2022
Souradeep Chakraborty, Ke Ma, Rajarsi Gupta, Beatrice Knudsen, Gregory J. Zelinsky, Joel H. Saltz, Dimitris Samaras

Figure 1 for Visual attention analysis of pathologists examining whole slide images of Prostate cancer
Figure 2 for Visual attention analysis of pathologists examining whole slide images of Prostate cancer
Figure 3 for Visual attention analysis of pathologists examining whole slide images of Prostate cancer
Figure 4 for Visual attention analysis of pathologists examining whole slide images of Prostate cancer

We study the attention of pathologists as they examine whole-slide images (WSIs) of prostate cancer tissue using a digital microscope. To the best of our knowledge, our study is the first to report in detail how pathologists navigate WSIs of prostate cancer as they accumulate information for their diagnoses. We collected slide navigation data (i.e., viewport location, magnification level, and time) from 13 pathologists in 2 groups (5 genitourinary (GU) specialists and 8 general pathologists) and generated visual attention heatmaps and scanpaths. Each pathologist examined five WSIs from the TCGA PRAD dataset, which were selected by a GU pathology specialist. We examined and analyzed the distributions of visual attention for each group of pathologists after each WSI was examined. To quantify the relationship between a pathologist's attention and evidence for cancer in the WSI, we obtained tumor annotations from a genitourinary specialist. We used these annotations to compute the overlap between the distribution of visual attention and annotated tumor region to identify strong correlations. Motivated by this analysis, we trained a deep learning model to predict visual attention on unseen WSIs. We find that the attention heatmaps predicted by our model correlate quite well with the ground truth attention heatmap and tumor annotations on a test set of 17 WSIs by using various spatial and temporal evaluation metrics.

* Accepted for oral presentation at ISBI 2022 
Viaarxiv icon

Multi-Class Cell Detection Using Spatial Context Representation

Oct 10, 2021
Shahira Abousamra, David Belinsky, John Van Arnam, Felicia Allard, Eric Yee, Rajarsi Gupta, Tahsin Kurc, Dimitris Samaras, Joel Saltz, Chao Chen

Figure 1 for Multi-Class Cell Detection Using Spatial Context Representation
Figure 2 for Multi-Class Cell Detection Using Spatial Context Representation
Figure 3 for Multi-Class Cell Detection Using Spatial Context Representation
Figure 4 for Multi-Class Cell Detection Using Spatial Context Representation

In digital pathology, both detection and classification of cells are important for automatic diagnostic and prognostic tasks. Classifying cells into subtypes, such as tumor cells, lymphocytes or stromal cells is particularly challenging. Existing methods focus on morphological appearance of individual cells, whereas in practice pathologists often infer cell classes through their spatial context. In this paper, we propose a novel method for both detection and classification that explicitly incorporates spatial contextual information. We use the spatial statistical function to describe local density in both a multi-class and a multi-scale manner. Through representation learning and deep clustering techniques, we learn advanced cell representation with both appearance and spatial context. On various benchmarks, our method achieves better performance than state-of-the-arts, especially on the classification task.

Viaarxiv icon