Alert button
Picture for Sang Hyun Park

Sang Hyun Park

Alert button

Feature Re-calibration based MIL for Whole Slide Image Classification

Jun 22, 2022
Philip Chikontwe, Soo Jeong Nam, Heounjeong Go, Meejeong Kim, Hyun Jung Sung, Sang Hyun Park

Figure 1 for Feature Re-calibration based MIL for Whole Slide Image Classification
Figure 2 for Feature Re-calibration based MIL for Whole Slide Image Classification
Figure 3 for Feature Re-calibration based MIL for Whole Slide Image Classification
Figure 4 for Feature Re-calibration based MIL for Whole Slide Image Classification

Whole slide image (WSI) classification is a fundamental task for the diagnosis and treatment of diseases; but, curation of accurate labels is time-consuming and limits the application of fully-supervised methods. To address this, multiple instance learning (MIL) is a popular method that poses classification as a weakly supervised learning task with slide-level labels only. While current MIL methods apply variants of the attention mechanism to re-weight instance features with stronger models, scant attention is paid to the properties of the data distribution. In this work, we propose to re-calibrate the distribution of a WSI bag (instances) by using the statistics of the max-instance (critical) feature. We assume that in binary MIL, positive bags have larger feature magnitudes than negatives, thus we can enforce the model to maximize the discrepancy between bags with a metric feature loss that models positive bags as out-of-distribution. To achieve this, unlike existing MIL methods that use single-batch training modes, we propose balanced-batch sampling to effectively use the feature loss i.e., (+/-) bags simultaneously. Further, we employ a position encoding module (PEM) to model spatial/morphological information, and perform pooling by multi-head self-attention (PSMA) with a Transformer encoder. Experimental results on existing benchmark datasets show our approach is effective and improves over state-of-the-art MIL methods.

* MICCAI 2022 
Viaarxiv icon

CAD: Co-Adapting Discriminative Features for Improved Few-Shot Classification

Mar 25, 2022
Philip Chikontwe, Soopil Kim, Sang Hyun Park

Figure 1 for CAD: Co-Adapting Discriminative Features for Improved Few-Shot Classification
Figure 2 for CAD: Co-Adapting Discriminative Features for Improved Few-Shot Classification
Figure 3 for CAD: Co-Adapting Discriminative Features for Improved Few-Shot Classification
Figure 4 for CAD: Co-Adapting Discriminative Features for Improved Few-Shot Classification

Few-shot classification is a challenging problem that aims to learn a model that can adapt to unseen classes given a few labeled samples. Recent approaches pre-train a feature extractor, and then fine-tune for episodic meta-learning. Other methods leverage spatial features to learn pixel-level correspondence while jointly training a classifier. However, results using such approaches show marginal improvements. In this paper, inspired by the transformer style self-attention mechanism, we propose a strategy to cross-attend and re-weight discriminative features for few-shot classification. Given a base representation of support and query images after global pooling, we introduce a single shared module that projects features and cross-attends in two aspects: (i) query to support, and (ii) support to query. The module computes attention scores between features to produce an attention pooled representation of features in the same class that is later added to the original representation followed by a projection head. This effectively re-weights features in both aspects (i & ii) to produce features that better facilitate improved metric-based meta-learning. Extensive experiments on public benchmarks show our approach outperforms state-of-the-art methods by 3%~5%.

* CVPR2022 
Viaarxiv icon

Uncertainty-Aware Semi-Supervised Few Shot Segmentation

Oct 18, 2021
Soopil Kim, Philip Chikontwe, Sang Hyun Park

Figure 1 for Uncertainty-Aware Semi-Supervised Few Shot Segmentation
Figure 2 for Uncertainty-Aware Semi-Supervised Few Shot Segmentation
Figure 3 for Uncertainty-Aware Semi-Supervised Few Shot Segmentation
Figure 4 for Uncertainty-Aware Semi-Supervised Few Shot Segmentation

Few shot segmentation (FSS) aims to learn pixel-level classification of a target object in a query image using only a few annotated support samples. This is challenging as it requires modeling appearance variations of target objects and the diverse visual cues between query and support images with limited information. To address this problem, we propose a semi-supervised FSS strategy that leverages additional prototypes from unlabeled images with uncertainty guided pseudo label refinement. To obtain reliable prototypes from unlabeled images, we meta-train a neural network to jointly predict segmentation and estimate the uncertainty of predictions. We employ the uncertainty estimates to exclude predictions with high degrees of uncertainty for pseudo label construction to obtain additional prototypes based on the refined pseudo labels. During inference, query segmentation is predicted using prototypes from both support and unlabeled images including low-level features of the query images. Our approach is end-to-end and can easily supplement existing approaches without the requirement of additional training to employ unlabeled samples. Extensive experiments on PASCAL-$5^i$ and COCO-$20^i$ demonstrate that our model can effectively remove unreliable predictions to refine pseudo labels and significantly improve upon state-of-the-art performances.

* 9 pages 
Viaarxiv icon

Data Generation using Texture Co-occurrence and Spatial Self-Similarity for Debiasing

Oct 15, 2021
Myeongkyun Kang, Dongkyu Won, Miguel Luna, Kyung Soo Hong, June Hong Ahn, Sang Hyun Park

Figure 1 for Data Generation using Texture Co-occurrence and Spatial Self-Similarity for Debiasing
Figure 2 for Data Generation using Texture Co-occurrence and Spatial Self-Similarity for Debiasing
Figure 3 for Data Generation using Texture Co-occurrence and Spatial Self-Similarity for Debiasing
Figure 4 for Data Generation using Texture Co-occurrence and Spatial Self-Similarity for Debiasing

Classification models trained on biased datasets usually perform poorly on out-of-distribution samples since biased representations are embedded into the model. Recently, adversarial learning methods have been proposed to disentangle biased representations, but it is challenging to discard only the biased features without altering other relevant information. In this paper, we propose a novel de-biasing approach that explicitly generates additional images using texture representations of oppositely labeled images to enlarge the training dataset and mitigate the effect of biases when training a classifier. Every new generated image contains similar spatial information from a source image while transferring textures from a target image of opposite label. Our model integrates a texture co-occurrence loss that determines whether a generated image's texture is similar to that of the target, and a spatial self-similarity loss that determines whether the spatial details between the generated and source images are well preserved. Both generated and original training images are further used to train a classifier that is able to avoid learning unknown bias representations. We employ three distinct artificially designed datasets with known biases to demonstrate the ability of our method to mitigate bias information, and report competitive performance over existing state-of-the-art methods.

Viaarxiv icon

First Demonstration of the Korean eLoran Accuracy in a Narrow Waterway Using Improved ASF Maps

Sep 28, 2021
Woohyun Kim, Pyo-Woong Son, Sul Gee Park, Sang Hyun Park, Jiwon Seo

Figure 1 for First Demonstration of the Korean eLoran Accuracy in a Narrow Waterway Using Improved ASF Maps
Figure 2 for First Demonstration of the Korean eLoran Accuracy in a Narrow Waterway Using Improved ASF Maps
Figure 3 for First Demonstration of the Korean eLoran Accuracy in a Narrow Waterway Using Improved ASF Maps
Figure 4 for First Demonstration of the Korean eLoran Accuracy in a Narrow Waterway Using Improved ASF Maps

The vulnerabilities of global navigation satellite systems (GNSSs) to radio frequency jamming and spoofing have attracted significant research attention. In particular, the large-scale jamming incidents that occurred in South Korea substantiate the practical importance of implementing a complementary navigation system. This letter briefly summarizes the efforts of South Korea to deploy an enhanced long-range navigation (eLoran) system, which is a terrestrial low-frequency radio navigation system that can complement GNSSs. After four years of research and development, the Korean eLoran testbed system has been recently deployed and is operational since June 1, 2021. Although its initial performance at sea is satisfactory, navigation through a narrow waterway is still challenging because a complete survey of the additional secondary factor (ASF), which is the largest source of error for eLoran, is practically difficult in a narrow waterway. This letter proposes an alternative way to survey the ASF in a narrow waterway and improve the ASF map generation methods. Moreover, the performance of the proposed approach was validated experimentally.

* Submitted to IEEE Transactions on Aerospace and Electronic Systems 
Viaarxiv icon

First Demonstration of Korean eLoran Accuracy in a Narrow Waterway using Improved ASF Maps

Sep 18, 2021
Woohyun Kim, Pyo-Woong Son, Sul Gee Park, Sang Hyun Park, Jiwon Seo

Figure 1 for First Demonstration of Korean eLoran Accuracy in a Narrow Waterway using Improved ASF Maps
Figure 2 for First Demonstration of Korean eLoran Accuracy in a Narrow Waterway using Improved ASF Maps
Figure 3 for First Demonstration of Korean eLoran Accuracy in a Narrow Waterway using Improved ASF Maps
Figure 4 for First Demonstration of Korean eLoran Accuracy in a Narrow Waterway using Improved ASF Maps

The vulnerabilities of global navigation satellite systems (GNSSs) to radio frequency jamming and spoofing have attracted significant research attention. In particular, the large-scale jamming incidents that occurred in South Korea substantiate the practical importance of implementing a complementary navigation system. This letter briefly summarizes the efforts of South Korea to deploy an enhanced long-range navigation (eLoran) system, which is a terrestrial low-frequency radio navigation system that can complement GNSSs. After four years of research and development, the Korean eLoran testbed system has been recently deployed and is operational since June 1, 2021. Although its initial performance at sea is satisfactory, navigation through a narrow waterway is still challenging because a complete survey of the additional secondary factor (ASF), which is the largest source of error for eLoran, is practically difficult in a narrow waterway. This letter proposes an alternative way to survey the ASF in a narrow waterway and improve the ASF map generation methods. Moreover, the performance of the proposed approach was validated experimentally.

* Submitted to IEEE Transactions on Aerospace and Electronic Systems 
Viaarxiv icon

A Meta-Learning Approach for Medical Image Registration

Apr 21, 2021
Heejung Park, Gyeong Min Lee, Soopil Kim, Ga Hyung Ryu, Areum Jeong, Sang Hyun Park, Min Sagong

Figure 1 for A Meta-Learning Approach for Medical Image Registration
Figure 2 for A Meta-Learning Approach for Medical Image Registration
Figure 3 for A Meta-Learning Approach for Medical Image Registration
Figure 4 for A Meta-Learning Approach for Medical Image Registration

Non-rigid registration is a necessary but challenging task in medical imaging studies. Recently, unsupervised registration models have shown good performance, but they often require a large-scale training dataset and long training times. Therefore, in real world application where only dozens to hundreds of image pairs are available, existing models cannot be practically used. To address these limitations, we propose a novel unsupervised registration model which is integrated with a gradient-based meta learning framework. In particular, we train a meta learner which finds an initialization point of parameters by utilizing a variety of existing registration datasets. To quickly adapt to various tasks, the meta learner was updated to get close to the center of parameters which are fine-tuned for each registration task. Thereby, our model can adapt to unseen domain tasks via a short fine-tuning process and perform accurate registration. To verify the superiority of our model, we train the model for various 2D medical image registration tasks such as retinal choroid Optical Coherence Tomography Angiography (OCTA), CT organs, and brain MRI scans and test on registration of retinal OCTA Superficial Capillary Plexus (SCP). In our experiments, the proposed model obtained significantly improved performance in terms of accuracy and training time compared to other registration models.

Viaarxiv icon

Self-Supervised Learning based CT Denoising using Pseudo-CT Image Pairs

Apr 06, 2021
Dongkyu Won, Euijin Jung, Sion An, Philip Chikontwe, Sang Hyun Park

Figure 1 for Self-Supervised Learning based CT Denoising using Pseudo-CT Image Pairs
Figure 2 for Self-Supervised Learning based CT Denoising using Pseudo-CT Image Pairs
Figure 3 for Self-Supervised Learning based CT Denoising using Pseudo-CT Image Pairs
Figure 4 for Self-Supervised Learning based CT Denoising using Pseudo-CT Image Pairs

Recently, Self-supervised learning methods able to perform image denoising without ground truth labels have been proposed. These methods create low-quality images by adding random or Gaussian noise to images and then train a model for denoising. Ideally, it would be beneficial if one can generate high-quality CT images with only a few training samples via self-supervision. However, the performance of CT denoising is generally limited due to the complexity of CT noise. To address this problem, we propose a novel self-supervised learning-based CT denoising method. In particular, we train pre-train CT denoising and noise models that can predict CT noise from Low-dose CT (LDCT) using available LDCT and Normal-dose CT (NDCT) pairs. For a given test LDCT, we generate Pseudo-LDCT and NDCT pairs using the pre-trained denoising and noise models and then update the parameters of the denoising model using these pairs to remove noise in the test LDCT. To make realistic Pseudo LDCT, we train multiple noise models from individual images and generate the noise using the ensemble of noise models. We evaluate our method on the 2016 AAPM Low-Dose CT Grand Challenge dataset. The proposed ensemble noise model can generate realistic CT noise, and thus our method significantly improves the denoising performance existing denoising models trained by supervised- and self-supervised learning.

Viaarxiv icon

Mixing-AdaSIN: Constructing a de-biased dataset using Adaptive Structural Instance Normalization and texture Mixing

Mar 26, 2021
Myeongkyun Kang, Philip Chikontwe, Miguel Luna, Kyung Soo Hong, June Hong Ahn, Sang Hyun Park

Figure 1 for Mixing-AdaSIN: Constructing a de-biased dataset using Adaptive Structural Instance Normalization and texture Mixing
Figure 2 for Mixing-AdaSIN: Constructing a de-biased dataset using Adaptive Structural Instance Normalization and texture Mixing
Figure 3 for Mixing-AdaSIN: Constructing a de-biased dataset using Adaptive Structural Instance Normalization and texture Mixing
Figure 4 for Mixing-AdaSIN: Constructing a de-biased dataset using Adaptive Structural Instance Normalization and texture Mixing

Following the pandemic outbreak, several works have proposed to diagnose COVID-19 with deep learning in computed tomography (CT); reporting performance on-par with experts. However, models trained/tested on the same in-distribution data may rely on the inherent data biases for successful prediction, failing to generalize on out-of-distribution samples or CT with different scanning protocols. Early attempts have partly addressed bias-mitigation and generalization through augmentation or re-sampling, but are still limited by collection costs and the difficulty of quantifying bias in medical images. In this work, we propose Mixing-AdaSIN; a bias mitigation method that uses a generative model to generate de-biased images by mixing texture information between different labeled CT scans with semantically similar features. Here, we use Adaptive Structural Instance Normalization (AdaSIN) to enhance de-biasing generation quality and guarantee structural consistency. Following, a classifier trained with the generated images learns to correctly predict the label without bias and generalizes better. To demonstrate the efficacy of our method, we construct a biased COVID-19 vs. bacterial pneumonia dataset based on CT protocols and compare with existing state-of-the-art de-biasing methods. Our experiments show that classifiers trained with de-biased generated images report improved in-distribution performance and generalization on an external COVID-19 dataset.

Viaarxiv icon

Bidirectional RNN-based Few Shot Learning for 3D Medical Image Segmentation

Nov 19, 2020
Soopil Kim, Sion An, Philip Chikontwe, Sang Hyun Park

Figure 1 for Bidirectional RNN-based Few Shot Learning for 3D Medical Image Segmentation
Figure 2 for Bidirectional RNN-based Few Shot Learning for 3D Medical Image Segmentation
Figure 3 for Bidirectional RNN-based Few Shot Learning for 3D Medical Image Segmentation
Figure 4 for Bidirectional RNN-based Few Shot Learning for 3D Medical Image Segmentation

Segmentation of organs of interest in 3D medical images is necessary for accurate diagnosis and longitudinal studies. Though recent advances using deep learning have shown success for many segmentation tasks, large datasets are required for high performance and the annotation process is both time consuming and labor intensive. In this paper, we propose a 3D few shot segmentation framework for accurate organ segmentation using limited training samples of the target organ annotation. To achieve this, a U-Net like network is designed to predict segmentation by learning the relationship between 2D slices of support data and a query image, including a bidirectional gated recurrent unit (GRU) that learns consistency of encoded features between adjacent slices. Also, we introduce a transfer learning method to adapt the characteristics of the target image and organ by updating the model before testing with arbitrary support and query data sampled from the support data. We evaluate our proposed model using three 3D CT datasets with annotations of different organs. Our model yielded significantly improved performance over state-of-the-art few shot segmentation models and was comparable to a fully supervised model trained with more target training data.

* Submitted to AAAI21 
Viaarxiv icon