Alert button
Picture for Manohar Karki

Manohar Karki

Alert button

Lesion Conditional Image Generation for Improved Segmentation of Intracranial Hemorrhage from CT Images

Mar 30, 2020
Manohar Karki, Junghwan Cho

Figure 1 for Lesion Conditional Image Generation for Improved Segmentation of Intracranial Hemorrhage from CT Images
Figure 2 for Lesion Conditional Image Generation for Improved Segmentation of Intracranial Hemorrhage from CT Images
Figure 3 for Lesion Conditional Image Generation for Improved Segmentation of Intracranial Hemorrhage from CT Images
Figure 4 for Lesion Conditional Image Generation for Improved Segmentation of Intracranial Hemorrhage from CT Images

Data augmentation can effectively resolve a scarcity of images when training machine-learning algorithms. It can make them more robust to unseen images. We present a lesion conditional Generative Adversarial Network LcGAN to generate synthetic Computed Tomography (CT) images for data augmentation. A lesion conditional image (segmented mask) is an input to both the generator and the discriminator of the LcGAN during training. The trained model generates contextual CT images based on input masks. We quantify the quality of the images by using a fully convolutional network (FCN) score and blurriness. We also train another classification network to select better synthetic images. These synthetic CT images are then augmented to our hemorrhagic lesion segmentation network. By applying this augmentation method on 2.5%, 10% and 25% of original data, segmentation improved by 12.8%, 6% and 1.6% respectively.

Viaarxiv icon

DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification

Nov 15, 2019
Qun Liu, Saikat Basu, Sangram Ganguly, Supratik Mukhopadhyay, Robert DiBiano, Manohar Karki, Ramakrishna Nemani

Figure 1 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification
Figure 2 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification
Figure 3 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification
Figure 4 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification

Satellite image classification is a challenging problem that lies at the crossroads of remote sensing, computer vision, and machine learning. Due to the high variability inherent in satellite data, most of the current object classification approaches are not suitable for handling satellite datasets. The progress of satellite image analytics has also been inhibited by the lack of a single labeled high-resolution dataset with multiple class labels. In a preliminary version of this work, we introduced two new high resolution satellite imagery datasets (SAT-4 and SAT-6) and proposed DeepSat framework for classification based on "handcrafted" features and a deep belief network (DBN). The present paper is an extended version, we present an end-to-end framework leveraging an improved architecture that augments a convolutional neural network (CNN) with handcrafted features (instead of using DBN-based architecture) for classification. Our framework, having access to fused spatial information obtained from handcrafted features as well as CNN feature maps, have achieved accuracies of 99.90% and 99.84% respectively, on SAT-4 and SAT-6, surpassing all the other state-of-the-art results. A statistical analysis based on Distribution Separability Criterion substantiates the robustness of our approach in learning better representations for satellite imagery.

* This is an Accepted Manuscript of an article published by Taylor & Francis Group in Remote Sensing Letters. arXiv admin note: text overlap with arXiv:1509.03602 
Viaarxiv icon

Pixel-level Reconstruction and Classification for Noisy Handwritten Bangla Characters

Jun 21, 2018
Manohar Karki, Qun Liu, Robert DiBiano, Saikat Basu, Supratik Mukhopadhyay

Figure 1 for Pixel-level Reconstruction and Classification for Noisy Handwritten Bangla Characters
Figure 2 for Pixel-level Reconstruction and Classification for Noisy Handwritten Bangla Characters
Figure 3 for Pixel-level Reconstruction and Classification for Noisy Handwritten Bangla Characters
Figure 4 for Pixel-level Reconstruction and Classification for Noisy Handwritten Bangla Characters

Classification techniques for images of handwritten characters are susceptible to noise. Quadtrees can be an efficient representation for learning from sparse features. In this paper, we improve the effectiveness of probabilistic quadtrees by using a pixel level classifier to extract the character pixels and remove noise from handwritten character images. The pixel level denoiser (a deep belief network) uses the map responses obtained from a pretrained CNN as features for reconstructing the characters eliminating noise. We experimentally demonstrate the effectiveness of our approach by reconstructing and classifying a noisy version of handwritten Bangla Numeral and Basic Character datasets.

* Paper was accepted at the 16th International Conference on Frontiers in Handwriting Recognition (ICFHR 2018) 
Viaarxiv icon

Core Sampling Framework for Pixel Classification

Dec 06, 2016
Manohar Karki, Robert DiBiano, Saikat Basu, Supratik Mukhopadhyay

Figure 1 for Core Sampling Framework for Pixel Classification
Figure 2 for Core Sampling Framework for Pixel Classification
Figure 3 for Core Sampling Framework for Pixel Classification
Figure 4 for Core Sampling Framework for Pixel Classification

The intermediate map responses of a Convolutional Neural Network (CNN) contain information about an image that can be used to extract contextual knowledge about it. In this paper, we present a core sampling framework that is able to use these activation maps from several layers as features to another neural network using transfer learning to provide an understanding of an input image. Our framework creates a representation that combines features from the test data and the contextual knowledge gained from the responses of a pretrained network, processes it and feeds it to a separate Deep Belief Network. We use this representation to extract more information from an image at the pixel level, hence gaining understanding of the whole image. We experimentally demonstrate the usefulness of our framework using a pretrained VGG-16 model to perform segmentation on the BAERI dataset of Synthetic Aperture Radar(SAR) imagery and the CAMVID dataset.

Viaarxiv icon

A Theoretical Analysis of Deep Neural Networks for Texture Classification

Jun 21, 2016
Saikat Basu, Manohar Karki, Robert DiBiano, Supratik Mukhopadhyay, Sangram Ganguly, Ramakrishna Nemani, Shreekant Gayaka

Figure 1 for A Theoretical Analysis of Deep Neural Networks for Texture Classification
Figure 2 for A Theoretical Analysis of Deep Neural Networks for Texture Classification
Figure 3 for A Theoretical Analysis of Deep Neural Networks for Texture Classification
Figure 4 for A Theoretical Analysis of Deep Neural Networks for Texture Classification

We investigate the use of Deep Neural Networks for the classification of image datasets where texture features are important for generating class-conditional discriminative representations. To this end, we first derive the size of the feature space for some standard textural features extracted from the input dataset and then use the theory of Vapnik-Chervonenkis dimension to show that hand-crafted feature extraction creates low-dimensional representations which help in reducing the overall excess error rate. As a corollary to this analysis, we derive for the first time upper bounds on the VC dimension of Convolutional Neural Network as well as Dropout and Dropconnect networks and the relation between excess error rate of Dropout and Dropconnect networks. The concept of intrinsic dimension is used to validate the intuition that texture-based datasets are inherently higher dimensional as compared to handwritten digits or other object recognition datasets and hence more difficult to be shattered by neural networks. We then derive the mean distance from the centroid to the nearest and farthest sampling points in an n-dimensional manifold and show that the Relative Contrast of the sample data vanishes as dimensionality of the underlying vector space tends to infinity.

* Accepted in International Joint Conference on Neural Networks, IJCNN 2016 
Viaarxiv icon

DeepSat - A Learning framework for Satellite Imagery

Sep 11, 2015
Saikat Basu, Sangram Ganguly, Supratik Mukhopadhyay, Robert DiBiano, Manohar Karki, Ramakrishna Nemani

Figure 1 for DeepSat - A Learning framework for Satellite Imagery
Figure 2 for DeepSat - A Learning framework for Satellite Imagery
Figure 3 for DeepSat - A Learning framework for Satellite Imagery
Figure 4 for DeepSat - A Learning framework for Satellite Imagery

Satellite image classification is a challenging problem that lies at the crossroads of remote sensing, computer vision, and machine learning. Due to the high variability inherent in satellite data, most of the current object classification approaches are not suitable for handling satellite datasets. The progress of satellite image analytics has also been inhibited by the lack of a single labeled high-resolution dataset with multiple class labels. The contributions of this paper are twofold - (1) first, we present two new satellite datasets called SAT-4 and SAT-6, and (2) then, we propose a classification framework that extracts features from an input image, normalizes them and feeds the normalized feature vectors to a Deep Belief Network for classification. On the SAT-4 dataset, our best network produces a classification accuracy of 97.95% and outperforms three state-of-the-art object recognition algorithms, namely - Deep Belief Networks, Convolutional Neural Networks and Stacked Denoising Autoencoders by ~11%. On SAT-6, it produces a classification accuracy of 93.9% and outperforms the other algorithms by ~15%. Comparative studies with a Random Forest classifier show the advantage of an unsupervised learning approach over traditional supervised learning techniques. A statistical analysis based on Distribution Separability Criterion and Intrinsic Dimensionality Estimation substantiates the effectiveness of our approach in learning better representations for satellite imagery.

* Paper was accepted at ACM SIGSPATIAL 2015 
Viaarxiv icon

Learning Sparse Feature Representations using Probabilistic Quadtrees and Deep Belief Nets

Sep 11, 2015
Saikat Basu, Manohar Karki, Sangram Ganguly, Robert DiBiano, Supratik Mukhopadhyay, Ramakrishna Nemani

Figure 1 for Learning Sparse Feature Representations using Probabilistic Quadtrees and Deep Belief Nets
Figure 2 for Learning Sparse Feature Representations using Probabilistic Quadtrees and Deep Belief Nets
Figure 3 for Learning Sparse Feature Representations using Probabilistic Quadtrees and Deep Belief Nets

Learning sparse feature representations is a useful instrument for solving an unsupervised learning problem. In this paper, we present three labeled handwritten digit datasets, collectively called n-MNIST. Then, we propose a novel framework for the classification of handwritten digits that learns sparse representations using probabilistic quadtrees and Deep Belief Nets. On the MNIST and n-MNIST datasets, our framework shows promising results and significantly outperforms traditional Deep Belief Networks.

* Published in the European Symposium on Artificial Neural Networks, ESANN 2015 
Viaarxiv icon