Alert button
Picture for Supratik Mukhopadhyay

Supratik Mukhopadhyay

Alert button

Collisionless Pattern Discovery in Robot Swarms Using Deep Reinforcement Learning

Sep 20, 2022
Nelson Sharma, Aswini Ghosh, Rajiv Misra, Supratik Mukhopadhyay, Gokarna Sharma

Figure 1 for Collisionless Pattern Discovery in Robot Swarms Using Deep Reinforcement Learning
Figure 2 for Collisionless Pattern Discovery in Robot Swarms Using Deep Reinforcement Learning
Figure 3 for Collisionless Pattern Discovery in Robot Swarms Using Deep Reinforcement Learning
Figure 4 for Collisionless Pattern Discovery in Robot Swarms Using Deep Reinforcement Learning

We present a deep reinforcement learning-based framework for automatically discovering patterns available in any given initial configuration of fat robot swarms. In particular, we model the problem of collision-less gathering and mutual visibility in fat robot swarms and discover patterns for solving them using our framework. We show that by shaping reward signals based on certain constraints like mutual visibility and safe proximity, the robots can discover collision-less trajectories leading to well-formed gathering and visibility patterns.

Viaarxiv icon

XOOD: Extreme Value Based Out-Of-Distribution Detection For Image Classification

Aug 01, 2022
Frej Berglind, Haron Temam, Supratik Mukhopadhyay, Kamalika Das, Md Saiful Islam Sajol, Sricharan Kumar, Kumar Kallurupalli

Figure 1 for XOOD: Extreme Value Based Out-Of-Distribution Detection For Image Classification
Figure 2 for XOOD: Extreme Value Based Out-Of-Distribution Detection For Image Classification
Figure 3 for XOOD: Extreme Value Based Out-Of-Distribution Detection For Image Classification
Figure 4 for XOOD: Extreme Value Based Out-Of-Distribution Detection For Image Classification

Detecting out-of-distribution (OOD) data at inference time is crucial for many applications of machine learning. We present XOOD: a novel extreme value-based OOD detection framework for image classification that consists of two algorithms. The first, XOOD-M, is completely unsupervised, while the second XOOD-L is self-supervised. Both algorithms rely on the signals captured by the extreme values of the data in the activation layers of the neural network in order to distinguish between in-distribution and OOD instances. We show experimentally that both XOOD-M and XOOD-L outperform state-of-the-art OOD detection methods on many benchmark data sets in both efficiency and accuracy, reducing false-positive rate (FPR95) by 50%, while improving the inferencing time by an order of magnitude.

Viaarxiv icon

Numerical Evaluation of a muon tomography system for imaging defects in concrete structures

Feb 17, 2021
Sridhar Tripathy, Jaydeep Datta, Nayana Majumdar, Supratik Mukhopadhyay

Figure 1 for Numerical Evaluation of a muon tomography system for imaging defects in concrete structures
Figure 2 for Numerical Evaluation of a muon tomography system for imaging defects in concrete structures
Figure 3 for Numerical Evaluation of a muon tomography system for imaging defects in concrete structures
Figure 4 for Numerical Evaluation of a muon tomography system for imaging defects in concrete structures

Among its numerous applications, deployment in civil structures has caught attraction of many recently. The detection of defects based on inherent physical quantities such as density and atomic number by probing naturally available cosmic muons makes MST a novel idea suitable for inexpensive and non-destructive imaging. In this work, capability of MST to detect concrete defects has been tested and evaluated in terms of two-dimensional imaging and statistical calculations. The imaging has been done on unique and critical defects causing degradation in civil structures. The capability and limitation of MST in this avenue have also been studied.

* First Version 
Viaarxiv icon

A One-Shot Learning Framework for Assessment of Fibrillar Collagen from Second Harmonic Generation Images of an Infarcted Myocardium

Jan 30, 2020
Qun Liu, Supratik Mukhopadhyay, Maria Ximena Bastidas Rodriguez, Xing Fu, Sushant Sahu, David Burk, Manas Gartia

Figure 1 for A One-Shot Learning Framework for Assessment of Fibrillar Collagen from Second Harmonic Generation Images of an Infarcted Myocardium
Figure 2 for A One-Shot Learning Framework for Assessment of Fibrillar Collagen from Second Harmonic Generation Images of an Infarcted Myocardium
Figure 3 for A One-Shot Learning Framework for Assessment of Fibrillar Collagen from Second Harmonic Generation Images of an Infarcted Myocardium
Figure 4 for A One-Shot Learning Framework for Assessment of Fibrillar Collagen from Second Harmonic Generation Images of an Infarcted Myocardium

Myocardial infarction (MI) is a scientific term that refers to heart attack. In this study, we infer highly relevant second harmonic generation (SHG) cues from collagen fibers exhibiting highly non-centrosymmetric assembly together with two-photon excited cellular autofluorescence in infarcted mouse heart to quantitatively probe fibrosis, especially targeted at an early stage after MI. We present a robust one-shot machine learning algorithm that enables determination of 2D assembly of collagen with high spatial resolution along with its structural arrangement in heart tissues post-MI with spectral specificity and sensitivity. Detection, evaluation, and precise quantification of fibrosis extent at early stage would guide one to develop treatment therapies that may prevent further progression and determine heart transplant needs for patient survival.

* Paper was accepted at the IEEE International Symposium on Biomedical Imaging (ISBI 2020) 
Viaarxiv icon

Context-Aware Design of Cyber-Physical Human Systems (CPHS)

Jan 07, 2020
Supratik Mukhopadhyay, Qun Liu, Edward Collier, Yimin Zhu, Ravindra Gudishala, Chanachok Chokwitthaya, Robert DiBiano, Alimire Nabijiang, Sanaz Saeidi, Subhajit Sidhanta, Arnab Ganguly

Figure 1 for Context-Aware Design of Cyber-Physical Human Systems (CPHS)
Figure 2 for Context-Aware Design of Cyber-Physical Human Systems (CPHS)
Figure 3 for Context-Aware Design of Cyber-Physical Human Systems (CPHS)

Recently, it has been widely accepted by the research community that interactions between humans and cyber-physical infrastructures have played a significant role in determining the performance of the latter. The existing paradigm for designing cyber-physical systems for optimal performance focuses on developing models based on historical data. The impacts of context factors driving human system interaction are challenging and are difficult to capture and replicate in existing design models. As a result, many existing models do not or only partially address those context factors of a new design owing to the lack of capabilities to capture the context factors. This limitation in many existing models often causes performance gaps between predicted and measured results. We envision a new design environment, a cyber-physical human system (CPHS) where decision-making processes for physical infrastructures under design are intelligently connected to distributed resources over cyberinfrastructure such as experiments on design features and empirical evidence from operations of existing instances. The framework combines existing design models with context-aware design-specific data involving human-infrastructure interactions in new designs, using a machine learning approach to create augmented design models with improved predictive powers.

* Paper was accepted at the 12th International Conference on Communication Systems and Networks (COMSNETS 2020) 
Viaarxiv icon

DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification

Nov 15, 2019
Qun Liu, Saikat Basu, Sangram Ganguly, Supratik Mukhopadhyay, Robert DiBiano, Manohar Karki, Ramakrishna Nemani

Figure 1 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification
Figure 2 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification
Figure 3 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification
Figure 4 for DeepSat V2: Feature Augmented Convolutional Neural Nets for Satellite Image Classification

Satellite image classification is a challenging problem that lies at the crossroads of remote sensing, computer vision, and machine learning. Due to the high variability inherent in satellite data, most of the current object classification approaches are not suitable for handling satellite datasets. The progress of satellite image analytics has also been inhibited by the lack of a single labeled high-resolution dataset with multiple class labels. In a preliminary version of this work, we introduced two new high resolution satellite imagery datasets (SAT-4 and SAT-6) and proposed DeepSat framework for classification based on "handcrafted" features and a deep belief network (DBN). The present paper is an extended version, we present an end-to-end framework leveraging an improved architecture that augments a convolutional neural network (CNN) with handcrafted features (instead of using DBN-based architecture) for classification. Our framework, having access to fused spatial information obtained from handcrafted features as well as CNN feature maps, have achieved accuracies of 99.90% and 99.84% respectively, on SAT-4 and SAT-6, surpassing all the other state-of-the-art results. A statistical analysis based on Distribution Separability Criterion substantiates the robustness of our approach in learning better representations for satellite imagery.

* This is an Accepted Manuscript of an article published by Taylor & Francis Group in Remote Sensing Letters. arXiv admin note: text overlap with arXiv:1509.03602 
Viaarxiv icon

PCGAN-CHAR: Progressively Trained Classifier Generative Adversarial Networks for Classification of Noisy Handwritten Bangla Characters

Aug 11, 2019
Qun Liu, Edward Collier, Supratik Mukhopadhyay

Figure 1 for PCGAN-CHAR: Progressively Trained Classifier Generative Adversarial Networks for Classification of Noisy Handwritten Bangla Characters
Figure 2 for PCGAN-CHAR: Progressively Trained Classifier Generative Adversarial Networks for Classification of Noisy Handwritten Bangla Characters
Figure 3 for PCGAN-CHAR: Progressively Trained Classifier Generative Adversarial Networks for Classification of Noisy Handwritten Bangla Characters
Figure 4 for PCGAN-CHAR: Progressively Trained Classifier Generative Adversarial Networks for Classification of Noisy Handwritten Bangla Characters

Due to the sparsity of features, noise has proven to be a great inhibitor in the classification of handwritten characters. To combat this, most techniques perform denoising of the data before classification. In this paper, we consolidate the approach by training an all-in-one model that is able to classify even noisy characters. For classification, we progressively train a classifier generative adversarial network on the characters from low to high resolution. We show that by learning the features at each resolution independently a trained model is able to accurately classify characters even in the presence of noise. We experimentally demonstrate the effectiveness of our approach by classifying noisy versions of MNIST, handwritten Bangla Numeral, and Basic Character datasets.

* Paper was accepted at the 21st International Conference on Asia-Pacific Digital Libraries (ICADL 2019) 
Viaarxiv icon

Improving Prediction Accuracy in Building Performance Models Using Generative Adversarial Networks (GANs)

Jun 14, 2019
Chanachok Chokwitthaya, Edward Collier, Yimin Zhu, Supratik Mukhopadhyay

Figure 1 for Improving Prediction Accuracy in Building Performance Models Using Generative Adversarial Networks (GANs)
Figure 2 for Improving Prediction Accuracy in Building Performance Models Using Generative Adversarial Networks (GANs)
Figure 3 for Improving Prediction Accuracy in Building Performance Models Using Generative Adversarial Networks (GANs)
Figure 4 for Improving Prediction Accuracy in Building Performance Models Using Generative Adversarial Networks (GANs)

Building performance discrepancies between building design and operation are one of the causes that lead many new designs fail to achieve their goals and objectives. One of main factors contributing to the discrepancy is occupant behaviors. Occupants responding to a new design are influenced by several factors. Existing building performance models (BPMs) ignore or partially address those factors (called contextual factors) while developing BPMs. To potentially reduce the discrepancies and improve the prediction accuracy of BPMs, this paper proposes a computational framework for learning mixture models by using Generative Adversarial Networks (GANs) that appropriately combining existing BPMs with knowledge on occupant behaviors to contextual factors in new designs. Immersive virtual environments (IVEs) experiments are used to acquire data on such behaviors. Performance targets are used to guide appropriate combination of existing BPMs with knowledge on occupant behaviors. The resulting model obtained is called an augmented BPM. Two different experiments related to occupant lighting behaviors are shown as case study. The results reveal that augmented BPMs significantly outperformed existing BPMs with respect to achieving specified performance targets. The case study confirms the potential of the computational framework for improving prediction accuracy of BPMs during design.

* 9 pages, 4 figures, The 2019 International Joint Conference on Neural Networks (IJCNN) 
Viaarxiv icon

Improving Route Choice Models by Incorporating Contextual Factors via Knowledge Distillation

Mar 27, 2019
Qun Liu, Supratik Mukhopadhyay, Yimin Zhu, Ravindra Gudishala, Sanaz Saeidi, Alimire Nabijiang

Figure 1 for Improving Route Choice Models by Incorporating Contextual Factors via Knowledge Distillation
Figure 2 for Improving Route Choice Models by Incorporating Contextual Factors via Knowledge Distillation
Figure 3 for Improving Route Choice Models by Incorporating Contextual Factors via Knowledge Distillation
Figure 4 for Improving Route Choice Models by Incorporating Contextual Factors via Knowledge Distillation

Route Choice Models predict the route choices of travelers traversing an urban area. Most of the route choice models link route characteristics of alternative routes to those chosen by the drivers. The models play an important role in prediction of traffic levels on different routes and thus assist in development of efficient traffic management strategies that result in minimizing traffic delay and maximizing effective utilization of transport system. High fidelity route choice models are required to predict traffic levels with higher accuracy. Existing route choice models do not take into account dynamic contextual conditions such as the occurrence of an accident, the socio-cultural and economic background of drivers, other human behaviors, the dynamic personal risk level, etc. As a result, they can only make predictions at an aggregate level and for a fixed set of contextual factors. For higher fidelity, it is highly desirable to use a model that captures significance of subjective or contextual factors in route choice. This paper presents a novel approach for developing high-fidelity route choice models with increased predictive power by augmenting existing aggregate level baseline models with information on drivers' responses to contextual factors obtained from Stated Choice Experiments carried out in an Immersive Virtual Environment through the use of knowledge distillation.

* Paper was accepted at the 2019 International Joint Conference on Neural Networks (IJCNN 2019) 
Viaarxiv icon