Alert button
Picture for Sameer Ambekar

Sameer Ambekar

Alert button

Learning Variational Neighbor Labels for Test-Time Domain Generalization

Jul 08, 2023
Sameer Ambekar, Zehao Xiao, Jiayi Shen, Xiantong Zhen, Cees G. M. Snoek

Figure 1 for Learning Variational Neighbor Labels for Test-Time Domain Generalization
Figure 2 for Learning Variational Neighbor Labels for Test-Time Domain Generalization
Figure 3 for Learning Variational Neighbor Labels for Test-Time Domain Generalization
Figure 4 for Learning Variational Neighbor Labels for Test-Time Domain Generalization

This paper strives for domain generalization, where models are trained exclusively on source domains before being deployed at unseen target domains. We follow the strict separation of source training and target testing but exploit the value of the unlabeled target data itself during inference. We make three contributions. First, we propose probabilistic pseudo-labeling of target samples to generalize the source-trained model to the target domain at test time. We formulate the generalization at test time as a variational inference problem by modeling pseudo labels as distributions to consider the uncertainty during generalization and alleviate the misleading signal of inaccurate pseudo labels. Second, we learn variational neighbor labels that incorporate the information of neighboring target samples to generate more robust pseudo labels. Third, to learn the ability to incorporate more representative target information and generate more precise and robust variational neighbor labels, we introduce a meta-generalization stage during training to simulate the generalization procedure. Experiments on six widely-used datasets demonstrate the benefits, abilities, and effectiveness of our proposal.

* Under review 
Viaarxiv icon

SKDCGN: Source-free Knowledge Distillation of Counterfactual Generative Networks using cGANs

Aug 10, 2022
Sameer Ambekar, Ankit Ankit, Diego van der Mast, Mark Alence, Matteo Tafuro, Christos Athanasiadis

Figure 1 for SKDCGN: Source-free Knowledge Distillation of Counterfactual Generative Networks using cGANs
Figure 2 for SKDCGN: Source-free Knowledge Distillation of Counterfactual Generative Networks using cGANs
Figure 3 for SKDCGN: Source-free Knowledge Distillation of Counterfactual Generative Networks using cGANs
Figure 4 for SKDCGN: Source-free Knowledge Distillation of Counterfactual Generative Networks using cGANs

With the usage of appropriate inductive biases, Counterfactual Generative Networks (CGNs) can generate novel images from random combinations of shape, texture, and background manifolds. These images can be utilized to train an invariant classifier, avoiding the wide spread problem of deep architectures learning spurious correlations rather than meaningful ones. As a consequence, out-of-domain robustness is improved. However, the CGN architecture comprises multiple over parameterized networks, namely BigGAN and U2-Net. Training these networks requires appropriate background knowledge and extensive computation. Since one does not always have access to the precise training details, nor do they always possess the necessary knowledge of counterfactuals, our work addresses the following question: Can we use the knowledge embedded in pre-trained CGNs to train a lower-capacity model, assuming black-box access (i.e., only access to the pretrained CGN model) to the components of the architecture? In this direction, we propose a novel work named SKDCGN that attempts knowledge transfer using Knowledge Distillation (KD). In our proposed architecture, each independent mechanism (shape, texture, background) is represented by a student 'TinyGAN' that learns from the pretrained teacher 'BigGAN'. We demonstrate the efficacy of the proposed method using state-of-the-art datasets such as ImageNet, and MNIST by using KD and appropriate loss functions. Moreover, as an additional contribution, our paper conducts a thorough study on the composition mechanism of the CGNs, to gain a better understanding of how each mechanism influences the classification accuracy of an invariant classifier. Code available at: https://github.com/ambekarsameer96/SKDCGN

* Accepted at ECCV 2022 Workshop VIPriors 
Viaarxiv icon

Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images

Feb 16, 2021
Kartikeya Badola, Sameer Ambekar, Himanshu Pant, Sumit Soman, Anuradha Sural, Rajiv Narang, Suresh Chandra, Jayadeva

Figure 1 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images
Figure 2 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images
Figure 3 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images
Figure 4 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images

The gold standard for COVID-19 is RT-PCR, testing facilities for which are limited and not always optimally distributed. Test results are delayed, which impacts treatment. Expert radiologists, one of whom is a co-author, are able to diagnose COVID-19 positivity from Chest X-Rays (CXR) and CT scans, that can facilitate timely treatment. Such diagnosis is particularly valuable in locations lacking radiologists with sufficient expertise and familiarity with COVID-19 patients. This paper has two contributions. One, we analyse literature on CXR based COVID-19 diagnosis. We show that popular choices of dataset selection suffer from data homogeneity, leading to misleading results. We compile and analyse a viable benchmark dataset from multiple existing heterogeneous sources. Such a benchmark is important for realistically testing models. Our second contribution relates to learning from imbalanced data. Datasets for COVID X-Ray classification face severe class imbalance, since most subjects are COVID -ve. Twin Support Vector Machines (Twin SVM) and Twin Neural Networks (Twin NN) have, in recent years, emerged as effective ways of handling skewed data. We introduce a state-of-the-art technique, termed as Twin Augmentation, for modifying popular pre-trained deep learning models. Twin Augmentation boosts the performance of a pre-trained deep neural network without requiring re-training. Experiments show, that across a multitude of classifiers, Twin Augmentation is very effective in boosting the performance of given pre-trained model for classification in imbalanced settings.

Viaarxiv icon

Unsupervised Domain Adaptation for Semantic Segmentation of NIR Images through Generative Latent Search

Jul 17, 2020
Prashant Pandey, Aayush Kumar Tyagi, Sameer Ambekar, Prathosh AP

Figure 1 for Unsupervised Domain Adaptation for Semantic Segmentation of NIR Images through Generative Latent Search
Figure 2 for Unsupervised Domain Adaptation for Semantic Segmentation of NIR Images through Generative Latent Search
Figure 3 for Unsupervised Domain Adaptation for Semantic Segmentation of NIR Images through Generative Latent Search
Figure 4 for Unsupervised Domain Adaptation for Semantic Segmentation of NIR Images through Generative Latent Search

Segmentation of the pixels corresponding to human skin is an essential first step in multiple applications ranging from surveillance to heart-rate estimation from remote-photoplethysmography. However, the existing literature considers the problem only in the visible-range of the EM-spectrum which limits their utility in low or no light settings where the criticality of the application is higher. To alleviate this problem, we consider the problem of skin segmentation from the Near-infrared images. However, Deep learning based state-of-the-art segmentation techniques demands large amounts of labelled data that is unavailable for the current problem. Therefore we cast the skin segmentation problem as that of target-independent Unsupervised Domain Adaptation (UDA) where we use the data from the Red-channel of the visible-range to develop skin segmentation algorithm on NIR images. We propose a method for target-independent segmentation where the 'nearest-clone' of a target image in the source domain is searched and used as a proxy in the segmentation network trained only on the source domain. We prove the existence of 'nearest-clone' and propose a method to find it through an optimization algorithm over the latent space of a Deep generative model based on variational inference. We demonstrate the efficacy of the proposed method for NIR skin segmentation over the state-of-the-art UDA segmentation methods on the two newly created skin segmentation datasets in NIR domain despite not having access to the target NIR data. Additionally, we report state-of-the-art results for adaption from Synthia to Cityscapes which is a popular setting in Unsupervised Domain Adaptation for semantic segmentation. The code and datasets are available at https://github.com/ambekarsameer96/GLSS.

* ECCV 2020 [Spotlight] 
Viaarxiv icon

Skin Segmentation from NIR Images using Unsupervised Domain Adaptation through Generative Latent Search

Jun 15, 2020
Prashant Pandey, Aayush Kumar Tyagi, Sameer Ambekar, Prathosh AP

Figure 1 for Skin Segmentation from NIR Images using Unsupervised Domain Adaptation through Generative Latent Search
Figure 2 for Skin Segmentation from NIR Images using Unsupervised Domain Adaptation through Generative Latent Search
Figure 3 for Skin Segmentation from NIR Images using Unsupervised Domain Adaptation through Generative Latent Search
Figure 4 for Skin Segmentation from NIR Images using Unsupervised Domain Adaptation through Generative Latent Search

Segmentation of the pixels corresponding to human skin is an essential first step in multiple applications ranging from surveillance to heart-rate estimation from remote-photoplethysmography. However, the existing literature considers the problem only in the visible-range of the EM-spectrum which limits their utility in low or no light settings where the criticality of the application is higher. To alleviate this problem, we consider the problem of skin segmentation from the Near-infrared images. However, Deep learning based state-of-the-art segmentation techniques demands large amounts of labelled data that is unavailable for the current problem. Therefore we cast the skin segmentation problem as that of target-independent unsupervised domain adaptation (UDA) where we use the data from the Red-channel of the visible-range to develop skin segmentation algorithm on NIR images. We propose a method for target-independent segmentation where the 'nearest-clone' of a target image in the source domain is searched and used as a proxy in the segmentation network trained only on the source domain. We prove the existence of 'nearest-clone' and propose a method to find it through an optimization algorithm over the latent space of a Deep generative model based on variational inference. We demonstrate the efficacy of the proposed method for NIR skin segmentation over the state-of-the-art UDA segmenation methods on the two newly created skin segmentation datasets in NIR domain despite not having access to the target NIR data.

Viaarxiv icon