Alert button
Picture for Ashwin De Silva

Ashwin De Silva

Alert button

Approximately optimal domain adaptation with Fisher's Linear Discriminant Analysis

Mar 14, 2023
Hayden S. Helm, Ashwin De Silva, Joshua T. Vogelstein, Carey E. Priebe, Weiwei Yang

Figure 1 for Approximately optimal domain adaptation with Fisher's Linear Discriminant Analysis
Figure 2 for Approximately optimal domain adaptation with Fisher's Linear Discriminant Analysis
Figure 3 for Approximately optimal domain adaptation with Fisher's Linear Discriminant Analysis
Figure 4 for Approximately optimal domain adaptation with Fisher's Linear Discriminant Analysis

We propose a class of models based on Fisher's Linear Discriminant (FLD) in the context of domain adaptation. The class is the convex combination of two hypotheses: i) an average hypothesis representing previously seen source tasks and ii) a hypothesis trained on a new target task. For a particular generative setting we derive the optimal convex combination of the two models under 0-1 loss, propose a computable approximation, and study the effect of various parameter settings on the relative risks between the optimal hypothesis, hypothesis i), and hypothesis ii). We demonstrate the effectiveness of the proposed optimal classifier in the context of EEG- and ECG-based classification settings and argue that the optimal classifier can be computed without access to direct information from any of the individual source tasks. We conclude by discussing further applications, limitations, and possible future directions.

Viaarxiv icon

The Value of Out-of-Distribution Data

Aug 23, 2022
Ashwin De Silva, Rahul Ramesh, Carey E. Priebe, Pratik Chaudhari, Joshua T. Vogelstein

Figure 1 for The Value of Out-of-Distribution Data
Figure 2 for The Value of Out-of-Distribution Data
Figure 3 for The Value of Out-of-Distribution Data
Figure 4 for The Value of Out-of-Distribution Data

More data helps us generalize to a task. But real datasets can contain out-of-distribution (OOD) data; this can come in the form of heterogeneity such as intra-class variability but also in the form of temporal shifts or concept drifts. We demonstrate a counter-intuitive phenomenon for such problems: generalization error of the task can be a non-monotonic function of the number of OOD samples; a small number of OOD samples can improve generalization but if the number of OOD samples is beyond a threshold, then the generalization error can deteriorate. We also show that if we know which samples are OOD, then using a weighted objective between the target and OOD samples ensures that the generalization error decreases monotonically. We demonstrate and analyze this issue using linear classifiers on synthetic datasets and medium-sized neural networks on CIFAR-10.

* To be presented as a short paper at the Out-of-Distribution Generalization in Computer Vision (OOD-CV) workshop, ECCV 2022, Tel Aviv, Israel 
Viaarxiv icon

Out-of-distribution and in-distribution posterior calibration using Kernel Density Polytopes

Feb 14, 2022
Jayanta Dey, Ashwin De Silva, Will LeVine, Jong M. Shin, Haoyin Xu, Ali Geisa, Tiffany Chu, Leyla Isik, Joshua T. Vogelstein

Figure 1 for Out-of-distribution and in-distribution posterior calibration using Kernel Density Polytopes
Figure 2 for Out-of-distribution and in-distribution posterior calibration using Kernel Density Polytopes
Figure 3 for Out-of-distribution and in-distribution posterior calibration using Kernel Density Polytopes
Figure 4 for Out-of-distribution and in-distribution posterior calibration using Kernel Density Polytopes

Any reasonable machine learning (ML) model should not only interpolate efficiently in between the training samples provided (in-distribution region), but also approach the extrapolative or out-of-distribution (OOD) region without being overconfident. Our experiment on human subjects justifies the aforementioned properties for human intelligence as well. Many state-of-the-art algorithms have tried to fix the overconfidence problem of ML models in the OOD region. However, in doing so, they have often impaired the in-distribution performance of the model. Our key insight is that ML models partition the feature space into polytopes and learn constant (random forests) or affine (ReLU networks) functions over those polytopes. This leads to the OOD overconfidence problem for the polytopes which lie in the training data boundary and extend to infinity. To resolve this issue, we propose kernel density methods that fit Gaussian kernel over the polytopes, which are learned using ML models. Specifically, we introduce two variants of kernel density polytopes: Kernel Density Forest (KDF) and Kernel Density Network (KDN) based on random forests and deep networks, respectively. Studies on various simulation settings show that both KDF and KDN achieve uniform confidence over the classes in the OOD region while maintaining good in-distribution accuracy compared to that of their respective parent models.

Viaarxiv icon

Out-of-distribution Detection Using Kernel Density Polytopes

Feb 06, 2022
Jayanta Dey, Ashwin De Silva, Will LeVine, Jong M. Shin, Haoyin Xu, Ali Geisa, Tiffany Chu, Leyla Isik, Joshua T. Vogelstein

Figure 1 for Out-of-distribution Detection Using Kernel Density Polytopes
Figure 2 for Out-of-distribution Detection Using Kernel Density Polytopes
Figure 3 for Out-of-distribution Detection Using Kernel Density Polytopes
Figure 4 for Out-of-distribution Detection Using Kernel Density Polytopes

Any reasonable machine learning (ML) model should not only interpolate efficiently in between the training samples provided (in-distribution region), but also approach the extrapolative or out-of-distribution (OOD) region without being overconfident. Our experiment on human subjects justifies the aforementioned properties for human intelligence as well. Many state-of-the-art algorithms have tried to fix the overconfidence problem of ML models in the OOD region. However, in doing so, they have often impaired the in-distribution performance of the model. Our key insight is that ML models partition the feature space into polytopes and learn constant (random forests) or affine (ReLU networks) functions over those polytopes. This leads to the OOD overconfidence problem for the polytopes which lie in the training data boundary and extend to infinity. To resolve this issue, we propose kernel density methods that fit Gaussian kernel over the polytopes, which are learned using ML models. Specifically, we introduce two variants of kernel density polytopes: Kernel Density Forest (KDF) and Kernel Density Network (KDN) based on random forests and deep networks, respectively. Studies on various simulation settings show that both KDF and KDN achieve uniform confidence over the classes in the OOD region while maintaining good in-distribution accuracy compared to that of their respective parent models.

Viaarxiv icon

Towards Accurate Cross-Domain In-Bed Human Pose Estimation

Oct 07, 2021
Mohamed Afham, Udith Haputhanthri, Jathurshan Pradeepkumar, Mithunjha Anandakumar, Ashwin De Silva, Chamira Edussooriya

Figure 1 for Towards Accurate Cross-Domain In-Bed Human Pose Estimation
Figure 2 for Towards Accurate Cross-Domain In-Bed Human Pose Estimation
Figure 3 for Towards Accurate Cross-Domain In-Bed Human Pose Estimation
Figure 4 for Towards Accurate Cross-Domain In-Bed Human Pose Estimation

Human behavioral monitoring during sleep is essential for various medical applications. Majority of the contactless human pose estimation algorithms are based on RGB modality, causing ineffectiveness in in-bed pose estimation due to occlusions by blankets and varying illumination conditions. Long-wavelength infrared (LWIR) modality based pose estimation algorithms overcome the aforementioned challenges; however, ground truth pose generations by a human annotator under such conditions are not feasible. A feasible solution to address this issue is to transfer the knowledge learned from images with pose labels and no occlusions, and adapt it towards real world conditions (occlusions due to blankets). In this paper, we propose a novel learning strategy comprises of two-fold data augmentation to reduce the cross-domain discrepancy and knowledge distillation to learn the distribution of unlabeled images in real world conditions. Our experiments and analysis show the effectiveness of our approach over multiple standard human pose estimation baselines.

* Code is available at https://github.com/MohamedAfham/CD_HPE 
Viaarxiv icon

A Thickness Sensitive Vessel Extraction Framework for Retinal and Conjunctival Vascular Tortuosity Analysis

Jan 02, 2021
Ashwin De Silva, Malsha V. Perera, Navodini Wijethilake, Saroj Jayasinghe, Nuwan D. Nanayakkara, Anjula De Silva

Figure 1 for A Thickness Sensitive Vessel Extraction Framework for Retinal and Conjunctival Vascular Tortuosity Analysis
Figure 2 for A Thickness Sensitive Vessel Extraction Framework for Retinal and Conjunctival Vascular Tortuosity Analysis
Figure 3 for A Thickness Sensitive Vessel Extraction Framework for Retinal and Conjunctival Vascular Tortuosity Analysis
Figure 4 for A Thickness Sensitive Vessel Extraction Framework for Retinal and Conjunctival Vascular Tortuosity Analysis

Systemic diseases such as diabetes, hypertension, atherosclerosis are among the leading causes of annual human mortality rate. It is suggested that retinal and conjunctival vascular tortuosity is a potential biomarker for such systemic diseases. Most importantly, it is observed that the tortuosity depends on the thickness of these vessels. Therefore, selective calculation of tortuosity within specific vessel thicknesses is required depending on the disease being analysed. In this paper, we propose a thickness sensitive vessel extraction framework that is primarily applicable for studies related to retinal and conjunctival vascular tortuosity. The framework uses a Convolutional Neural Network based on the IterNet architecture to obtain probability maps of the entire vasculature. They are then processed by a multi-scale vessel enhancement technique that exploits both fine and coarse structural vascular details of these probability maps in order to extract vessels of specified thicknesses. We evaluated the proposed framework on four datasets including DRIVE and SBVPI, and obtained Matthew's Correlation Coefficient values greater than 0.71 for all the datasets. In addition, the proposed framework was utilized to determine the association of diabetes with retinal and conjunctival vascular tortuosity. We observed that retinal vascular tortuosity (Eccentricity based Tortuosity Index) of the diabetic group was significantly higher (p < .05) than that of the non-diabetic group and that conjunctival vascular tortuosity (Total Curvature normalized by Arc Length) of diabetic group was significantly lower (p < .05) than that of the non-diabetic group. These observations were in agreement with the literature, strengthening the suitability of the proposed framework.

* Submitted for Reviewing 
Viaarxiv icon

A Joint Convolutional and Spatial Quad-Directional LSTM Network for Phase Unwrapping

Oct 26, 2020
Malsha V. Perera, Ashwin De Silva

Figure 1 for A Joint Convolutional and Spatial Quad-Directional LSTM Network for Phase Unwrapping
Figure 2 for A Joint Convolutional and Spatial Quad-Directional LSTM Network for Phase Unwrapping
Figure 3 for A Joint Convolutional and Spatial Quad-Directional LSTM Network for Phase Unwrapping
Figure 4 for A Joint Convolutional and Spatial Quad-Directional LSTM Network for Phase Unwrapping

Phase unwrapping is a classical ill-posed problem which aims to recover the true phase from wrapped phase. In this paper, we introduce a novel Convolutional Neural Network (CNN) that incorporates a Spatial Quad-Directional Long Short Term Memory (SQD-LSTM) for phase unwrapping, by formulating it as a regression problem. Incorporating SQD-LSTM can circumvent the typical CNNs' inherent difficulty of learning global spatial dependencies which are vital when recovering the true phase. Furthermore, we employ a problem specific composite loss function to train this network. The proposed network is found to be performing better than the existing methods under severe noise conditions (Normalized Root Mean Square Error of 1.3 % at SNR = 0 dB) while spending a significantly less computational time (0.054 s). The network also does not require a large scale dataset during training, thus making it ideal for applications with limited data that require fast and accurate phase unwrapping.

* Under Review 
Viaarxiv icon