Alert button
Picture for Jayadeva

Jayadeva

Alert button

MaScQA: A Question Answering Dataset for Investigating Materials Science Knowledge of Large Language Models

Aug 17, 2023
Mohd Zaki, Jayadeva, Mausam, N. M. Anoop Krishnan

Figure 1 for MaScQA: A Question Answering Dataset for Investigating Materials Science Knowledge of Large Language Models
Figure 2 for MaScQA: A Question Answering Dataset for Investigating Materials Science Knowledge of Large Language Models
Figure 3 for MaScQA: A Question Answering Dataset for Investigating Materials Science Knowledge of Large Language Models
Figure 4 for MaScQA: A Question Answering Dataset for Investigating Materials Science Knowledge of Large Language Models

Information extraction and textual comprehension from materials literature are vital for developing an exhaustive knowledge base that enables accelerated materials discovery. Language models have demonstrated their capability to answer domain-specific questions and retrieve information from knowledge bases. However, there are no benchmark datasets in the materials domain that can evaluate the understanding of the key concepts by these language models. In this work, we curate a dataset of 650 challenging questions from the materials domain that require the knowledge and skills of a materials student who has cleared their undergraduate degree. We classify these questions based on their structure and the materials science domain-based subcategories. Further, we evaluate the performance of GPT-3.5 and GPT-4 models on solving these questions via zero-shot and chain of thought prompting. It is observed that GPT-4 gives the best performance (~62% accuracy) as compared to GPT-3.5. Interestingly, in contrast to the general observation, no significant improvement in accuracy is observed with the chain of thought prompting. To evaluate the limitations, we performed an error analysis, which revealed conceptual errors (~64%) as the major contributor compared to computational errors (~36%) towards the reduced performance of LLMs. We hope that the dataset and analysis performed in this work will promote further research in developing better materials science domain-specific LLMs and strategies for information extraction.

Viaarxiv icon

Discovering Symbolic Laws Directly from Trajectories with Hamiltonian Graph Neural Networks

Jul 11, 2023
Suresh Bishnoi, Ravinder Bhattoo, Jayadeva, Sayan Ranu, N M Anoop Krishnan

Figure 1 for Discovering Symbolic Laws Directly from Trajectories with Hamiltonian Graph Neural Networks
Figure 2 for Discovering Symbolic Laws Directly from Trajectories with Hamiltonian Graph Neural Networks
Figure 3 for Discovering Symbolic Laws Directly from Trajectories with Hamiltonian Graph Neural Networks
Figure 4 for Discovering Symbolic Laws Directly from Trajectories with Hamiltonian Graph Neural Networks

The time evolution of physical systems is described by differential equations, which depend on abstract quantities like energy and force. Traditionally, these quantities are derived as functionals based on observables such as positions and velocities. Discovering these governing symbolic laws is the key to comprehending the interactions in nature. Here, we present a Hamiltonian graph neural network (HGNN), a physics-enforced GNN that learns the dynamics of systems directly from their trajectory. We demonstrate the performance of HGNN on n-springs, n-pendulums, gravitational systems, and binary Lennard Jones systems; HGNN learns the dynamics in excellent agreement with the ground truth from small amounts of data. We also evaluate the ability of HGNN to generalize to larger system sizes, and to hybrid spring-pendulum system that is a combination of two original systems (spring and pendulum) on which the models are trained independently. Finally, employing symbolic regression on the learned HGNN, we infer the underlying equations relating the energy functionals, even for complex systems such as the binary Lennard-Jones liquid. Our framework facilitates the interpretable discovery of interaction laws directly from physical system trajectories. Furthermore, this approach can be extended to other systems with topology-dependent dynamics, such as cells, polydisperse gels, or deformable bodies.

Viaarxiv icon

Graph Neural Stochastic Differential Equations for Learning Brownian Dynamics

Jun 20, 2023
Suresh Bishnoi, Jayadeva, Sayan Ranu, N. M. Anoop Krishnan

Figure 1 for Graph Neural Stochastic Differential Equations for Learning Brownian Dynamics
Figure 2 for Graph Neural Stochastic Differential Equations for Learning Brownian Dynamics
Figure 3 for Graph Neural Stochastic Differential Equations for Learning Brownian Dynamics
Figure 4 for Graph Neural Stochastic Differential Equations for Learning Brownian Dynamics

Neural networks (NNs) that exploit strong inductive biases based on physical laws and symmetries have shown remarkable success in learning the dynamics of physical systems directly from their trajectory. However, these works focus only on the systems that follow deterministic dynamics, for instance, Newtonian or Hamiltonian dynamics. Here, we propose a framework, namely Brownian graph neural networks (BROGNET), combining stochastic differential equations (SDEs) and GNNs to learn Brownian dynamics directly from the trajectory. We theoretically show that BROGNET conserves the linear momentum of the system, which in turn, provides superior performance on learning dynamics as revealed empirically. We demonstrate this approach on several systems, namely, linear spring, linear spring with binary particle types, and non-linear spring systems, all following Brownian dynamics at finite temperatures. We show that BROGNET significantly outperforms proposed baselines across all the benchmarked Brownian systems. In addition, we demonstrate zero-shot generalizability of BROGNET to simulate unseen system sizes that are two orders of magnitude larger and to different temperatures than those used during training. Altogether, our study contributes to advancing the understanding of the intricate dynamics of Brownian motion and demonstrates the effectiveness of graph neural networks in modeling such complex systems.

Viaarxiv icon

Cementron: Machine Learning the Constituent Phases in Cement Clinker from Optical Images

Nov 06, 2022
Mohd Zaki, Siddhant Sharma, Sunil Kumar Gurjar, Raju Goyal, Jayadeva, N. M. Anoop Krishnan

Figure 1 for Cementron: Machine Learning the Constituent Phases in Cement Clinker from Optical Images
Figure 2 for Cementron: Machine Learning the Constituent Phases in Cement Clinker from Optical Images
Figure 3 for Cementron: Machine Learning the Constituent Phases in Cement Clinker from Optical Images
Figure 4 for Cementron: Machine Learning the Constituent Phases in Cement Clinker from Optical Images

Cement is the most used construction material. The performance of cement hydrate depends on the constituent phases, viz. alite, belite, aluminate, and ferrites present in the cement clinker, both qualitatively and quantitatively. Traditionally, clinker phases are analyzed from optical images relying on a domain expert and simple image processing techniques. However, the non-uniformity of the images, variations in the geometry and size of the phases, and variabilities in the experimental approaches and imaging methods make it challenging to obtain the phases. Here, we present a machine learning (ML) approach to detect clinker microstructure phases automatically. To this extent, we create the first annotated dataset of cement clinker by segmenting alite and belite particles. Further, we use supervised ML methods to train models for identifying alite and belite regions. Specifically, we finetune the image detection and segmentation model Detectron-2 on the cement microstructure to develop a model for detecting the cement phases, namely, Cementron. We demonstrate that Cementron, trained only on literature data, works remarkably well on new images obtained from our experiments, demonstrating its generalizability. We make Cementron available for public use.

Viaarxiv icon

Predicting Oxide Glass Properties with Low Complexity Neural Network and Physical and Chemical Descriptors

Oct 19, 2022
Suresh Bishnoi, Skyler Badge, Jayadeva, N. M. Anoop Krishnan

Figure 1 for Predicting Oxide Glass Properties with Low Complexity Neural Network and Physical and Chemical Descriptors
Figure 2 for Predicting Oxide Glass Properties with Low Complexity Neural Network and Physical and Chemical Descriptors
Figure 3 for Predicting Oxide Glass Properties with Low Complexity Neural Network and Physical and Chemical Descriptors
Figure 4 for Predicting Oxide Glass Properties with Low Complexity Neural Network and Physical and Chemical Descriptors

Due to their disordered structure, glasses present a unique challenge in predicting the composition-property relationships. Recently, several attempts have been made to predict the glass properties using machine learning techniques. However, these techniques have the limitations, namely, (i) predictions are limited to the components that are present in the original dataset, and (ii) predictions towards the extreme values of the properties, important regions for new materials discovery, are not very reliable due to the sparse datapoints in this region. To address these challenges, here we present a low complexity neural network (LCNN) that provides improved performance in predicting the properties of oxide glasses. In addition, we combine the LCNN with physical and chemical descriptors that allow the development of universal models that can provide predictions for components beyond the training set. By training on a large dataset (~50000) of glass components, we show the LCNN outperforms state-of-the-art algorithms such as XGBoost. In addition, we interpret the LCNN models using Shapely additive explanations to gain insights into the role played by the descriptors in governing the property. Finally, we demonstrate the universality of the LCNN models by predicting the properties for glasses with new components that were not present in the original training set. Altogether, the present approach provides a promising direction towards accelerated discovery of novel glass compositions.

* 15 pages, 3 figures 
Viaarxiv icon

A Novel Meta-predictor based Algorithm for Testing VLSI Circuits

Jul 22, 2022
Shruti Pandey, Jayadeva, Smruti R. Sarangi

Figure 1 for A Novel Meta-predictor based Algorithm for Testing VLSI Circuits
Figure 2 for A Novel Meta-predictor based Algorithm for Testing VLSI Circuits
Figure 3 for A Novel Meta-predictor based Algorithm for Testing VLSI Circuits
Figure 4 for A Novel Meta-predictor based Algorithm for Testing VLSI Circuits

Testing of integrated circuits (IC) is a highly expensive process but also the most important one in determining the defect level of an IC. Manufacturing defects in the IC are modeled using stuck-at-fault models. Stuck-at-fault models cover most of the physical faults that occur during the manufacturing process. With decreasing feature sizes due to the advancement of semiconductor technology, the defects are also getting smaller in size. Tests for these hard-to-detect defects are generated using deterministic test generation (DTG) algorithms. Our work aims at reducing the cost of Path Oriented Decision Making: PODEM (a DTG algorithm) without compromising the test quality. We trained a meta predictor to choose the best model given the circuit and the target net. This ensemble chooses the best probability prediction model with a 95% accuracy. This leads to a reduced number of backtracking decisions and much better performance of PODEM in terms of its CPU time. We show that our ML- guided PODEM algorithm with a meta predictor outperforms the baseline PODEM by 34% and other state-of-the-art ML-guided algorithms by at least 15% for ISCAS85 benchmark circuits.

* 7 pages, 8 figures and 4 tables 
Viaarxiv icon

Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images

Feb 16, 2021
Kartikeya Badola, Sameer Ambekar, Himanshu Pant, Sumit Soman, Anuradha Sural, Rajiv Narang, Suresh Chandra, Jayadeva

Figure 1 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images
Figure 2 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images
Figure 3 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images
Figure 4 for Twin Augmented Architectures for Robust Classification of COVID-19 Chest X-Ray Images

The gold standard for COVID-19 is RT-PCR, testing facilities for which are limited and not always optimally distributed. Test results are delayed, which impacts treatment. Expert radiologists, one of whom is a co-author, are able to diagnose COVID-19 positivity from Chest X-Rays (CXR) and CT scans, that can facilitate timely treatment. Such diagnosis is particularly valuable in locations lacking radiologists with sufficient expertise and familiarity with COVID-19 patients. This paper has two contributions. One, we analyse literature on CXR based COVID-19 diagnosis. We show that popular choices of dataset selection suffer from data homogeneity, leading to misleading results. We compile and analyse a viable benchmark dataset from multiple existing heterogeneous sources. Such a benchmark is important for realistically testing models. Our second contribution relates to learning from imbalanced data. Datasets for COVID X-Ray classification face severe class imbalance, since most subjects are COVID -ve. Twin Support Vector Machines (Twin SVM) and Twin Neural Networks (Twin NN) have, in recent years, emerged as effective ways of handling skewed data. We introduce a state-of-the-art technique, termed as Twin Augmentation, for modifying popular pre-trained deep learning models. Twin Augmentation boosts the performance of a pre-trained deep neural network without requiring re-training. Experiments show, that across a multitude of classifiers, Twin Augmentation is very effective in boosting the performance of given pre-trained model for classification in imbalanced settings.

Viaarxiv icon

Complexity Controlled Generative Adversarial Networks

Nov 20, 2020
Himanshu Pant, Jayadeva, Sumit Soman

Figure 1 for Complexity Controlled Generative Adversarial Networks
Figure 2 for Complexity Controlled Generative Adversarial Networks
Figure 3 for Complexity Controlled Generative Adversarial Networks
Figure 4 for Complexity Controlled Generative Adversarial Networks

One of the issues faced in training Generative Adversarial Nets (GANs) and their variants is the problem of mode collapse, wherein the training stability in terms of the generative loss increases as more training data is used. In this paper, we propose an alternative architecture via the Low-Complexity Neural Network (LCNN), which attempts to learn models with low complexity. The motivation is that controlling model complexity leads to models that do not overfit the training data. We incorporate the LCNN loss function for GANs, Deep Convolutional GANs (DCGANs) and Spectral Normalized GANs (SNGANs), in order to develop hybrid architectures called the LCNN-GAN, LCNN-DCGAN and LCNN-SNGAN respectively. On various large benchmark image datasets, we show that the use of our proposed models results in stable training while avoiding the problem of mode collapse, resulting in better training stability. We also show how the learning behavior can be controlled by a hyperparameter in the LCNN functional, which also provides an improved inception score.

* 11 pages 
Viaarxiv icon

Enhash: A Fast Streaming Algorithm For Concept Drift Detection

Nov 07, 2020
Aashi Jindal, Prashant Gupta, Debarka Sengupta, Jayadeva

Figure 1 for Enhash: A Fast Streaming Algorithm For Concept Drift Detection
Figure 2 for Enhash: A Fast Streaming Algorithm For Concept Drift Detection
Figure 3 for Enhash: A Fast Streaming Algorithm For Concept Drift Detection
Figure 4 for Enhash: A Fast Streaming Algorithm For Concept Drift Detection

We propose Enhash, a fast ensemble learner that detects \textit{concept drift} in a data stream. A stream may consist of abrupt, gradual, virtual, or recurring events, or a mixture of various types of drift. Enhash employs projection hash to insert an incoming sample. We show empirically that the proposed method has competitive performance to existing ensemble learners in much lesser time. Also, Enhash has moderate resource requirements. Experiments relevant to performance comparison were performed on 6 artificial and 4 real data sets consisting of various types of drifts.

Viaarxiv icon

Guided Random Forest and its application to data approximation

Sep 02, 2019
Prashant Gupta, Aashi Jindal, Jayadeva, Debarka Sengupta

Figure 1 for Guided Random Forest and its application to data approximation
Figure 2 for Guided Random Forest and its application to data approximation
Figure 3 for Guided Random Forest and its application to data approximation
Figure 4 for Guided Random Forest and its application to data approximation

We present a new way of constructing an ensemble classifier, named the Guided Random Forest (GRAF) in the sequel. GRAF extends the idea of building oblique decision trees with localized partitioning to obtain a global partitioning. We show that global partitioning bridges the gap between decision trees and boosting algorithms. We empirically demonstrate that global partitioning reduces the generalization error bound. Results on 115 benchmark datasets show that GRAF yields comparable or better results on a majority of datasets. We also present a new way of approximating the datasets in the framework of random forests.

Viaarxiv icon