Alert button
Picture for C. Bayan Bruss

C. Bayan Bruss

Alert button

Center for Machine Learning, Capital One

A Performance-Driven Benchmark for Feature Selection in Tabular Deep Learning

Nov 10, 2023
Valeriia Cherepanova, Roman Levin, Gowthami Somepalli, Jonas Geiping, C. Bayan Bruss, Andrew Gordon Wilson, Tom Goldstein, Micah Goldblum

Academic tabular benchmarks often contain small sets of curated features. In contrast, data scientists typically collect as many features as possible into their datasets, and even engineer new features from existing ones. To prevent overfitting in subsequent downstream modeling, practitioners commonly use automated feature selection methods that identify a reduced subset of informative features. Existing benchmarks for tabular feature selection consider classical downstream models, toy synthetic datasets, or do not evaluate feature selectors on the basis of downstream performance. Motivated by the increasing popularity of tabular deep learning, we construct a challenging feature selection benchmark evaluated on downstream neural networks including transformers, using real datasets and multiple methods for generating extraneous features. We also propose an input-gradient-based analogue of Lasso for neural networks that outperforms classical feature selection methods on challenging problems such as selecting from corrupted or second-order features.

* Conference on Neural Information Processing Systems 2023  
Viaarxiv icon

From Explanation to Action: An End-to-End Human-in-the-loop Framework for Anomaly Reasoning and Management

Apr 06, 2023
Xueying Ding, Nikita Seleznev, Senthil Kumar, C. Bayan Bruss, Leman Akoglu

Figure 1 for From Explanation to Action: An End-to-End Human-in-the-loop Framework for Anomaly Reasoning and Management
Figure 2 for From Explanation to Action: An End-to-End Human-in-the-loop Framework for Anomaly Reasoning and Management
Figure 3 for From Explanation to Action: An End-to-End Human-in-the-loop Framework for Anomaly Reasoning and Management
Figure 4 for From Explanation to Action: An End-to-End Human-in-the-loop Framework for Anomaly Reasoning and Management

Anomalies are often indicators of malfunction or inefficiency in various systems such as manufacturing, healthcare, finance, surveillance, to name a few. While the literature is abundant in effective detection algorithms due to this practical relevance, autonomous anomaly detection is rarely used in real-world scenarios. Especially in high-stakes applications, a human-in-the-loop is often involved in processes beyond detection such as verification and troubleshooting. In this work, we introduce ALARM (for Analyst-in-the-Loop Anomaly Reasoning and Management); an end-to-end framework that supports the anomaly mining cycle comprehensively, from detection to action. Besides unsupervised detection of emerging anomalies, it offers anomaly explanations and an interactive GUI for human-in-the-loop processes -- visual exploration, sense-making, and ultimately action-taking via designing new detection rules -- that help close ``the loop'' as the new rules complement rule-based supervised detection, typical of many deployed systems in practice. We demonstrate \method's efficacy through a series of case studies with fraud analysts from the financial industry.

Viaarxiv icon

BASED-XAI: Breaking Ablation Studies Down for Explainable Artificial Intelligence

Jul 12, 2022
Isha Hameed, Samuel Sharpe, Daniel Barcklow, Justin Au-Yeung, Sahil Verma, Jocelyn Huang, Brian Barr, C. Bayan Bruss

Figure 1 for BASED-XAI: Breaking Ablation Studies Down for Explainable Artificial Intelligence
Figure 2 for BASED-XAI: Breaking Ablation Studies Down for Explainable Artificial Intelligence
Figure 3 for BASED-XAI: Breaking Ablation Studies Down for Explainable Artificial Intelligence
Figure 4 for BASED-XAI: Breaking Ablation Studies Down for Explainable Artificial Intelligence

Explainable artificial intelligence (XAI) methods lack ground truth. In its place, method developers have relied on axioms to determine desirable properties for their explanations' behavior. For high stakes uses of machine learning that require explainability, it is not sufficient to rely on axioms as the implementation, or its usage, can fail to live up to the ideal. As a result, there exists active research on validating the performance of XAI methods. The need for validation is especially magnified in domains with a reliance on XAI. A procedure frequently used to assess their utility, and to some extent their fidelity, is an ablation study. By perturbing the input variables in rank order of importance, the goal is to assess the sensitivity of the model's performance. Perturbing important variables should correlate with larger decreases in measures of model capability than perturbing less important features. While the intent is clear, the actual implementation details have not been studied rigorously for tabular data. Using five datasets, three XAI methods, four baselines, and three perturbations, we aim to show 1) how varying perturbations and adding simple guardrails can help to avoid potentially flawed conclusions, 2) how treatment of categorical variables is an important consideration in both post-hoc explainability and ablation studies, and 3) how to identify useful baselines for XAI methods and viable perturbations for ablation studies.

* 6 pages, accepted by the KDD 2022 Workshop on Machine Learning for Finance (KDD MLF) 
Viaarxiv icon

Transfer Learning with Deep Tabular Models

Jun 30, 2022
Roman Levin, Valeriia Cherepanova, Avi Schwarzschild, Arpit Bansal, C. Bayan Bruss, Tom Goldstein, Andrew Gordon Wilson, Micah Goldblum

Figure 1 for Transfer Learning with Deep Tabular Models
Figure 2 for Transfer Learning with Deep Tabular Models
Figure 3 for Transfer Learning with Deep Tabular Models
Figure 4 for Transfer Learning with Deep Tabular Models

Recent work on deep learning for tabular data demonstrates the strong performance of deep tabular models, often bridging the gap between gradient boosted decision trees and neural networks. Accuracy aside, a major advantage of neural models is that they learn reusable features and are easily fine-tuned in new domains. This property is often exploited in computer vision and natural language applications, where transfer learning is indispensable when task-specific training data is scarce. In this work, we demonstrate that upstream data gives tabular neural networks a decisive advantage over widely used GBDT models. We propose a realistic medical diagnosis benchmark for tabular transfer learning, and we present a how-to guide for using upstream data to boost performance with a variety of tabular neural network architectures. Finally, we propose a pseudo-feature method for cases where the upstream and downstream feature sets differ, a tabular-specific problem widespread in real-world applications. Our code is available at https://github.com/LevinRoman/tabular-transfer-learning .

Viaarxiv icon

Counterfactual Explanations via Latent Space Projection and Interpolation

Dec 02, 2021
Brian Barr, Matthew R. Harrington, Samuel Sharpe, C. Bayan Bruss

Figure 1 for Counterfactual Explanations via Latent Space Projection and Interpolation
Figure 2 for Counterfactual Explanations via Latent Space Projection and Interpolation
Figure 3 for Counterfactual Explanations via Latent Space Projection and Interpolation
Figure 4 for Counterfactual Explanations via Latent Space Projection and Interpolation

Counterfactual explanations represent the minimal change to a data sample that alters its predicted classification, typically from an unfavorable initial class to a desired target class. Counterfactuals help answer questions such as "what needs to change for this application to get accepted for a loan?". A number of recently proposed approaches to counterfactual generation give varying definitions of "plausible" counterfactuals and methods to generate them. However, many of these methods are computationally intensive and provide unconvincing explanations. Here we introduce SharpShooter, a method for binary classification that starts by creating a projected version of the input that classifies as the target class. Counterfactual candidates are then generated in latent space on the interpolation line between the input and its projection. We then demonstrate that our framework translates core characteristics of a sample to its counterfactual through the use of learned representations. Furthermore, we show that SharpShooter is competitive across common quality metrics on tabular and image datasets while being orders of magnitude faster than two comparable methods and excels at measures of realism, making it well-suited for high velocity machine learning applications which require timely explanations.

* 10 pages, 6 figures 
Viaarxiv icon

MetaBalance: High-Performance Neural Networks for Class-Imbalanced Data

Jun 17, 2021
Arpit Bansal, Micah Goldblum, Valeriia Cherepanova, Avi Schwarzschild, C. Bayan Bruss, Tom Goldstein

Figure 1 for MetaBalance: High-Performance Neural Networks for Class-Imbalanced Data
Figure 2 for MetaBalance: High-Performance Neural Networks for Class-Imbalanced Data
Figure 3 for MetaBalance: High-Performance Neural Networks for Class-Imbalanced Data
Figure 4 for MetaBalance: High-Performance Neural Networks for Class-Imbalanced Data

Class-imbalanced data, in which some classes contain far more samples than others, is ubiquitous in real-world applications. Standard techniques for handling class-imbalance usually work by training on a re-weighted loss or on re-balanced data. Unfortunately, training overparameterized neural networks on such objectives causes rapid memorization of minority class data. To avoid this trap, we harness meta-learning, which uses both an ''outer-loop'' and an ''inner-loop'' loss, each of which may be balanced using different strategies. We evaluate our method, MetaBalance, on image classification, credit-card fraud detection, loan default prediction, and facial recognition tasks with severely imbalanced data, and we find that MetaBalance outperforms a wide array of popular re-sampling strategies.

Viaarxiv icon

SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training

Jun 02, 2021
Gowthami Somepalli, Micah Goldblum, Avi Schwarzschild, C. Bayan Bruss, Tom Goldstein

Figure 1 for SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training
Figure 2 for SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training
Figure 3 for SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training
Figure 4 for SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training

Tabular data underpins numerous high-impact applications of machine learning from fraud detection to genomics and healthcare. Classical approaches to solving tabular problems, such as gradient boosting and random forests, are widely used by practitioners. However, recent deep learning methods have achieved a degree of performance competitive with popular techniques. We devise a hybrid deep learning approach to solving tabular data problems. Our method, SAINT, performs attention over both rows and columns, and it includes an enhanced embedding method. We also study a new contrastive self-supervised pre-training method for use when labels are scarce. SAINT consistently improves performance over previous deep learning methods, and it even outperforms gradient boosting methods, including XGBoost, CatBoost, and LightGBM, on average over a variety of benchmark tasks.

Viaarxiv icon

Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations

Dec 16, 2020
Rachana Balasubramanian, Samuel Sharpe, Brian Barr, Jason Wittenbach, C. Bayan Bruss

Figure 1 for Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations
Figure 2 for Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations
Figure 3 for Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations
Figure 4 for Latent-CF: A Simple Baseline for Reverse Counterfactual Explanations

In the environment of fair lending laws and the General Data Protection Regulation (GDPR), the ability to explain a model's prediction is of paramount importance. High quality explanations are the first step in assessing fairness. Counterfactuals are valuable tools for explainability. They provide actionable, comprehensible explanations for the individual who is subject to decisions made from the prediction. It is important to find a baseline for producing them. We propose a simple method for generating counterfactuals by using gradient descent to search in the latent space of an autoencoder and benchmark our method against approaches that search for counterfactuals in feature space. Additionally, we implement metrics to concretely evaluate the quality of the counterfactuals. We show that latent space counterfactual generation strikes a balance between the speed of basic feature gradient descent methods and the sparseness and authenticity of counterfactuals generated by more complex feature space oriented techniques.

Viaarxiv icon

DLGNet-Task: An End-to-end Neural Network Framework for Modeling Multi-turn Multi-domain Task-Oriented Dialogue

Oct 06, 2020
Oluwatobi O. Olabiyi, Prarthana Bhattarai, C. Bayan Bruss, Zachary Kulis

Figure 1 for DLGNet-Task: An End-to-end Neural Network Framework for Modeling Multi-turn Multi-domain Task-Oriented Dialogue
Figure 2 for DLGNet-Task: An End-to-end Neural Network Framework for Modeling Multi-turn Multi-domain Task-Oriented Dialogue
Figure 3 for DLGNet-Task: An End-to-end Neural Network Framework for Modeling Multi-turn Multi-domain Task-Oriented Dialogue
Figure 4 for DLGNet-Task: An End-to-end Neural Network Framework for Modeling Multi-turn Multi-domain Task-Oriented Dialogue

Task oriented dialogue (TOD) requires the complex interleaving of a number of individually controllable components with strong guarantees for explainability and verifiability. This has made it difficult to adopt the multi-turn multi-domain dialogue generation capabilities of streamlined end-to-end open-domain dialogue systems. In this paper, we present a new framework, DLGNet-Task, a unified task-oriented dialogue system which employs autoregressive transformer networks such as DLGNet and GPT-2/3 to complete user tasks in multi-turn multi-domain conversations. Our framework enjoys the controllable, verifiable, and explainable outputs of modular approaches, and the low development, deployment and maintenance cost of end-to-end systems. Treating open-domain system components as additional TOD system modules allows DLGNet-Task to learn the joint distribution of the inputs and outputs of all the functional blocks of existing modular approaches such as, natural language understanding (NLU), state tracking, action policy, as well as natural language generation (NLG). Rather than training the modules individually, as is common in real-world systems, we trained them jointly with appropriate module separations. When evaluated on the MultiWOZ2.1 dataset, DLGNet-Task shows comparable performance to the existing state-of-the-art approaches. Furthermore, using DLGNet-Task in conversational AI systems reduces the level of effort required for developing, deploying, and maintaining intelligent assistants at scale.

Viaarxiv icon