Alert button
Picture for Andrei Liviu Nicolicioiu

Andrei Liviu Nicolicioiu

Alert button

Environment-biased Feature Ranking for Novelty Detection Robustness

Sep 21, 2023
Stefan Smeu, Elena Burceanu, Emanuela Haller, Andrei Liviu Nicolicioiu

We tackle the problem of robust novelty detection, where we aim to detect novelties in terms of semantic content while being invariant to changes in other, irrelevant factors. Specifically, we operate in a setup with multiple environments, where we determine the set of features that are associated more with the environments, rather than to the content relevant for the task. Thus, we propose a method that starts with a pretrained embedding and a multi-env setup and manages to rank the features based on their environment-focus. First, we compute a per-feature score based on the feature distribution variance between envs. Next, we show that by dropping the highly scored ones, we manage to remove spurious correlations and improve the overall performance by up to 6%, both in covariance and sub-population shift cases, both for a real and a synthetic benchmark, that we introduce for this task.

* ICCV 2024 - Workshop on Out Of Distribution Generalization in Computer Vision 
Viaarxiv icon

Learning Diverse Features in Vision Transformers for Improved Generalization

Aug 30, 2023
Armand Mihai Nicolicioiu, Andrei Liviu Nicolicioiu, Bogdan Alexe, Damien Teney

Deep learning models often rely only on a small set of features even when there is a rich set of predictive signals in the training data. This makes models brittle and sensitive to distribution shifts. In this work, we first examine vision transformers (ViTs) and find that they tend to extract robust and spurious features with distinct attention heads. As a result of this modularity, their performance under distribution shifts can be significantly improved at test time by pruning heads corresponding to spurious features, which we demonstrate using an "oracle selection" on validation data. Second, we propose a method to further enhance the diversity and complementarity of the learned features by encouraging orthogonality of the attention heads' input gradients. We observe improved out-of-distribution performance on diagnostic benchmarks (MNIST-CIFAR, Waterbirds) as a consequence of the enhanced diversity of features and the pruning of undesirable heads.

* 2023 ICML Workshop on Spurious Correlations, Invariance and Stability 
Viaarxiv icon

Spuriosity Didn't Kill the Classifier: Using Invariant Predictions to Harness Spurious Features

Jul 19, 2023
Cian Eastwood, Shashank Singh, Andrei Liviu Nicolicioiu, Marin Vlastelica, Julius von Kügelgen, Bernhard Schölkopf

To avoid failures on out-of-distribution data, recent works have sought to extract features that have a stable or invariant relationship with the label across domains, discarding the "spurious" or unstable features whose relationship with the label changes across domains. However, unstable features often carry complementary information about the label that could boost performance if used correctly in the test domain. Our main contribution is to show that it is possible to learn how to use these unstable features in the test domain without labels. In particular, we prove that pseudo-labels based on stable features provide sufficient guidance for doing so, provided that stable and unstable features are conditionally independent given the label. Based on this theoretical insight, we propose Stable Feature Boosting (SFB), an algorithm for: (i) learning a predictor that separates stable and conditionally-independent unstable features; and (ii) using the stable-feature predictions to adapt the unstable-feature predictions in the test domain. Theoretically, we prove that SFB can learn an asymptotically-optimal predictor without test-domain labels. Empirically, we demonstrate the effectiveness of SFB on real and synthetic data.

Viaarxiv icon

Env-Aware Anomaly Detection: Ignore Style Changes, Stay True to Content!

Oct 06, 2022
Stefan Smeu, Elena Burceanu, Andrei Liviu Nicolicioiu, Emanuela Haller

Figure 1 for Env-Aware Anomaly Detection: Ignore Style Changes, Stay True to Content!
Figure 2 for Env-Aware Anomaly Detection: Ignore Style Changes, Stay True to Content!
Figure 3 for Env-Aware Anomaly Detection: Ignore Style Changes, Stay True to Content!
Figure 4 for Env-Aware Anomaly Detection: Ignore Style Changes, Stay True to Content!

We introduce a formalization and benchmark for the unsupervised anomaly detection task in the distribution-shift scenario. Our work builds upon the iWildCam dataset, and, to the best of our knowledge, we are the first to propose such an approach for visual data. We empirically validate that environment-aware methods perform better in such cases when compared with the basic Empirical Risk Minimization (ERM). We next propose an extension for generating positive samples for contrastive methods that considers the environment labels when training, improving the ERM baseline score by 8.7%.

Viaarxiv icon

DCI-ES: An Extended Disentanglement Framework with Connections to Identifiability

Oct 01, 2022
Cian Eastwood, Andrei Liviu Nicolicioiu, Julius von Kügelgen, Armin Kekić, Frederik Träuble, Andrea Dittadi, Bernhard Schölkopf

Figure 1 for DCI-ES: An Extended Disentanglement Framework with Connections to Identifiability
Figure 2 for DCI-ES: An Extended Disentanglement Framework with Connections to Identifiability
Figure 3 for DCI-ES: An Extended Disentanglement Framework with Connections to Identifiability
Figure 4 for DCI-ES: An Extended Disentanglement Framework with Connections to Identifiability

In representation learning, a common approach is to seek representations which disentangle the underlying factors of variation. Eastwood & Williams (2018) proposed three metrics for quantifying the quality of such disentangled representations: disentanglement (D), completeness (C) and informativeness (I). In this work, we first connect this DCI framework to two common notions of linear and nonlinear identifiability, thus establishing a formal link between disentanglement and the closely-related field of independent component analysis. We then propose an extended DCI-ES framework with two new measures of representation quality - explicitness (E) and size (S) - and point out how D and C can be computed for black-box predictors. Our main idea is that the functional capacity required to use a representation is an important but thus-far neglected aspect of representation quality, which we quantify using explicitness or ease-of-use (E). We illustrate the relevance of our extensions on the MPI3D and Cars3D datasets.

Viaarxiv icon

Mining for meaning: from vision to language through multiple networks consensus

Sep 18, 2018
Iulia Duta, Andrei Liviu Nicolicioiu, Simion-Vlad Bogolin, Marius Leordeanu

Figure 1 for Mining for meaning: from vision to language through multiple networks consensus
Figure 2 for Mining for meaning: from vision to language through multiple networks consensus
Figure 3 for Mining for meaning: from vision to language through multiple networks consensus
Figure 4 for Mining for meaning: from vision to language through multiple networks consensus

Describing visual data into natural language is a very challenging task, at the intersection of computer vision, natural language processing and machine learning. Language goes well beyond the description of physical objects and their interactions and can convey the same abstract idea in many ways. It is both about content at the highest semantic level as well as about fluent form. Here we propose an approach to describe videos in natural language by reaching a consensus among multiple encoder-decoder networks. Finding such a consensual linguistic description, which shares common properties with a larger group, has a better chance to convey the correct meaning. We propose and train several network architectures and use different types of image, audio and video features. Each model produces its own description of the input video and the best one is chosen through an efficient, two-phase consensus process. We demonstrate the strength of our approach by obtaining state of the art results on the challenging MSR-VTT dataset.

* Accepted at BMVC 2018 
Viaarxiv icon