Alert button
Picture for Evgeny Burnaev

Evgeny Burnaev

Alert button

Disentanglement Learning via Topology

Aug 24, 2023
Nikita Balabin, Daria Voronkova, Ilya Trofimov, Evgeny Burnaev, Serguei Barannikov

Figure 1 for Disentanglement Learning via Topology
Figure 2 for Disentanglement Learning via Topology
Figure 3 for Disentanglement Learning via Topology
Figure 4 for Disentanglement Learning via Topology

We propose TopDis (Topological Disentanglement), a method for learning disentangled representations via adding multi-scale topological loss term. Disentanglement is a crucial property of data representations substantial for the explainability and robustness of deep learning models and a step towards high-level cognition. The state-of-the-art method based on VAE minimizes the total correlation of the joint distribution of latent variables. We take a different perspective on disentanglement by analyzing topological properties of data manifolds. In particular, we optimize the topological similarity for data manifolds traversals. To the best of our knowledge, our paper is the first one to propose a differentiable topological loss for disentanglement. Our experiments have shown that the proposed topological loss improves disentanglement scores such as MIG, FactorVAE score, SAP score and DCI disentanglement score with respect to state-of-the-art results. Our method works in an unsupervised manner, permitting to apply it for problems without labeled factors of variation. Additionally, we show how to use the proposed topological loss to find disentangled directions in a trained GAN.

Viaarxiv icon

Anomaly Detection in Networks via Score-Based Generative Models

Jun 27, 2023
Dmitrii Gavrilev, Evgeny Burnaev

Figure 1 for Anomaly Detection in Networks via Score-Based Generative Models
Figure 2 for Anomaly Detection in Networks via Score-Based Generative Models
Figure 3 for Anomaly Detection in Networks via Score-Based Generative Models
Figure 4 for Anomaly Detection in Networks via Score-Based Generative Models

Node outlier detection in attributed graphs is a challenging problem for which there is no method that would work well across different datasets. Motivated by the state-of-the-art results of score-based models in graph generative modeling, we propose to incorporate them into the aforementioned problem. Our method achieves competitive results on small-scale graphs. We provide an empirical analysis of the Dirichlet energy, and show that generative models might struggle to accurately reconstruct it.

* 16 pages, 8 figures, ICML workshop on Structured Probabilistic Inference & Generative Modeling 
Viaarxiv icon

Building the Bridge of Schrödinger: A Continuous Entropic Optimal Transport Benchmark

Jun 16, 2023
Nikita Gushchin, Alexander Kolesov, Petr Mokrov, Polina Karpikova, Andrey Spiridonov, Evgeny Burnaev, Alexander Korotin

Figure 1 for Building the Bridge of Schrödinger: A Continuous Entropic Optimal Transport Benchmark
Figure 2 for Building the Bridge of Schrödinger: A Continuous Entropic Optimal Transport Benchmark
Figure 3 for Building the Bridge of Schrödinger: A Continuous Entropic Optimal Transport Benchmark
Figure 4 for Building the Bridge of Schrödinger: A Continuous Entropic Optimal Transport Benchmark

Over the last several years, there has been a significant progress in developing neural solvers for the Schr\"odinger Bridge (SB) problem and applying them to generative modeling. This new research field is justifiably fruitful as it is interconnected with the practically well-performing diffusion models and theoretically-grounded entropic optimal transport (EOT). Still the area lacks non-trivial tests allowing a researcher to understand how well do the methods solve SB or its equivalent continuous EOT problem. We fill this gap and propose a novel way to create pairs of probability distributions for which the ground truth OT solution in known by the construction. Our methodology is generic and works for a wide range of OT formulations, in particular, it covers the EOT which is equivalent to SB (the main interest of our study). This development allows us to create continuous benchmark distributions with the known EOT and SB solution on high-dimensional spaces such as spaces of images. As an illustration, we use these benchmark pairs to test how well do existing neural EOT/SB solvers actually compute the EOT solution. The benchmark is available via the link: https://github.com/ngushchin/EntropicOTBenchmark.

Viaarxiv icon

Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts

Jun 07, 2023
Eduard Tulchinskii, Kristian Kuznetsov, Laida Kushnareva, Daniil Cherniavskii, Serguei Barannikov, Irina Piontkovskaya, Sergey Nikolenko, Evgeny Burnaev

Figure 1 for Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts
Figure 2 for Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts
Figure 3 for Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts
Figure 4 for Intrinsic Dimension Estimation for Robust Detection of AI-Generated Texts

Rapidly increasing quality of AI-generated content makes it difficult to distinguish between human and AI-generated texts, which may lead to undesirable consequences for society. Therefore, it becomes increasingly important to study the properties of human texts that are invariant over text domains and various proficiency of human writers, can be easily calculated for any language, and can robustly separate natural and AI-generated texts regardless of the generation model and sampling method. In this work, we propose such an invariant of human texts, namely the intrinsic dimensionality of the manifold underlying the set of embeddings of a given text sample. We show that the average intrinsic dimensionality of fluent texts in natural language is hovering around the value $9$ for several alphabet-based languages and around $7$ for Chinese, while the average intrinsic dimensionality of AI-generated texts for each language is $\approx 1.5$ lower, with a clear statistical separation between human-generated and AI-generated distributions. This property allows us to build a score-based artificial text detector. The proposed detector's accuracy is stable over text domains, generator models, and human writer proficiency levels, outperforming SOTA detectors in model-agnostic and cross-domain scenarios by a significant margin.

Viaarxiv icon

Factored-NeuS: Reconstructing Surfaces, Illumination, and Materials of Possibly Glossy Objects

May 29, 2023
Yue Fan, Ivan Skorokhodov, Oleg Voynov, Savva Ignatyev, Evgeny Burnaev, Peter Wonka, Yiqun Wang

Figure 1 for Factored-NeuS: Reconstructing Surfaces, Illumination, and Materials of Possibly Glossy Objects
Figure 2 for Factored-NeuS: Reconstructing Surfaces, Illumination, and Materials of Possibly Glossy Objects
Figure 3 for Factored-NeuS: Reconstructing Surfaces, Illumination, and Materials of Possibly Glossy Objects
Figure 4 for Factored-NeuS: Reconstructing Surfaces, Illumination, and Materials of Possibly Glossy Objects

We develop a method that recovers the surface, materials, and illumination of a scene from its posed multi-view images. In contrast to prior work, it does not require any additional data and can handle glossy objects or bright lighting. It is a progressive inverse rendering approach, which consists of three stages. First, we reconstruct the scene radiance and signed distance function (SDF) with our novel regularization strategy for specular reflections. Our approach considers both the diffuse and specular colors, which allows for handling complex view-dependent lighting effects for surface reconstruction. Second, we distill light visibility and indirect illumination from the learned SDF and radiance field using learnable mapping functions. Third, we design a method for estimating the ratio of incoming direct light represented via Spherical Gaussians reflected in a specular manner and then reconstruct the materials and direct illumination of the scene. Experimental results demonstrate that the proposed method outperforms the current state-of-the-art in recovering surfaces, materials, and lighting without relying on any additional data.

* 12 pages, 10 figures. Project page: https://authors-hub.github.io/Factored-NeuS 
Viaarxiv icon

Energy-guided Entropic Neural Optimal Transport

Apr 12, 2023
Petr Mokrov, Alexander Korotin, Evgeny Burnaev

Figure 1 for Energy-guided Entropic Neural Optimal Transport
Figure 2 for Energy-guided Entropic Neural Optimal Transport
Figure 3 for Energy-guided Entropic Neural Optimal Transport
Figure 4 for Energy-guided Entropic Neural Optimal Transport

Energy-Based Models (EBMs) are known in the Machine Learning community for the decades. Since the seminal works devoted to EBMs dating back to the noughties there have been appearing a lot of efficient methods which solve the generative modelling problem by means of energy potentials (unnormalized likelihood functions). In contrast, the realm of Optimal Transport (OT) and, in particular, neural OT solvers is much less explored and limited by few recent works (excluding WGAN based approaches which utilize OT as a loss function and do not model OT maps themselves). In our work, we bridge the gap between EBMs and Entropy-regularized OT. We present the novel methodology which allows utilizing the recent developments and technical improvements of the former in order to enrich the latter. We validate the applicability of our method on toy 2D scenarios as well as standard unpaired image-to-image translation problems. For the sake of simplicity, we choose simple short- and long- run EBMs as a backbone of our Energy-guided Entropic OT method, leaving the application of more sophisticated EBMs for future research.

Viaarxiv icon

Partial Neural Optimal Transport

Mar 14, 2023
Milena Gazdieva, Alexander Korotin, Evgeny Burnaev

Figure 1 for Partial Neural Optimal Transport
Figure 2 for Partial Neural Optimal Transport
Figure 3 for Partial Neural Optimal Transport

We propose a novel neural method to compute partial optimal transport (OT) maps, i.e., OT maps between parts of measures of the specified masses. We test our partial neural optimal transport algorithm on synthetic examples.

Viaarxiv icon

Neural Gromov-Wasserstein Optimal Transport

Mar 10, 2023
Maksim Nekrashevich, Alexander Korotin, Evgeny Burnaev

Figure 1 for Neural Gromov-Wasserstein Optimal Transport
Figure 2 for Neural Gromov-Wasserstein Optimal Transport
Figure 3 for Neural Gromov-Wasserstein Optimal Transport
Figure 4 for Neural Gromov-Wasserstein Optimal Transport

We present a scalable neural method to solve the Gromov-Wasserstein (GW) Optimal Transport (OT) problem with the inner product cost. In this problem, given two distributions supported on (possibly different) spaces, one has to find the most isometric map between them. Our proposed approach uses neural networks and stochastic mini-batch optimization which allows to overcome the limitations of existing GW methods such as their poor scalability with the number of samples and the lack of out-of-sample estimation. To demonstrate the effectiveness of our proposed method, we conduct experiments on the synthetic data and explore the practical applicability of our method to the popular task of the unsupervised alignment of word embeddings.

Viaarxiv icon

Label Attention Network for sequential multi-label classification

Mar 01, 2023
Elizaveta Kovtun, Galina Boeva, Artem Zabolotnyi, Evgeny Burnaev, Martin Spindler, Alexey Zaytsev

Figure 1 for Label Attention Network for sequential multi-label classification
Figure 2 for Label Attention Network for sequential multi-label classification
Figure 3 for Label Attention Network for sequential multi-label classification
Figure 4 for Label Attention Network for sequential multi-label classification

Multi-label classification is a natural problem statement for sequential data. We might be interested in the items of the next order by a customer, or types of financial transactions that will occur tomorrow. Most modern approaches focus on transformer architecture for multi-label classification, introducing self-attention for the elements of a sequence with each element being a multi-label vector and supplementary information. However, in this way we loose local information related to interconnections between particular labels. We propose instead to use a self-attention mechanism over labels preceding the predicted step. Conducted experiments suggest that such architecture improves the model performance and provides meaningful attention between labels. The metric such as micro-AUC of our label attention network is $0.9847$ compared to $0.7390$ for vanilla transformers benchmark.

Viaarxiv icon

Learning Topology-Preserving Data Representations

Feb 15, 2023
Ilya Trofimov, Daniil Cherniavskii, Eduard Tulchinskii, Nikita Balabin, Evgeny Burnaev, Serguei Barannikov

Figure 1 for Learning Topology-Preserving Data Representations
Figure 2 for Learning Topology-Preserving Data Representations
Figure 3 for Learning Topology-Preserving Data Representations
Figure 4 for Learning Topology-Preserving Data Representations

We propose a method for learning topology-preserving data representations (dimensionality reduction). The method aims to provide topological similarity between the data manifold and its latent representation via enforcing the similarity in topological features (clusters, loops, 2D voids, etc.) and their localization. The core of the method is the minimization of the Representation Topology Divergence (RTD) between original high-dimensional data and low-dimensional representation in latent space. RTD minimization provides closeness in topological features with strong theoretical guarantees. We develop a scheme for RTD differentiation and apply it as a loss term for the autoencoder. The proposed method "RTD-AE" better preserves the global structure and topology of the data manifold than state-of-the-art competitors as measured by linear correlation, triplet distance ranking accuracy, and Wasserstein distance between persistence barcodes.

* 11th International Conference on Learning Representations (ICLR 2023)  
Viaarxiv icon