Alert button
Picture for Ivan Oseledets

Ivan Oseledets

Alert button

Unsupervised evaluation of GAN sample quality: Introducing the TTJac Score

Aug 31, 2023
Egor Sevriugov, Ivan Oseledets

Evaluation metrics are essential for assessing the performance of generative models in image synthesis. However, existing metrics often involve high memory and time consumption as they compute the distance between generated samples and real data points. In our study, the new evaluation metric called the "TTJac score" is proposed to measure the fidelity of individual synthesized images in a data-free manner. The study first establishes a theoretical approach to directly evaluate the generated sample density. Then, a method incorporating feature extractors and discrete function approximation through tensor train is introduced to effectively assess the quality of generated samples. Furthermore, the study demonstrates that this new metric can be used to improve the fidelity-variability trade-off when applying the truncation trick. The experimental results of applying the proposed metric to StyleGAN 2 and StyleGAN 2 ADA models on FFHQ, AFHQ-Wild, LSUN-Cars, and LSUN-Horse datasets are presented. The code used in this research will be made publicly available online for the research community to access and utilize.

* 11 pages, 7 figures 
Viaarxiv icon

Robust GAN inversion

Aug 31, 2023
Egor Sevriugov, Ivan Oseledets

Recent advancements in real image editing have been attributed to the exploration of Generative Adversarial Networks (GANs) latent space. However, the main challenge of this procedure is GAN inversion, which aims to map the image to the latent space accurately. Existing methods that work on extended latent space $W+$ are unable to achieve low distortion and high editability simultaneously. To address this issue, we propose an approach which works in native latent space $W$ and tunes the generator network to restore missing image details. We introduce a novel regularization strategy with learnable coefficients obtained by training randomized StyleGAN 2 model - WRanGAN. This method outperforms traditional approaches in terms of reconstruction quality and computational efficiency, achieving the lowest distortion with 4 times fewer parameters. Furthermore, we observe a slight improvement in the quality of constructing hyperplanes corresponding to binary image attributes. We demonstrate the effectiveness of our approach on two complex datasets: Flickr-Faces-HQ and LSUN Church.

* 22 pages, 28 figures 
Viaarxiv icon

Quantization Aware Factorization for Deep Neural Network Compression

Aug 08, 2023
Daria Cherniuk, Stanislav Abukhovich, Anh-Huy Phan, Ivan Oseledets, Andrzej Cichocki, Julia Gusak

Figure 1 for Quantization Aware Factorization for Deep Neural Network Compression
Figure 2 for Quantization Aware Factorization for Deep Neural Network Compression
Figure 3 for Quantization Aware Factorization for Deep Neural Network Compression
Figure 4 for Quantization Aware Factorization for Deep Neural Network Compression

Tensor decomposition of convolutional and fully-connected layers is an effective way to reduce parameters and FLOP in neural networks. Due to memory and power consumption limitations of mobile or embedded devices, the quantization step is usually necessary when pre-trained models are deployed. A conventional post-training quantization approach applied to networks with decomposed weights yields a drop in accuracy. This motivated us to develop an algorithm that finds tensor approximation directly with quantized factors and thus benefit from both compression techniques while keeping the prediction quality of the model. Namely, we propose to use Alternating Direction Method of Multipliers (ADMM) for Canonical Polyadic (CP) decomposition with factors whose elements lie on a specified quantization grid. We compress neural network weights with a devised algorithm and evaluate it's prediction quality and performance. We compare our approach to state-of-the-art post-training quantization methods and demonstrate competitive results and high flexibility in achiving a desirable quality-performance tradeoff.

Viaarxiv icon

Efficient GPT Model Pre-training using Tensor Train Matrix Representation

Jun 05, 2023
Viktoriia Chekalina, Georgii Novikov, Julia Gusak, Ivan Oseledets, Alexander Panchenko

Figure 1 for Efficient GPT Model Pre-training using Tensor Train Matrix Representation
Figure 2 for Efficient GPT Model Pre-training using Tensor Train Matrix Representation
Figure 3 for Efficient GPT Model Pre-training using Tensor Train Matrix Representation
Figure 4 for Efficient GPT Model Pre-training using Tensor Train Matrix Representation

Large-scale transformer models have shown remarkable performance in language modelling tasks. However, such models feature billions of parameters, leading to difficulties in their deployment and prohibitive training costs from scratch. To reduce the number of the parameters in the GPT-2 architecture, we replace the matrices of fully-connected layers with the corresponding Tensor Train Matrix~(TTM) structure. Finally, we customize forward and backward operations through the TTM-based layer for simplicity and the stableness of further training. % The resulting GPT-2-based model stores up to 40% fewer parameters, showing the perplexity comparable to the original model. On the downstream tasks, including language understanding and text summarization, the model performs similarly to the original GPT-2 model. The proposed tensorized layers could be used to efficiently pre-training other Transformer models.

Viaarxiv icon

Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning

May 31, 2023
Marina Munkhoeva, Ivan Oseledets

Figure 1 for Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning
Figure 2 for Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning
Figure 3 for Bridging Spectral Embedding and Matrix Completion in Self-Supervised Learning

Self-supervised methods received tremendous attention thanks to their seemingly heuristic approach to learning representations that respect the semantics of the data without any apparent supervision in the form of labels. A growing body of literature is already being published in an attempt to build a coherent and theoretically grounded understanding of the workings of a zoo of losses used in modern self-supervised representation learning methods. In this paper, we attempt to provide an understanding from the perspective of a Laplace operator and connect the inductive bias stemming from the augmentation process to a low-rank matrix completion problem. To this end, we leverage the results from low-rank matrix completion to provide theoretical analysis on the convergence of modern SSL methods and a key property that affects their downstream performance.

* 13 pages, 3 figures 
Viaarxiv icon

TT-SDF2PC: Registration of Point Cloud and Compressed SDF Directly in the Memory-Efficient Tensor Train Domain

Apr 11, 2023
Alexey I. Boyko, Anastasiia Kornilova, Rahim Tariverdizadeh, Mirfarid Musavian, Larisa Markeeva, Ivan Oseledets, Gonzalo Ferrer

Figure 1 for TT-SDF2PC: Registration of Point Cloud and Compressed SDF Directly in the Memory-Efficient Tensor Train Domain
Figure 2 for TT-SDF2PC: Registration of Point Cloud and Compressed SDF Directly in the Memory-Efficient Tensor Train Domain
Figure 3 for TT-SDF2PC: Registration of Point Cloud and Compressed SDF Directly in the Memory-Efficient Tensor Train Domain
Figure 4 for TT-SDF2PC: Registration of Point Cloud and Compressed SDF Directly in the Memory-Efficient Tensor Train Domain

This paper addresses the following research question: ``can one compress a detailed 3D representation and use it directly for point cloud registration?''. Map compression of the scene can be achieved by the tensor train (TT) decomposition of the signed distance function (SDF) representation. It regulates the amount of data reduced by the so-called TT-ranks. Using this representation we have proposed an algorithm, the TT-SDF2PC, that is capable of directly registering a PC to the compressed SDF by making use of efficient calculations of its derivatives in the TT domain, saving computations and memory. We compare TT-SDF2PC with SOTA local and global registration methods in a synthetic dataset and a real dataset and show on par performance while requiring significantly less resources.

Viaarxiv icon

Translate your gibberish: black-box adversarial attack on machine translation systems

Mar 20, 2023
Andrei Chertkov, Olga Tsymboi, Mikhail Pautov, Ivan Oseledets

Figure 1 for Translate your gibberish: black-box adversarial attack on machine translation systems
Figure 2 for Translate your gibberish: black-box adversarial attack on machine translation systems
Figure 3 for Translate your gibberish: black-box adversarial attack on machine translation systems
Figure 4 for Translate your gibberish: black-box adversarial attack on machine translation systems

Neural networks are deployed widely in natural language processing tasks on the industrial scale, and perhaps the most often they are used as compounds of automatic machine translation systems. In this work, we present a simple approach to fool state-of-the-art machine translation tools in the task of translation from Russian to English and vice versa. Using a novel black-box gradient-free tensor-based optimizer, we show that many online translation tools, such as Google, DeepL, and Yandex, may both produce wrong or offensive translations for nonsensical adversarial input queries and refuse to translate seemingly benign input phrases. This vulnerability may interfere with understanding a new language and simply worsen the user's experience while using machine translation systems, and, hence, additional improvements of these tools are required to establish better translation.

Viaarxiv icon

Multiparticle Kalman filter for object localization in symmetric environments

Mar 14, 2023
Roman Korkin, Ivan Oseledets, Aleksandr Katrutsa

Figure 1 for Multiparticle Kalman filter for object localization in symmetric environments
Figure 2 for Multiparticle Kalman filter for object localization in symmetric environments
Figure 3 for Multiparticle Kalman filter for object localization in symmetric environments
Figure 4 for Multiparticle Kalman filter for object localization in symmetric environments

This study considers the object localization problem and proposes a novel multiparticle Kalman filter to solve it in complex and symmetric environments. Two well-known classes of filtering algorithms to solve the localization problem are Kalman filter-based methods and particle filter-based methods. We consider these classes, demonstrate their complementary properties, and propose a novel filtering algorithm that takes the best from two classes. We evaluate the multiparticle Kalman filter in symmetric and noisy environments. Such environments are especially challenging for both classes of classical methods. We compare the proposed approach with the particle filter since only this method is feasible if the initial state is unknown. In the considered challenging environments, our method outperforms the particle filter in terms of both localization error and runtime.

Viaarxiv icon

Machine learning methods for prediction of breakthrough curves in reactive porous media

Jan 12, 2023
Daria Fokina, Pavel Toktaliev, Oleg Iliev, Ivan Oseledets

Figure 1 for Machine learning methods for prediction of breakthrough curves in reactive porous media
Figure 2 for Machine learning methods for prediction of breakthrough curves in reactive porous media
Figure 3 for Machine learning methods for prediction of breakthrough curves in reactive porous media
Figure 4 for Machine learning methods for prediction of breakthrough curves in reactive porous media

Reactive flows in porous media play an important role in our life and are crucial for many industrial, environmental and biomedical applications. Very often the concentration of the species at the inlet is known, and the so-called breakthrough curves, measured at the outlet, are the quantities which could be measured or computed numerically. The measurements and the simulations could be time-consuming and expensive, and machine learning and Big Data approaches can help to predict breakthrough curves at lower costs. Machine learning (ML) methods, such as Gaussian processes and fully-connected neural networks, and a tensor method, cross approximation, are well suited for predicting breakthrough curves. In this paper, we demonstrate their performance in the case of pore scale reactive flow in catalytic filters.

Viaarxiv icon

Mitigating Human and Computer Opinion Fraud via Contrastive Learning

Jan 08, 2023
Yuliya Tukmacheva, Ivan Oseledets, Evgeny Frolov

Figure 1 for Mitigating Human and Computer Opinion Fraud via Contrastive Learning
Figure 2 for Mitigating Human and Computer Opinion Fraud via Contrastive Learning
Figure 3 for Mitigating Human and Computer Opinion Fraud via Contrastive Learning

We introduce the novel approach towards fake text reviews detection in collaborative filtering recommender systems. The existing algorithms concentrate on detecting the fake reviews, generated by language models and ignore the texts, written by dishonest users, mostly for monetary gains. We propose the contrastive learning-based architecture, which utilizes the user demographic characteristics, along with the text reviews, as the additional evidence against fakes. This way, we are able to account for two different types of fake reviews spamming and make the recommendation system more robust to biased reviews.

* 15 pages, 3 figures, 1 table 
Viaarxiv icon