Alert button
Picture for Arip Asadulaev

Arip Asadulaev

Alert button

Adversarial Training Improves Joint Energy-Based Generative Modelling

Jul 18, 2022
Rostislav Korst, Arip Asadulaev

Figure 1 for Adversarial Training Improves Joint Energy-Based Generative Modelling
Figure 2 for Adversarial Training Improves Joint Energy-Based Generative Modelling

We propose the novel framework for generative modelling using hybrid energy-based models. In our method we combine the interpretable input gradients of the robust classifier and Langevin Dynamics for sampling. Using the adversarial training we improve not only the training stability, but robustness and generative modelling of the joint energy-based models.

Viaarxiv icon

Multi-step domain adaptation by adversarial attack to $\mathcal{H} Δ\mathcal{H}$-divergence

Jul 18, 2022
Arip Asadulaev, Alexander Panfilov, Andrey Filchenkov

Figure 1 for Multi-step domain adaptation by adversarial attack to $\mathcal{H} Δ\mathcal{H}$-divergence
Figure 2 for Multi-step domain adaptation by adversarial attack to $\mathcal{H} Δ\mathcal{H}$-divergence

Adversarial examples are transferable between different models. In our paper, we propose to use this property for multi-step domain adaptation. In unsupervised domain adaptation settings, we demonstrate that replacing the source domain with adversarial examples to $\mathcal{H} \Delta \mathcal{H}$-divergence can improve source classifier accuracy on the target domain. Our method can be connected to most domain adaptation techniques. We conducted a range of experiments and achieved improvement in accuracy on Digits and Office-Home datasets.

Viaarxiv icon

Easy Batch Normalization

Jul 18, 2022
Arip Asadulaev, Alexander Panfilov, Andrey Filchenkov

Figure 1 for Easy Batch Normalization
Figure 2 for Easy Batch Normalization
Figure 3 for Easy Batch Normalization

It was shown that adversarial examples improve object recognition. But what about their opposite side, easy examples? Easy examples are samples that the machine learning model classifies correctly with high confidence. In our paper, we are making the first step toward exploring the potential benefits of using easy examples in the training procedure of neural networks. We propose to use an auxiliary batch normalization for easy examples for the standard and robust accuracy improvement.

Viaarxiv icon

Connecting adversarial attacks and optimal transport for domain adaptation

Jun 04, 2022
Arip Asadulaev, Vitaly Shutov, Alexander Korotin, Alexander Panfilov, Andrey Filchenkov

Figure 1 for Connecting adversarial attacks and optimal transport for domain adaptation
Figure 2 for Connecting adversarial attacks and optimal transport for domain adaptation
Figure 3 for Connecting adversarial attacks and optimal transport for domain adaptation
Figure 4 for Connecting adversarial attacks and optimal transport for domain adaptation

We present a novel algorithm for domain adaptation using optimal transport. In domain adaptation, the goal is to adapt a classifier trained on the source domain samples to the target domain. In our method, we use optimal transport to map target samples to the domain named source fiction. This domain differs from the source but is accurately classified by the source domain classifier. Our main idea is to generate a source fiction by c-cyclically monotone transformation over the target domain. If samples with the same labels in two domains are c-cyclically monotone, the optimal transport map between these domains preserves the class-wise structure, which is the main goal of domain adaptation. To generate a source fiction domain, we propose an algorithm that is based on our finding that adversarial attacks are a c-cyclically monotone transformation of the dataset. We conduct experiments on Digits and Modern Office-31 datasets and achieve improvement in performance for simple discrete optimal transport solvers for all adaptation tasks.

Viaarxiv icon

Neural Optimal Transport with General Cost Functionals

May 30, 2022
Arip Asadulaev, Alexander Korotin, Vage Egiazarian, Evgeny Burnaev

Figure 1 for Neural Optimal Transport with General Cost Functionals
Figure 2 for Neural Optimal Transport with General Cost Functionals
Figure 3 for Neural Optimal Transport with General Cost Functionals
Figure 4 for Neural Optimal Transport with General Cost Functionals

We present a novel neural-networks-based algorithm to compute optimal transport (OT) plans and maps for general cost functionals. The algorithm is based on a saddle point reformulation of the OT problem and generalizes prior OT methods for weak and strong cost functionals. As an application, we construct a functional to map data distributions with preserving the class-wise structure of data.

Viaarxiv icon

Stabilizing Transformer-Based Action Sequence Generation For Q-Learning

Oct 23, 2020
Gideon Stein, Andrey Filchenkov, Arip Asadulaev

Figure 1 for Stabilizing Transformer-Based Action Sequence Generation For Q-Learning
Figure 2 for Stabilizing Transformer-Based Action Sequence Generation For Q-Learning
Figure 3 for Stabilizing Transformer-Based Action Sequence Generation For Q-Learning
Figure 4 for Stabilizing Transformer-Based Action Sequence Generation For Q-Learning

Since the publication of the original Transformer architecture (Vaswani et al. 2017), Transformers revolutionized the field of Natural Language Processing. This, mainly due to their ability to understand timely dependencies better than competing RNN-based architectures. Surprisingly, this architecture change does not affect the field of Reinforcement Learning (RL), even though RNNs are quite popular in RL, and time dependencies are very common in RL. Recently, (Parisotto et al. 2019) conducted the first promising research of Transformers in RL. To support the findings of this work, this paper seeks to provide an additional example of a Transformer-based RL method. Specifically, the goal is a simple Transformer-based Deep Q-Learning method that is stable over several environments. Due to the unstable nature of Transformers and RL, an extensive method search was conducted to arrive at a final method that leverages developments around Transformers as well as Q-learning. The proposed method can match the performance of classic Q-learning on control environments while showing potential on some selected Atari benchmarks. Furthermore, it was critically evaluated to give additional insights into the relation between Transformers and RL.

* Transformers, Reinforcement Learning, 8 pages, AAAI format 
Viaarxiv icon

Wasserstein-2 Generative Networks

Sep 28, 2019
Alexander Korotin, Vage Egiazarian, Arip Asadulaev, Evgeny Burnaev

Figure 1 for Wasserstein-2 Generative Networks
Figure 2 for Wasserstein-2 Generative Networks
Figure 3 for Wasserstein-2 Generative Networks
Figure 4 for Wasserstein-2 Generative Networks

Modern generative learning is mainly associated with Generative Adversarial Networks (GANs). Training such networks is always hard due to the minimax nature of the optimization objective. In this paper we propose a novel algorithm for training generative models, which gets rid of mini-max GAN objective, thus significantly simplified model training. The proposed algorithm uses the variational approximation of Wasserstein-2 distances by Input Convex Neural Networks. We also provide the results of computational experiments, which confirms the efficiency of our algorithm in application to latent spaces optimal transport and image-to-image style transfer.

* 7 pages, 8 figures 
Viaarxiv icon

Backronym

Aug 08, 2019
Arip Asadulaev

The field of Machine Learning research is divided into subject areas, where each area tries to solve a specific problem, using specific methods. In recent years, borders have almost been erased, and many areas inherit methods from other areas. This trend leads to better results and the number of papers in the field is growing every year. The problem is that the amount of information is also growing, and many methods remain unknown in a large number of papers. In this work, we propose the concept of inheritance between machine learning models, which allows conducting research, processing much less information, and pay attention to previously unnoticed models. We hope that this project will allow researchers to find ways to improve their ideas. In addition, it can be used by researchers to publish their methods too. Project is available by link: https://www.infornopolitan.xyz/backronym

Viaarxiv icon

Jacobian Policy Optimizations

Jun 13, 2019
Arip Asadulaev, Gideon Stein, Igor Kuznetsov, Andrey Filchenkov

Figure 1 for Jacobian Policy Optimizations
Figure 2 for Jacobian Policy Optimizations
Figure 3 for Jacobian Policy Optimizations
Figure 4 for Jacobian Policy Optimizations

Recently, natural policy gradient algorithms gained widespread recognition due to their strong performance in reinforcement learning tasks. However, their major drawback is the need to secure the policy being in a ``trust region'' and meanwhile allowing for sufficient exploration. The main objective of this study was to present an approach which models dynamical isometry of agents policies by estimating conditioning of its Jacobian at individual points in the environment space. We present a Jacobian Policy Optimization algorithm for policy optimization, which dynamically adapts the trust interval with respect to policy conditioning. The suggested approach was tested across a range of Atari environments. This paper offers some important insights into an improvement of policy optimization in reinforcement learning tasks.

Viaarxiv icon