Alert button
Picture for Andrei-Timotei Ardelean

Andrei-Timotei Ardelean

Alert button

High-Fidelity Zero-Shot Texture Anomaly Localization Using Feature Correspondence Analysis

Apr 13, 2023
Andrei-Timotei Ardelean, Tim Weyrich

Figure 1 for High-Fidelity Zero-Shot Texture Anomaly Localization Using Feature Correspondence Analysis
Figure 2 for High-Fidelity Zero-Shot Texture Anomaly Localization Using Feature Correspondence Analysis
Figure 3 for High-Fidelity Zero-Shot Texture Anomaly Localization Using Feature Correspondence Analysis
Figure 4 for High-Fidelity Zero-Shot Texture Anomaly Localization Using Feature Correspondence Analysis

We propose a novel method for Zero-Shot Anomaly Localization that leverages a bidirectional mapping derived from the 1-dimensional Wasserstein Distance. The proposed approach allows pinpointing the anomalous regions in a texture with increased precision by aggregating the contribution of a pixel to the errors of all nearby patches. We validate our solution on several datasets and obtain more than a 40% reduction in error over the previous state of the art on the MVTec AD dataset in a zero-shot setting.

Viaarxiv icon

NPBG++: Accelerating Neural Point-Based Graphics

Mar 24, 2022
Ruslan Rakhimov, Andrei-Timotei Ardelean, Victor Lempitsky, Evgeny Burnaev

Figure 1 for NPBG++: Accelerating Neural Point-Based Graphics
Figure 2 for NPBG++: Accelerating Neural Point-Based Graphics
Figure 3 for NPBG++: Accelerating Neural Point-Based Graphics
Figure 4 for NPBG++: Accelerating Neural Point-Based Graphics

We present a new system (NPBG++) for the novel view synthesis (NVS) task that achieves high rendering realism with low scene fitting time. Our method efficiently leverages the multiview observations and the point cloud of a static scene to predict a neural descriptor for each point, improving upon the pipeline of Neural Point-Based Graphics in several important ways. By predicting the descriptors with a single pass through the source images, we lift the requirement of per-scene optimization while also making the neural descriptors view-dependent and more suitable for scenes with strong non-Lambertian effects. In our comparisons, the proposed system outperforms previous NVS approaches in terms of fitting and rendering runtimes while producing images of similar quality.

* Accepted to CVPR 2022. The project page: https://rakhimovv.github.io/npbgpp/ 
Viaarxiv icon

Multi-sensor large-scale dataset for multi-view 3D reconstruction

Mar 11, 2022
Oleg Voynov, Gleb Bobrovskikh, Pavel Karpyshev, Andrei-Timotei Ardelean, Arseniy Bozhenko, Saveliy Galochkin, Ekaterina Karmanova, Pavel Kopanev, Yaroslav Labutin-Rymsho, Ruslan Rakhimov, Aleksandr Safin, Valerii Serpiva, Alexey Artemov, Evgeny Burnaev, Dzmitry Tsetserukou, Denis Zorin

Figure 1 for Multi-sensor large-scale dataset for multi-view 3D reconstruction
Figure 2 for Multi-sensor large-scale dataset for multi-view 3D reconstruction
Figure 3 for Multi-sensor large-scale dataset for multi-view 3D reconstruction
Figure 4 for Multi-sensor large-scale dataset for multi-view 3D reconstruction

We present a new multi-sensor dataset for 3D surface reconstruction. It includes registered RGB and depth data from sensors of different resolutions and modalities: smartphones, Intel RealSense, Microsoft Kinect, industrial cameras, and structured-light scanner. The data for each scene is obtained under a large number of lighting conditions, and the scenes are selected to emphasize a diverse set of material properties challenging for existing algorithms. In the acquisition process, we aimed to maximize high-resolution depth data quality for challenging cases, to provide reliable ground truth for learning algorithms. Overall, we provide over 1.4 million images of 110 different scenes acquired at 14 lighting conditions from 100 viewing directions. We expect our dataset will be useful for evaluation and training of 3D reconstruction algorithms of different types and for other related tasks. Our dataset and accompanying software will be available online.

Viaarxiv icon

Pose Manipulation with Identity Preservation

Apr 20, 2020
Andrei-Timotei Ardelean, Lucian Mircea Sasu

Figure 1 for Pose Manipulation with Identity Preservation
Figure 2 for Pose Manipulation with Identity Preservation
Figure 3 for Pose Manipulation with Identity Preservation
Figure 4 for Pose Manipulation with Identity Preservation

This paper describes a new model which generates images in novel poses e.g. by altering face expression and orientation, from just a few instances of a human subject. Unlike previous approaches which require large datasets of a specific person for training, our approach may start from a scarce set of images, even from a single image. To this end, we introduce Character Adaptive Identity Normalization GAN (CainGAN) which uses spatial characteristic features extracted by an embedder and combined across source images. The identity information is propagated throughout the network by applying conditional normalization. After extensive adversarial training, CainGAN receives figures of faces from a certain individual and produces new ones while preserving the person's identity. Experimental results show that the quality of generated images scales with the size of the input set used during inference. Furthermore, quantitative measurements indicate that CainGAN performs better compared to other methods when training data is limited.

* International Journal of Computers Communications & Control, Vol 15, Nr 2, 3862, 2020  
* 9 pages, journal article 
Viaarxiv icon