Alert button
Picture for Huy H. Nguyen

Huy H. Nguyen

Alert button

Face Forgery Detection Based on Facial Region Displacement Trajectory Series

Dec 07, 2022
YuYang Sun, ZhiYong Zhang, Isao Echizen, Huy H. Nguyen, ChangZhen Qiu, Lu Sun

Figure 1 for Face Forgery Detection Based on Facial Region Displacement Trajectory Series
Figure 2 for Face Forgery Detection Based on Facial Region Displacement Trajectory Series
Figure 3 for Face Forgery Detection Based on Facial Region Displacement Trajectory Series
Figure 4 for Face Forgery Detection Based on Facial Region Displacement Trajectory Series

Deep-learning-based technologies such as deepfakes ones have been attracting widespread attention in both society and academia, particularly ones used to synthesize forged face images. These automatic and professional-skill-free face manipulation technologies can be used to replace the face in an original image or video with any target object while maintaining the expression and demeanor. Since human faces are closely related to identity characteristics, maliciously disseminated identity manipulated videos could trigger a crisis of public trust in the media and could even have serious political, social, and legal implications. To effectively detect manipulated videos, we focus on the position offset in the face blending process, resulting from the forced affine transformation of the normalized forged face. We introduce a method for detecting manipulated videos that is based on the trajectory of the facial region displacement. Specifically, we develop a virtual-anchor-based method for extracting the facial trajectory, which can robustly represent displacement information. This information was used to construct a network for exposing multidimensional artifacts in the trajectory sequences of manipulated videos that is based on dual-stream spatial-temporal graph attention and a gated recurrent unit backbone. Testing of our method on various manipulation datasets demonstrated that its accuracy and generalization ability is competitive with that of the leading detection methods.

Viaarxiv icon

Analysis of Master Vein Attacks on Finger Vein Recognition Systems

Oct 18, 2022
Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen

Figure 1 for Analysis of Master Vein Attacks on Finger Vein Recognition Systems
Figure 2 for Analysis of Master Vein Attacks on Finger Vein Recognition Systems
Figure 3 for Analysis of Master Vein Attacks on Finger Vein Recognition Systems
Figure 4 for Analysis of Master Vein Attacks on Finger Vein Recognition Systems

Finger vein recognition (FVR) systems have been commercially used, especially in ATMs, for customer verification. Thus, it is essential to measure their robustness against various attack methods, especially when a hand-crafted FVR system is used without any countermeasure methods. In this paper, we are the first in the literature to introduce master vein attacks in which we craft a vein-looking image so that it can falsely match with as many identities as possible by the FVR systems. We present two methods for generating master veins for use in attacking these systems. The first uses an adaptation of the latent variable evolution algorithm with a proposed generative model (a multi-stage combination of beta-VAE and WGAN-GP models). The second uses an adversarial machine learning attack method to attack a strong surrogate CNN-based recognition system. The two methods can be easily combined to boost their attack ability. Experimental results demonstrated that the proposed methods alone and together achieved false acceptance rates up to 73.29% and 88.79%, respectively, against Miura's hand-crafted FVR system. We also point out that Miura's system is easily compromised by non-vein-looking samples generated by a WGAN-GP model with false acceptance rates up to 94.21%. The results raise the alarm about the robustness of such systems and suggest that master vein attacks should be considered an important security measure.

* Accepted to be Published in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) 2023 
Viaarxiv icon

Rethinking Adversarial Examples for Location Privacy Protection

Jun 28, 2022
Trung-Nghia Le, Ta Gu, Huy H. Nguyen, Isao Echizen

Figure 1 for Rethinking Adversarial Examples for Location Privacy Protection
Figure 2 for Rethinking Adversarial Examples for Location Privacy Protection
Figure 3 for Rethinking Adversarial Examples for Location Privacy Protection
Figure 4 for Rethinking Adversarial Examples for Location Privacy Protection

We have investigated a new application of adversarial examples, namely location privacy protection against landmark recognition systems. We introduce mask-guided multimodal projected gradient descent (MM-PGD), in which adversarial examples are trained on different deep models. Image contents are protected by analyzing the properties of regions to identify the ones most suitable for blending in adversarial examples. We investigated two region identification strategies: class activation map-based MM-PGD, in which the internal behaviors of trained deep models are targeted; and human-vision-based MM-PGD, in which regions that attract less human attention are targeted. Experiments on the Places365 dataset demonstrated that these strategies are potentially effective in defending against black-box landmark recognition systems without the need for much image manipulation.

Viaarxiv icon

Closer Look at the Transferability of Adversarial Examples: How They Fool Different Models Differently

Dec 29, 2021
Futa Waseda, Sosuke Nishikawa, Trung-Nghia Le, Huy H. Nguyen, Isao Echizen

Figure 1 for Closer Look at the Transferability of Adversarial Examples: How They Fool Different Models Differently
Figure 2 for Closer Look at the Transferability of Adversarial Examples: How They Fool Different Models Differently
Figure 3 for Closer Look at the Transferability of Adversarial Examples: How They Fool Different Models Differently
Figure 4 for Closer Look at the Transferability of Adversarial Examples: How They Fool Different Models Differently

Deep neural networks are vulnerable to adversarial examples (AEs), which have adversarial transferability: AEs generated for the source model can mislead another (target) model's predictions. However, the transferability has not been understood from the perspective of to which class target model's predictions were misled (i.e., class-aware transferability). In this paper, we differentiate the cases in which a target model predicts the same wrong class as the source model ("same mistake") or a different wrong class ("different mistake") to analyze and provide an explanation of the mechanism. First, our analysis shows (1) that same mistakes correlate with "non-targeted transferability" and (2) that different mistakes occur between similar models regardless of the perturbation size. Second, we present evidence that the difference in same and different mistakes can be explained by non-robust features, predictive but human-uninterpretable patterns: different mistakes occur when non-robust features in AEs are used differently by models. Non-robust features can thus provide consistent explanations for the class-aware transferability of AEs.

* 6 pages, 8 figures 
Viaarxiv icon

Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio

Dec 03, 2021
Khanh-Duy Nguyen, Huy H. Nguyen, Trung-Nghia Le, Junichi Yamagishi, Isao Echizen

Figure 1 for Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio
Figure 2 for Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio
Figure 3 for Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio
Figure 4 for Effectiveness of Detection-based and Regression-based Approaches for Estimating Mask-Wearing Ratio

Estimating the mask-wearing ratio in public places is important as it enables health authorities to promptly analyze and implement policies. Methods for estimating the mask-wearing ratio on the basis of image analysis have been reported. However, there is still a lack of comprehensive research on both methodologies and datasets. Most recent reports straightforwardly propose estimating the ratio by applying conventional object detection and classification methods. It is feasible to use regression-based approaches to estimate the number of people wearing masks, especially for congested scenes with tiny and occluded faces, but this has not been well studied. A large-scale and well-annotated dataset is still in demand. In this paper, we present two methods for ratio estimation that leverage either a detection-based or regression-based approach. For the detection-based approach, we improved the state-of-the-art face detector, RetinaFace, used to estimate the ratio. For the regression-based approach, we fine-tuned the baseline network, CSRNet, used to estimate the density maps for masked and unmasked faces. We also present the first large-scale dataset, the ``NFM dataset,'' which contains 581,108 face annotations extracted from 18,088 video frames in 17 street-view videos. Experiments demonstrated that the RetinaFace-based method has higher accuracy under various situations and that the CSRNet-based method has a shorter operation time thanks to its compactness.

Viaarxiv icon

Master Face Attacks on Face Recognition Systems

Sep 08, 2021
Huy H. Nguyen, Sébastien Marcel, Junichi Yamagishi, Isao Echizen

Figure 1 for Master Face Attacks on Face Recognition Systems
Figure 2 for Master Face Attacks on Face Recognition Systems
Figure 3 for Master Face Attacks on Face Recognition Systems
Figure 4 for Master Face Attacks on Face Recognition Systems

Face authentication is now widely used, especially on mobile devices, rather than authentication using a personal identification number or an unlock pattern, due to its convenience. It has thus become a tempting target for attackers using a presentation attack. Traditional presentation attacks use facial images or videos of the victim. Previous work has proven the existence of master faces, i.e., faces that match multiple enrolled templates in face recognition systems, and their existence extends the ability of presentation attacks. In this paper, we perform an extensive study on latent variable evolution (LVE), a method commonly used to generate master faces. We run an LVE algorithm for various scenarios and with more than one database and/or face recognition system to study the properties of the master faces and to understand in which conditions strong master faces could be generated. Moreover, through analysis, we hypothesize that master faces come from some dense areas in the embedding spaces of the face recognition systems. Last but not least, simulated presentation attacks using generated master faces generally preserve the false-matching ability of their original digital forms, thus demonstrating that the existence of master faces poses an actual threat.

* This paper is an extension of the IJCB paper published in 2019 (Generating Master Faces for Use in Performing Wolf Attacks on Face Recognition Systems) and its first version was initially submitted to T-BIOM journal on Dec 25, 2020 
Viaarxiv icon

OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild

Jul 30, 2021
Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen

Figure 1 for OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild
Figure 2 for OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild
Figure 3 for OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild
Figure 4 for OpenForensics: Large-Scale Challenging Dataset For Multi-Face Forgery Detection And Segmentation In-The-Wild

The proliferation of deepfake media is raising concerns among the public and relevant authorities. It has become essential to develop countermeasures against forged faces in social media. This paper presents a comprehensive study on two new countermeasure tasks: multi-face forgery detection and segmentation in-the-wild. Localizing forged faces among multiple human faces in unrestricted natural scenes is far more challenging than the traditional deepfake recognition task. To promote these new tasks, we have created the first large-scale dataset posing a high level of challenges that is designed with face-wise rich annotations explicitly for face forgery detection and segmentation, namely OpenForensics. With its rich annotations, our OpenForensics dataset has great potentials for research in both deepfake prevention and general human face detection. We have also developed a suite of benchmarks for these tasks by conducting an extensive evaluation of state-of-the-art instance detection and segmentation methods on our newly constructed dataset in various scenarios. The dataset, benchmark results, codes, and supplementary materials will be publicly available on our project page: https://sites.google.com/view/ltnghia/research/openforensics

* Accepted to ICCV 2021. Project page: https://sites.google.com/view/ltnghia/research/openforensics 
Viaarxiv icon

Fashion-Guided Adversarial Attack on Person Segmentation

Apr 20, 2021
Marc Treu, Trung-Nghia Le, Huy H. Nguyen, Junichi Yamagishi, Isao Echizen

Figure 1 for Fashion-Guided Adversarial Attack on Person Segmentation
Figure 2 for Fashion-Guided Adversarial Attack on Person Segmentation
Figure 3 for Fashion-Guided Adversarial Attack on Person Segmentation
Figure 4 for Fashion-Guided Adversarial Attack on Person Segmentation

This paper presents the first adversarial example based method for attacking human instance segmentation networks, namely person segmentation networks in short, which are harder to fool than classification networks. We propose a novel Fashion-Guided Adversarial Attack (FashionAdv) framework to automatically identify attackable regions in the target image to minimize the effect on image quality. It generates adversarial textures learned from fashion style images and then overlays them on the clothing regions in the original image to make all persons in the image invisible to person segmentation networks. The synthesized adversarial textures are inconspicuous and appear natural to the human eye. The effectiveness of the proposed method is enhanced by robustness training and by jointly attacking multiple components of the target network. Extensive experiments demonstrated the effectiveness of FashionAdv in terms of robustness to image manipulations and storage in cyberspace as well as appearing natural to the human eye. The code and data are publicly released on our project page https://github.com/nii-yamagishilab/fashion_adv

* CVPR Workshops 2021  
* Accepted to Workshop on Media Forensics, CVPR 2021. Project page: https://github.com/nii-yamagishilab/fashion_adv 
Viaarxiv icon