Asymmetric modality translation for face presentation attack detection

Face presentation attack detection (PAD) is an essentialmeasure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the gener...

Full description

Saved in:
Bibliographic Details
Main Authors: Li, Zhi, Li, Haoliang, Luo, Xin, Hu, Yongjian, Lam, Kwok-Yan, Kot, Alex Chichung
Other Authors: School of Computer Science and Engineering
Format: Article
Language:English
Published: 2023
Subjects:
Online Access:https://hdl.handle.net/10356/168025
Tags: Add Tag
No Tags, Be the first to tag this record!
Institution: Nanyang Technological University
Language: English
id sg-ntu-dr.10356-168025
record_format dspace
spelling sg-ntu-dr.10356-1680252023-05-19T15:36:23Z Asymmetric modality translation for face presentation attack detection Li, Zhi Li, Haoliang Luo, Xin Hu, Yongjian Lam, Kwok-Yan Kot, Alex Chichung School of Computer Science and Engineering School of Electrical and Electronic Engineering China-Singapore International Joint Research Institute Engineering::Computer science and engineering Face Presentation Attack Detection Asymmetric Modality Translation Face presentation attack detection (PAD) is an essentialmeasure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the generalization issue of face presentation attack detection under cross-domain settings (e.g., the setting of unseen attacks and varying illumination) remains to be solved. In this paper, we propose a novel framework based on asymmetric modality translation for face presentation attack detection in bi-modality scenarios. Under the framework, we establish connections between two modality images of genuine faces. Specifically, a novel modality fusion scheme is presented that the image of one modality is translated to the other one through an asymmetric modality translator, then fused with its corresponding paired image. The fusion result is fed as the input to a discriminator for inference. The training of the translator is supervised by an asymmetric modality translation loss. Besides, an illumination normalization module based on Pattern of Local Gravitational Force (PLGF) representation is used to reduce the impact of illumination variation. We conduct extensive experiments on three public datasets, which validate that our method is effective in detecting various types of attacks and achieves state-of-the-art performance under different evaluation protocols. Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This work was supported in part by the NTU-PKU\penalty -\@M Joint Research Institute (a collaboration between the Nanyang Technological University and Peking University that is sponsored by a donation from the Ng Teng Fong Charitable Foundation), in part by the Science and Technology Foundation of Guangzhou Huangpu Development District under Grant 2019GH16, in part by the China-Singapore International Joint Research Institute under Grant 206-A018001, and in part by the National Research Foundation, PrimeMinister.s Office, Singapore under its Strategic Capability Research Centres Funding Initiative. 2023-05-19T02:41:16Z 2023-05-19T02:41:16Z 2021 Journal Article Li, Z., Li, H., Luo, X., Hu, Y., Lam, K. & Kot, A. C. (2021). Asymmetric modality translation for face presentation attack detection. IEEE Transactions On Multimedia, 25, 62-76. https://dx.doi.org/10.1109/TMM.2021.3121140 1520-9210 https://hdl.handle.net/10356/168025 10.1109/TMM.2021.3121140 2-s2.0-85118251519 25 62 76 en 2019GH16 206-A018001 IEEE Transactions on Multimedia © 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/TMM.2021.3121140. application/pdf
institution Nanyang Technological University
building NTU Library
continent Asia
country Singapore
Singapore
content_provider NTU Library
collection DR-NTU
language English
topic Engineering::Computer science and engineering
Face Presentation Attack Detection
Asymmetric Modality Translation
spellingShingle Engineering::Computer science and engineering
Face Presentation Attack Detection
Asymmetric Modality Translation
Li, Zhi
Li, Haoliang
Luo, Xin
Hu, Yongjian
Lam, Kwok-Yan
Kot, Alex Chichung
Asymmetric modality translation for face presentation attack detection
description Face presentation attack detection (PAD) is an essentialmeasure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the generalization issue of face presentation attack detection under cross-domain settings (e.g., the setting of unseen attacks and varying illumination) remains to be solved. In this paper, we propose a novel framework based on asymmetric modality translation for face presentation attack detection in bi-modality scenarios. Under the framework, we establish connections between two modality images of genuine faces. Specifically, a novel modality fusion scheme is presented that the image of one modality is translated to the other one through an asymmetric modality translator, then fused with its corresponding paired image. The fusion result is fed as the input to a discriminator for inference. The training of the translator is supervised by an asymmetric modality translation loss. Besides, an illumination normalization module based on Pattern of Local Gravitational Force (PLGF) representation is used to reduce the impact of illumination variation. We conduct extensive experiments on three public datasets, which validate that our method is effective in detecting various types of attacks and achieves state-of-the-art performance under different evaluation protocols.
author2 School of Computer Science and Engineering
author_facet School of Computer Science and Engineering
Li, Zhi
Li, Haoliang
Luo, Xin
Hu, Yongjian
Lam, Kwok-Yan
Kot, Alex Chichung
format Article
author Li, Zhi
Li, Haoliang
Luo, Xin
Hu, Yongjian
Lam, Kwok-Yan
Kot, Alex Chichung
author_sort Li, Zhi
title Asymmetric modality translation for face presentation attack detection
title_short Asymmetric modality translation for face presentation attack detection
title_full Asymmetric modality translation for face presentation attack detection
title_fullStr Asymmetric modality translation for face presentation attack detection
title_full_unstemmed Asymmetric modality translation for face presentation attack detection
title_sort asymmetric modality translation for face presentation attack detection
publishDate 2023
url https://hdl.handle.net/10356/168025
_version_ 1772826996682784768