Asymmetric modality translation for face presentation attack detection
Face presentation attack detection (PAD) is an essentialmeasure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the gener...
Saved in:
Main Authors: | , , , , , |
---|---|
Other Authors: | |
Format: | Article |
Language: | English |
Published: |
2023
|
Subjects: | |
Online Access: | https://hdl.handle.net/10356/168025 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Institution: | Nanyang Technological University |
Language: | English |
id |
sg-ntu-dr.10356-168025 |
---|---|
record_format |
dspace |
spelling |
sg-ntu-dr.10356-1680252023-05-19T15:36:23Z Asymmetric modality translation for face presentation attack detection Li, Zhi Li, Haoliang Luo, Xin Hu, Yongjian Lam, Kwok-Yan Kot, Alex Chichung School of Computer Science and Engineering School of Electrical and Electronic Engineering China-Singapore International Joint Research Institute Engineering::Computer science and engineering Face Presentation Attack Detection Asymmetric Modality Translation Face presentation attack detection (PAD) is an essentialmeasure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the generalization issue of face presentation attack detection under cross-domain settings (e.g., the setting of unseen attacks and varying illumination) remains to be solved. In this paper, we propose a novel framework based on asymmetric modality translation for face presentation attack detection in bi-modality scenarios. Under the framework, we establish connections between two modality images of genuine faces. Specifically, a novel modality fusion scheme is presented that the image of one modality is translated to the other one through an asymmetric modality translator, then fused with its corresponding paired image. The fusion result is fed as the input to a discriminator for inference. The training of the translator is supervised by an asymmetric modality translation loss. Besides, an illumination normalization module based on Pattern of Local Gravitational Force (PLGF) representation is used to reduce the impact of illumination variation. We conduct extensive experiments on three public datasets, which validate that our method is effective in detecting various types of attacks and achieves state-of-the-art performance under different evaluation protocols. Nanyang Technological University National Research Foundation (NRF) Submitted/Accepted version This work was supported in part by the NTU-PKU\penalty -\@M Joint Research Institute (a collaboration between the Nanyang Technological University and Peking University that is sponsored by a donation from the Ng Teng Fong Charitable Foundation), in part by the Science and Technology Foundation of Guangzhou Huangpu Development District under Grant 2019GH16, in part by the China-Singapore International Joint Research Institute under Grant 206-A018001, and in part by the National Research Foundation, PrimeMinister.s Office, Singapore under its Strategic Capability Research Centres Funding Initiative. 2023-05-19T02:41:16Z 2023-05-19T02:41:16Z 2021 Journal Article Li, Z., Li, H., Luo, X., Hu, Y., Lam, K. & Kot, A. C. (2021). Asymmetric modality translation for face presentation attack detection. IEEE Transactions On Multimedia, 25, 62-76. https://dx.doi.org/10.1109/TMM.2021.3121140 1520-9210 https://hdl.handle.net/10356/168025 10.1109/TMM.2021.3121140 2-s2.0-85118251519 25 62 76 en 2019GH16 206-A018001 IEEE Transactions on Multimedia © 2021 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works. The published version is available at: https://doi.org/10.1109/TMM.2021.3121140. application/pdf |
institution |
Nanyang Technological University |
building |
NTU Library |
continent |
Asia |
country |
Singapore Singapore |
content_provider |
NTU Library |
collection |
DR-NTU |
language |
English |
topic |
Engineering::Computer science and engineering Face Presentation Attack Detection Asymmetric Modality Translation |
spellingShingle |
Engineering::Computer science and engineering Face Presentation Attack Detection Asymmetric Modality Translation Li, Zhi Li, Haoliang Luo, Xin Hu, Yongjian Lam, Kwok-Yan Kot, Alex Chichung Asymmetric modality translation for face presentation attack detection |
description |
Face presentation attack detection (PAD) is an essentialmeasure to protect face recognition systems from being spoofed by malicious users and has attracted great attention from both academia and industry. Although most of the existing methods can achieve desired performance to some extent, the generalization issue of face presentation attack detection under cross-domain settings (e.g., the setting of unseen attacks and varying illumination) remains to be solved. In this paper, we propose a novel framework based on asymmetric modality translation for face presentation attack detection in bi-modality scenarios. Under the framework, we establish connections between two modality images of genuine faces. Specifically, a novel modality fusion scheme is presented that the image of one modality is translated to the other one through an asymmetric modality translator, then fused with its corresponding paired image. The fusion result is fed as the input to a discriminator for inference. The training of the translator is supervised by an asymmetric modality translation loss. Besides, an illumination normalization module based on Pattern of Local Gravitational Force (PLGF) representation is used to reduce the impact of illumination variation. We conduct extensive experiments on three public datasets, which validate that our method is effective in detecting various types of attacks and achieves state-of-the-art performance under different evaluation protocols. |
author2 |
School of Computer Science and Engineering |
author_facet |
School of Computer Science and Engineering Li, Zhi Li, Haoliang Luo, Xin Hu, Yongjian Lam, Kwok-Yan Kot, Alex Chichung |
format |
Article |
author |
Li, Zhi Li, Haoliang Luo, Xin Hu, Yongjian Lam, Kwok-Yan Kot, Alex Chichung |
author_sort |
Li, Zhi |
title |
Asymmetric modality translation for face presentation attack detection |
title_short |
Asymmetric modality translation for face presentation attack detection |
title_full |
Asymmetric modality translation for face presentation attack detection |
title_fullStr |
Asymmetric modality translation for face presentation attack detection |
title_full_unstemmed |
Asymmetric modality translation for face presentation attack detection |
title_sort |
asymmetric modality translation for face presentation attack detection |
publishDate |
2023 |
url |
https://hdl.handle.net/10356/168025 |
_version_ |
1772826996682784768 |