Improving Cross-database Face Presentation Attack Detection via Adversarial Domain Adaptation
Guoqing Wang1,3, Hu Han∗,1,2, Shiguang Shan1,2,3,4, and Xilin Chen1,3
1Key Laboratory of Intelligent Information Processing of Chinese Academy of Sciences (CAS),
Institute of Computing Technology, CAS, Beijing 100190, China
2Peng Cheng Laboratory, Shenzhen, China 3University of Chinese Academy of Sciences, Beijing 100049, China 4CAS Center for Excellence in Brain Science and Intelligence Technology, Shanghai, China
{guoqing.wang}@vipl.ict.ac.cn, {hanhu, sgshan, xlchen}@ict.ac.cn
Abstract
Face recognition (FR) is being widely used in many ap- plications from access control to smartphone unlock. As a result, face presentation attack detection (PAD) has drawn increasing attentions to secure the FR systems. Tradi- tional approaches for PAD mainly assume that training and testing scenarios are similar in imaging conditions (illu- mination, scene, camera sensor, etc.), and thus may lack good generalization capability into new application sce-
- narios. In this work, we propose an end-to-end learning
approach to improve PAD generalization capability by uti- lizing prior knowledge from source domain via adversarial domain adaptation. We first build a source domain PAD model optimized with triplet loss. Subsequently, we perform adversarial domain adaptation w.r.t. the target domain to learn a shared embedding space by both the source and tar- get domain models, in which the discriminator cannot reli- ably predict whether a sample is from the source or target
- domain. Finally, PAD in the target domain is performed
with k-nearest neighbors (k-NN) classifier in the embedding
- space. The proposed approach shows promising general-
ization capability in a number of public-domain face PAD databases.
- 1. Introduction
Biometric technologies such as FR are widely used in
- ur daily life, e.g., in smartphone unlock, access control,
and payment. It is well known that most of existing FR
∗Corresponding author.
978-1-7281-3640-0/19/$31.00 c 2019 IEEE
Figure 1. 2D visualization of the genuine and spoof face images from CASIA [35] and Idiap [6] with deeply learned features by ResNet-18 [12]. (a) The model trained on CASIA is tested (used for feature extraction) on CASIA (intra-database testing). (b) The model trained on CASIA is tested (used for feature extraction) on Idiap (cross-database testing). We observe that a model trained on the source domain does not generalize well to the target domain.
systems [28] are vulnerable to face presentation attacks (PA), e.g., a printed face on paper (print attack), replay- ing a face video on a screen (replay attack), wearing a face mask (3D mask attack), etc. Since an authorized user’s face images can be easily obtained with a smartphone camera
- r from social media, which can be used for launching at-