Multi-modal face anti-spoofing based on central difference networks
Yu, Zitong; Qin, Yunxiao; Li, Xiaobai; Wang, Zezheng; Zhao, Chenxu; Lei, Zhen; Zhao, Guoying (2020-07-28)
Z. Yu et al., "Multi-Modal Face Anti-Spoofing Based on Central Difference Networks," 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 2020, pp. 2766-2774, doi: 10.1109/CVPRW50498.2020.00333
© 2020 IEEE. Personal use of this material is permitted. Permission from IEEE must be obtained for all other uses, in any current or future media, including reprinting/republishing this material for advertising or promotional purposes, creating new collective works, for resale or redistribution to servers or lists, or reuse of any copyrighted component of this work in other works.
https://rightsstatements.org/vocab/InC/1.0/
https://urn.fi/URN:NBN:fi-fe202102195358
Tiivistelmä
Abstract
Face anti-spoofing (FAS) plays a vital role in securing face recognition systems from presentation attacks. Existing multi-modal FAS methods rely on stacked vanilla convolutions, which is weak in describing detailed intrinsic information from modalities and easily being ineffective when the domain shifts (e.g., cross attack and cross ethnicity). In this paper, we extend the central difference convolutional networks (CDCN) [39] to a multimodal version, intending to capture intrinsic spoofing patterns among three modalities (RGB, depth and infrared). Meanwhile, we also give an elaborate study about singlemodal based CDCN. Our approach won the first place in "Track Multi-Modal" as well as the second place in “Track Single-Modal (RGB)” of ChaLearn Face Antispoofing Attack Detection Challenge@CVPR2020 [20]. Our final submission obtains 1.02±0.59% and 4.84±1.79% ACER in “Track Multi-Modal” and “Track Single-Modal (RGB)”, respectively. The codes are available at https://github.com/ZitongYu/CDCN.
Kokoelmat
- Avoin saatavuus [32130]