학술논문

Domain Invariant Vision Transformer Learning for Face Anti-spoofing
Document Type
Conference
Source
2023 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV) WACV Applications of Computer Vision (WACV), 2023 IEEE/CVF Winter Conference on. :6087-6096 Jan, 2023
Subject
Computing and Processing
Engineering Profession
Computer vision
Protocols
Computational modeling
Aggregates
Transformers
Feature extraction
Data models
Algorithms: Biometrics
face
gesture
body pose
Image recognition and understanding (object detection
categorization
segmentation
scene modeling
visual reasoning)
Language
ISSN
2642-9381
Abstract
Existing face anti-spoofing (FAS) models have achieved high performance on specific datasets. However, for the application of real-world systems, the FAS model should generalize to the data from unknown domains rather than only achieve good results on a single baseline. As vision transformer models have demonstrated astonishing performance and strong capability in learning discriminative information, we investigate applying transformers to distinguish the face presentation attacks over unknown domains. In this work, we propose the Domain-invariant Vision Transformer (DiVT) for FAS, which adopts two losses to improve the generalizability of the vision transformer. First, a concentration loss is employed to learn a domain-invariant representation that aggregates the features of real face data. Second, a separation loss is utilized to union each type of attack from different domains. The experimental results show that our proposed method achieves state-of-the-art performance on the protocols of domain-generalized FAS tasks. Compared to previous domain generalization FAS models, our proposed method is simpler but more effective.