학술논문

Face-Periocular Cross-Identification via Contrastive Hybrid Attention Vision Transformer
Document Type
Periodical
Source
IEEE Signal Processing Letters IEEE Signal Process. Lett. Signal Processing Letters, IEEE. 30:254-258 2023
Subject
Signal Processing and Analysis
Computing and Processing
Communication, Networking and Broadcast Technologies
Face recognition
Probes
Faces
Transformers
Training
Feature extraction
Biological system modeling
Biometrics cross-identification
face-periocular contrastive learning
conv-based attention mechanism
vision transformer
Language
ISSN
1070-9908
1558-2361
Abstract
Traditional biometrics identification performs matching between probe and gallery that may involve the same single or multiple biometric modalities. This letter presents a cross-matching scenario where the probe and gallery are from two distinctive biometrics, i.e., face and periocular, coined as face-periocular cross-identification (FPCI). We propose a novel contrastive loss tailored for face-periocular cross-matching to learn a joint embedding, which can be used as a gallery or a probe regardless of the biometric modality. On the other hand, a hybrid attention vision transformer is devised. The hybrid attention module performs depth-wise convolution and conv-based multi-head self-attention in parallel to aggregate global and local features of the face and periocular biometrics. Extensive experiments on three benchmark datasets demonstrate that our model sufficiently improves the performance of FPCI. Besides that, a new face-periocular dataset in the wild, the Cross-modal Face-periocular dataset, is developed for the FPCI models training.