TY - JOUR
T1 - Multimodal soft biometrics
T2 - combining ear and face biometrics for age and gender classification
AU - Yaman, Dogucan
AU - Eyiokur, Fevziye Irem
AU - Ekenel, Hazım Kemal
N1 - Publisher Copyright:
© 2021, The Author(s), under exclusive licence to Springer Science+Business Media, LLC part of Springer Nature.
PY - 2022/7
Y1 - 2022/7
N2 - In this paper, we present a multimodal, multitask deep convolutional neural network framework for age and gender classification. In the developed framework, we have employed two different biometric modalities: ear and profile face. We have explored three different fusion methods, namely, data, feature, and score fusion, to combine the information extracted from ear and profile face images. In the framework, we have utilized VGG-16 and ResNet-50 models with center loss to obtain more discriminative features. Moreover, we have performed two-stage fine-tuning to increase the representation capacity of the models. To assess the performance of the proposed approach, we have conducted extensive experiments on the FERET, UND-F, and UND-J2 datasets. Experimental results indicate that ear and profile face images contain useful features to extract soft biometric traits. We have shown that when frontal face view of the subject is not available, use of ear and profile face images can be a good alternative for the soft biometric recognition systems. The presented multimodal system achieves very high age and gender classification accuracies, matching the ones obtained by using frontal face images. The multimodal approach has outperformed both the unimodal approaches and the previous state-of-the-art profile face image or ear image-based age and gender classification methods, significantly in both tasks.
AB - In this paper, we present a multimodal, multitask deep convolutional neural network framework for age and gender classification. In the developed framework, we have employed two different biometric modalities: ear and profile face. We have explored three different fusion methods, namely, data, feature, and score fusion, to combine the information extracted from ear and profile face images. In the framework, we have utilized VGG-16 and ResNet-50 models with center loss to obtain more discriminative features. Moreover, we have performed two-stage fine-tuning to increase the representation capacity of the models. To assess the performance of the proposed approach, we have conducted extensive experiments on the FERET, UND-F, and UND-J2 datasets. Experimental results indicate that ear and profile face images contain useful features to extract soft biometric traits. We have shown that when frontal face view of the subject is not available, use of ear and profile face images can be a good alternative for the soft biometric recognition systems. The presented multimodal system achieves very high age and gender classification accuracies, matching the ones obtained by using frontal face images. The multimodal approach has outperformed both the unimodal approaches and the previous state-of-the-art profile face image or ear image-based age and gender classification methods, significantly in both tasks.
KW - Age estimation
KW - Convolutional neural networks
KW - Gender classification
KW - Multimodal learning
KW - Multitask learning
KW - Soft biometrics
UR - http://www.scopus.com/inward/record.url?scp=85102815470&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85102815470&partnerID=8YFLogxK
U2 - 10.1007/s11042-021-10630-8
DO - 10.1007/s11042-021-10630-8
M3 - Article
AN - SCOPUS:85102815470
SN - 1380-7501
VL - 81
SP - 22695
EP - 22713
JO - Multimedia Tools and Applications
JF - Multimedia Tools and Applications
IS - 16
ER -