TY - GEN
T1 - Analyzing the Feature Extractor Networks for Face Image Synthesis
AU - Saritas, Erdi
AU - Ekenel, Hazim Kemal
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Advancements like Generative Adversarial Networks have attracted the attention of researchers toward face image synthesis to generate ever more realistic images. Thereby, the need for the evaluation criteria to assess the realism of the generated images has become apparent. While FID utilized with InceptionV3 is one of the primary choices for benchmarking, concerns about InceptionV3 's limitations for face images have emerged. This study investigates the behavior of diverse feature extractors - InceptionV3, CLIP, DINOv2, and ArcFace - considering a variety of metrics - FID, KID, Precision&Recall. While the FFHQ dataset is used as the target domain, as the source domains, the CelebA-HQ dataset and the synthetic datasets generated using Style-GAN2 and Projected FastGAN are used. Experiments include deep-down analysis of the features: L2 normalization, model attention during extraction, and domain distributions in the feature space. We aim to give valuable insights into the behavior of feature extractors for evaluating face image synthesis methodologies. The code is publicly available at https://github.com/ThEnded32/AnalyzingFeatureExtractors.
AB - Advancements like Generative Adversarial Networks have attracted the attention of researchers toward face image synthesis to generate ever more realistic images. Thereby, the need for the evaluation criteria to assess the realism of the generated images has become apparent. While FID utilized with InceptionV3 is one of the primary choices for benchmarking, concerns about InceptionV3 's limitations for face images have emerged. This study investigates the behavior of diverse feature extractors - InceptionV3, CLIP, DINOv2, and ArcFace - considering a variety of metrics - FID, KID, Precision&Recall. While the FFHQ dataset is used as the target domain, as the source domains, the CelebA-HQ dataset and the synthetic datasets generated using Style-GAN2 and Projected FastGAN are used. Experiments include deep-down analysis of the features: L2 normalization, model attention during extraction, and domain distributions in the feature space. We aim to give valuable insights into the behavior of feature extractors for evaluating face image synthesis methodologies. The code is publicly available at https://github.com/ThEnded32/AnalyzingFeatureExtractors.
UR - http://www.scopus.com/inward/record.url?scp=85199485481&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85199485481&partnerID=8YFLogxK
U2 - 10.1109/FG59268.2024.10581922
DO - 10.1109/FG59268.2024.10581922
M3 - Conference contribution
AN - SCOPUS:85199485481
T3 - 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition, FG 2024
BT - 2024 IEEE 18th International Conference on Automatic Face and Gesture Recognition, FG 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 18th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2024
Y2 - 27 May 2024 through 31 May 2024
ER -