LIPSNNA Light Intrusion-Proving Siamese Neural Network Model for Facial Verification

  1. Asier Alcaide 1
  2. Miguel A. Patricio 2
  3. Antonio Berlanga 2
  4. Angel Arroyo 3
  5. Juan J. Cuadrado Gallego 4
  1. 1 Ultra Tendency International GmbH, Germany
  2. 2 Universidad Carlos III de Madrid
    info

    Universidad Carlos III de Madrid

    Madrid, España

    ROR https://ror.org/03ths8210

  3. 3 Universidad Politécnica de Madrid
    info

    Universidad Politécnica de Madrid

    Madrid, España

    ROR https://ror.org/03n6nwv02

  4. 4 Universidad de Alcalá
    info

    Universidad de Alcalá

    Alcalá de Henares, España

    ROR https://ror.org/04pmn0e78

Revista:
IJIMAI

ISSN: 1989-1660

Año de publicación: 2022

Volumen: 7

Número: 4

Páginas: 121-131

Tipo: Artículo

DOI: 10.9781/IJIMAI.2021.11.003 DIALNET GOOGLE SCHOLAR lock_openDialnet editor

Otras publicaciones en: IJIMAI

Resumen

Facial verification has experienced a breakthrough in recent years, not only due to the improvement in accuracy of the verification systems but also because of their increased use. One of the main reasons for this has been the appearance and use of new models of Deep Learning to address this problem. This extension in the use of facial verification has had a high impact due to the importance of its applications, especially on security, but the extension of its use could be significantly higher if the problem of the required complex calculations needed by the Deep Learning models, that usually need to be executed on machines with specialised hardware, were solved. That would allow the use of facial verification to be extended, making it possible to run this software on computers with low computing resources, such as Smartphones or tablets. To solve this problem, this paper presents the proposal of a new neural model, called Light Intrusion-Proving Siamese Neural Network, LIPSNN. This new light model, which is based on Siamese Neural Networks, is fully presented from the description of its two block architecture, going through its development, including its training with the well- known dataset Labeled Faces in the Wild, LFW; to its benchmarking with other traditional and deep learning models for facial verification in order to compare its performance for its use in low computing resources systems for facial recognition. For this comparison the attribute parameters, storage, accuracy and precision have been used, and from the results obtained it can be concluded that the LIPSNN can be an alternative to the existing models to solve the facet problem of running facial verification in low computing resource devices

Referencias bibliográficas

  • A. Suruliandi, A. Kasthuris, S. P. Raja, “Deep feature representation and similarity matrix based noise label refinement method for efficient face annotation,” International Journal of Interactive Multimedia and Artificial Intelligence, In Press, 2021, doi: 10.9781/ijimai.2021.05.001.
  • Imatest, “Automated image quality analysis,” 2020. [Online]. Available: http://www.imatest.com/products/test-charts-sfrplus/.
  • F. Schroff, D. Kalenichenko, J. Philbin, “FaceNet: A unified embedding for face recognition and clustering,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, vol. 07- 12-June-2015, 2015.
  • Y. Wen, K. Zhang, Z. Li, Y. Qiao, “A discriminative feature learning approach for deep face recognition,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 9911 LNCS, 2016.
  • F. Wang, X. Xiang, J. Cheng, A. L. Yuille, “NormFace: L2 hypersphere embedding for face verification,” in MM 2017 - Proceedings of the 2017 ACM Multimedia Conference, 2017.
  • W. Liu, Y. Wen, Z. Yu, M. Li, B. Raj, L. Song, “SphereFace: Deep hypersphere embedding for face recognition,” in Proceedings - 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, vol. 2017-January, 2017.
  • Y. Taigman, M. Yang, M. Ranzato, L. Wolf, “DeepFace: Closing the gap to human-level performance in face verification,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2014.
  • G. B. Huang, M. Ramesh, T. Berg, E. Learned-Miller, “Labeled faces in the wild: A database for studying face recognition in unconstrained environments,” University of Massachusetts, Amherst, October 2007.
  • G. B. Huang, E. Learned-miller, “Labeled faces in the wild: Updates and new reporting procedures,” University of Massachusetts Amherst Technical Report, 2014.
  • J. Deng, J. Guo, N. Xue, S. Zafeiriou, “Arcface: Additive angular margin loss for deep face recognition,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2019, pp. 4690–4699.
  • Y. Sun, C. Cheng, Y. Zhang, C. Zhang, L. Zheng, Z. Wang, Y. Wei, “Circle loss: A unified perspective of pair similarity optimization,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 6398–6407.
  • G. G. Chrysos, S. Moschoglou, G. Bouritsas, J. Deng, Y. Panagakis, S. P. Zafeiriou, “Deep polynomial neural networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, 2021.
  • U. Jayaraman, P. Gupta, S. Gupta, G. Arora, K. Tiwari, “Recent development in face recognition,” Neurocomputing, 2020, doi: 10.1016/j. neucom.2019.08.110.
  • M. Rastegari, V. Ordonez, J. Redmon, A. Farhadi, “XNOR- Net : ImageNet Classification Using Binary,” Eccv2016, 2016.
  • T. Simons, D. J. Lee, “A review of binarized neural networks,” 2019. doi: 10.3390/electronics8060661.
  • Z. Wang, F. Li, G. Shi, X. Xie, F. Wang, “Network pruning using sparse learning and genetic algorithm,” Neurocomputing, vol. 404, 2020, doi: 10.1016/j.neucom.2020.03.082.
  • F. Tung, G. Mori, “Deep Neural Network Compression by In-Parallel Pruning-Quantization,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 3, 2020, doi: 10.1109/TPAMI.2018.2886192.
  • Z. Liu, J. Li, Z. Shen, G. Huang, S. Yan, C. Zhang, “Learning Efficient Convolutional Networks through Network Slimming,” in Proceedings of the IEEE International Conference on Computer Vision, vol. 2017-October, 2017.
  • Q. Li, S. Jin, J. Yan, “Mimicking Very Efficient Network for Object Detection,” in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2017, pp. 7341–7349.
  • Y. Wei, X. Pan, H. Qin, W. Ouyang, J. Yan, “Quantization Mimic: Towards Very Tiny CNN for Object Detection,” in Computer Vision – ECCV 2018, Cham, 2018, pp. 274–290, Springer International Publishing.
  • S. Chopra, R. Hadsell, Y. LeCun, “Learning a similarity metric discriminatively, with application to face verification,” in Proceedings - 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, CVPR 2005, vol. I, 2005.
  • X. Wang, Y. Zhou, D. Kong, J. Currey, D. Li, J. Zhou, “Unleash the Black Magic in Age: A Multi-Task Deep Neural Network Approach for CrossAge Face Verification,” in 2017 12th IEEE International Conference on Automatic Face Gesture Recognition (FG 2017), 2017, pp. 596–603.
  • C. Reale, N. M. Nasrabadi, H. Kwon, R. Chellappa, “Seeing the Forest from the Trees: A Holistic Approach to Near-Infrared Heterogeneous Face Recognition,” in IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, 2016.
  • C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, Z. Wojna, “Rethinking the inception architecture for computer vision. arxiv 2015,” arXiv preprint arXiv:1512.00567, vol. 1512, 2015.
  • K. Z. He, K. Zhang, “X., ren, s. & sun j. deep residual learning for image recognition,” Preprint at https://arxiv. org/abs/1512.03385, 2015.
  • C. Szegedy, S. Ioffe, V. Vanhoucke, A. A. Alemi, “Inception- v4, inceptionResNet and the impact of residual connections on learning,” in 31st AAAI Conference on Artificial Intelligence, AAAI 2017, 2017.
  • M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, L. C. Chen, “MobileNetV2: Inverted Residuals and Linear Bottlenecks,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2018.
  • A. Howard, M. Sandler, B. Chen, W. Wang, L. C. Chen, M. Tan, G. Chu, V. Vasudevan, Y. Zhu, R. Pang, Q. Le, H. Adam, “Searching for mobileNetV3,” in Proceedings of the IEEE International Conference on Computer Vision, vol. 2019- October, 2019.
  • Z. Mian, W. Hong, “Face verification using gabor wavelets and AdaBoost,” in Proceedings - International Conference on Pattern Recognition, vol. 1, 2006.
  • D. Sandberg, “Face Recognition using Tensorflow.” https://github.com/ davidsandberg/facenet, 2018.
  • Q. Cao, L. Shen, W. Xie, O. M. Parkhi, A. Zisserman, “Vggface2: A dataset for recognising faces across pose and age,” in 2018 13th IEEE international conference on automatic face & gesture recognition (FG 2018), 2018, pp. 67–74, IEEE.
  • N. Silberman, S. Guadarrama, “TensorFlow-Slim image classification model library.” https://github.com/ tensorflow/models/tree/master/ research/slim, 2016.
  • J. Deng, A. Berg, S. Satheesh, H. Su, A. Khosla, F. Li, “Large scale visual recognition challenge 2012 (ilsvrc-2012),” 2012.
  • R. Gomez, “Understanding categorical cross-entropy loss, binary crossentropy loss, softmax loss, logistic loss, focal loss and all those confusing names,” URL: https://gombru.github.io/2018/05/23/cross_entropy_loss/ (visited on 29/03/2019), 2018.
  • S. Bianco, R. Cadene, L. Celona, P. Napoletano, “Benchmark analysis of representative deep neural network architectures,” IEEE Access, vol. 6, 2018, doi: 10.1109/ACCESS.2018.2877890.
  • A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand, M. Andreetto, H. Adam, “MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications,” 2017.
  • J. Hu, L. Shen, S. Albanie, G. Sun, E. Wu, “Squeeze-and-Excitation Networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 42, no. 8, 2020, doi: 10.1109/TPAMI.2019.2913372.
  • G. B. Huang, M. Ramesh, T. Berg, E. Learned-Miller, “Labeled faces in the wild: A database for studying face recognition in unconstrained environments,” University of Massachusetts, Amherst, October 2007.
  • H. Junlin, J. Lu, Y. Junsong, T. Yap-Peng, “Large Margin Multi-metric Learning for Face and Kinship Verification in the Wild,” in Computer Vision – ACCV 2014, Cham, 2015, pp. 252–267, Springer International Publishing.
  • O. Barkan, J. Weill, L. Wolf, H. Aronowitz, “Fast high dimensional vector multiplication face recognition,” in Proceedings of the IEEE International Conference on Computer Vision, 2013.
  • Q. Cao, Y. Ying, P. Li, “Similarity metric learning for face recognition,” in Proceedings of the IEEE International Conference on Computer Vision, 2013.
  • “12th Chinese Conference on Biometric Recognition, CCBR 2017,” 2017.
  • O. M. Parkhi, A. Vedaldi, A. Zisserman, “Deep face recognition,” in Proceedings of the British Machine Vision Conference (BMVC), September 2015, pp. 41.1–41.12, BMVA Press.
  • X. Wu, R. He, Z. Sun, T. Tan, “A light CNN for deep face representation with noisy labels,” IEEE Transactions on Information Forensics and Security, vol. 13, no. 11, 2018, doi: 10.1109/TIFS.2018.2833032.
  • K. Simonyan, A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
  • M. Xi, L. Chen, D. Polajnar, W. Tong, “Local binary pattern network: A deep learning approach for face recognition,” in Proceedings -International Conference on Image Processing, ICIP, vol. 2016-August, 2016.
  • D. Chen, X. Cao, F. Wen, J. Sun, “Blessing of dimensionality: Highdimensional feature and its efficient compression for face verification,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2013.
  • Y. Sun, Y. Chen, X. Wang, X. Tang, “Deep learning face representation by joint identification-verification,” in Advances in Neural Information Processing Systems 27, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, K. Q. Weinberger Eds., Curran Associates, Inc., 2014, pp. 1988–1996