Analyzing ConvNets Depth for Deep Face Recognition

  • Mohan Raj
  • I. Gogul
  • M. Deepan Raj
  • V. Sathiesh Kumar
  • V. Vaidehi
  • S. Sibi Chakkaravarthy
Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 703)

Abstract

Deep convolutional neural networks are becoming increasingly popular in large-scale image recognition, classification, localization, and detection. In this paper, the performance of state-of-the-art convolution neural networks (ConvNets) models of the ImageNet challenge (ILSVRC), namely VGG16, VGG19, OverFeat, ResNet50, and Inception-v3 which achieved top-5 error rates up to 4.2% are analyzed in the context of face recognition. Instead of using handcrafted feature extraction techniques which requires a domain-level understanding, ConvNets have the advantages of automatically learning complex features, more training time, and less evaluation time. These models are benchmarked on AR and Extended Yale B face dataset with five performance metrics, namely Precision, Recall, F1-score, Rank-1 accuracy, and Rank-5 accuracy. It is found that GoogleNet ConvNets model with Inception-v3 architecture outperforms than other four architectures with a Rank-1 accuracy of 98.46% on AR face dataset and 97.94% accuracy on Extended Yale B face dataset. It confirms that deep CNN architectures are suitable for real-time face recognition in the future.

Keywords

Deep learning Face recognition Convolutional neural networks Computer vision 

Notes

Acknowledgements

This research project is supported by DAE-BRNS, Department of Atomic Energy, Government of India. The authors would like to extend their sincere thanks to DAE-BRNS for their support.

References

  1. 1.
    T. Ojala, M. Pietikäinen and T. Mäenpää, “Multiresolution Gray Scale and Rotation Invariant Texture Classification with Local Binary Patterns”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 7, pp. 971–987, 2002.Google Scholar
  2. 2.
    D.G. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints”, International Journal of Computer Vision, 91–110.  https://doi.org/10.1023/b:visi.0000029664.99615.94, 2004.
  3. 3.
    N. Dalal and B. Triggs, “Histograms of Oriented Gradients for Human Detection”, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Vol. 1, 886–893, 2005.Google Scholar
  4. 4.
    J. Yang, Y.G. Jiang, A.G. Hauptmann, and C.W. Ngo, “Evaluating Bag-of-Visual Words Representations in Scene Classification”, Proceedings of the International Workshop on Multimedia Information Retrieval (ACM), 197–206, 2007.Google Scholar
  5. 5.
    A. Krizhevsky, I. Sutskever, and G.E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks”, Conference on Neural Information Processing Systems (NIPS), pp. 1106–1114, 2012.Google Scholar
  6. 6.
    K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition”, International Conference on Learning Representations (ICLR), 2015.Google Scholar
  7. 7.
    C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going Deeper with Convolutions”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015.Google Scholar
  8. 8.
    P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun, “Overfeat: Integrated Recognition, Localization and Detection using Convolutional Networks”, International Conference on Learning Representations (ICLR), 2014.Google Scholar
  9. 9.
    K. He, X. Zhang, S. Ren, J. Sun, “Deep Residual Learning for Image Recognition”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.Google Scholar
  10. 10.
    Z. Cao, Q. Yin, X. Tang, and J. Sun, “Face Recognition with Learning based Descriptor”, Proceedings of Computer Vision and Pattern Recognition (CVPR), 2010.Google Scholar
  11. 11.
    P. Li, S. Prince, Y. Fu, U. Mohammed, and J. Elder. “Probabilistic Models for Inference about Identity”, IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 2012.Google Scholar
  12. 12.
    T. Berg and P. Belhumeur, “Tom-vs-Pete Classifiers and Identity preserving Alignment for Face Verification”, Proceedings of British Machine Vision Conference (BMVC), 2012.Google Scholar
  13. 13.
    F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A Unified Embedding for Face Recognition and Clustering”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015.Google Scholar
  14. 14.
    Y. Taigman, M. Yang, M. Ranzato, and L. Wolf, “Deep-Face: Closing the Gap to Human Level Performance in Face Verification”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2014.Google Scholar
  15. 15.
    A. Romero, N. Ballas, S.E. Kahou, A. Chassang, C. GattaandY. Bengio, “Fitnets: Hints for Thin Deep Nets”, arXiv:1412.6550, 2014.
  16. 16.
    J. Ba and R. Caruana, “Do deep nets really need to be deep?”, Advances in Neural Information Processing Systems 27, arXiv:1312.6184, 2014.
  17. 17.
    M.D. Zeiler, and R. Fergus, “Visualizing and Understanding Convolutional Networks”, European Conference on Computer Vision (ECCV), 2014.Google Scholar
  18. 18.
    A.M. Martinez and R. Benavente, “The AR Face Database”, CVC Technical Report #24, June 1998.Google Scholar
  19. 19.
    “The Extended Yale Face Database B”, available online: http://vision.ucsd.edu/~leekc/ExtYaleDatabase/ExtYaleB.html.

Copyright information

© Springer Nature Singapore Pte Ltd. 2018

Authors and Affiliations

  • Mohan Raj
    • 1
  • I. Gogul
    • 1
  • M. Deepan Raj
    • 1
  • V. Sathiesh Kumar
    • 1
  • V. Vaidehi
    • 2
  • S. Sibi Chakkaravarthy
    • 1
  1. 1.Department of Electronics EngineeringMadras Institute of Technology Campus, Anna UniversityChennaiIndia
  2. 2.School of Computing Science and EngineeringVIT UniversityChennaiIndia

Personalised recommendations