Abstract
Deep convolutional neural networks are becoming increasingly popular in large-scale image recognition, classification, localization, and detection. In this paper, the performance of state-of-the-art convolution neural networks (ConvNets) models of the ImageNet challenge (ILSVRC), namely VGG16, VGG19, OverFeat, ResNet50, and Inception-v3 which achieved top-5 error rates up to 4.2% are analyzed in the context of face recognition. Instead of using handcrafted feature extraction techniques which requires a domain-level understanding, ConvNets have the advantages of automatically learning complex features, more training time, and less evaluation time. These models are benchmarked on AR and Extended Yale B face dataset with five performance metrics, namely Precision, Recall, F1-score, Rank-1 accuracy, and Rank-5 accuracy. It is found that GoogleNet ConvNets model with Inception-v3 architecture outperforms than other four architectures with a Rank-1 accuracy of 98.46% on AR face dataset and 97.94% accuracy on Extended Yale B face dataset. It confirms that deep CNN architectures are suitable for real-time face recognition in the future.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
T. Ojala, M. Pietikäinen and T. Mäenpää, “Multiresolution Gray Scale and Rotation Invariant Texture Classification with Local Binary Patterns”, IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 7, pp. 971–987, 2002.
D.G. Lowe, “Distinctive Image Features from Scale-Invariant Keypoints”, International Journal of Computer Vision, 91–110. https://doi.org/10.1023/b:visi.0000029664.99615.94, 2004.
N. Dalal and B. Triggs, “Histograms of Oriented Gradients for Human Detection”, IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR), Vol. 1, 886–893, 2005.
J. Yang, Y.G. Jiang, A.G. Hauptmann, and C.W. Ngo, “Evaluating Bag-of-Visual Words Representations in Scene Classification”, Proceedings of the International Workshop on Multimedia Information Retrieval (ACM), 197–206, 2007.
A. Krizhevsky, I. Sutskever, and G.E. Hinton, “ImageNet Classification with Deep Convolutional Neural Networks”, Conference on Neural Information Processing Systems (NIPS), pp. 1106–1114, 2012.
K. Simonyan and A. Zisserman, “Very Deep Convolutional Networks for Large-Scale Image Recognition”, International Conference on Learning Representations (ICLR), 2015.
C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan, V. Vanhoucke, and A. Rabinovich, “Going Deeper with Convolutions”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
P. Sermanet, D. Eigen, X. Zhang, M. Mathieu, R. Fergus, and Y. LeCun, “Overfeat: Integrated Recognition, Localization and Detection using Convolutional Networks”, International Conference on Learning Representations (ICLR), 2014.
K. He, X. Zhang, S. Ren, J. Sun, “Deep Residual Learning for Image Recognition”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2016.
Z. Cao, Q. Yin, X. Tang, and J. Sun, “Face Recognition with Learning based Descriptor”, Proceedings of Computer Vision and Pattern Recognition (CVPR), 2010.
P. Li, S. Prince, Y. Fu, U. Mohammed, and J. Elder. “Probabilistic Models for Inference about Identity”, IEEE Transactions on Pattern Analysis and Machine Intelligence (PAMI), 2012.
T. Berg and P. Belhumeur, “Tom-vs-Pete Classifiers and Identity preserving Alignment for Face Verification”, Proceedings of British Machine Vision Conference (BMVC), 2012.
F. Schroff, D. Kalenichenko, and J. Philbin, “Facenet: A Unified Embedding for Face Recognition and Clustering”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2015.
Y. Taigman, M. Yang, M. Ranzato, and L. Wolf, “Deep-Face: Closing the Gap to Human Level Performance in Face Verification”, IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 2014.
A. Romero, N. Ballas, S.E. Kahou, A. Chassang, C. GattaandY. Bengio, “Fitnets: Hints for Thin Deep Nets”, arXiv:1412.6550, 2014.
J. Ba and R. Caruana, “Do deep nets really need to be deep?”, Advances in Neural Information Processing Systems 27, arXiv:1312.6184, 2014.
M.D. Zeiler, and R. Fergus, “Visualizing and Understanding Convolutional Networks”, European Conference on Computer Vision (ECCV), 2014.
A.M. Martinez and R. Benavente, “The AR Face Database”, CVC Technical Report #24, June 1998.
“The Extended Yale Face Database B”, available online: http://vision.ucsd.edu/~leekc/ExtYaleDatabase/ExtYaleB.html.
Acknowledgements
This research project is supported by DAE-BRNS, Department of Atomic Energy, Government of India. The authors would like to extend their sincere thanks to DAE-BRNS for their support.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Raj, M., Gogul, I., Deepan Raj, M., Sathiesh Kumar, V., Vaidehi, V., Sibi Chakkaravarthy, S. (2018). Analyzing ConvNets Depth for Deep Face Recognition. In: Chaudhuri, B., Kankanhalli, M., Raman, B. (eds) Proceedings of 2nd International Conference on Computer Vision & Image Processing . Advances in Intelligent Systems and Computing, vol 703. Springer, Singapore. https://doi.org/10.1007/978-981-10-7895-8_25
Download citation
DOI: https://doi.org/10.1007/978-981-10-7895-8_25
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-10-7894-1
Online ISBN: 978-981-10-7895-8
eBook Packages: EngineeringEngineering (R0)