Journal of Real-Time Image Processing

, Volume 16, Issue 1, pp 15–29 | Cite as

Deepgender: real-time gender classification using deep learning for smartphones

  • Khurram Zeeshan Haider
  • Kaleem Razzaq MalikEmail author
  • Shehzad Khalid
  • Tabassam Nawaz
  • Sohail Jabbar
Special Issue Paper


Face recognition, expression identification, age determination, racial binding and gender classification are common examples of image processing computerization. Gender classification is very straightforward for us like we can tell by the person’s hair, nose, eyes, mouth and skin whether that person is a male or female with a relatively high degree of confidence and accuracy; however, can we program a computer to perform just as well at gender classification? The very problem is the main focus of this research. The conventional sequence for recent real-time facial image processing consists of five steps: face detection, noise removal, face alignment, feature representation and classification. With the aim of human gender classification, face alignment and feature vector extraction stages have been re-examined keeping in view the application of the system on smartphones. Face alignment has been made by spotting out 83 facial landmarks and 3-D facial model with the purpose of applying affine transformation. Furthermore, ‘feature representation’ is prepared through proposed modification in multilayer deep neural network, and hence we name it Deepgender. This convolutional deep neural network consists of some locally connected hidden layers without common weights of kernels as previously followed in legacy layered architecture. This specific case study involves deep learning as four convolutional layers, three max-pool layers (for downsizing of unrelated data), two fully connected layers (connection of outcome to all inputs) and a single layer of ‘multinomial logistic regression.’ Training has been made using CAS-PEAL and FEI which contain 99,594 face images of 1040 people and 2800 face images of 200 individuals, respectively. These images are either in different poses or taken under uncontrolled conditions which are close to real-time input facial image for gender classification application. The proposed system ‘Deepgender’ has registered 98% accuracy by combined use of both databases with the specific preprocess procedure, i.e., exhibiting alignment before resizing. Experiments suggest that accuracy is nearly 100% with frontal and nonblurred facial images. State-of-the-art steps have been taken to overcome memory and battery constraints in mobiles.


Real-time systems Machine learning Image processing Deep learning Gender classification Advanced mobile applications 


  1. 1.
    Abuarqoub, A., Hammoudeh, M., Alsboui, T.: An overview of information extraction from mobile wireless sensor networks. In: Andreev S., Balandin S., Koucheryavy Y. (eds.) Internet of Things, Smart Spaces, and Next Generation Networking, vol 7496, pp. 95–106. Springer, Berlin (2012)Google Scholar
  2. 2.
    Khalid, S., Sajjad, S., Jabbar, S., Chang, H.: Accurate and efficient shape matching approach using vocabularies of multi-feature space representations. J. Real-Time Image Process. 1–17 (2015)Google Scholar
  3. 3.
    Khalid, S., Arshad, S., Jabbar, S., Rho, S.: Robust framework to combine diverse classifiers assigning distributed confidence to individual classifiers at class level. Sci. World J. 2014, 492387 (2014). doi: 10.1155/2014/492387
  4. 4.
    Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., Fei-Fei, L.: Large-scale video classification with convolutional neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, 2014. IEEE (2014)Google Scholar
  5. 5.
    Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: NIPS'12 Proceedings of the 25th International Conference on Neural Information Processing Systems, pp. 1097–1105. Curran Associates Inc, USA (2012)Google Scholar
  6. 6.
    Song, I., Kim, H.-J., and Jeon, P.B.: Deep learning for real-time robust facial expression recognition on a smartphone. In: IEEE International Conference on Consumer Electronics (ICCE), 2014. IEEE (2014)Google Scholar
  7. 7.
    Meseguer, J.E., Calafate, C.T., Cano, J.C., Manzoni, P.: Drivingstyles: a smartphone application to assess driver behavior. In: IEEE Symposium on Computers and Communications (ISCC), 2013. IEEE (2013)Google Scholar
  8. 8.
    Kim, Y.-D., Park, E., Yoo, S., Choi, T., Yang, L., Shin, D.: Compression of deep convolutional neural networks for fast and low power mobile applications. arXiv preprint arXiv:1511.06530, 2015
  9. 9.
    Awais, M., Palmerini, L., Chiari, L.: Physical activity classification using body-worn inertial sensors in a multi-sensor setup. In: IEEE 2nd International Forum on Research and Technologies for Society and Industry Leveraging a better tomorrow (RTSI), 2016. IEEE (2016)Google Scholar
  10. 10.
    Del Rosario, M.B., Wang, K., Wang, J., Liu, Y., Brodie, M., Delbaere, K., Lovell, N.H., Lord, S.R., Redmond, S.J.: A comparison of activity classification in younger and older cohorts using a smartphone. Physiol. Meas. 35(11), 2269 (2014)CrossRefGoogle Scholar
  11. 11.
    Osadchy, M., Cun, Y.L., Miller, M.L.: Synergistic face detection and pose estimation with energy-based models. J. Mach. Learn. Res. 8(May), 1197–1215 (2007)Google Scholar
  12. 12.
    Sun, Y., Wang, X., Tang, X.: Deep convolutional network cascade for facial point detection. In: 2013 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE (2013)Google Scholar
  13. 13.
    Huang, G.B.: Learning hierarchical representations for face verification with convolutional deep belief networks. In: CVPR '12 Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 2518–2525. IEEE Computer Society, Washington, DC (2012)Google Scholar
  14. 14.
    Fang, Y., Wang, Z.: Improving LBP features for gender classification. In: International Conference on Wavelet Analysis and Pattern Recognition, 2008. ICWAPR '08. IEEE (2008)Google Scholar
  15. 15.
    Zhang, N., Paluri, M., Ranzato, M.A., Darrell, T., Bourdev, L.: Panda: pose aligned networks for deep attribute modeling. (2014).
  16. 16.
    Hu, S.Y.D., Jou, B., Jaech, A., Savvides, M.: Fusion of region-based representations for gender identification. In: International Joint Conference on Biometrics (IJCB), 2011. IEEE (2011)Google Scholar
  17. 17.
    Rai, P., Khanna, P.: A gender classification system robust to occlusion using Gabor features based (2D) 2 PCA. J. Vis. Commun. Image Represent. 25(5), 1118–1129 (2014)CrossRefGoogle Scholar
  18. 18.
    Fazl-Ersi, E., Mousa-Pasandi, M.E., Laganiere, R., Awad, M.: Age and gender recognition using informative features of various types. In: IEEE International Conference on Image Processing (ICIP), 2014. IEEE (2014)Google Scholar
  19. 19.
    Leng, X., Wang, Y.: Improving generalization for gender classification. In: 15th IEEE International Conference on Image Processing, 2008. ICIP 2008. IEEE (2008)Google Scholar
  20. 20.
    Chen, D.-Y., Lin, K.-Y.: Robust gender recognition for uncontrolled environment of real-life images. IEEE Trans. Consum. Electron. 56(3) (2010)Google Scholar
  21. 21.
    Juefei-Xu, F., Verma, E., Goel, P., Cherodian, A., Savvides, M.: Deepgender: occlusion and low resolution robust facial gender classification via progressively trained convolutional neural networks with attention. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), 2016. IEEE (2016)Google Scholar
  22. 22.
    Biswas, S., Sil, J.: Gender recognition using fusion of spatial and temporal features. In: Kumar Kundu M., Mohapatra D., Konar A., Chakraborty A. (eds.) Advanced Computing, Networking and Informatics, vol. 1, pp. 109–116. Springer, Berlin (2014)Google Scholar
  23. 23.
    Mansanet, J., Albiol, A., Paredes, R.: Local deep neural networks for gender recognition. Pattern Recogn. Lett. 70, 80–86 (2016)CrossRefGoogle Scholar
  24. 24.
    Zhang, K., Tan, L., Li, Z., Qiao, Y.: Gender and smile classification using deep convolutional neural networks. In: Book Gender and Smile Classification Using Deep Convolutional Neural Networks, pp. 34–38. (2016)Google Scholar
  25. 25.
    Jiang, Y., Li, S., Liu, P., Dai, Q.: Multi-feature deep learning for face gender recognition. In: Book Multi-Feature Deep Learning for Face Gender Recognition, pp. 507–511. (2014)Google Scholar
  26. 26.
    Barkan, O., Weill, J., Wolf, L., Aronowitz, H.: Fast high dimensional vector multiplication face recognition. In: Book Fast High Dimensional Vector Multiplication Face Recognition, pp. 1960–1967. (2013)Google Scholar
  27. 27.
    Hassner, T.: Viewing real-world faces in 3D. In: Book Viewing Real-World Faces in 3D, pp. 3607–3614. (2013)Google Scholar
  28. 28.
    Berg, T., Belhumeur, P.N.: Tom-vs-Pete classifiers and identity-preserving alignment for face verification. In: Book Tom-vs-Pete Classifiers and Identity-Preserving Alignment for Face Verification, pp. 7. (2012)Google Scholar
  29. 29.
    Huang, G., Mattar, M., Lee, H., Learned-Miller, E.G.: Learning to align from scratch. In: Book Learning to Align from Scratch, pp. 764–772. (2012)Google Scholar
  30. 30.
    Zhu, X., Ramanan, D.: Face detection, pose estimation, and landmark localization in the wild. In: Book Face Detection, Pose Estimation, and Landmark Localization in the Wild, pp. 2879–2886. (2012)Google Scholar
  31. 31.
    Cao, X., Wei, Y., Wen, F., Sun, J.: Face alignment by explicit shape regression. Int. J. Comput. vis. 107(2), 177–190 (2014)MathSciNetCrossRefGoogle Scholar
  32. 32.
    Chen, H., Li, J., Zhang, F., Li, Y., Wang, H.: 3D model-based continuous emotion recognition. In: Book 3D Model-Based Continuous Emotion Recognition, pp. 1836–1845. (2015)Google Scholar
  33. 33.
    Levi, G., Hassner, T.: Age and gender classification using convolutional neural networks. In: Book Age and Gender Classification Using Convolutional Neural Networks, pp. 34–42. (2015)Google Scholar
  34. 34.
    Hammoudeh, M., Newman, R.: Information extraction from sensor networks using the Watershed transform algorithm. Inf. Fusion 22, 39–49 (2015)CrossRefGoogle Scholar
  35. 35.
    Abuarqoub, A., Hammoudeh, M., Adebisi, B., Jabbar, S., Bounceur, A., Al-Bashar, H.: Dynamic clustering and management of mobile wireless sensor networks. Comput. Netw. 117, 62–75 (2017)CrossRefGoogle Scholar
  36. 36.
    Gao, W., Cao, B., Shan, S., Chen, X., Zhou, D., Zhang, X., Zhao, D.: The CAS-PEAL large-scale Chinese face database and baseline evaluations. IEEE Trans. Syst. Man Cybern. A Syst. Hum. 38(1), 149–161 (2008)CrossRefGoogle Scholar
  37. 37.
    Thomaz, C.E., Giraldi, G.A.: A new ranking method for principal components analysis and its application to face image analysis. Image Vis. Comput. 28(6), 902–913 (2010)CrossRefGoogle Scholar
  38. 38.
    Rahimi, M.R., Ren, J., Liu, C.H., Vasilakos, A.V., Venkatasubramanian, N.: Mobile cloud computing: a survey, state of art and future directions. Mobile Netw. Appl. 19(2), 133–143 (2014)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag GmbH Germany 2017

Authors and Affiliations

  • Khurram Zeeshan Haider
    • 1
  • Kaleem Razzaq Malik
    • 2
    Email author
  • Shehzad Khalid
    • 3
  • Tabassam Nawaz
    • 4
  • Sohail Jabbar
    • 5
  1. 1.Government College University FaisalabadFaisalabadPakistan
  2. 2.Department of Computer ScienceCOMSATS Institute of Information TechnologySahiwalPakistan
  3. 3.Department of Computer EngineeringBahria UniversityIslamabadPakistan
  4. 4.University of Engineering and TechnologyTaxilaPakistan
  5. 5.Department of Computer ScienceNational Textile UniversityFaisalabadPakistan

Personalised recommendations