Skip to main content

Employing Kaze Features for the Purpose of Emotion Recognition

  • Conference paper
  • First Online:
Progress in Computing, Analytics and Networking

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 710))

Abstract

In this research, a novel approach for emotion detection is exploited by taking the Accelerated Kaze (A-Kaze) features for emotion recognition. The Kaze Features work in a way such that object boundaries can be preserved by making blurring locally adaptive to the image data without severely affecting the noise-reducing capability of the Gaussian blurring, thereby increasing the accuracy of the system. After extracting the Kaze features, GMM is constructed and thus a Fisher Vector representation is made. The extracted features are passed through an SVM detector. An efficiency of 87.5% has been shown thus proving that Kaze can also be used effectively in the field of facial image processing.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. L. S. Chen. Joint processing of audio-visual information for the recognition of emotional expressions in human-computer interaction. PhD thesis, University of Illinois at Urbana- Champaign, Dept. of Electrical Engineering, 2000.

    Google Scholar 

  2. L. S. Chen, H. Tao, T. S. Huang, T. Miyasato, and R. Nakatsu. Emotion recognition from audiovisual information. In Proc. IEEE Workshop on Multimedia Signal Processing, pages 83–88, Los Angeles, CA, USA, Dec. 7–9, 1998. s.

    Google Scholar 

  3. Kaltwang, S., Rudovic, O., & Pantic, M. (2012). Continuous pain intensity estimation from facial expressions. Advances in visual computing, 368–377.

    Google Scholar 

  4. Cohn, Jeffrey F., et al. “Detecting depression from facial actions and vocal prosody.” Affective Computing and Intelligent Interaction and Workshops, 2009. ACII 2009. 3rd International Conference on. IEEE, 2009.

    Google Scholar 

  5. Ekman, P., & Friesen, W. V. (1971). Constants across cultures in the face and emotion. Journal of personality and social psychology, 17(2), 124.

    Google Scholar 

  6. Lucey, P., Cohn, J. F., Kanade, T., Saragih, J., Ambadar, Z., & Matthews, I. (2010, June). The Extended Cohn-Kanade Dataset (CK+): A complete dataset for action unit and emotion-specified expression. In Computer Vision and Pattern Recognition Workshops (CVPRW), 2010 IEEE Computer Society Conference on (pp. 94–101). IEEE.

    Google Scholar 

  7. Viola, Paul, and Michael Jones. “Rapid object detection using a boosted cascade of simple features.” Computer Vision and Pattern Recognition, 2001. CVPR 2001. Proceedings of the 2001 IEEE Computer Society Conference on. Vol. 1. IEEE, 2001.

    Google Scholar 

  8. Rathee, N., Vaish, A., & Gupta, S. (2016, April). Adaptive system to learn and recognize emotional state of mind. In Computing, Communication and Automation (ICCCA), 2016 International Conference on (pp. 32–36). IEEE.

    Google Scholar 

  9. Bay, H., Tuytelaars, T., & Van Gool, L. (2006). Surf: Speeded up robust features. Computer vision–ECCV 2006, 404–417.

    Google Scholar 

  10. Lowe, D. G. (2004). Distinctive image features from scale-invariant keypoints. International journal of computer vision, 60(2), 91–110.

    Google Scholar 

  11. Alcantarilla, P., Bartoli, A., & Davison, A. (2012). KAZE features. Computer Vision–ECCV 2012, 214–227.

    Google Scholar 

  12. Alcantarilla, P. F., & Solutions, T. (2011). Fast explicit diffusion for accelerated features in nonlinear scale spaces. IEEE Trans. Patt. Anal. Mach. Intell, 34(7), 1281–1298.

    Google Scholar 

  13. Alcantarilla, P. F., Bergasa, L. M., Jiménez, P., Sotelo, M. A., Parra, I., Fernandez, D., & Mayoral, S. S. (2008, June). Night time vehicle detection for driving assistance lightbeam controller. In Intelligent Vehicles Symposium, 2008 IEEE (pp. 291–296). IEEE.

    Google Scholar 

  14. Alcantarilla, P. F., Oh, S. M., Mariottini, G. L., Bergasa, L. M., & Dellaert, F. (2010, May). Learning visibility of landmarks for vision-based localization. In Robotics and Automation (ICRA), 2010 IEEE International Conference on (pp. 4881–4888). IEEE.

    Google Scholar 

  15. Li, Wei, et al. “The application of KAZE features to the classification echocardiogram videos.” Multimodal Retrieval in the Medical Domain. Springer International Publishing, 2015. 61–72.

    Google Scholar 

  16. Weickert, Joachim, BM Ter Haar Romeny, and Max A. Viergever. “Efficient and reliable schemes for nonlinear diffusion filtering.” IEEE transactions on image processing 7.3 (1998): 398–410.

    Google Scholar 

  17. Weickert, Joachim. “Efficient image segmentation using partial differential equations and morphology.” Pattern Recognition 34.9 (2001): 1813–1824.

    Google Scholar 

  18. Qiu, Zhen, Lei Yang, and Weiping Lu. “A New Feature-preserving Nonlinear Anisotropic Diffusion Method for Image Denoising.” BMVC. 2011.

    Google Scholar 

  19. Yacoob, Y., and Davis, L., “Recognizing human facial expressions from long image sequences using optical flow,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 18, pp. 636–642, June 1996.

    Google Scholar 

  20. Chang, C. C., & Lin, C. J. (2011). LIBSVM: A library for support vector machines. ACM Transactions on Intelligent Systems and Technology (TIST), 2(3), 27.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Sagar Gupta .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2018 Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Vaish, A., Gupta, S. (2018). Employing Kaze Features for the Purpose of Emotion Recognition. In: Pattnaik, P., Rautaray, S., Das, H., Nayak, J. (eds) Progress in Computing, Analytics and Networking. Advances in Intelligent Systems and Computing, vol 710. Springer, Singapore. https://doi.org/10.1007/978-981-10-7871-2_65

Download citation

  • DOI: https://doi.org/10.1007/978-981-10-7871-2_65

  • Published:

  • Publisher Name: Springer, Singapore

  • Print ISBN: 978-981-10-7870-5

  • Online ISBN: 978-981-10-7871-2

  • eBook Packages: EngineeringEngineering (R0)

Publish with us

Policies and ethics