Skip to main content

Person Invariant Classification of Subtle Facial Expressions Using Coded Movement Direction of Keypoints

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNIP,volume 10165))

Abstract

This paper describes a person invariant method for classifying subtle facial expressions. The method uses keypoints detected by using a face tracking tool called Face Tracker. It describes features such as coded movements of keypoints and uses them for classification. Its classification accuracy was evaluated using the facial images of unlearned people. The results showed the average F-measure was 0.93 for neutral (expressionless) facial images, 0.73 for subtle smile images, and 0.92 for exaggerated smile images. Also, person invariant accuracy was evaluated by using F-measure frequency of unlearned people. The results revealed that the proposed method has higher person invariant accuracy than the previous methods.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Kothe, C.A., Makeig, S., Onton, J.A.: Emotion recognition from EEG during self-paced emotional imagery. In: Humaine Association Conference on Affective Computing and Intelligent Interaction (ACII2013), pp. 855–858 (2013)

    Google Scholar 

  2. Chen, J., Chen, Z., Chi, Z., Fu, H.: Facial expression recognition based on facial components detection and HOG features. In: Scientific Cooperations International Workshops on Electrical and Computer Engineering Subfields, pp. 64–69 (2014)

    Google Scholar 

  3. Shan, C., Gong, S., McOwan, P.: Facial expression recognition based on local binary patterns: a comprehensive study. Image Vis. Comput. 27(6), 803–816 (2009)

    Article  Google Scholar 

  4. Owusu, E., Zhan, Y., Mao, Q.R.: A neural-AdaBoost based facial expression recognition system. Expect Syst. Appl. 41(7), 3383–3390 (2014)

    Article  Google Scholar 

  5. Majumder, A., Behera, L., Subramanian, V.K.: Emotion recognition from geometric facial features using self-organizing map. Pattern Recogn. 47(3), 1282–1293 (2014)

    Article  Google Scholar 

  6. Kotsia, I., Pitas, I.: Facial expression recognition in image sequences using geometric deformation features and support vector machines. IEEE Trans. Image Process. 16, 172–187 (2007)

    Article  MathSciNet  Google Scholar 

  7. Matsuhisa, H., Hashimoto, M.: Identifying subtle facial expression changes using optimized gabor features. J. Inst. Image Inform. Telev. Eng. 68(6), J252–J255 (2014). (in Japanese)

    Google Scholar 

  8. Nomiya, H., Hochin, T.: Efficient emotional video scene detection based on ensemble learning. IEICE Trans. Inform. Syst. J95-D(2), 193–205 (2012). (in Japanese)

    Google Scholar 

  9. Ahmed, F., Bari, H., Hossain, E.: Person-independent facial expression recognition based on compound local binary pattern (CLBP). IAJIT 11(2), 195–203 (2014)

    Google Scholar 

  10. Ekman, P., Friesen, W.V.: Unmasking the Face. Malor Books

    Google Scholar 

  11. Saragih, J., Lucey, S., Cohn, J.: Deformable model fitting by regularized landmark mean-shifts. Int. J. Comput. Vision 91(1), 200–215 (2011)

    Article  MathSciNet  MATH  Google Scholar 

  12. Kennedy, B., Balint, A.: EmotionNet. github. https://github.com/co60ca/EmotionNet

Download references

Acknowledgments

This research is partially supported by the Center of Innovation Program from Japan Science and Technology Agency, JST.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Kosuke Sasaki .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2017 Springer International Publishing AG

About this paper

Cite this paper

Sasaki, K., Hashimoto, M., Nagata, N. (2017). Person Invariant Classification of Subtle Facial Expressions Using Coded Movement Direction of Keypoints. In: Nasrollahi, K., et al. Video Analytics. Face and Facial Expression Recognition and Audience Measurement. VAAM FFER 2016 2016. Lecture Notes in Computer Science(), vol 10165. Springer, Cham. https://doi.org/10.1007/978-3-319-56687-0_6

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-56687-0_6

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-56686-3

  • Online ISBN: 978-3-319-56687-0

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics