Skip to main content

Human Action Recognition in Video via Fused Optical Flow and Moment Features – Towards a Hierarchical Approach to Complex Scenario Recognition

  • Conference paper
MultiMedia Modeling (MMM 2014)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 8326))

Included in the following conference series:

Abstract

This paper explores using motion features for human action recognition in video, as the first step towards hierarchical complex event detection for surveillance and security. We compensate for the low resolution and noise, characteristic of many CCTV modalities, by generating optical flow feature descriptors which view motion vectors as a global representation of the scene as opposed to a set of pixel-wise measurements. Specifically, we combine existing optical flow features with a set of moment-based features which not only capture the orientation of motion within each video scene, but incorporate spatial information regarding the relative locations of directed optical flow magnitudes. Our evaluation, using a benchmark dataset, considers their diagnostic capability when recognizing human actions under varying feature set parameterizations and signal-to-noise ratios. The results show that human actions can be recognized with mean accuracy across all actions of 93.3%. Furthermore, we illustrate that precision degrades less in low signal-to -noise images when our moments-based features are utilized.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. The SAVASA project, http://www.savasa.eu

  2. Yang, J.B., Liu, J., Sii, H.S., Wang, H.W.: Belief rule-base inference methodology using the evidential reasoning approach—RIMER. IEEE Trans. Syst. Man Cybern. 36(2), 266–284 (2006)

    Google Scholar 

  3. Efros, A., Berg, A.C., Mori, G., Malik, J.: Recognizing action at a distance. In: Proceedings of the Ninth IEEE International Conference on Computer Vision, pp. 726–733 (2003)

    Google Scholar 

  4. Aggarwal, J., Ryoo, M.S.: Human Activity Analysis: A Review. ACM Computing Surveys 43(3), 1–43 (2011)

    Article  Google Scholar 

  5. Regazzoni, C., Cavallaro, A., Wu, Y., Konrad, J., Hampapur, A.: Video analytics for surveillance: theory and practice. IEEE Signal Processing Magazine 5, 16–17 (2010)

    Article  Google Scholar 

  6. Ziani, A., Motamed, C.: Temporal Bayesian Networks for Scenario Recognition. In: Ersbøll, B.K., Pedersen, K.S. (eds.) SCIA 2007. LNCS, vol. 4522, pp. 689–698. Springer, Heidelberg (2007)

    Chapter  Google Scholar 

  7. Ziani, A., Motamed, C., Noyer, J.: Temporal reasoning for scenario recognition in video-surveillance using Bayesian networks. Computer Vision 2(2), 99–107 (2008)

    Google Scholar 

  8. Vu, V., Bremond, F., Thonnat, M.: Automatic video interpretation: a novel algorithm for temporal scenario recognition. In: Proceedings of the 18th International Joint Conference on Artificial Intelligence (IJCAI 2003), Acapulco, Mexico, pp. 523–533 (2003)

    Google Scholar 

  9. Rodriguez, M., Ahamed, J., Shah, M.: Action MACH: A spatio-temporal maximum average correlation height filter for action recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, Los Alamitos (2008)

    Google Scholar 

  10. Wang, H., Klaser, A., Schmid, C., Liu, C.: Action recognition using dense trajectories. In: IEEE Conference on Computer Vision & Pattern Recognition, pp. 3169–3176 (2011)

    Google Scholar 

  11. Hongeng, S., Nevatia, R.: Large scale event detection using semi-hidden Markov models. In: Proceedings of the International Conference on Computer Vision, vol. 2, pp. 1455–1462 (2003)

    Google Scholar 

  12. Szarvas, M., Sakai, U., Ogata, J.: Real-time pedestrian detection using LIDAR and convolutional neural networks. In: IEEE Intelligent Vehicles Symposium, pp. 213–218 (2006)

    Google Scholar 

  13. Lucas, B., Kanade, T.: An Iterative Image Registration Technique with an Application to Stereo Vision. In: Proceedings of 7th International Conference on Artificial Intelligence, pp. 121–130

    Google Scholar 

  14. Wu, H., Sankaranarayanan, A., Chellappa, R.: Online Empirical Evaluation of Tracking Algorithms. IEEE Transactions on Pattern Analysis and Machine Intelligence 32(8), 1443–1458 (2009)

    Article  Google Scholar 

  15. Tomasi, C., Kanade, T.: Detection and tracking of point Features, Technical Report CMU-CS-91-132, Carnegie Mellon University (1991)

    Google Scholar 

  16. Hu, M.: Visual pattern recognition by moment invariants. IRE Transaction Information Theory IT-8(2), 179–187 (1962)

    Google Scholar 

  17. Mukundan, R., Rmakrishnan, K.: Moments functions in image analysis theory and applications. World Scientific Publishing, Singapore (1998)

    Book  Google Scholar 

  18. The, C., Chin, R.: On image analysis by the methods of moments. IEEE Transactions on Patten Analysis Machine Intelligence 10(4), 16–19 (2004)

    Google Scholar 

  19. Schuldt, C., Laptev, I., Caputo, B.: Recognizing human actions: a local svm approach. In: Proceedings of the International Conference on Pattern Recognition, Cambridge (2004)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2014 Springer International Publishing Switzerland

About this paper

Cite this paper

Clawson, K., Jing, M., Scotney, B., Wang, H., Liu, J. (2014). Human Action Recognition in Video via Fused Optical Flow and Moment Features – Towards a Hierarchical Approach to Complex Scenario Recognition. In: Gurrin, C., Hopfgartner, F., Hurst, W., Johansen, H., Lee, H., O’Connor, N. (eds) MultiMedia Modeling. MMM 2014. Lecture Notes in Computer Science, vol 8326. Springer, Cham. https://doi.org/10.1007/978-3-319-04117-9_10

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-04117-9_10

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-04116-2

  • Online ISBN: 978-3-319-04117-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics