Skip to main content

Human Activity Identification in Smart Daily Environments

  • Chapter
  • First Online:
Smart Assisted Living

Abstract

Research in human activity recognition (HAR) benefits many applications such as intelligent surveillance systems to track humans’ abnormal activities. It could also be applied to robots to understand human activity, which improves smart home efficiency and usability. This chapter aims to accurately recognize different sports types in the Sports Video in the Wild (SVW) dataset employing transfer learning. The dataset consists of noisy and similar classes shot in daily environments, not in controlled laboratory environments. Heretofore, different methods have been used and developed for this purpose. Transfer learning is the process of using pre-trained neural networks. The experimental results on different splits of the dataset, size, and pre-trained models show that accuracy of 80.7% is achievable. In another experiment, we have used the famous UCF101 dataset which is collected from YouTube and trained a convolutional neural network (CNN) with Batch Normalization (BN). The achieved accuracy for the test dataset is around 91.2%. One application of the proposed system is to integrate it with a smart home platform to identify sports activities of individuals and track their progress.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 79.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 99.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD 139.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Aggarwal J, Cai Q (1999) Human motion analysis: a review. Comput Vis Image Underst 73:428–440

    Article  Google Scholar 

  2. Aggarwal JK, Ryoo M (2011) Human activity analysis: a review. ACM Comput Surv (CSUR) 43:1–43

    Article  Google Scholar 

  3. Asperger J, Poore A (2017) Convolutional neural networks for classification of noisy sports videos. s.l., s.n.

    Google Scholar 

  4. Bilbao I, Bilbao J (2017) Overfitting problem and the over-training in the era of data: particularly for artificial neural networks. s.l., s.n., pp 173–177

    Google Scholar 

  5. Bilen H et al (2016) Dynamic image networks for action recognition. s.l., s.n., pp 3034–3042

    Google Scholar 

  6. Bishop C (2006) Pattern recognition and machine learning. NJ Springer, Secaucus

    MATH  Google Scholar 

  7. Blank M (2007) Actions as space-time shapes. Trans Pattern Anal Mach Intell 29:2247–2253

    Article  Google Scholar 

  8. Carreira J, Zisserman A (2017) Quo vadis, action recognition? A new model and the kinetics dataset. s.l., s.n., pp 6299–6308

    Google Scholar 

  9. Chen L, Hoey J, Nugent CD, Cook DJ, Yu Z (2012) Sensor-based activity recognition. IEEE Trans Syst Man Cybern Part C (Appl Rev) 42(6):790–808

    Article  Google Scholar 

  10. Chen L et al (2017) Why batch normalization works? A buckling perspective. s.l., s.n., pp 1184–1189

    Google Scholar 

  11. Chen X (2017) Deep manifold learning combined with convolutional neural networks for action recognition. IEEE Trans Neural Netw Learn Syst 99:1–15

    Google Scholar 

  12. Dalal N, Triggs B, Schmid C (2016) Human detection using oriented histograms of flow and appearance. s.l., s.n., pp 428–441

    Chapter  Google Scholar 

  13. Feichtenhofer C, Pinz A, Wildes R (2016) Spatiotemporal residual networks for video action recognition. s.l., s.n., pp 3468–3476

    Google Scholar 

  14. Feichtenhofer C, Pinz A, Zisserman A (2016) Convolutional two-stream network fusion for video action recognition. s.l., s.n., pp 1933–1941

    Google Scholar 

  15. Gavrila D (1999) The visual analysis of human movement: a survey. Comput Vis Image Underst 73:82–98

    Article  Google Scholar 

  16. Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT Press, s.l.

    MATH  Google Scholar 

  17. Hongeng S, Nevatia R, Bremond F (2004) Video-based event recognition: activity representation and probabilistic recognition methods. Comput Vis Image Underst 96(2):129–162

    Article  Google Scholar 

  18. Ikizler N, Duygulu P (2007) Human action recognition using distribution of oriented rectangular patches. In: Proceedings of th 2nd conference on human motion: understanding modelling, capture and animation, vol 2, pp 271–284

    Google Scholar 

  19. Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167

  20. Jaimes A, Sebe N (2005) Multimodal human computer interaction: a survey. ICCV-HCI, s.l.

    Google Scholar 

  21. Krizhevsky A, Sutskever I, Hinton G (2012) ImageNet classification with deep convolutional neural networks. In: Advances in neural information processing systems, vol 25

    Google Scholar 

  22. KukaÄŤka J, Golkov V, Cremers D (2017) Regularization for deep learning: a taxonomy. arXiv preprint arXiv:1710.10686

  23. Lin Q, Zhang D, Chen L, Ni H, Zhou X (2014) Managing elders’ wandering behavior using sensors-based solutions: a survey. Int J Gerontol 2:49–55

    Article  Google Scholar 

  24. Liu J, Kuipers B, Savarese S (2011) Recognizing human actions by attributes. In: CVPR, pp 3337–3344

    Google Scholar 

  25. Lowe D (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60:91–110

    Article  Google Scholar 

  26. Marszałek M, Laptev I, Schmid C (2009) Actions in context. s.l., s.n., pp 2929–2936

    Google Scholar 

  27. Ni B, Wang G, Moulin P (2011) Rgbd-hudaact: a color-depth video database for human daily activity recognition. In: IEEE international conference on computer vision workshops (ICCV workshops), s.l.

    Google Scholar 

  28. Ojala T, Pietikainen M, Maenpaa T (2002) Multiresolution gray scale and rotation invariant texture classification with local binary patterns. IEEE Trans Pattern Anal Mach Intell 24:971–987

    Article  Google Scholar 

  29. Pantic M, Pentland A, Nijholt A, Huang T (2007) Human computing and machine understanding of human behavior: a survey. Lect Notes Artif Intell 4451:47–71

    Google Scholar 

  30. Poppe R (2010) A survey on vision-based human action recognition. Image Vis Comput 28:976–990

    Article  Google Scholar 

  31. Rachmadi RF, Koutaki G, Uchimura F (2016) Combined convolutional neural network for event recognition. In: The Korea-Japan joint workshop on Frontiers of Computer Vision (FCV)

    Google Scholar 

  32. Rachmadi RF, Koutaki G, Uchimura N (2016) Video classification using compacted dataset based on selected key frames. In: IEEE region 10 conference (TENCON)

    Google Scholar 

  33. Refaeilzadeh P, Tang L, Liu H (2009) Cross-validation. In: Encyclopedia of database systems, pp 532–538

    Google Scholar 

  34. Safdarnejad SM (2015) Sport videos in the wild (SVW): a video dataset for sports analysis. In: Proceedings of international conference on automatic face and gesture recognition

    Google Scholar 

  35. Schuldt C, Laptev I, Caputo B (2004) Recognizing human actions: a local SVM approach. s.l., s.n.

    Google Scholar 

  36. Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. In: Computer vision and pattern recognition

    Google Scholar 

  37. Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: ICLR

    Google Scholar 

  38. Soomro K, Zamir AR, Shah M (2012) UFC101: a dataset of 101 human action classes from videos in the wild. CRCV-TR-12-01

    Google Scholar 

  39. Srivastava N et al (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15:1929–1958

    MathSciNet  MATH  Google Scholar 

  40. Szegedy C, Vanhoucke V, Loffe S, Shlens J (2015) Rethinking the inception architecture for computer vision

    Google Scholar 

  41. Tran D et al (2015) Learning spatiotemporal features with 3d convolutional networks. s.l., s.n., pp 4489–4497

    Google Scholar 

  42. Vrigkas M, Nikou C, Kakadiaris L (2015) A review of human activity recognition methods. Front Robot AI 2:1–28

    Article  Google Scholar 

  43. Wang H, Schmid C (2013) Action recognition with improved trajectories. s.l., s.n., pp 3551–3558

    Google Scholar 

  44. Wang J, Chen Y, Hao S, Peng X, Hu L (2019) Deep learning for sensor-based activity recognition: a survey. Pattern Recognit Lett 3–11

    Article  Google Scholar 

  45. Wang J, Liu Z, Wu Y, Yuan J (2012) Mining actionlet ensemble for action recognition with depth cameras. In: IEEE conference on computer vision and pattern recognition, pp 1290–1297

    Google Scholar 

  46. Wang L, Qiao Y, Tang X (2015) Action recognition with trajectory-pooled deep-convolutional descriptors. s.l., s.n., pp 4305–4314

    Google Scholar 

  47. Wang L et al (2016) Temporal segment networks: towards good practices for deep action recognition. s.l., s.n., pp 20–36

    Google Scholar 

  48. Wrzalik M, Krechel D (2017) Human action recognition using optical flow and convolutional neural networks. In: 16th IEEE international conference on machine learning and applications, pp 801–805

    Google Scholar 

  49. Yamato J, Ohya J, Ishii K (1992) Recognizing human action in time-sequential images using hidden Markov model. In: Proceedings of IEEE conference on computer vision and pattern recognition, pp 379–385

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hossein Malekmohamadi .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2020 Springer Nature Switzerland AG

About this chapter

Check for updates. Verify currency and authenticity via CrossMark

Cite this chapter

Malekmohamadi, H., Pattanjak, N., Bom, R. (2020). Human Activity Identification in Smart Daily Environments. In: Chen, F., García-Betances, R., Chen, L., Cabrera-Umpiérrez, M., Nugent, C. (eds) Smart Assisted Living. Computer Communications and Networks. Springer, Cham. https://doi.org/10.1007/978-3-030-25590-9_5

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-25590-9_5

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-25589-3

  • Online ISBN: 978-3-030-25590-9

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics