Abstract
Research in human activity recognition (HAR) benefits many applications such as intelligent surveillance systems to track humans’ abnormal activities. It could also be applied to robots to understand human activity, which improves smart home efficiency and usability. This chapter aims to accurately recognize different sports types in the Sports Video in the Wild (SVW) dataset employing transfer learning. The dataset consists of noisy and similar classes shot in daily environments, not in controlled laboratory environments. Heretofore, different methods have been used and developed for this purpose. Transfer learning is the process of using pre-trained neural networks. The experimental results on different splits of the dataset, size, and pre-trained models show that accuracy of 80.7% is achievable. In another experiment, we have used the famous UCF101 dataset which is collected from YouTube and trained a convolutional neural network (CNN) with Batch Normalization (BN). The achieved accuracy for the test dataset is around 91.2%. One application of the proposed system is to integrate it with a smart home platform to identify sports activities of individuals and track their progress.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Aggarwal J, Cai Q (1999) Human motion analysis: a review. Comput Vis Image Underst 73:428–440
Aggarwal JK, Ryoo M (2011) Human activity analysis: a review. ACM Comput Surv (CSUR) 43:1–43
Asperger J, Poore A (2017) Convolutional neural networks for classification of noisy sports videos. s.l., s.n.
Bilbao I, Bilbao J (2017) Overfitting problem and the over-training in the era of data: particularly for artificial neural networks. s.l., s.n., pp 173–177
Bilen H et al (2016) Dynamic image networks for action recognition. s.l., s.n., pp 3034–3042
Bishop C (2006) Pattern recognition and machine learning. NJ Springer, Secaucus
Blank M (2007) Actions as space-time shapes. Trans Pattern Anal Mach Intell 29:2247–2253
Carreira J, Zisserman A (2017) Quo vadis, action recognition? A new model and the kinetics dataset. s.l., s.n., pp 6299–6308
Chen L, Hoey J, Nugent CD, Cook DJ, Yu Z (2012) Sensor-based activity recognition. IEEE Trans Syst Man Cybern Part C (Appl Rev) 42(6):790–808
Chen L et al (2017) Why batch normalization works? A buckling perspective. s.l., s.n., pp 1184–1189
Chen X (2017) Deep manifold learning combined with convolutional neural networks for action recognition. IEEE Trans Neural Netw Learn Syst 99:1–15
Dalal N, Triggs B, Schmid C (2016) Human detection using oriented histograms of flow and appearance. s.l., s.n., pp 428–441
Feichtenhofer C, Pinz A, Wildes R (2016) Spatiotemporal residual networks for video action recognition. s.l., s.n., pp 3468–3476
Feichtenhofer C, Pinz A, Zisserman A (2016) Convolutional two-stream network fusion for video action recognition. s.l., s.n., pp 1933–1941
Gavrila D (1999) The visual analysis of human movement: a survey. Comput Vis Image Underst 73:82–98
Goodfellow I, Bengio Y, Courville A (2016) Deep learning. MIT Press, s.l.
Hongeng S, Nevatia R, Bremond F (2004) Video-based event recognition: activity representation and probabilistic recognition methods. Comput Vis Image Underst 96(2):129–162
Ikizler N, Duygulu P (2007) Human action recognition using distribution of oriented rectangular patches. In: Proceedings of th 2nd conference on human motion: understanding modelling, capture and animation, vol 2, pp 271–284
Ioffe S, Szegedy C (2015) Batch normalization: accelerating deep network training by reducing internal covariate shift. arXiv preprint arXiv:1502.03167
Jaimes A, Sebe N (2005) Multimodal human computer interaction: a survey. ICCV-HCI, s.l.
Krizhevsky A, Sutskever I, Hinton G (2012) ImageNet classification with deep convolutional neural networks. In: Advances in neural information processing systems, vol 25
KukaÄŤka J, Golkov V, Cremers D (2017) Regularization for deep learning: a taxonomy. arXiv preprint arXiv:1710.10686
Lin Q, Zhang D, Chen L, Ni H, Zhou X (2014) Managing elders’ wandering behavior using sensors-based solutions: a survey. Int J Gerontol 2:49–55
Liu J, Kuipers B, Savarese S (2011) Recognizing human actions by attributes. In: CVPR, pp 3337–3344
Lowe D (2004) Distinctive image features from scale-invariant keypoints. Int J Comput Vis 60:91–110
Marszałek M, Laptev I, Schmid C (2009) Actions in context. s.l., s.n., pp 2929–2936
Ni B, Wang G, Moulin P (2011) Rgbd-hudaact: a color-depth video database for human daily activity recognition. In: IEEE international conference on computer vision workshops (ICCV workshops), s.l.
Ojala T, Pietikainen M, Maenpaa T (2002) Multiresolution gray scale and rotation invariant texture classification with local binary patterns. IEEE Trans Pattern Anal Mach Intell 24:971–987
Pantic M, Pentland A, Nijholt A, Huang T (2007) Human computing and machine understanding of human behavior: a survey. Lect Notes Artif Intell 4451:47–71
Poppe R (2010) A survey on vision-based human action recognition. Image Vis Comput 28:976–990
Rachmadi RF, Koutaki G, Uchimura F (2016) Combined convolutional neural network for event recognition. In: The Korea-Japan joint workshop on Frontiers of Computer Vision (FCV)
Rachmadi RF, Koutaki G, Uchimura N (2016) Video classification using compacted dataset based on selected key frames. In: IEEE region 10 conference (TENCON)
Refaeilzadeh P, Tang L, Liu H (2009) Cross-validation. In: Encyclopedia of database systems, pp 532–538
Safdarnejad SM (2015) Sport videos in the wild (SVW): a video dataset for sports analysis. In: Proceedings of international conference on automatic face and gesture recognition
Schuldt C, Laptev I, Caputo B (2004) Recognizing human actions: a local SVM approach. s.l., s.n.
Simonyan K, Zisserman A (2014) Two-stream convolutional networks for action recognition in videos. In: Computer vision and pattern recognition
Simonyan K, Zisserman A (2015) Very deep convolutional networks for large-scale image recognition. In: ICLR
Soomro K, Zamir AR, Shah M (2012) UFC101: a dataset of 101 human action classes from videos in the wild. CRCV-TR-12-01
Srivastava N et al (2014) Dropout: a simple way to prevent neural networks from overfitting. J Mach Learn Res 15:1929–1958
Szegedy C, Vanhoucke V, Loffe S, Shlens J (2015) Rethinking the inception architecture for computer vision
Tran D et al (2015) Learning spatiotemporal features with 3d convolutional networks. s.l., s.n., pp 4489–4497
Vrigkas M, Nikou C, Kakadiaris L (2015) A review of human activity recognition methods. Front Robot AI 2:1–28
Wang H, Schmid C (2013) Action recognition with improved trajectories. s.l., s.n., pp 3551–3558
Wang J, Chen Y, Hao S, Peng X, Hu L (2019) Deep learning for sensor-based activity recognition: a survey. Pattern Recognit Lett 3–11
Wang J, Liu Z, Wu Y, Yuan J (2012) Mining actionlet ensemble for action recognition with depth cameras. In: IEEE conference on computer vision and pattern recognition, pp 1290–1297
Wang L, Qiao Y, Tang X (2015) Action recognition with trajectory-pooled deep-convolutional descriptors. s.l., s.n., pp 4305–4314
Wang L et al (2016) Temporal segment networks: towards good practices for deep action recognition. s.l., s.n., pp 20–36
Wrzalik M, Krechel D (2017) Human action recognition using optical flow and convolutional neural networks. In: 16th IEEE international conference on machine learning and applications, pp 801–805
Yamato J, Ohya J, Ishii K (1992) Recognizing human action in time-sequential images using hidden Markov model. In: Proceedings of IEEE conference on computer vision and pattern recognition, pp 379–385
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Malekmohamadi, H., Pattanjak, N., Bom, R. (2020). Human Activity Identification in Smart Daily Environments. In: Chen, F., GarcĂa-Betances, R., Chen, L., Cabrera-UmpiĂ©rrez, M., Nugent, C. (eds) Smart Assisted Living. Computer Communications and Networks. Springer, Cham. https://doi.org/10.1007/978-3-030-25590-9_5
Download citation
DOI: https://doi.org/10.1007/978-3-030-25590-9_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-25589-3
Online ISBN: 978-3-030-25590-9
eBook Packages: Computer ScienceComputer Science (R0)