Abstract
Video-based human activity recognition has fascinated researchers of computer vision community due to its critical challenges and wide variety of applications in surveillance domain. Thus, the development of techniques related to human activity recognition has accelerated. There is now a trend towards implementing deep learning-based activity recognition systems because of performance improvement and automatic feature learning capabilities. This paper implements fusion-based dual-stream deep model for activity recognition with emphasis on minimizing amount of pre-processing required along with fine-tuning of pre-trained model. The architecture is trained and evaluated using standard video actions benchmarks of UCF101. The proposed approach not only provides results comparable with state-of-the-art methods but is also better at exploiting pre-trained model and image data.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Poppe, R. 2010). A survey on vision-based human action recognition. Image and Vision Computing, 28(6), 976–990.
Bobick, A. F., Davis, J. W. (2001). The recognition of human movement using temporal templates. IEEE Transactions on Pattern Analysis and Machine Intelligence, 23(3), 257–267.
Laptev, I. (2005). On space-time interest points. Int. Journal of Computer Vision, 64(2), 107–123.
Soomro, K., Roshan Zamir, A., & Shah, M. (2012). UCF101: A dataset of 101 human action classes from videos in the wild CRCV-TR-12-01, 1, 2, 3, 5.
Dobhal, T., et al. (2015). Human activity recognition using binary motion image and deep learning. Procedia Computer Science, 58, 178–185.
Wang, P., Zhang, J., & Ogunbona, P. O. (2015). Action recognition from depth maps using deep convolutional neural networks. IEEE Transactions on Human-Machine Systems.
Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., Fei-Fei, L., 2014. Large-scale video classification with convolutional neural networks. In: Proc. IEEE Conference on Computer Vision and Pattern Recognition (CVPR). pp. 1725–1732.
Simonyan, K., & Zisserman, A.. (2014). Two-stream convolutional networks for action recognition in videos. In Proceedings of the Advances in Neural Information Processing Systems (NIPS) (pp. 568–576).
Feichtenhofer, C., Pinz, A., & Zisserman, A.. (2016). Convolutional two-stream network fusion for video action recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 1933–1941).
Zolfaghari, M., Oliveira, G. L., Sedaghat, N., & Brox, T. Chained Multi-stream Networks Exploiting Pose, Motion, and Appearance for Action Classification and Detection. https://arxiv.org/abs/1704.00616.
Tran, D., Bourdev, L., Fergus, R., Torresani, L., & Paluri, M. (2015) .Learning spatiotemporal features with 3D convolutional networks. In ICCV.
Ji, S., Xu, W., Yang, M., & Yu, K. (2010). 3D convolutional neural networks for human action recognition. In ICML.
Taylor, G. W., Fergus, R., LeCun, Y., & Bregler, C. (2010). Convolutional learning of spatio-temporal features. In ECCV.
Baccouche, M., Mamalet, F., Wolf, C., Garcia, C., & Baskurt A.. (2011). Sequential deep learning for human action recognition, A.. A. Salah & B. Lepri (Eds.) HBU, LNCS 7065 (pp. 29–39).
Varol, G., Laptev, I., & Schmid, C. (2016). Long-term Temporal Convolutions for Action Recognition. arXiv:1604.04494.
Deng, J., Dong, W., Socher, R., Li, L., Li, K., & Li, F. (2009). ImageNet: a large-scale hierarchical image database. In CVPR (pp. 248–255).
He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (pp. 770–778), 1, 2, 3, 4, 5.
Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., et al. (2015). Going deeper with convolutions. In CVPR (pp. 1–9).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Chandni, Khurana, R., Kushwaha, A.K.S. (2019). Delving Deeper with Dual-Stream CNN for Activity Recognition. In: Khare, A., Tiwary, U., Sethi, I., Singh, N. (eds) Recent Trends in Communication, Computing, and Electronics. Lecture Notes in Electrical Engineering, vol 524. Springer, Singapore. https://doi.org/10.1007/978-981-13-2685-1_32
Download citation
DOI: https://doi.org/10.1007/978-981-13-2685-1_32
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-2684-4
Online ISBN: 978-981-13-2685-1
eBook Packages: EngineeringEngineering (R0)