Abstract
Human activity recognition (HAR) is a broad area of research which solves the problem of determining a user’s activity from a set of observations recorded on video or low-level sensors (accelerometer, gyroscope, etc.) HAR has important applications in medical care and entertainment. In this paper, we address sensor-based HAR, because it could be deployed on a smartphone and eliminates the need to use additional equipment. Using machine learning methods for HAR is common. However, such, methods are vulnerable to changes in the domain of training and test data. More specifically, a model trained on data collected by one user loses accuracy when utilised by another user, because of the domain gap (differences in devices and movement pattern results in differences in sensors’ readings.) Despite significant results achieved in HAR, it is not well-investigated from domain adaptation (DA) perspective. In this paper, we implement a CNN-LSTM based architecture along with several classical machine learning methods for HAR and conduct a series of cross-domain tests. The result of this work is a collection of statistics on the performance of our model under DA task. We believe that our findings will serve as a foundation for future research in solving DA problem for HAR.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
Sussex-Huawei Locomotion Dataset, http://www.shl-dataset.org/dataset/.
- 2.
Sensors Activity Recognition Dataset, https://www.researchgate.net/publication/266384007_Sensors_Activity_Recognition_DataSet.
References
Jasim, W.N., Harfash, E.J.: Human activity recognition system to benefit healthcare field by using hog and harris techniques with K-NN model. Int. J. Comput. Appl. 975, 8887 (2018)
Barman, D., Sharma, U.M.: A study on human activity recognition from video. In: 2016 3rd International Conference on Computing for Sustainable Global Development (INDIACom), pp. 2832–2835. IEEE (2016)
Sharma, C.M., Kushwaha, A.K.S., Nigam, S., Khare, A.: On human activity recognition in video sequences. In: 2011 2nd International Conference on Computer and Communication Technology, ICCCT 2011, pp. 152–158. IEEE (2011)
Wang, Y., Huang, K., Tan, T.: Human activity recognition based on r transform. In: 2007 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–8. IEEE (2007)
Giannakopoulos, T., Siantikos, G.: A ROS framework for audio-based activity recognition. In: Proceedings of the 9th ACM International Conference on PErvasive Technologies Related to Assistive Environments, p. 41. ACM (2016)
Asghari, P., Nazerfard, E.: Activity recognition using hierarchical hidden Markov models on streaming sensor data. In: 2018 9th International Symposium on Telecommunications (IST), pp. 416–420. IEEE (2018)
Chereshnev, R., Kertész-Farkas, A.: RapidHARe: a computationally inexpensive method for real-time human activity recognition from wearable sensors. J. Ambient Intell. Smart Environ. 10(5), 377–391 (2018)
Khan, A.M., Tufail, A., Khattak, A.M., Laine, T.H.: Activity recognition on smartphones via sensor-fusion and KDA-based SVMs. Int. J. Distrib. Sens. Netw. 10(5), 503291 (2014)
Khan, A.M., Lee, Y.-K., Lee, S.-Y., Kim, T.-S.: Human activity recognition via an accelerometer-enabled-smartphone using Kernel discriminant analysis. In: 2010 5th International Conference on Future Information Technology, pp. 1–6. IEEE (2010)
Saputri, T.R.D., Khan, A.M., Lee, S.-W.: User-independent activity recognition via three-stage GA-based feature selection. Int. J. Distrib. Sens. Netw. 10(3), 706287 (2014)
Wang, J., Chen, Y., Hu, L., Peng, X., Philip, S.Yu.: Stratified transfer learning for cross-domain activity recognition. In: 2018 IEEE International Conference on Pervasive Computing and Communications (PerCom), pp. 1–10. IEEE (2018)
Soleimani, E., Nazerfard, E.: Cross-subject transfer learning in human activity recognition systems using generative adversarial networks. arXiv preprint arXiv:1903.12489 (2019)
Zeng, M., Yu, T., Wang, X., Nguyen, L.T., Mengshoel, O.J., Lane, I.: Semi-supervised convolutional neural networks for human activity recognition. In: 2017 IEEE International Conference on Big Data (Big Data), pp. 522–529. IEEE (2017)
Wang, J., Chen, Y., Hao, S., Peng, X., Lisha, H.: Deep learning for sensor-based activity recognition: a survey. Pattern Recogn. Lett. 119, 3–11 (2019)
Bulling, A., Blanke, U., Schiele, B.: A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. (CSUR) 46(3), 33 (2014)
Khan, A.M., Lee, Y.-K., Lee, S.Y., Kim, T.-S.: A triaxial accelerometer-based physical-activity recognition via augmented-signal features and a hierarchical recognizer. IEEE Trans. Inf. Technol. Biomed. 14(5), 1166–1172 (2010)
Khan, A., Siddiqi, M., Lee, S.-W.: Exploratory data analysis of acceleration signals to select light-weight and accurate features for real-time activity recognition on smartphones. Sensors 13(10), 13099–13122 (2013)
Preece, S.J., Goulermas, J.Y., Kenney, L.P.J., Howard, D.: A comparison of feature extraction methods for the classification of dynamic activities from accelerometer data. IEEE Trans. Biomed. Eng. 56(3), 871–879 (2008)
Chathuramali, K.G.M., Rodrigo, R.: Faster human activity recognition with SVM. In: International Conference on Advances in ICT for Emerging Regions (ICTer2012), pp. 197–203. IEEE (2012)
Fan, L., Wang, Z., Wang, H.: Human activity recognition model based on decision tree. In: 2013 International Conference on Advanced Cloud and Big Data, pp. 64–68. IEEE (2013)
Paul, P., George, T.: An effective approach for human activity recognition on smartphone. In: 2015 IEEE International Conference on Engineering and Technology (Icetech), pp. 1–3. IEEE (2015)
Chorowski, J.K., Bahdanau, D., Serdyuk, D., Cho, K., Bengio, Y.: Attention-based models for speech recognition. In: Advances in Neural Information Processing Systems, pp. 577–585 (2015)
Pandey, H.M., Windridge, D.: A comprehensive classification of deep learning libraries. In: Yang, X.-S., Sherratt, S., Dey, N., Joshi, A. (eds.) Third International Congress on Information and Communication Technology. AISC, vol. 797, pp. 427–435. Springer, Singapore (2019). https://doi.org/10.1007/978-981-13-1165-9_40
Szegedy, C., et al.: Going deeper with convolutions. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–9 (2015)
Plötz, T., Hammerla, N.Y., Olivier, P.L.: Feature learning for activity recognition in ubiquitous computing. In: Twenty-Second International Joint Conference on Artificial Intelligence (2011)
Ronao, C.A., Cho, S.-B.: Deep convolutional neural networks for human activity recognition with smartphone sensors. In: Arik, S., Huang, T., Lai, W.K., Liu, Q. (eds.) ICONIP 2015. LNCS, vol. 9492, pp. 46–53. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-26561-2_6
Xue, L., et al.: Understanding and improving deep neural network for activity recognition. arXiv preprint arXiv:1805.07020 (2018)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Ordóñez, F., Roggen, D.: Deep convolutional and lstm recurrent neural networks for multimodal wearable activity recognition. Sensors 16(1), 115 (2016)
Hammerla, N.Y., Halloran, S., Plötz, T.: Deep, convolutional, and recurrent models for human activity recognition using wearables. arXiv preprint arXiv:1604.08880 (2016)
Guan, Y., Plötz, T.: Ensembles of deep LSTM learners for activity recognition using wearables. Proc. ACM Interact. Mob. Wear. Ubiquit. Technol. 1(2), 11 (2017)
Zeng, M., et al.: Understanding and improving recurrent networks for human activity recognition by continuous attention. In: Proceedings of the 2018 ACM International Symposium on Wearable Computers, pp. 56–63. ACM, (2018)
Khan, A.M., Lee, Y.-K., Kim, T.-S.: Accelerometer signal-based human activity recognition using augmented autoregressive model coefficients and artificial neural nets. In: 2008 30th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 5172–5175. IEEE (2008)
Chatterjee, S.K., et al.: Comparison of decision tree based classification strategies to detect external chemical stimuli from raw and filtered plant electrical response. Sens. Actuators B: Chem. 249, 278–295 (2017)
Xiaowei, L.: Application of decision tree classification method based on information entropy to web marketing. In: 2014 Sixth International Conference on Measuring Technology and Mechatronics Automation, pp. 121–127. IEEE (2014)
Brzezinski, J.R., Knafl, G.J.: Logistic regression modeling for context-based classification. In: Proceedings of Tenth International Workshop on Database and Expert Systems Applications, DEXA 1999, pp. 755–759. IEEE (1999)
Kang, K., Gao, F., Feng, J.: A new multi-layer classification method based on logistic regression. In: 2018 13th International Conference on Computer Science & Education (ICCSE), pp. 1–4. IEEE (2018)
Shoaib, M., Bosch, S., Incel, O., Scholten, H., Havinga, P.: Fusion of smartphone motion sensors for physical activity recognition. Sensors 14(6), 10146–10176 (2014)
Howard, A.G., et al.: MobileNets: efficient convolutional neural networks for mobile vision applications. arXiv preprint arXiv:1704.04861 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Switzerland AG
About this paper
Cite this paper
Gurov, N., Khan, A., Hussain, R., Khattak, A. (2019). Human Activity Recognition Using Deep Models and Its Analysis from Domain Adaptation Perspective. In: Mazzara, M., Bruel, JM., Meyer, B., Petrenko, A. (eds) Software Technology: Methods and Tools. TOOLS 2019. Lecture Notes in Computer Science(), vol 11771. Springer, Cham. https://doi.org/10.1007/978-3-030-29852-4_15
Download citation
DOI: https://doi.org/10.1007/978-3-030-29852-4_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-29851-7
Online ISBN: 978-3-030-29852-4
eBook Packages: Computer ScienceComputer Science (R0)