Abstract
Human Activity Recognition (HAR) is the process of predicting the activity of a person based on a trace of their movement using sensors. Wearable smart devices such as body-worn sensors and smart phones are currently used to detect human activity recognition. By wearing these devices, it is easy to collect data and perform further analysis to obtain necessary information on human activities. The main goal of this project is to recognize patterns from the raw data and extract useful information about the daily activities of the human. Existing strategies, like deep learning, made progress for explicit recognition, yet the recognition of the transition from one activity to the other is poor. In this project, the data corresponding to human activity recognition is presented to convolutional neural network (CNN) model and transitions between activities is presented to the long short-term memory (LSTM) model. The combined model of CNN and LSTM is implemented to improve the HAR accuracy. The HAPT and HAR datasets are used to train the model. The performance analysis of the model is evaluated by tuning the model hyperparameters, such as number of neurons, batch size, learning rate, batch normalization, and dropout. Further the evaluation metrics namely, precision, F1-score and recall are computed. It is observed from the results that better HAR accuracy is obtained with optimized hyperparameter values.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Golestani N, Moghaddam M (2020) Human activity recognition using magnetic induction-based motion signals and deep recurrent neural networks. Nat Commun 11(1):1–11
Jain A, Kanhangad V (2017) Human activity classification in smartphones using accelerometer and gyroscope sensors. IEEE Sens J 18(3):1169–1177
LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436. https://doi.org/10.1038/nature14539
Wang H, Zhao J, Li J, Tian L, Tu P, Cao T, An Y, Wang K, Li S (2020) Wearable sensor-based human activity recognition using hybrid deep learning techniques. Secur Commun Netw
Taylor GW, Fergus R, LeCun Y, Bregler C (2010) Convolutional learning of spatio-temporal features. In: European conference on computer vision. Springer, Berlin, pp 140–153
Ji S, Xu W, Yang M, Yu K (2012) 3D convolutional neural networks for human action recognition. IEEE Trans Pattern Anal Mach Intell 35(1):221–231
Liu C, Liu J, He Z, Zhai Y, Hu Q, Huang Y (2016) Convolutional neural random fields for action recognition. Pattern Recogn 59:213–224
Cho K, Van Merriënboer B, Gulcehre C, Bahdanau D, Bougares F, Schwenk H, Bengio Y (2014) Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv preprint arXiv:1406.1078
Zeng M, Le TN, Bo Y et al (2015) Convolutional neural networks for human activity recognition using mobile sensors. In: Proceedings of the 2014 6th international conference on mobile computing, applications and services, Austin, TX, USA, pp 197–205
Jiang W, Yin Z (2015) Human activity recognition using wearable sensors by deep convolutional neural networks. In: Proceedings of the 23rd ACM international conference on Multimedia, pp 1307–1310
Chen Y, Xue Y (2015) A deep learning approach to human activity recognition based on single accelerometer. In: 2015 IEEE international conference on systems, man, and cybernetics. IEEE, pp 1488–1492
Ronao CA, Cho SB (2016) Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst Appl 59:235–244
Murad A, Pyun JY (2017) Deep recurrent neural networks for human activity recognition. Sensors 17(11):2556
Zhou J, Sun J, Cong P et al (2019) Security-critical energy-aware task scheduling for heterogeneous real-time MPSoCs in IoT. IEEE Trans Serv Comput (TSC) 12:99
Guan Y, Plötz T (2017) Ensembles of deep lstm learners for activity recognition using wearables. Proc ACM Interact Mob Wearable Ubiquitous Technol 1(2):1–28
Qi L, Zhang X, Dou W, Hu C, Yang C, Chen J (2018) A two-stage locality-sensitive hashing based approach for privacy-preserving mobile service recommendation in cross-platform edge environment. Futur Gener Comput Syst 88:636–643
Aloysius N, Geetha M (2020) A scale space model of weighted average CNN ensemble for ASL finger spelling recognition. Int J Comput Sci Eng 22(1):154–161
Jayadeep G, Vishnupriya NV, Venugopal V, Vishnu S, Geetha M (2020) Mudra: convolutional neural network based Indian sign language translator for banks. In: 2020 4th International conference on intelligent computing and control systems (ICICCS). IEEE, pp 1228–1232
Aparna C, Geetha M (2019) CNN and stacked LSTM model for Indian sign language recognition. In: Symposium on machine learning and metaheuristics algorithms, and applications. Springer, Singapore, pp 126–134
Weiss GM, Lockhart JW, Pulickal TT, McHugh PT, Ronan IH, Timko JL (2016) Actitracker: a smartphone-based activity recognition system for improving health and well-being. In: 2016 IEEE international conference on data science and advanced analytics (DSAA). IEEE, pp 682–688
Anguita D, Ghio A, Oneto L, Parra X, Reyes-Ortiz JL (2013) A public domain dataset for human activity recognition using smartphones. In: Esann, vol 3, p 3
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Krishna, K.S., Paneerselvam, S. (2022). An Implementation of Hybrid CNN-LSTM Model for Human Activity Recognition. In: Sengodan, T., Murugappan, M., Misra, S. (eds) Advances in Electrical and Computer Technologies. ICAECT 2021. Lecture Notes in Electrical Engineering, vol 881. Springer, Singapore. https://doi.org/10.1007/978-981-19-1111-8_63
Download citation
DOI: https://doi.org/10.1007/978-981-19-1111-8_63
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-1110-1
Online ISBN: 978-981-19-1111-8
eBook Packages: Computer ScienceComputer Science (R0)