Abstract
Human Activity Recognition (HAR) is a challenging classification task. In the past, it traditionally involved the identification of the movement and activities of a person based on sensor inputs, apply signal processing to receive features and fit the features into a machine learning model. In recent times, deep learning methods have shown good results in automatic Human Activity Recognition. In this paper, we propose a pre-trained CNN (Inception-v3) and LSTM based methodology for Human Activity Recognition. The proposed methodology is evaluated on the publicly available UCF-101 dataset. The results show that the proposed methodology outperforms recent state-of-art methods in terms of accuracy (79.21%) and top-5 accuracy (92.92%) on the HAR task.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Ahmad, Z., Illanko, K., Khan, N., Androutsos, D.: Human action recognition using convolutional neural network and depth sensor data. In: Proceedings of the 2019 International Conference on Information Technology and Computer Communications, pp. 1–5 (2019)
Avilés-Cruz, C., Ferreyra-Ramírez, A., Zúñiga-López, A., Villegas-Cortéz, J.: Coarse-fine convolutional deep-learning strategy for human activity recognition. Sensors 19(7), 1556 (2019)
Banjarey, K., Sahu, S.P., Dewangan, D.K.: A survey on human activity recognition using sensors and deep learning methods. In: 2021 5th International Conference on Computing Methodologies and Communication (ICCMC), pp. 1610–1617. IEEE (2021)
Choutas, V., Weinzaepfel, P., Revaud, J., Schmid, C.: Potion: Pose motion representation for action recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp. 7024–7033 (2018)
Das, Srijan, Thonnat, Monique, Sakhalkar, Kaustubh, Koperski, Michal, Bremond, Francois, Francesca, Gianpiero: A new hybrid architecture for human activity recognition from RGB-D videos. In: Kompatsiaris, Ioannis, Huet, Benoit, Mezaris, Vasileios, Gurrin, Cathal, Cheng, Wen-Huang., Vrochidis, Stefanos (eds.) MMM 2019. LNCS, vol. 11296, pp. 493–505. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-05716-9_40
El-Ghaish, H., Hussien, M.E., Shoukry, A., Onai, R.: Human action recognition based on integrating body pose, part shape, and motion. IEEE Access 6, 49040–49055 (2018)
Geng, C., Song, J.: Human action recognition based on convolutional neural networks with a convolutional auto-encoder. In: 2015 5th International Conference on Computer Sciences and Automation Engineering (ICCSAE 2015), pp. 933–938. Atlantis Press (2016)
Karpathy, A., Toderici, G., Shetty, S., Leung, T., Sukthankar, R., Fei-Fei, L.: Large-scale video classification with convolutional neural networks. In: Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, pp. 1725–1732 (2014)
Khan, S., et al.: Human action recognition: a paradigm of best deep learning features selection and serial based extended fusion. Sensors 21(23), 7941 (2021)
Khattar, L., Kapoor, C., Aggarwal, G.: Analysis of human activity recognition using deep learning. In: 2021 11th International Conference on Cloud Computing, Data Science & Engineering (Confluence), pp. 100–104. IEEE (2021)
Kong, Y., Fu, Y.: Human action recognition and prediction: A survey. arXiv preprint arXiv:1806.11230 (2018)
Kopuklu, O., Kose, N., Gunduz, A., Rigoll, G.: Resource efficient 3d convolutional neural networks. In: Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops (2019)
Mazari, A., Sahbi, H.: Mlgcn: Multi-laplacian graph convolutional networks for human action recognition. In: The British Machine Vision Conference (BMVC) (2019)
Moussa, M.M., Hamayed, E., Fayek, M.B., El Nemr, H.A.: An enhanced method for human action recognition. J. Adv. Res. 6(2), 163–169 (2015)
Orozco, C.I., Xamena, E., Buemi, M.E., Berlles, J.J.: Human action recognition in videos using a robust cnn lstm approach. Ciencia y Tecnologí 23–36 (2020)
Özyer, T., Ak, D.S., Alhajj, R.: Human action recognition approaches with video datasets-a survey. Knowledge-Based Systems 222, 106995 (2021)
Pan, T., Song, Y., Yang, T., Jiang, W., Liu, W.: Videomoco: Contrastive video representation learning with temporally adversarial examples. In: Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 11205–11214 (2021)
Pareek, P., Thakkar, A.: A survey on video-based human action recognition: recent updates, datasets, challenges, and applications. Artif. Intell. Rev. 54(3), 2259–2322 (2021)
Pienaar, S.W., Malekian, R.: Human activity recognition using lstm-rnn deep neural network architecture. In 2019 IEEE 2nd Wireless Africa Conference (WAC), pp. 1–5. IEEE (2019)
Roshan Singh, Alok Kumar Singh Kushwaha, Rajeev Srivastava, et al. Recent trends in human activity recognition-a comparative study. Cognitive Systems Research, 2022
Khurram Soomro, Amir Roshan Zamir, and Mubarak Shah. Ucf101: A dataset of 101 human actions classes from videos in the wild. arXiv preprint arXiv:1212.0402, 2012
Sultani, W., Shah, M.: Human action recognition in drone videos using a few aerial training examples. Comput. Vis. Image Underst. 206, 103186 (2021)
Ankit Vijayvargiya, Nidhi Kumari, Palak Gupta, and Rajesh Kumar. Implementation of machine learning algorithms for human activity recognition. In 2021 3rd International Conference on Signal Processing and Communication (ICPSC), pages 440–444. IEEE, 2021
Michalis Vrigkas, Christophoros Nikou, and Ioannis A Kakadiaris. A review of human activity recognition methods. Frontiers in Robotics and AI, 2:28, 2015
Wang, L., Yangyang, X., Cheng, J., Xia, H., Yin, J., Jiaji, W.: Human action recognition by learning spatio-temporal features with deep neural networks. IEEE access 6, 17913–17922 (2018)
Xia, K., Huang, J., Wang, H.: Lstm-cnn architecture for human activity recognition. IEEE Access 8, 56855–56866 (2020)
Joe Yue-Hei Ng, Matthew Hausknecht, Sudheendra Vijayanarasimhan, Oriol Vinyals, Rajat Monga, and George Toderici. Beyond short snippets: Deep networks for video classification. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 4694–4702, 2015
Yi Zhu, Yang Long, Yu Guan, Shawn Newsam, and Ling Shao. Towards universal representation for unseen action recognition. In Proceedings of the IEEE conference on computer vision and pattern recognition, pages 9436–9445, 2018
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Kumar, M., Rana, A., Ankita, Yadav, A.K., Yadav, D. (2023). Human Activity Recognition in Videos Using Deep Learning. In: Patel, K.K., Santosh, K.C., Patel, A., Ghosh, A. (eds) Soft Computing and Its Engineering Applications. icSoftComp 2022. Communications in Computer and Information Science, vol 1788. Springer, Cham. https://doi.org/10.1007/978-3-031-27609-5_23
Download citation
DOI: https://doi.org/10.1007/978-3-031-27609-5_23
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-27608-8
Online ISBN: 978-3-031-27609-5
eBook Packages: Computer ScienceComputer Science (R0)