Abstract
Human activity recognition is an active and interesting field in computer vision from past decades. The objective of the system is to identify human activities using different sensors such as cameras, wearable devices, motion and location sensors, and smartphones. The human actions are automatically identified through their physical activities in human–computer interaction. Determining the human action in an uncontrolled environment is a challenging task in human activity recognition system. In this paper, a novel approach is proposed to recognize human actions effectively in an uncontrolled environment. A frame for the video segment is selected by temporal superpixel, which acts as the input image for the model. Convolutional neural network techniques are applied to extract the features and recognize the human activities from the image. The proposed method has experimented on KTH database and it shows the performance of the method in terms of accuracy. However, the proposed method has attained better accuracy when compared to the existing methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bevilacqua, A., MacDonald, K., Rangarej, A., Widjaya, V., Caulfield, B., Kechadi, T.: Human activity recognition with convolutional neural networks. In: Joint European Conference on Machine Learning and Knowledge Discovery in Databases, pp. 541–552. Springer (2018)
Chang, J., Wei, D., Fisher, J.W.: A video representation using temporal superpixels. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 2051–2058 (2013)
Cheng, G., Wan, Y., Saudagar, A.N., Namuduri, K., Buckles, B.P.: Advances in human action recognition: a survey. arXiv:1501.05964 (2015)
Ibrahim, M.S., Muralidharan, S., Deng, Z., Vahdat, A., Mori, G.: A hierarchical deep temporal model for group activity recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1971–1980 (2016)
Jiang, Y.G., Dai, Q., Liu, W., Xue, X., Ngo, C.W.: Human action recognition in unconstrained videos by explicit motion modeling. IEEE Trans. Image Process. 24(11), 3781–3795 (2015)
Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Advances in Neural Information Processing Systems, pp. 1097–1105 (2012)
Lara, O.D., Labrador, M.A.: A survey on human activity recognition using wearable sensors. IEEE Commun. Surv. Tutor. 15(3), 1192–1209 (2013)
Le, Q.V., Zou, W.Y., Yeung, S.Y., Ng, A.Y.: Learning hierarchical invariant spatio-temporal features for action recognition with independent subspace analysis (2011)
Lin, L., Wang, K., Zuo, W., Wang, M., Luo, J., Zhang, L.: A deep structured model with radius-margin bound for 3D human activity recognition. Int. J. Comput. Vis. 118(2), 256–273 (2016)
Patel, C.I., Garg, S., Zaveri, T., Banerjee, A., Patel, R.: Human action recognition using fusion of features for unconstrained video sequences. Comput. Electr. Eng. 70, 284–301 (2018)
Schuldt, C., Laptev, I., Caputo, B.: Recognizing human actions: a local SVM approach. In: 2004 Proceedings of the 17th International Conference on Pattern Recognition ICPR 2004, vol. 3, pp. 32–36. IEEE (2004)
Thakkar, S., Joshi, M.: Classification of human actions using 3-d convolutional neural networks: a hierarchical approach. In: Computer Vision, Pattern Recognition, Image Processing, and Graphics: 6th National Conference, NCVPRIPG 2017, Mandi, India, 16–19 December 2017, Revised Selected Papers 6, pp. 14–23. Springer (2018)
Wang, H., Oneata, D., Verbeek, J., Schmid, C.: A robust and efficient video representation for action recognition. Int. J. Comput. Vis. 119(3), 219–238 (2016)
Wang, L., Qiao, Y., Tang, X.: MoFAP: a multi-level representation for action recognition. Int. J. Comput. Vis. 119(3), 254–271 (2016)
Zeng, M., Nguyen, L.T., Yu, B., Mengshoel, O.J., Zhu, J., Wu, P., Zhang, J.: Convolutional neural networks for human activity recognition using mobile sensors. In: 6th International Conference on Mobile Computing, Applications and Services, pp. 197–205. IEEE (2014)
Zhao, F., Huang, Y., Wang, L., Xiang, T., Tan, T.: Learning relevance restricted Boltzmann machine for unstructured group activity and event understanding. Int. J. Comput. Vis. 119(3), 329–345 (2016)
Acknowledgements
The authors from Vellore Institute of Technology, Vellore thank the management for providing VIT SEED GRANT for carrying out this research work. Further, all the authors thank the NVIDIA for providing the Titan Xp GPU card under the NVIDIA GPU Grant scheme.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Lakshmi Priya, G.G., Jain, M., Srinivasa Perumal, R., Chandra Mouli, P.V.S.S.R. (2020). Human Action Recognition in Unconstrained Videos Using Deep Learning Techniques. In: Bhateja, V., Satapathy, S., Zhang, YD., Aradhya, V. (eds) Intelligent Computing and Communication. ICICC 2019. Advances in Intelligent Systems and Computing, vol 1034. Springer, Singapore. https://doi.org/10.1007/978-981-15-1084-7_72
Download citation
DOI: https://doi.org/10.1007/978-981-15-1084-7_72
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-1083-0
Online ISBN: 978-981-15-1084-7
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)