Abstract
Action recognition is one of the most important fields in computer vision. Hence, there is an open question of the high accuracy of complex background of human activities. A deep learning approach has recently been used to increase recognition validity with different application areas such as video surveillance, entertainment, autonomous driving vehicles, and human–machine interactions, etc. The aim of this research is to recognize human religious actions that differ in different activities. In our study, we have created our dataset from religious praying videos collected from YouTube, which has been classified into four different classes in terms of religion. We have applied a deep convolutional neural network using the Resnet-50 model for identifying human activity recognition (HAR) and we have got 98.79% accuracy. This research will help to cover more human action recognition tasks of daily activities.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Huang Y, Yang H, Huang P (2012) Action recognition using hog feature in different resolution video sequences. In: 2012 ınternational conference on Computer distributed control and ıntelligent environmental monitoring (CDCIEM). IEEE. pp 85–88
Sadanand S, Corso J (2012) Action bank: a high-level representation of activity in video. IEEE, pp 1234–1241
Wang H, Kläser A, Schmid C et al (2013) Dense trajectories and motion boundary descriptors for action recognition. Int J Comput Vision 103(1):60–79
LeCun Y, Bottou L, Bengio Y et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324
Ijjina EP, Mohan CK (2014) Human action recognition based on recognition of linear patterns inaction bank features using convolutional neural networks. In: 2014 13th International Conference on Machine learning and applications (ICMLA). IEEE, pp 178–182
Jung M, Hwang J, Tani J (2014) Multiple spatio-temporal scales neural network for contextual visual recognition of human actions. In: 2014 Joint IEEE ınternational conferences on development and learning and epigenetic robotics (ICDL-Epirob). IEEE, pp 235–241
Zhang N, Paluri M, Ranzato MA et al Panda: Pose aligned networks for deep
Soomro K, Zamir AR, Shah M (2012) UCF101: a dataset of 101 human action classes from videos in the wild. CRCV-TR-12–01
Ji S, Xu W, Yang M, Yu K (2013) 3d convolutional neural networks for human action recognition. IEEE T-PAMI 35(1):221–231
Wang J, Cherian A, Porikli F (2017) Ordered pooling of optical flow sequences for action recognition. In WACV
Song S, Lan C, Xing J, Zeng W, Liu J (2017) An end-to-end spatial-temporal attention model for human action recognition from skeleton data. In: AAAI
Ran L, Zhang Y, Zhang Q, Yang T (2017) Convolutional neural network-based robot navigation using uncelebrated spherical images. Sensors 17(6)
Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semanticsegmentation. In: CVPR, pp 3431–3440
Feichtenhofer C, Pinz A, Wildes R (2016) Spatiotemporal residual networks for video action recognition. In NIPS
Abu-El-Haija S, Kothari N, Lee J, Natsev P, Toderici G, Varadarajan B, Vijayanarasimhan S (2016) YouTube-8M: a large-scale video classification benchmark. arXiv:1609.08675
Minhas RA, Javed A, Irtaza A, Mahmood MT, Joo YB (2019) Shot classification of field sports videos using AlexNet convolutional neural network. Appl Sci 9(3):483. https://doi.org/10.3390/app9030483
Russo MA, Kurnianggoro L, Jo KH (2019) Classification of sports videos with combination of deep learning models and transfer learning. In: Proceedings of the 2nd ınternational conference on electrical, computer and communication engineering. Cox’sBazar, Bangladesh. https://doi.org/10.1109/ECACE.2019.8679371. [CrossRef]
Ng YHJ, Hausknecht M, Vijayanarasimhan S, Vinyals O, Monga R, Toderici G (2015) Beyond short snippets: deep networks for video classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition. Boston, MA, USA
Baccouche M, Mamalet F, Wolf C, Garcia C, Baskurt A (2011) Sequential deep learning for human action recognition. In: International workshop on human behavior understanding. Springer, Berlin, Heidelberg, pp 29–39
Ramanan D (2007) Learning to parse images of articulated bodies. Advances in neural information processing systems
Soomro K, Zamir AR (2014) Action recognition in realistic sports videos. In: Computer vision in sports. Springer, Cham, pp 181–208
Chaquet JM, Carmona EJ, Fernández-Caballero A (2013) A survey of video datasets for human action and activity recognition. Comput Vis Image Underst 117(6):633–659
Rafiq M et al (2020) Scene classification for sports video summarization using transfer learning. Sensors 20.6:1702
He K et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition
Wu J et al (2018) Error compensated quantized SGD and its applications to largescale distributed optimization. arXiv:1806.08054
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Hossain, E., Arman, S., Hossain, S.S., Hasan, A., Jahan, M.R., Hossen, A. (2022). Human Praying Structure Classifıcation with Transfer Learning. In: Yang, XS., Sherratt, S., Dey, N., Joshi, A. (eds) Proceedings of Sixth International Congress on Information and Communication Technology. Lecture Notes in Networks and Systems, vol 217. Springer, Singapore. https://doi.org/10.1007/978-981-16-2102-4_19
Download citation
DOI: https://doi.org/10.1007/978-981-16-2102-4_19
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-16-2101-7
Online ISBN: 978-981-16-2102-4
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)