Skip to main content

Human Praying Structure Classifıcation with Transfer Learning

  • Conference paper
  • First Online:
Proceedings of Sixth International Congress on Information and Communication Technology

Abstract

Action recognition is one of the most important fields in computer vision. Hence, there is an open question of the high accuracy of complex background of human activities. A deep learning approach has recently been used to increase recognition validity with different application areas such as video surveillance, entertainment, autonomous driving vehicles, and human–machine interactions, etc. The aim of this research is to recognize human religious actions that differ in different activities. In our study, we have created our dataset from religious praying videos collected from YouTube, which has been classified into four different classes in terms of religion. We have applied a deep convolutional neural network using the Resnet-50 model for identifying human activity recognition (HAR) and we have got 98.79% accuracy. This research will help to cover more human action recognition tasks of daily activities.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 189.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 249.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Similar content being viewed by others

References

  1. Huang Y, Yang H, Huang P (2012) Action recognition using hog feature in different resolution video sequences. In: 2012 ınternational conference on Computer distributed control and ıntelligent environmental monitoring (CDCIEM). IEEE. pp 85–88

    Google Scholar 

  2. Sadanand S, Corso J (2012) Action bank: a high-level representation of activity in video. IEEE, pp 1234–1241

    Google Scholar 

  3. Wang H, Kläser A, Schmid C et al (2013) Dense trajectories and motion boundary descriptors for action recognition. Int J Comput Vision 103(1):60–79

    Article  MathSciNet  Google Scholar 

  4. LeCun Y, Bottou L, Bengio Y et al (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324

    Article  Google Scholar 

  5. Ijjina EP, Mohan CK (2014) Human action recognition based on recognition of linear patterns inaction bank features using convolutional neural networks. In: 2014 13th International Conference on Machine learning and applications (ICMLA). IEEE, pp 178–182

    Google Scholar 

  6. Jung M, Hwang J, Tani J (2014) Multiple spatio-temporal scales neural network for contextual visual recognition of human actions. In: 2014 Joint IEEE ınternational conferences on development and learning and epigenetic robotics (ICDL-Epirob). IEEE, pp 235–241

    Google Scholar 

  7. Zhang N, Paluri M, Ranzato MA et al Panda: Pose aligned networks for deep

    Google Scholar 

  8. Soomro K, Zamir AR, Shah M (2012) UCF101: a dataset of 101 human action classes from videos in the wild. CRCV-TR-12–01

    Google Scholar 

  9. Ji S, Xu W, Yang M, Yu K (2013) 3d convolutional neural networks for human action recognition. IEEE T-PAMI 35(1):221–231

    Article  Google Scholar 

  10. Wang J, Cherian A, Porikli F (2017) Ordered pooling of optical flow sequences for action recognition. In WACV

    Google Scholar 

  11. Song S, Lan C, Xing J, Zeng W, Liu J (2017) An end-to-end spatial-temporal attention model for human action recognition from skeleton data. In: AAAI

    Google Scholar 

  12. Ran L, Zhang Y, Zhang Q, Yang T (2017) Convolutional neural network-based robot navigation using uncelebrated spherical images. Sensors 17(6)

    Google Scholar 

  13. Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semanticsegmentation. In: CVPR, pp 3431–3440

    Google Scholar 

  14. Feichtenhofer C, Pinz A, Wildes R (2016) Spatiotemporal residual networks for video action recognition. In NIPS

    Google Scholar 

  15. Abu-El-Haija S, Kothari N, Lee J, Natsev P, Toderici G, Varadarajan B, Vijayanarasimhan S (2016) YouTube-8M: a large-scale video classification benchmark. arXiv:1609.08675

  16. Minhas RA, Javed A, Irtaza A, Mahmood MT, Joo YB (2019) Shot classification of field sports videos using AlexNet convolutional neural network. Appl Sci 9(3):483. https://doi.org/10.3390/app9030483

    Article  Google Scholar 

  17. Russo MA, Kurnianggoro L, Jo KH (2019) Classification of sports videos with combination of deep learning models and transfer learning. In: Proceedings of the 2nd ınternational conference on electrical, computer and communication engineering. Cox’sBazar, Bangladesh. https://doi.org/10.1109/ECACE.2019.8679371. [CrossRef]

  18. Ng YHJ, Hausknecht M, Vijayanarasimhan S, Vinyals O, Monga R, Toderici G (2015) Beyond short snippets: deep networks for video classification. In: Proceedings of the IEEE conference on computer vision and pattern recognition. Boston, MA, USA

    Google Scholar 

  19. Baccouche M, Mamalet F, Wolf C, Garcia C, Baskurt A (2011) Sequential deep learning for human action recognition. In: International workshop on human behavior understanding. Springer, Berlin, Heidelberg, pp 29–39

    Google Scholar 

  20. Ramanan D (2007) Learning to parse images of articulated bodies. Advances in neural information processing systems

    Google Scholar 

  21. Soomro K, Zamir AR (2014) Action recognition in realistic sports videos. In: Computer vision in sports. Springer, Cham, pp 181–208

    Google Scholar 

  22. Chaquet JM, Carmona EJ, Fernández-Caballero A (2013) A survey of video datasets for human action and activity recognition. Comput Vis Image Underst 117(6):633–659

    Article  Google Scholar 

  23. Rafiq M et al (2020) Scene classification for sports video summarization using transfer learning. Sensors 20.6:1702

    Google Scholar 

  24. He K et al (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition

    Google Scholar 

  25. Wu J et al (2018) Error compensated quantized SGD and its applications to largescale distributed optimization. arXiv:1806.08054

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Syeda Sumbul Hossain .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Hossain, E., Arman, S., Hossain, S.S., Hasan, A., Jahan, M.R., Hossen, A. (2022). Human Praying Structure Classifıcation with Transfer Learning. In: Yang, XS., Sherratt, S., Dey, N., Joshi, A. (eds) Proceedings of Sixth International Congress on Information and Communication Technology. Lecture Notes in Networks and Systems, vol 217. Springer, Singapore. https://doi.org/10.1007/978-981-16-2102-4_19

Download citation

Publish with us

Policies and ethics