Skip to main content

Crowd Estimation of Real-Life Images with Different View-Points

  • Conference paper
  • First Online:
International Conference on Innovative Computing and Communications

Part of the book series: Advances in Intelligent Systems and Computing ((AISC,volume 1166))

  • 980 Accesses

Abstract

This paper found out the problem and suggested a solution for crowd estimation of real-life images. In general, the camera position is fixed at the public places and capture the top view of the images. Most of the recent neural networks are developed with these images. Recently, the CSRNet model was developed for the ShanghaiTech dataset. This model achieved better accuracy than state-of-the-art methods. It is difficult to capture the top view of images where the crowd is gathered at random such as strike, and riot. Therefore, we capture both the top and the front view of images to deal with such circumstances. In this work, the CSRNet model is evaluated using two different test cases consisting of either only top view images or front view images. The mean absolute error (MAE) and mean squared error (MSE) values of the front view images are higher than the top view images. The relative MAE and MSE of the CSRNet model for the front view images are 28.64 and 47.86%, respectively, higher than the top view images. It is noted that higher MAE and MSE means lower performance. This issue can be resolved using the suggested GANN network, which can project the front view images into the top view images. After that, these images can be evaluated using the CSRNet model.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 169.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 219.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. P.F. Felzenszwalb, R.B. Girshick, D. McAllester, D. Ramanan, Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1627–1645 (2009)

    Article  Google Scholar 

  2. J. Gall, V. Lempitsky, Class-specific hough forests for object detection, in Decision Forests for Computer Vision and Medical Image Analysis (Springer, 2013), pp. 143–157

    Google Scholar 

  3. P. Gardziński, K. Kowalak, Ł. Kamiński, S. Maćkowiak, Crowd density estimation based on voxel model in multi-view surveillance systems, in 2015 International Conference on Systems, Signals, and Image Processing (IWSSIP) (IEEE, 2015), pp. 216–219

    Google Scholar 

  4. A.S. Rao, J. Gubbi, S. Marusic, M. Palaniswami, Estimation of crowd density by clustering motion cues. Vis. Comput. 31(11), 1533–1552 (2015)

    Article  Google Scholar 

  5. N. Hussain, H.S.M. Yatim, N.L. Hussain, J.L.S. Yan, F. Haron, Cdes: a pixel-based crowd density estimation system for masjid al-haram. Saf. Sci. 49(6), 824–833 (2011)

    Article  Google Scholar 

  6. X. Xu, D. Zhang, H. Zheng, Crowd density estimation of scenic spots based on multifeature ensemble learning. J. Electr. Comput. Eng. 2017, (2017)

    Google Scholar 

  7. G. Shrivastava, K. Sharma, M. Khari, S.E. Zohora, Role of cyber security and cyber forensics in India, in Handbook of Research on Network Forensics and Analysis Techniques (IGI Global, 2018), pp. 143–161

    Google Scholar 

  8. R. Ghosh, S. Thakre, P. Kumar, A vehicle number plate recognition system using region-of-interest based filtering method, in2018 Conference on Information and Communication Technology (CICT) (IEEE, 2018), pp. 1–6

    Google Scholar 

  9. V.A. Sindagi, V.M. Patel, A survey of recent advances in cnn-based single image crowd counting and density estimation. Pattern Recogn. Lett. 107, 3–16 (2018)

    Article  Google Scholar 

  10. D.B. Sam, S. Surya, R.V. Babu, Switching convolutional neural network for crowd counting, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (IEEE, 2017), pp. 4031–4039

    Google Scholar 

  11. E. Walach, L. Wolf, Learning to count with cnn boosting, in European Conference on Computer Vision (Springer, 2016), pp. 660–676

    Google Scholar 

  12. C. Shang, H. Ai, B. Bai, End-to-end crowd counting via joint learning local and global count, in 2016 IEEE International Conference on Image Processing (ICIP) (IEEE, 2016), pp. 1215–1219

    Google Scholar 

  13. V.A. Sindagi, V.M. Patel, Generating high-quality crowd density maps using contextual pyramid cnns, in Proceedings of the IEEE International Conference on Computer Vision (2017), pp. 1861–1870

    Google Scholar 

  14. Cong Zhang, Kai Kang, Hongsheng Li, Xiaogang Wang, Rong Xie, Xiaokang Yang, Data-driven crowd understanding: A baseline for a large-scale crowd dataset. IEEE Trans. Multimedia 18(6), 1048–1061 (2016)

    Article  Google Scholar 

  15. M. Marsden, K. McGuinness, S. Little, N.E. O’Connor, Fully convolutional crowd counting on highly congested scenes. arXiv preprint arXiv:1612.00220 (2016)

  16. Y. Zhang, D. Zhou, S. Chen, S. Gao, Y. Ma, Single-image crowd counting via multi-column convolutional neural network, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016), pp. 589–597

    Google Scholar 

  17. Y. Li, X. Zhang, D. Chen, CSRNet: dilated convolutional neural networks for understanding the highly congested scenes, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018), pp. 1091–1100

    Google Scholar 

  18. S. Han, J. Pool, J. Tran, W. Dally, Learning both weights and connections for efficient neural network, in Advances in Neural Information Processing Systems (2015), pp. 1135–1143

    Google Scholar 

  19. L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, A.L. Yuille, Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834–848 (2017)

    Article  Google Scholar 

  20. H. Noh, S. Hong, B. Han, Learning deconvolution network for semantic segmentation, in Proceedings of the IEEE International Conference on Computer Vision (2015), pp. 1520–1528

    Google Scholar 

  21. S. Yu, H. Chen, G. Reyes, B. Edel, N. Poh, Gaitgan: invariant gait feature extraction using generative adversarial networks, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2017), pp. 30–37

    Google Scholar 

  22. I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, Y. Bengio, Generative adversarial nets, in Advances in Neural Information Processing Systems (2014), pp. 2672–2680

    Google Scholar 

  23. S. Sahu, H.V. Singh, B. Kumar, A.K. Singh, P. Kumar, Image processing based automated glaucoma detection techniques and role of de-noising: a technical survey, in Handbook of Multimedia Information Security: Techniques and Applications (Springer, Cham, 2019), pp. 359–375

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Akshay Deepak .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2021 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Fahad, M.S., Deepak, A. (2021). Crowd Estimation of Real-Life Images with Different View-Points. In: Gupta, D., Khanna, A., Bhattacharyya, S., Hassanien, A.E., Anand, S., Jaiswal, A. (eds) International Conference on Innovative Computing and Communications. Advances in Intelligent Systems and Computing, vol 1166. Springer, Singapore. https://doi.org/10.1007/978-981-15-5148-2_90

Download citation

Publish with us

Policies and ethics