Abstract
This paper found out the problem and suggested a solution for crowd estimation of real-life images. In general, the camera position is fixed at the public places and capture the top view of the images. Most of the recent neural networks are developed with these images. Recently, the CSRNet model was developed for the ShanghaiTech dataset. This model achieved better accuracy than state-of-the-art methods. It is difficult to capture the top view of images where the crowd is gathered at random such as strike, and riot. Therefore, we capture both the top and the front view of images to deal with such circumstances. In this work, the CSRNet model is evaluated using two different test cases consisting of either only top view images or front view images. The mean absolute error (MAE) and mean squared error (MSE) values of the front view images are higher than the top view images. The relative MAE and MSE of the CSRNet model for the front view images are 28.64 and 47.86%, respectively, higher than the top view images. It is noted that higher MAE and MSE means lower performance. This issue can be resolved using the suggested GANN network, which can project the front view images into the top view images. After that, these images can be evaluated using the CSRNet model.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
P.F. Felzenszwalb, R.B. Girshick, D. McAllester, D. Ramanan, Object detection with discriminatively trained part-based models. IEEE Trans. Pattern Anal. Mach. Intell. 32(9), 1627–1645 (2009)
J. Gall, V. Lempitsky, Class-specific hough forests for object detection, in Decision Forests for Computer Vision and Medical Image Analysis (Springer, 2013), pp. 143–157
P. Gardziński, K. Kowalak, Ł. Kamiński, S. Maćkowiak, Crowd density estimation based on voxel model in multi-view surveillance systems, in 2015 International Conference on Systems, Signals, and Image Processing (IWSSIP) (IEEE, 2015), pp. 216–219
A.S. Rao, J. Gubbi, S. Marusic, M. Palaniswami, Estimation of crowd density by clustering motion cues. Vis. Comput. 31(11), 1533–1552 (2015)
N. Hussain, H.S.M. Yatim, N.L. Hussain, J.L.S. Yan, F. Haron, Cdes: a pixel-based crowd density estimation system for masjid al-haram. Saf. Sci. 49(6), 824–833 (2011)
X. Xu, D. Zhang, H. Zheng, Crowd density estimation of scenic spots based on multifeature ensemble learning. J. Electr. Comput. Eng. 2017, (2017)
G. Shrivastava, K. Sharma, M. Khari, S.E. Zohora, Role of cyber security and cyber forensics in India, in Handbook of Research on Network Forensics and Analysis Techniques (IGI Global, 2018), pp. 143–161
R. Ghosh, S. Thakre, P. Kumar, A vehicle number plate recognition system using region-of-interest based filtering method, in2018 Conference on Information and Communication Technology (CICT) (IEEE, 2018), pp. 1–6
V.A. Sindagi, V.M. Patel, A survey of recent advances in cnn-based single image crowd counting and density estimation. Pattern Recogn. Lett. 107, 3–16 (2018)
D.B. Sam, S. Surya, R.V. Babu, Switching convolutional neural network for crowd counting, in 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (IEEE, 2017), pp. 4031–4039
E. Walach, L. Wolf, Learning to count with cnn boosting, in European Conference on Computer Vision (Springer, 2016), pp. 660–676
C. Shang, H. Ai, B. Bai, End-to-end crowd counting via joint learning local and global count, in 2016 IEEE International Conference on Image Processing (ICIP) (IEEE, 2016), pp. 1215–1219
V.A. Sindagi, V.M. Patel, Generating high-quality crowd density maps using contextual pyramid cnns, in Proceedings of the IEEE International Conference on Computer Vision (2017), pp. 1861–1870
Cong Zhang, Kai Kang, Hongsheng Li, Xiaogang Wang, Rong Xie, Xiaokang Yang, Data-driven crowd understanding: A baseline for a large-scale crowd dataset. IEEE Trans. Multimedia 18(6), 1048–1061 (2016)
M. Marsden, K. McGuinness, S. Little, N.E. O’Connor, Fully convolutional crowd counting on highly congested scenes. arXiv preprint arXiv:1612.00220 (2016)
Y. Zhang, D. Zhou, S. Chen, S. Gao, Y. Ma, Single-image crowd counting via multi-column convolutional neural network, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2016), pp. 589–597
Y. Li, X. Zhang, D. Chen, CSRNet: dilated convolutional neural networks for understanding the highly congested scenes, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (2018), pp. 1091–1100
S. Han, J. Pool, J. Tran, W. Dally, Learning both weights and connections for efficient neural network, in Advances in Neural Information Processing Systems (2015), pp. 1135–1143
L.-C. Chen, G. Papandreou, I. Kokkinos, K. Murphy, A.L. Yuille, Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 40(4), 834–848 (2017)
H. Noh, S. Hong, B. Han, Learning deconvolution network for semantic segmentation, in Proceedings of the IEEE International Conference on Computer Vision (2015), pp. 1520–1528
S. Yu, H. Chen, G. Reyes, B. Edel, N. Poh, Gaitgan: invariant gait feature extraction using generative adversarial networks, in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops (2017), pp. 30–37
I. Goodfellow, J. Pouget-Abadie, M. Mirza, B. Xu, D. Warde-Farley, S. Ozair, A. Courville, Y. Bengio, Generative adversarial nets, in Advances in Neural Information Processing Systems (2014), pp. 2672–2680
S. Sahu, H.V. Singh, B. Kumar, A.K. Singh, P. Kumar, Image processing based automated glaucoma detection techniques and role of de-noising: a technical survey, in Handbook of Multimedia Information Security: Techniques and Applications (Springer, Cham, 2019), pp. 359–375
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Fahad, M.S., Deepak, A. (2021). Crowd Estimation of Real-Life Images with Different View-Points. In: Gupta, D., Khanna, A., Bhattacharyya, S., Hassanien, A.E., Anand, S., Jaiswal, A. (eds) International Conference on Innovative Computing and Communications. Advances in Intelligent Systems and Computing, vol 1166. Springer, Singapore. https://doi.org/10.1007/978-981-15-5148-2_90
Download citation
DOI: https://doi.org/10.1007/978-981-15-5148-2_90
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-5147-5
Online ISBN: 978-981-15-5148-2
eBook Packages: Intelligent Technologies and RoboticsIntelligent Technologies and Robotics (R0)