Abstract
The human visual system has no difficulty to detect moving object. To design an automated method for detecting foreground in videos captured in a variety and complicated scenes is a challenge. The topic has attracted much research due to its wide range of video-based applications. We propose a foveated model that mimics the human visual system for the detection of foreground in image sequence. It is a two-step framework simulating the awareness of motion followed by the extraction of detailed information. In the first step, region proposals are extracted based on similarity of intensity and motion features with respect to the pre-generated archetype. Through integration of the similarity measures, each image frame is segregated into background and foreground points. Large foreground regions are preserved as region proposals (RPs). In the second step, analysis is performed on each RP in order to obtain the accurate shape of moving object. Photometric and textural features are extracted and matched with another archetype. We propose a probabilistic refinement scheme. If the RP contains a point initially labeled as background, it can be converted to a foreground point if its features are more similar to neighboring foreground points than neighboring background points. Both archetypes are updated immediately based on the segregation result. We compare our method with some well-known and recently proposed algorithms using various video datasets.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Akbas, E., Eckstein, M.P.: Object detection through search with a foveated visual system. PLOS Comput. Biol. 13(10), e1005743 (2017)
Benoit, A., Caplier, A., Durette, B., Herault, J.: Using human visual system modeling for bio-inspired low level image processing. Comput. Vis. Image Underst. 114(7), 758–773 (2010)
Hsieh, J.-W., Hsu, Y.-T., Liao, H.-Y.M., Chen, C.-C.: Video-based human movement analysis and its application to surveillance systems. IEEE Trans. Multimedia 10(3), 372–384 (2008)
Lu, W., Tan, Y.-P.: A vision-based approach to early detection of drowning incidents in swimming pools. IEEE Trans. Circuits Syst. Video Technol. 14(2), 159–178 (2004)
Visser, R., Sebe, N., Bakker, E.: Object recognition for video retrieval. In: Lew, M.S., Sebe, N., Eakins, John P. (eds.) CIVR 2002. LNCS, vol. 2383, pp. 262–270. Springer, Heidelberg (2002). https://doi.org/10.1007/3-540-45479-9_28
Sobral, A., Vacavant, A.: A comprehensive review of background subtraction algorithms evaluated with synthetic and real videos. Comput. Vis. Image Underst. 122, 4–21 (2014)
Stauffer, C., Grimson, W.E.L.: Learning patterns of activity using real-time tracking. IEEE Trans. Pattern Anal. Mach. Intell. 22(8), 747–757 (2000)
Elgammal, A., Duraiswami, R., Harwood, D., Davis, L.S.: Background and foreground modeling using nonparametric kernel density estimation for visual surveillance. Proc. IEEE 90(7), 1151–1163 (2002)
Barnich, O., Van Droogenbroeck, M.: ViBe: a powerful random technique to estimate the background in video sequences. In: Proceedings of International Conference on Acoustics, Speech and Signal Processing, pp. 945–948 (2009)
Heikkilä, M., Pietikäinen, M.: A texture-based method for modeling the background and detecting moving objects. IEEE Trans. Pattern Anal. Mach. Intell. 28(4), 657–662 (2006)
Liao, S., Zhao, G., Kellokumpu, V., Pietikäinen, M., Li, S.Z.: Modeling pixel process with scale invariant local patterns for background subtraction in complex scenes. In: Proceedings of IEEE Conference on Computer Vision and Pattern Recognition, pp. 1301–1306 (2010)
Van Droogenbroeck, M., Paquot, O.: Background subtraction: experiments and improvements for ViBe. In: Proceedings of IEEE Workshop on Change Detection at IEEE Conference on Computer Vision and Pattern Recognition, pp. 32–37 (2012)
Kim, S.W., Yun, K., Yi, K.M., Kim, S.J., Choi, J.Y.: Detection of moving objects with a moving camera using non-panoramic background model. Mach. Vis. Appl. 24, 1015–1028 (2013)
Heikkilä, M., Pietikäinen, M., Heikkilä, J.: A texture-based method for detecting moving objects. In: Proceedings of British Machine Vision Conference, pp. 187–196 (2004)
Ma, F., Sang, N.: Background subtraction based on multi-channel SILTP. In: Park, J.-I., Kim, J. (eds.) ACCV 2012. LNCS, vol. 7728, pp. 73–84. Springer, Heidelberg (2013). https://doi.org/10.1007/978-3-642-37410-4_7
Chan, K.L.: Saliency detection in video sequences using perceivable change encoded local pattern. SIViP 12(5), 975–982 (2018)
Toyama, K., Krumm, J., Brumitt, B., Meyers, B.: Wallflower: principles and practice of background maintenance. In: Proceedings of International Conference on Computer Vision, pp. 255–261 (1999)
Li, L., Huang, W., Gu, I.Y.-H., Tian, Q.: Statistical modelling of complex backgrounds for foreground object detection. IEEE Trans. Image Process. 13(11), 1459–1472 (2004)
Goyette, N., Jodoin, P.-M., Porikli, F., Konrad, J., Ishwar, P.: Changedetection.net: a new change detection benchmark dataset. In: Proceedings of IEEE Workshop on Change Detection at IEEE Conference on Computer Vision and Pattern Recognition, pp. 16–21 (2012)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Chan, K.L. (2020). Segmentation of Foreground in Image Sequence with Foveated Vision Concept. In: Palaiahnakote, S., Sanniti di Baja, G., Wang, L., Yan, W. (eds) Pattern Recognition. ACPR 2019. Lecture Notes in Computer Science(), vol 12046. Springer, Cham. https://doi.org/10.1007/978-3-030-41404-7_62
Download citation
DOI: https://doi.org/10.1007/978-3-030-41404-7_62
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-41403-0
Online ISBN: 978-3-030-41404-7
eBook Packages: Computer ScienceComputer Science (R0)