Abstract
Presented framework provides a method for adaptive background change detection in video from monocular static cameras. A background change constitutes of objects left in the scene and objects moved or taken from the scene. This framework may be applied to luggage left behind in public places, to asses the damage and theft of public property, or to detect minute changes in the scene. The key elements of the framework include spatiotemporal motion detection, texture classification of non-moving regions, and spatial clustering of detected background changes. Motion detection based on local variation of spatiotemporal texture separates the foreground and background regions. Local background dissimilarity measurement is based on wavelet decomposition of localized texture maps. Dynamic threshold of the normalized dissimilarity measurement identifies changed local background blocks, and spatial clustering isolates the regions of interest. The results are demonstrated on the PETS 2006 video sequences.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Cheung, S.C.S., Kamath, C.: Robust background subtraction with foreground validation for urban traffic video. EURASIP J. Appl. Signal Process 2005(1), 2330–2340 (2005)
Hall, D., Nascimento, J., Ribeiro, P., Andrade, E., Moreno, P., Pesnel, S., List, T., Emonet, R., Fisher, R.B., Victor, J.S., Crowley, J.L.: Comparison of target detection algorithms using adaptive background models. In: ICCCN 2005: Proceedings of the 14th International Conference on Computer Communications and Networks, pp. 113–120. IEEE Computer Society, Washington, DC (2005)
Piccardi, M.: Background subtraction techniques: a review. In: Proceedings of the IEEE International Conference on Systems, Man and Cybernetics, vol. 4, pp. 3099–3104 (2004)
Tian, Y.L., Lu, M., Hampapur, A.: Robust and efficient foreground analysis for real-time video surveillance. In: Proc. IEEE Conf. on Computer Vision and Pattern Recognition, vol. 1, pp. 1182–1187 (2005)
Oliver, N.M., Rosario, B., Pentland, A.P.: A bayesian computer vision system for modeling human interactions. IEEE Trans. Pattern Anal. Mach. Intell. 22(8), 831–843 (2000)
Stauffer, C., Grimson, W.E.L.: Adaptive background mixture models for real-time tracking, vol. 2, pp. 246–252 (1999)
Grabner, H., Roth, P.M., Grabner, M., Bischof, H.: Autonomous learning of a robust background model for change detection. In: Proceedings 9th IEEE International Workshop on PETS (2006)
Latecki, L.J., Miezianko, R., Pokrajac, D.: Motion detection based on local variation of spatiotemporal texture. In: 2004 Conference on Computer Vision and Pattern Recognition Workshop, vol. 8, pp. 135–141 (2004)
PETS2006: Performance Evaluation of Tracking and Surveillance 2006 Benchmark Data, http://www.cvg.rdg.ac.uk/PETS2006/
Latecki, L.J., Miezianko, R., Pokrajac, D.: Activity and motion detection based on measuring texture change. In: International Conference on Machine Learning and Data Mining, pp. 476–486 (2005)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Miezianko, R., Pokrajac, D. (2008). Texture Dissimilarity Measures for Background Change Detection. In: Campilho, A., Kamel, M. (eds) Image Analysis and Recognition. ICIAR 2008. Lecture Notes in Computer Science, vol 5112. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-69812-8_67
Download citation
DOI: https://doi.org/10.1007/978-3-540-69812-8_67
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-69811-1
Online ISBN: 978-3-540-69812-8
eBook Packages: Computer ScienceComputer Science (R0)