Abstract
In the human brain, independent components of optical flows from the medial superior temporal area are speculated for motion cognition. Inspired by this hypothesis, a novel approach combining independent component analysis (ICA) with principal component analysis (PCA) is proposed in this paper for multiple moving objects detection in complex scenes—a major real-time challenge as bad weather or dynamic background can seriously influence the results of motion detection. In the proposed approach, by taking advantage of ICA’s capability of separating the statistically independent features from signals, the ICA algorithm is initially employed to analyze the optical flows of consecutive visual image frames. As a result, the optical flows of background and foreground can be approximately separated. Since there are still many disturbances in the foreground optical flows in the complex scene, PCA is then applied to the optical flows of foreground components so that major optical flows corresponding to multiple moving objects can be enhanced effectively and the motions resulted from the changing background and small disturbances are relatively suppressed at the same time. Comparative experimental results with existing popular motion detection methods for challenging imaging sequences demonstrate that our proposed biologically inspired vision-based approach can extract multiple moving objects effectively in a complex scene.
Similar content being viewed by others
References
Paragios N, Deriche R. Geodesic active contours and level sets for the detection and tracking of moving objects. IEEE Trans Pattern Anal Mach Interface. 2000;22(3):266–80.
Edelstein A, Rabbat M. Background subtraction for online calibration of baseline RSS in RF sensing networks. IEEE Trans Mob Comput. 2013;12(12):2386–98.
Heikkila M, Pietikainen M. A texture-based method for modeling the background and detecting moving objects. IEEE Trans Pattern Anal Mach Intell. 2006;28(4):657–62.
Cavallaro A, Ebrahimi T. Video object extraction based on adaptive background and statistical change detection. In: Proceedings of SPIE VCIP, 2000.
Shimada A, Yoshinaga S, Taniguchi R. Adaptive background modeling for paused object regions. ACCV2010 workshops, 2010; vol 6468, p. 12–22.
Horn BKP, Schunck BG. Determining optical flow. Artif Intell. 1981;17(1–3):185–203.
Lucas B, Kanade T. An iterative image registration technique with an application to stereo vision. In: Proceedings of the 7th IJCAI, 1981; p. 674-679.
Sun D, Roth S, Black MJ. Secrets of optical flow estimation and their principles. In: 2010 IEEE conference on computer vision and pattern recognition (CVPR). 2010; p. 2432–39.
Baker S, Scharstein D, Lewis J, Roth S, Black MJ, Szeliski R. A database and evaluation methodology for optical flow. Int J Comput Vision. 2011;92(11):1–31.
Bouwmans T. Subspace learning for background modeling: a survey. Recent Pat Comput Sci. 2009;2(3):223–34.
Liu K, Yang H, Ma B, Du Q. A joint optical flow and principal component analysis approach for motion detection. In: 2010 IEEE international conference on acoustics speech and signal processing (ICASSP). March 2010; p. 1178–81.
Candes E, Li X, Ma Y, Wright J. Robust principal component analysis?. ACM, May 2011; vol 58, no. 3.
Guyon C, Bouwmans T, Zahzah E. Foreground detection by robust PCA solved via a linearized alternating direction method. In: International conference on image analysis and recognition (ICIAR) 2012; p. 115–122.
Bouwmans T, Zahzah E. Robust PCA via principal component pursuit: a review for a comparative evaluation in video surveillance. Special issue on background models challenge, computer vision and image understanding (CVIU). 2014.
Tong Y, Cheikh FA, Guraya FFE, Konik H, Trémeau A. A spatio-temporal saliency model for video surveillance. Cogn Comput. 2011;3(1):241–63.
de Croon GCHE, Postma EO, van den Herik HJ. Adaptive gaze control for object detection. Cogn Comput. 2011;3(1):264–78.
Stauffer C, Grimson E. Learning patterns of activity using real time tracking. IEEE Trans Pattern Anal Mach Intell. 2000;22(8):747–57.
Mukherjee D, Wu QMJ, Nguyen TM. Multiresolution based gaussian mixture model for background suppression. IEEE Trans Image Process. 2013;22(12):5022–35.
Lee D. Effective Gaussian mixture learning for video background subtraction. IEEE Trans Pattern Anal Mach Intell. 2005;27(5):827–32.
Zivkovic Z. Improved adaptive Gaussian mixture model for background subtraction. In: Proceedings of the 17th international conference on pattern recognition. 2004; vol 2, p. 28–31.
Barnich O, Van Droogenbroeck M. ViBe: a universal background subtraction algorithm for video sequences. IEEE Trans Image Process. 2011;20(6):1709–24.
Park KY, Jabri M, Lee SY, Sejnowski TJ. Independent components of optical flows have MSTd-like receptive fields. In: Proceedings of the 2nd international workshop on ICA and blind signal separation. 2000; p. 597–601.
Jabri MA, Park KY, Lee SY, Sejnowski TJ. Properties of independent components of self-motion optical flow. In: Proceedings of IEEE international symposium on multiple-valued logic, 2000; p. 355–362.
Ohnishi N, Imiya A. Dominant plane detection from optical flow for robot navigation. Pattern Recognit Lett. 2006;27:1009–21.
Ohnishi N, Imiya A. Independent component analysis of optical flow for robot navigation. Neurocomputing. 2008;71(10–12):2140–63.
Vaina LM, Beardsley SA, Rushton SK. Optic flow and beyond. Dordrecht: Kluwer Academic Publishers; 2004.
Hyvarinen A, Oja E. Independent component analysis: algorithms and application. Neural Netw. 2000;13:411–30.
Zemel RS, Sejnowski TJ. A model for encoding multiple object motions and self-motion in area MST of primate visual cortex. Neuroscience. 1998;18:531–47.
Bishop CM, Nasrabadi NM. Pattern recognition and machine learning. New York: Springer; 2006.
Acknowledgments
This research is supported by The Royal Society of Edinburgh (RSE) and The National Natural Science Foundation of China (NNSFC) under the RSE-NNSFC Joint Project (2012–2014) [Grant Number 61211130309] with Anhui University, China, and the ‘Sino-UK Higher Education Research Partnership for Ph.D. Studies’ Joint Project (2013–2015) funded by the British Council China and The China Scholarship Council (CSC). Amir Hussain and Erfu Yang are also funded, in part, by the UK Engineering and Physical Sciences Research Council (EPSRC) [Grant Number EP/I009310/1] and the RSE-NNSFC joint project (2012–2014) [Grant Number 61211130210] with Beihang University, China. We also thank Dr. Andrew Abel for providing some good suggestions for this paper.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Tu, Z., Zheng, A., Yang, E. et al. A Biologically Inspired Vision-Based Approach for Detecting Multiple Moving Objects in Complex Outdoor Scenes. Cogn Comput 7, 539–551 (2015). https://doi.org/10.1007/s12559-015-9318-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12559-015-9318-z