Decision Fusion for Object Detection and Tracking Using Mobile Cameras

  • Luis David López Gutiérrez
  • Leopoldo Altamirano Robles
Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 3287)


In this paper an approach to the automatic target detection and tracking using multisensor image sequences with the presence of camera motion is presented. The approach consists of three parts. The first part uses a motion segmentation method for targets detection in the visible images sequence. The second part uses a background model for detecting objects presented in the infrared sequence, which is preprocessed to eliminate the camera motion. The third part combines the individual results of the detection systems; it extends the Joint Probabilistic Data Association (JPDA) algorithm to handle an arbitrary number of sensors. Our approach is tested using image sequences with high clutter on dynamic environments. Experimental results show that the system detects 99% of the targets in the scene, and the fusion module removes 90% of the false detections.


Target Detection Camera Motion Motion Segmentation Decision Fusion False Target 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


  1. 1.
    Odobez, J., Bouthemy, P.: Direct incremental model-based image motion segmentation analysis for video analysis. Signal Processing 66, 143–155 (1998)zbMATHCrossRefGoogle Scholar
  2. 2.
    Odobez, J., Bouthemy, P.: Robust multiresolution estimation of parametric motion models. JVCIR 6(4), 348–365 (1995)CrossRefGoogle Scholar
  3. 3.
    Hubert, P.J.: Robust statistics. Wiley, Chichester (1981)CrossRefGoogle Scholar
  4. 4.
    Horn, S.: Determining optical flow. Artificial Intelligence 17, 185–203 (1981)CrossRefGoogle Scholar
  5. 5.
    Stauffer, C.: Adaptive background mixture models for real-time tracking. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 246–252 (1999)Google Scholar
  6. 6.
    Bar-Shalom, Fortmann, T.: Tracking and data association. Academic Press, San Diego (1988)zbMATHGoogle Scholar
  7. 7.
    Waltz, E., Llinas, J.: Handbook of multisensor data fusion. CRC Press, Boca Raton (2001)Google Scholar
  8. 8.
    Barron, J., Fleet, D., Bauchemin, S.: Performance of optical flow techniques. International Journal of Computer Vision 12(1), 43–77 (1994)CrossRefGoogle Scholar
  9. 9.
    Irani, M., Rousso, B., Peleg, S.: Computing occluding and transparent motion. Intern. J. Comput. Vis. 12(1), 5–16 (1994)CrossRefGoogle Scholar
  10. 10.
    Stauffer, C., Grimson, W.E.L.: Learning patterns of activity using real time tracking. IEEE trans. PAMI 22(8), 747–757 (2000)Google Scholar
  11. 11.
    Pao, L., O’Neil, S.: Multisensor Fusion algorithms for tracking. In: Proc. of American Control Conference, pp. 859–863 (1993)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2004

Authors and Affiliations

  • Luis David López Gutiérrez
    • 1
  • Leopoldo Altamirano Robles
    • 1
  1. 1.National Institute of Astrophysics Optics and ElectronicsSanta Maria TonantzintlaMéxico

Personalised recommendations