Multi-camera Vehicle Tracking Using Local Image Features and Neural Networks

  • Piotr Dalka
Part of the Communications in Computer and Information Science book series (CCIS, volume 287)

Abstract

A method for tracking moving objects crossing fields of view of multiple cameras is presented. The algorithm utilizes Artificial Neural Networks (ANNs). Each ANN is trained to recognize images of one moving object acquired by a single camera. Local image features calculated in the vicinity of automatically detected interest points are used as object image parameters. Next, ANNs are employed to identify the same objects captured by other cameras. Object tracking is supplemented by spatial and temporal constraints defining possible transitions between cameras’ fields of view. Experiments carried out were focused on identification of the same vehicles in different cameras. The results achieved prove that the algorithm is sufficiently effective for multi-camera object tracking provided that the cameras’ orientations with respect to moving objects and to the ground are similar.

Keywords

multi-camera object tracking moving object segmentation local image features SURF object identification 

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Javed, O., Safique, K., Rasheed, Z., Shah, M.: Modeling inter camera space-time and apperacnce relationships for tracking across non-overlapping views. Computer Vision and Image Understanding 109, 146–162 (2008)CrossRefGoogle Scholar
  2. 2.
    Du, W., Piater, J.: Multi-camera People Tracking by Collaborative Particle Filters and Principal Axis-Based Integration. In: Yagi, Y., Kang, S.B., Kweon, I.S., Zha, H. (eds.) ACCV 2007, Part I. LNCS, vol. 4843, pp. 365–374. Springer, Heidelberg (2007)CrossRefGoogle Scholar
  3. 3.
    Chilgunde, A., Kumar, P., Ranganath, S., WeiMin, H.: Multi-Camera Target Tracking in Blind Regions of Cameras with Non-overlapping Fields of View. In: British Machine Vision Conference BMVC, Kingston, September 7-9 (2004)Google Scholar
  4. 4.
    Cai, Q., Aggarwal, J.K.: Tracking human motion in structured environments using a distributed camera system. IEEE Trans. Pattern Anal. Mach. Intell. 2(11), 1241–1247 (1999)CrossRefGoogle Scholar
  5. 5.
    Jeong, K., Jaynes, C.: Object matching in disjoint cameras using a color transfer approach. Machine Vision and Application 19, 443–455 (2008)MATHCrossRefGoogle Scholar
  6. 6.
    Czyzewski, A., Dalka, P.: Moving object detection and tracking for the purpose of multimodal surveillance system in urban areas. In: Tsihrintzis, G.A., et al. (eds.) New Directions in Intelligent Interactive Multimedia. SCI, vol. 142, pp. 75–84. Springer, Heidelberg (2008)CrossRefGoogle Scholar
  7. 7.
    Czyzewski, A., Dalka, P.: Examining Kalman filters applied to tracking objects in motion. In: 9th Int. Workshop on Image Analysis for Mult. Interact. Services, pp. 175–178 (2008)Google Scholar
  8. 8.
    Bay, H., Tuytelaars, T., Van Gool, L.: SURF: Speeded Up Robust Features. In: Leonardis, A., Bischof, H., Pinz, A. (eds.) ECCV 2006, Part I. LNCS, vol. 3951, pp. 404–417. Springer, Heidelberg (2006)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2012

Authors and Affiliations

  • Piotr Dalka
    • 1
  1. 1.Multimedia Systems DepartmentGdansk University of TechnologyGdanskPoland

Personalised recommendations