Abstract
With the use of large video networks, there is a need to coordinate and interpret the video imagery for decision support systems with the goal of reducing the cognitive and perceptual overload of human operators. We present computer vision strategies that enable efficient control and management of cameras to effectively monitor wide-coverage areas, and examine the framework within an actual multi-camera outdoor urban video surveillance network. First, we construct a robust and precise camera control model for commercial pan-tilt-zoom (PTZ) video cameras. In addition to providing a complete functional control mapping for PTZ repositioning, the model can be used to generate wide-view spherical panoramic viewspaces for the cameras. Using the individual camera control models, we next individually map the spherical panoramic viewspace of each camera to a large aerial orthophotograph of the scene. The result provides a unified geo-referenced map representation to permit automatic (and manual) video control and exploitation of cameras in a coordinated manner. The combined framework provides new capabilities for video sensor networks that are of significance and benefit to the broad surveillance/security community.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Barreto, J., Araujo, H.: A general framework for the selection of world coordinate systems in perspective and catadioptric imaging applications. Int. J. Comput. Vis. 57, 23–47 (2004)
Basu, A., Ravi, K.: Active camera calibration using pan, tilt and roll. IEEE Trans. Syst. Man Cybern. 27(3), 559–566 (1997)
Bernardin, K., van de Camp, F., Stiefelhagen, R.: Automatic person detection and tracking using fuzzy controlled active cameras. In: Proc. Comp. Vis. and Pattern Rec. (2007)
Black, J., Ellis, T.: Multi camera image tracking. Image Vis. Comput. 24(11), 1256–1267 (2006)
Collins, R., Tsin, Y.: Calibration of an outdoor active camera system. In: Proc. Comp. Vis. and Pattern Rec., pp. 528–534 (1999)
Davis, J., Chen, X.: Calibrating pan-tilt cameras in wide-area surveillance networks. In: Proc. Int. Conf. Comp. Vis. (2003)
Fry, S., Bichsel, M., Muller, P., Robert, D.: Tracking of flying insects using pan-tilt cameras. J. Neurosci. Methods 101, 59–67 (2000)
Gaspar, J., Santos-Victor, J.: Visual path following with a catadioptric panoramic camera. In: International Symposium on Intelligent Robotic Systems—SIRS, pp. 139–147 (1999)
Jain, A., Kopell, D., Kakligian, K., Wang, Y.: Using stationary-dynamic camera assemblies for wide-area video surveillance and selective attention. In: Proc. Comp. Vis. and Pattern Rec. (2006)
Jethwa, M., Zisserman, A., Fitzgibbon, A.: Real-time panoramic mosaics and augmented reality. In: Brit. Mach. Vis. Conf. (1998)
Khan, S., Shah, M.: Consistent labeling of tracked objects in multiple cameras with overlapping fields of view. IEEE Trans. Pattern. Anal. Mach. Intell. 25(10), 1355–1360 (2003)
Krumm, J., Harris, S., Meyers, B., Brumitt, B., Hale, M., Shafer, S.: Multi-camera multi-person tracking for EasyLiving. In: IEEE International Workshop on Visual Surveillance, pp. 3–10 (2000)
Lee, L., Romano, R., Stein, G.: Monitoring activities from multiple video streams: Establishing a common coordinate frame. IEEE Trans. Pattern. Anal. Mach. Intell. 22(8), 758–767 (2000)
Lowe, D.: Distinctive image features from scale-invariant keypoints. Int. J. Comput. Vis. 60, 91–110 (2004)
Miyamoto, K.: Fisheye lens. J. Opt. Soc. Am., 1060–1061 (1964)
OGRIP: Ohio Geographically Referenced Information Program (2009). http://ogrip.oit.ohio.gov/
Porikli, F., Tuzel, O., Meer, P.: Covariance tracking using model update based on means on Riemannian manifolds. In: Proc. Comp. Vis. and Pattern Rec. (2006)
Sankaranarayanan, K., Davis, J.: An efficient active camera model for video surveillance. In: Proc. Wkshp. Applications of Comp. Vis. (2008)
Sankaranarayanan, K., Davis, J.: A fast linear registration framework for multi-camera GIS coordination. In: Proc. Advanced Video and Signal Based Surveillance (2008)
Weiss, Y.: Deriving intrinsic images from image sequences. In: Proc. Int. Conf. Comp. Vis. (2001)
Woo, D., Capson, D.: 3D visual tracking using a network of low-cost pan/tilt cameras. In: Canadian Conference on Electrical and Computer Engineering, vol. 2, pp. 884–889 (2000)
Acknowledgements
The author wishes to thank graduate students Karthik Sankaranarayanan and Matt Nedrich for their contribution to the work presented, and for partial support provided by the National Science Foundation under grant No. 0236653.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2011 Springer-Verlag London Limited
About this chapter
Cite this chapter
Davis, J.W. (2011). Camera Control and Geo-Registration for Video Sensor Networks. In: Bhanu, B., Ravishankar, C., Roy-Chowdhury, A., Aghajan, H., Terzopoulos, D. (eds) Distributed Video Sensor Networks. Springer, London. https://doi.org/10.1007/978-0-85729-127-1_17
Download citation
DOI: https://doi.org/10.1007/978-0-85729-127-1_17
Publisher Name: Springer, London
Print ISBN: 978-0-85729-126-4
Online ISBN: 978-0-85729-127-1
eBook Packages: Computer ScienceComputer Science (R0)