Abstract
We discuss the role of spatial representations and visual geometries in vision-based navigation. To a large extent, these choices determine the complexity and robustness of a given navigation strategy. For instance, navigation systems relying on a geometric representation of the environment, use most of the available computational resources for localization rather than for “progressing” towards the final destination. In most cases, however, the localization requirements can be alleviated and different (e.g. topological) representations used. In addition, these representations should be adapted to the robot’s perceptual capabilities.
Another aspect that strongly influences the success/complexity of a navigation system is the geometry of the visual system itself. Biological vision systems display alternative ocular geometries that proved successful in different (and yet demanding and challenging) navigation tasks. The compound eyes of insects or the human foveated retina are clear examples. Similarly, the choice of the particular geometry of the vision system and image sampling scheme, are important design options when building a navigation system.
We provide a number of examples in vision based navigation, where special spatial representations and visual geometries have been taken in consideration, resulting in added simplicity and robustness of the resulting system.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Baker S, Nayar SK (1998) A theory of catadioptric image formation. In Proc. ICCV, pp 35–42.
Bernardino A, Santos-Victor J (99) Binocular visual tracking: Integration of perception and control. IEEE Trans. Rob. Automation, 15(6).
Bernardino A, Santos-Victor J, Sandini G (2002) Foveated active tracking with redundant 2D motion parameters. Robotics and Autonomous Systems.
Brooks R (1985) Visual map making for a mobile robot. In IEEE Conf. Robotics and Automation.
Capurro C, Panerai F, Sandini G (1997) Dynamic vergence using log-polar images. IJCV, 24(1):79–94.
Chahl JS, Srinivasan MV (1997) Reflective surfaces for panoramic imaging. Applied Optics, 36(31):8275–8285.
Fermüller C, Aloimonos Y (1993) The role of fixation in visual motion analysis. IJCV, 11(2):165–186.
Gaspar J, Winters N, Santos-Victor J (2000) Vision-based navigation and environmental representations with an omni-directional camera. IEEE Trans. on Robotics and Automation, 16(6):890–898.
Gracias N, Santos-Victor J (2000) Underwater video mosaics as visual navigation maps. CVIU, 79(1):66–91.
Hicks A, Bajcsy R (1999) Reflective surfaces as computational sensors. In IEEE Workshop on Perception for Mobile Agents, CVPR 99, pp 82–86.
Kato K, Tsuji S, Ishiguro H (1998) Representing environment through target-guided navigation. In Proc. ICPR, pp 1794–1798.
Kuipers B (1978) Modeling spatial knowledge. Cognitive Science, 2:129–153.
Lin LJ, Hancock TR, Judd JS (1998) A robust landmark-based system for vehicle location using low-bandwidth vision. Robotics and Autonomous Systems, 25:19–32.
Murase H, Nayar SK (1995) Visual learning and recognition of 3D objects from appearance. Int. J. Computer Vision, 14(1):5–24.
Santos-Victor J, Sandini G (1997) Visual behaviors for docking. CVIU, 67(3).
Santos-Victor J, Sandini G, Curotto F, Garibaldi S (1995) Divergent stereo in autonomous navigation: From bees to robots. Int. J. Computer Vision, 14(2): 159–177.
Schwartz E (1977) Spatial mapping in the primate sensory projection: Analytic structure and relevance to perception. Biol. Cyb., 25:181–194.
Srinivasan M, Lehrer M, Kirchner W, Zhang S (1991) Range perception through apparent image speed in freely flying honeybees. Visual Neuroscience, 6:519–535.
Wehner R, Wehner S (1990) Insect navigation: use of maps or ariadne’s thread? Ethology, Ecology, Evolution, 2:27–48.
Weiman C, Chaikin G (1979) Logarithmic spiral grids for image processing and display. Comp Graphics and Image Proc, 11:197–226.
Yagi Y, Nishizawa Y, Yachida M (1995) Map-based navigation for mobile robot with omnidirectional image sensor COPIS. IEEE Trans. Robotics and Automation, 11(5): 634–648.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Santos-Victor, J., Bernardino, A. (2003). Vision-based Navigation, Environmental Representations and Imaging Geometries. In: Jarvis, R.A., Zelinsky, A. (eds) Robotics Research. Springer Tracts in Advanced Robotics, vol 6. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-36460-9_23
Download citation
DOI: https://doi.org/10.1007/3-540-36460-9_23
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-00550-6
Online ISBN: 978-3-540-36460-3
eBook Packages: Springer Book Archive