Abstract
This paper proposes a novel method to construct a stationary environment map and estimate the ego-motion of a sensor system from unknown planar motion by using an omni-directional vision sensor. Most environments where sensor moves to obtain maps are limited to two-dimensional space. However, conventional ”Structure from Motion (SFM)” algorithms cannot be applied to planar motion and one-dimensional measurements because they use the epipolar geometry. We propose an algorithm that can be applied to two-dimensional space. Since the number of parameters to be estimated is reduced, computational advantages can be obtained for large map reconstruction. Proposed algorithm exploits the azimuths of features which are obtained from an omni-directional vision sensor and gives robust results against the noise of image information by taking advantage of large field of view. A relation between observed azimuth and motion parameters of a vision sensor are constrained by a nonlinear equation. The proposed method obtains closed form solutions to all the motion parameters and an environment map through a two-step procedure. These estimation results can be used as a good initial seed for the incremental reconstruction of a large map.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Azarbayejani, A., Pentland, A.P.: Recursive estimation of motion, structure, and focal length. IEEE Trans. Pattern Anal. Machine Intell. 17, 562–575 (1995)
Hartley, R., Zisserman, A.: Multiple View Geometry in Computer Vision, 2nd edn. Cambridge University Press, Cambridge (2003)
Pollefeys, M., Gool, L.V., Vergauwen, M., Verbiest, F., Cornelis, K., Tops, J., Koch, R.: Visual modeling with a hand-held camera. Int. J. Comput. Vision 59, 207–232 (2004)
Davison, A.J., Reid, I.D., Molton, N.D., Stasse, O.: MonoSLAM: real-time single camera SLAM. IEEE Trans. Pattern Anal. Machine Intell. 29, 1052–1067 (2007)
Klein, G., Murray, D.: Parallel tracking and mapping for small ar workspaces. In: Proc. IEEE International Symposium on Mixed and Augmented Reality, Nara, Japan, pp. 225–234 (2007)
Bunschoten, R., Kröse, B.: Robust scene reconstruction from an omnidirectional vision system. IEEE Trans. Robot. Automat. 19, 351–357 (2003)
Corke, P., Strelow, D., Singh, S.: Omnidirectional visual odometry for a planetary rover. In: Proc. IEEE/RSJ International Conference on Intelligent Robots and Systems, Sendai, Japan, pp. 4007–4012 (2004)
Shimshoni, I.: On mobile robot localization from landmark bearings. IEEE Trans. Robot. Automat. 18, 971–976 (2002)
Chang, P., Hebert, M.: Omni-directional structure from motion. In: Proc. IEEE Workshop on Omnidirectional Vision, pp. 127–133 (2000)
Sato, T., Ikeda, S., Yokoya, N.: Extrinsic camera parameter recovery from multiple image sequences captured by an omni-directional multi-camera system. In: Pajdla, T., Matas, J. (eds.) ECCV 2004. LNCS, vol. 3022, pp. 326–340. Springer, Heidelberg (2004)
Point Grey Research Inc.: Ladybug5 (2013)
Baker, S., Matthews, I.: Lucas-kanade 20 years on: A unifying framework: Part 1. Technical Report CMU-RI-TR-02-16, Robotics Institute, Carnegie Mellon University, MA (2002)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Kim, JH., Choi, J.S. (2014). Initial Closed-Form Solution to Mapping from Unknown Planar Motion of an Omni-directional Vision Sensor. In: Bebis, G., et al. Advances in Visual Computing. ISVC 2014. Lecture Notes in Computer Science, vol 8888. Springer, Cham. https://doi.org/10.1007/978-3-319-14364-4_59
Download citation
DOI: https://doi.org/10.1007/978-3-319-14364-4_59
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-14363-7
Online ISBN: 978-3-319-14364-4
eBook Packages: Computer ScienceComputer Science (R0)