Abstract
This paper presents a novel approach to reconstruct illumination environment by omnidirectional camera calibration. The camera positions are estimated by our method considering the inlier distribution. The light positions are computed with the intersection points of the rays starting from the camera positions toward the corresponding points between two images. In addition, our method can integrate various synthetic objects in the real photographs by using the distributed ray tracing and HDR (High Dynamic Range) radiance map. Simulation results showed that we can generate photo-realistic image synthesis in the reconstructed illumination environment.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Fournier, A., Gunawan, A., Romanzin, C.: Common illumination between real and computer generated scenes. In: Proc. of Graphics Interface, pp. 254–262 (1993)
Debevec, P.: Rendering synthetic objects into real scenes: bridging traditional and image-based graphics with global illumination and high dynamic range photography. In: Proc. of Siggraph, pp. 189–198 (1998)
Gibson, S., Murta, A.: Interactive rendering with real-world illumination. In: Proc. of the 11th Eurographics Workshop on Rendering, pp. 365–376 (2000)
Xiong, Y., Turkowski, K.: Creating image based VR using a self-calibrating fisheye lens. In: Proc. of Computer Vision and Pattern Recognition, pp. 237–243 (1997)
Nene, S.A., Nayar, S.K.: Stereo with mirrors. In: Proc. of Int. Conf. on Computer Vision, pp. 1087–1094 (1998)
Sato, I., Sato, Y., Ikeuchi, K.: Acquiring a radiance distribution to superimpose virtual objects onto a real scene. IEEE Trans. on Visualization and Computer Graphics 5(1), 1–12 (1999)
Micusik, B., Pajdla, T.: Estimation of omnidiretional camera model from epipolar geometry. In: Proc. of Computer Vision and Pattern Recognition, pp. 485–490 (2003)
Micusik, B., Martinec, D., Pajdla, T.: 3D Metric reconstruction from uncalibrated omnidirectional Images. In: Proc. of Asian Conf. on Computer Vision, pp. 545–550 (2004)
Micusik, B., Pajdla, T.: Omnidirectional camera model and epipolar estimateion by RANSAC with bucketing. In: IEEE Scandinavian Conf. Image Analysis, pp. 83–90 (2003)
Hwang, Y., Hong, H.: Estimation of Omnidirectional Camera Model with One Parametric Projection. Lecture Notes in Control and Information Sciences, vol. 345, pp. 827–833 (2006)
Hwang, Y., Hong, H.: 3D Analysis of Omnidirectional Images for Illumination Environment Reconstruction. In: Greco, S., Hata, Y., Hirano, S., Inuiguchi, M., Miyamoto, S., Nguyen, H.S., Słowiński, R. (eds.) RSCTC 2006. LNCS (LNAI), vol. 4259, Springer, Heidelberg (2006)
Glassner, A.S.: Graphics Gems. AP Professional, Cambridge (1990)
Agarwal, S., Ramamoorthi, R., Belongie, S., Jensen, H.: Structured Importance Sampling of Environment Maps. In: Proc. of Siggraph, pp. 605–612 (2003)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Hwang, YH., Hong, HK. (2006). Reconstructing Illumination Environment by Omnidirectional Camera Calibration. In: Sattar, A., Kang, Bh. (eds) AI 2006: Advances in Artificial Intelligence. AI 2006. Lecture Notes in Computer Science(), vol 4304. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11941439_57
Download citation
DOI: https://doi.org/10.1007/11941439_57
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-49787-5
Online ISBN: 978-3-540-49788-2
eBook Packages: Computer ScienceComputer Science (R0)