The Visual Computer

, Volume 22, Issue 7, pp 456–467

View-dependent textured splatting

Regular Paper

DOI: 10.1007/s00371-006-0015-5

Cite this article as:
Yang, R., Guinnip, D. & Wang, L. Visual Comput (2006) 22: 456. doi:10.1007/s00371-006-0015-5

Abstract

We present a novel approach to render low resolution point clouds with multiple high resolution textures – the type of data typical from passive vision systems. The low precision, noisy, and sometimes incomplete nature of such data sets is not suitable for existing point-based rendering techniques that are designed to work with high precision and high density point clouds. Our new algorithm – view-dependent textured splatting (VDTS) – combines traditional splatting with a view-dependent texturing strategy to reduce rendering artifacts caused by imprecision or noise in the input data.

VDTS requires no pre-processing of input data, addresses texture aliasing, and most importantly, processes texture visibility on-the-fly. The combination of these characteristics makes VDTS well-suited for interactive rendering of dynamic scenes. Towards this end, we present a real-time view acquisition and rendering system to demonstrate the effectiveness of VDTS. In addition, we show that VDTS can produce high quality rendering when the texture images are augmented with per-pixel depth. In this scenario, VDTS is a reasonable alternative for interactive rendering of large CG models.

Keywords

Point renderingPicture/image generationMulti-textureReal-time

Copyright information

© Springer-Verlag 2006

Authors and Affiliations

  1. 1.Center for Visualization and Virtual EnvironmentUniversity of KentuckyLexingtonUSA