Next Best View Algorithms for Interior and Exterior Model Acquisition

  • Bradley D. Null
  • Eric D. Sinzinger
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4292)


Rapid determination of optimal laser range finder placement is becoming of increased importance in both computer vision and robotics. The need of next-best-view algorithms for intelligent decisions in where to place the laser range finder is important for scanning both objects and landscapes to assure that the scene is fully realized and can be registered accurately. Presented here are methods for determining sensor placement to maximize the amount of information that is learned about a scene or object. Using occupancy grids of voxels and ray tracing, the placement of the sensor can be optimized for maximum collection of new data. This work provides fast algorithms that determine optimal sensor placement both outside an object of interest and inside of a closed environment. These methods take into consideration the limitations of the sensor, the interaction of the sensor to the environment, and its placement in only known areas to restrict the search space.


Sensor Position Sensor Placement Laser Range Finder Occupancy Grid West Wall 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Scott, W.R., Roth, G., Rivest, J.F.: View planning for automated three-dimensional object reconstruction and inspection. ACM Computing Surveys 35, 64–96 (2003)CrossRefGoogle Scholar
  2. 2.
    Wong, L.M., Dumont, C., Abidi, M.A.: Next best view system in a 3-D object. In: International Symposium on Computational Intelligence in Robotics and Automation Proceedings, pp. 306–311 (1999)Google Scholar
  3. 3.
    Banta, J.E., Abidi, M.A.: Autonomous placement of a range sensor for acquisition of optimal 3-D models. In: Proceedings of the 1996 IEEE IECON 22nd International Conference on Industrial Electronics, Control, and Instrumentation, pp. 1583–1588 (1996)Google Scholar
  4. 4.
    Pito, R.: A sensor-based solution to the “next best view” problem. In: 13th International Conference on Pattern Recognition, vol. 1, pp. 941–945 (1996)Google Scholar
  5. 5.
    Garcia, M.A., Velaquez, S., Sappa, A.D.: A two-stage algorithm for planning the next view from range images. In: 9th British Machine Vision Conference, pp. 720–729 (1998)Google Scholar
  6. 6.
    Dissanayake, M.W.M.G., Newman, P., Clark, S., Durant-Whyte, H.F., Csorba, M.: A solution to the simultaneous localization and map building (SLAM) problem. IEEE Transactions on Robotics and Automation 17, 229–241 (2001)CrossRefGoogle Scholar
  7. 7.
    Surmann, H., Nuchter, A., Hertzberg, J.: An autonomous mobile robot with a 3-D laser range finder for 3-D exploration and digitalization of indoor environments. Robotics and Autonomous Systems, 181–198 (2003)Google Scholar
  8. 8.
    Gonzalez-Banos, H.H., Latombe, J.C.: Navigation strategies for exploring indoor environments. The International Journal of Robotics Research 21, 829–848 (2002)CrossRefGoogle Scholar
  9. 9.
    Sequeira, V., Ng, K., Wolfart, E., Goncalves, J., Hogg, D.: Automated 3D reconstruction of interiors with multiple scan-views. In: SPIE 11th AnnualSymposium (1999)Google Scholar
  10. 10.
    Papadopoulos-Orfanos, D., Schmitt, F.: Automatic 3-D digitization using a laser rangefinder with a small field of view. In: First International Convference on Recent Advances in 3-D Digital Imagaging and Modeling, pp. 60–67 (1997)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2006

Authors and Affiliations

  • Bradley D. Null
    • 1
  • Eric D. Sinzinger
    • 1
  1. 1.Department of Computer ScienceTexas Tech UniversityLubbockUSA

Personalised recommendations