Romansy 16 pp 431-438 | Cite as

Visual Target Detection in Unstructured Environments — A Novel Technique for Robotic Navigation

  • Andrzej Śluzek
  • Md Saiful Islam
Part of the CISM Courses and Lectures book series (CISM, volume 487)


We report a novel method of visual modeling and detecting 3D rigid objects randomly located in complex cluttered environments. Models are built (in reference scale) using visual saliencies (interest points) in template images presenting the object of interest from various viewpoints. By matching interest point detected in camera-captured images (relative scale is used there) to visual saliencies from the database, the target objects are detected and verified. If a stereovision system is available, the procedure is separately performed for both cameras. Subsequently, the corresponding matches from both images can be used to solve the fundamental correspondence problem. The paper briefly discusses methodology and presents exemplary experimental results in vision-based robotic aplications.


Target Object Reference Image Interest Point Visual Saliency Moment Invariant 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. Ballard, D.H. (1981). Generalizing the Hough transform to detect arbitrary shapes. Pattern Recognition 13(2): 111–122.MATHCrossRefGoogle Scholar
  2. Hausler, G., and Ritter, D. (1999). Feature-based object recognition and localization in 3D-space using a single video image. Computer Vision and Image Understanding 73(1): 64–81.CrossRefGoogle Scholar
  3. Hu, M.K. (1962). Visual pattern recognition by moment invariants. IRE Transactions on Information Theory 8: 179–187.Google Scholar
  4. Islam, M.S. (2005). Relative scale method to recognize and locate objects for robotic applications. PhD thesis, Nanyang Technological University, SingaporeGoogle Scholar
  5. Joseph, S.H. (1999). Optimal pose estimation in two and three dimension. Computer Vision and Image Understanding 73(2): 215–231.MATHCrossRefGoogle Scholar
  6. Jurie, F. and Schmid, C. (2004). Scale-invariant shape features for recognition of object categories. Proceedings of IEEE Conf. on Computer Vision and Pattern Recognition 2: II.90–II.96.Google Scholar
  7. Kim, D. and Nevada, R. (1999). Symbolic navigation with a generic map. Autonomous Robots 6(1): 69–88.CrossRefGoogle Scholar
  8. Lowe, D. G. (2004). Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision 60(2): 91–110.CrossRefGoogle Scholar
  9. Maitra, S. (1979). Moment invariants. Proceedings of IEEE 67(4): 697–699.CrossRefGoogle Scholar
  10. Mikolajczyk, K. and Schmid, C. (2004). Scale & affine invariant interest point detectors. International Journal of Computer Vision 60(1): 63–86.CrossRefGoogle Scholar
  11. Schmid, C., Mohr, R., and Bauckhage, C. (2000). Evaluation of interest point detectors. International Journal of Computer Vision 37(2): 151–172.MATHCrossRefGoogle Scholar
  12. Stein, F. and Medioni, G. (1992). Structural indexing: efficient 3-D object recognition. IEEE Transactions on Pattern Analysis and Machine Intelligence 14(2): 125–145.CrossRefGoogle Scholar
  13. Ulrich, M., Steger, C., and Baumgartner, A. (2003). Real-time object recognition using a modified generalized Hough transformation. Pattern Recognition 36: 2557–2570.MATHCrossRefGoogle Scholar
  14. Wong, A.K.C., Rong, L., and Liang, X. (1998). Robotic vision: 3D object recognition and pose determination. Proceedings of IEEE/RSJ Int. Conf. on Intelligent Robots and Systems, 1202–1209.Google Scholar

Copyright information

© CISM, Udine 2006

Authors and Affiliations

  • Andrzej Śluzek
    • 1
    • 2
  • Md Saiful Islam
    • 1
  1. 1.School of Computer EngineeringNTUSingapore
  2. 2.SWPSWarsawPoland

Personalised recommendations