Advertisement

Cognitive Recognition Under Occlusion for Visually Guided Robotic Errand Service

Conference paper
Part of the Advances in Intelligent Systems and Computing book series (AISC, volume 302)

Abstract

A reliable vision system for running robotic errand service in an unstructured indoor environment such as homes is difficult to construct. Many visual challenges, such as perspective, clutter, illumination, and occlusion, need to be handled appropriately. While most of previous researches addressed these problems from the contexts of either object recognition or object searching, our proposed approach relies on a solution that combines these two as one. We are proposing a “Cognitive Recognition” System. In the proposed cognitive recognition system, information gathered from scene recognition helps deciding the next optimal perspective, and environmental parameters measurements determine the uncertainty in recognition measurements and thus the proper probability map update used in object search. We show particularly in this paper how this approach provides a practical solution to cluttered and occluded environments. And we demonstrate the results with our HomeMate Service Robot.

Keywords

Object recognition Evidence collection Object searching  Occlusion management Cluttered environment Illumination variation  Visually-guided service robot Errand service for elderly 

Notes

Acknowledgments

This work is supported in part by MEGA Science R&D Project, funded by Ministry of Science ICT and Future Planning (NRF-2013MIA3A3A02042335), in part by Technology Innovation Program (10048320) funded by Ministry of Trade, Industry and Energy, and in part by Basic Science Research Program through NRF (NRF-2010-0020210) funded by Ministry of Education.

References

  1. 1.
    Kemp, C.C., Edsinger, A., Torres-Jara, E., “Challenges for robot manipulation in human environments [Grand Challenges of Robotics],” March 2007 Robotics & Automation Magazine, IEEE (Volume:14, Issue: 1).Google Scholar
  2. 2.
    M. Bertsche, T. Fromm, and W. Ertel, “BOR3D: A Use-Case-Oriented Software Framework for 3-D Object Recognition,” 2012 IEEE Conference on Technologies for Practical Robot Applications, Woburn.Google Scholar
  3. 3.
    A. Collet, M. Martinez, and S. S. Srinivasa, “The MOPED framework: Object Recognition and Pose Estimation for Manipulation,” Sep. 2011 the International Journal of Robotics Research, vol. 30, no. 10, pp. 1284–1306.Google Scholar
  4. 4.
    F. A. Pavel, Z. Wang, and D. D. Feng, “Reliable Object Recognition using SIFT Features,” MMSP’09, October 5–7, 2009, Rio de Janeiro, Brazil.Google Scholar
  5. 5.
    S. Lee, S. Lee, J. Lee, D. Moon, E. Kim, and J. Seo, “Robust Recognition and Pose Estimation of 3D Objects Based on Evidence Fusion in a Sequence of Images,” 10–14 April 2007 IEEE International Conference on Robotics and Automation Roma, Italy.Google Scholar
  6. 6.
    S. Lee, E. Kim, and Y. Park, “3D Object Recognition using Multiple Features for Robotic Manipulation,” May 2006 IEEE International Conference on Robotics and Automation, Orlando, Florida.Google Scholar
  7. 7.
    S. Lee, Z. Lu, and H. Kim, “Probabilistic 3D Object Recognition with Both Positive and Negative Evidences,” 2011 IEEE International Conference on Computer Vision.Google Scholar
  8. 8.
    H. Kim, J. Lee and S. Lee, “Environment Adaptive 3D Object Recognition and Pose Estimation by Cognitive Perception Engine,” 2009 IEEE International Symposium on Computational Intelligence in Robotics and Automation (CIRA).Google Scholar
  9. 9.
    Y. Ye, J.K. Tsotsos, “Sensor planning for 3d object search, Comput. Vis. Image Understand,” 73 (2) (1999) 145–168.Google Scholar
  10. 10.
    T.D. Garvey, “Perceptual strategies for purposive vision, Technical report, SRI International,” 117, 1976.Google Scholar
  11. 11.
    L. Wixson, D. Ballard, “Using intermediate object to improve efficiency of visual search,” Int. J. Comput. Vis. 18 (3) (1994) 209–230.Google Scholar
  12. 12.
    K. Sj öö, D. G’alvez-L’opez, C. Paul, P. Jensfelt, and D. Kragic, “Object search and localization for an indoor mobile robot,” J. Computing and IT, vol. 17, no. 1, pp. 67–80, 2009.Google Scholar
  13. 13.
    A. Aydemir, K. Sj öö, J. Folkesson, A. Pronobis, and P. Jensfelt, “Search in the real world: Active visual object search based on spatial relations,” in ICRA, 2011.Google Scholar
  14. 14.
    T. Kollar and N. Roy, “Utilizing object-object and object-scene context when planning to find things,” in ICRA, 2009.Google Scholar
  15. 15.
    HomeMate errand service demo: http://www.youtube.com/watch?v=4ENFcHP1GaI
  16. 16.
    Ahmed M.Naguib, Sukhan Lee, “Adaptive Bayesian Recognition with Multiple Evidences,” 2014 The 4th International Conference on Multimedia Computing and Systems (ICMCS) (To be published by April 2014).Google Scholar
  17. 17.
    Xi Chen, Sukhan Lee, “Visual search of an object in cluttered environments for robotic errand service,” 2013 IEEE International Conference on Systems, Man, and Cybernetics (SMC).Google Scholar

Copyright information

© Springer International Publishing Switzerland 2016

Authors and Affiliations

  1. 1.School of Information and Communication EngineeringSungkyunkwan UniversitySuwonSouth Korea

Personalised recommendations