Abstract
In this paper, we present a reactive vision based autonomous behavior on a humanoid robot. This perception action scheme enables the robot to walk toward a desired position or to track a given target and grasp it.
The proposed solution uses visual servoing as a reactive sensor based technique to address the problem of perception, thus to ensure reachability of the target object by generating appropriate motion commands. An omnidirectional velocity tracking gait generator is considered to solve the locomotion and balance problem. It is based on the framework developed by Herdt et al. (Advanced Robotics 24(5–6): 719–737, 2010; 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 190–195, 2010), which has been extended for instantaneous rotational velocity in this work. The redundancy problem due to the high number of degrees of freedom associated with the mobility in operational space is handled by a task sequencing technique. This allows the improvement of reachability and manipulability while accounting for joints limits and obstacles avoidance. Finally, simulations on the humanoid robot NAO are shown to evaluate the effectiveness of the proposed scheme.
Keywords
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
A. Herdt, H. Diedam, P.-B. Wieber, D. Dimitrov, K. Mombaur, and M. Diehl, “Online walking motion generation with automatic footstep placement,” Advanced Robotics, vol. 24, no. 5–6, pp. 719–737, 2010.
A. Herdt, N. Perrin, and P.-B. Wieber, “Walking without thinking about it,” 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 190–195, Oct. 2010.
G. Taylor and L. Kleeman, “Hybrid position-based visual servoing with online calibration for a humanoid robot,” 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No.04CH37566), vol. 1, pp. 686–691, 2004.
Y. Pang, Q. Huang, D. Jia, Y. Tian, J. Gao, and W. Zhang, “Object manipulation of a humanoid robot based on visual Servoing,” 2007 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 1124–1129, Oct. 2007.
F. Chaumette and S. Hutchinson, “Visual servo control. i. basic approaches,” Robotics Automation Magazine, IEEE, vol. 13, no. 4, pp. 82– 90, 2006.
A. Moughlbay, Amine, E. Cervera, and M. Philippe, “Real-time model based visual servoing tasks on a humanoid robot,” in Intelligent Autonomous Systems 12, Advances in Intelligent Systems and Computing, pp. pp 321–333, Springer Berlin Heidelberg, 2013.
N. Mansard, O. Stasse, F. Chaumette, and K. Yokoi, “Visually-guided grasping while walking on a humanoid robot,” in Robotics and Automation, 2007 IEEE International Conference on, pp. 3041–3047, 2007.
D. J. Agravante, J. Pagès, and F. Chaumette, “Visual Servoing for the REEM Humanoid Robot’s Upper Body,” in IEEE Int. Conf. on Robotics and Automation, ICRA’13, (Karlsruhe, Allemagne), pp. 5233–5238, May 2013.
M. Marey and F. Chaumette, “A new large projection operator for the redundancy framework,” in Robotics and Automation (ICRA), 2010 IEEE International Conference on, pp. 3727–3732, 2010.
N. Courty, E. Marchand, and B. Arnaldi, “Through-the-eyes control of a virtual humanoid,” in Computer Animation, 2001. The Fourteenth Conference on Computer Animation. Proceedings, pp. 74–83, 2001.
P. Michel, J. Chestnutt, J. Kuffner, and T. Kanade, “Vision-guided humanoid footstep planning for dynamic environments,” 5th IEEE-RAS International Conference on Humanoid Robots, 2005., pp. 13–18.
C. Dune, a. Herdt, O. Stasse, P.-B. Wieber, K. Yokoi, and E. Yoshida, “Cancelling the sway motion of dynamic walking in visual servoing,” 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems, pp. 3175–3180, Oct. 2010.
C. Dune, A. Herdt, E. Marchand, O. Stasse, P.-B. Wieber, and E. Yoshida, “Vision based control for Humanoid robots,” in IROS Workshop on Visual Control of Mobile Robots (ViCoMoR), (San Francisco, USA, États-Unis), pp. 19–26, 2011.
B. Espiau, F. Chaumette, and P. Rives, “A new approach to visual servoing in robotics,” IEEE Transactions on Robotics and Automation, vol. 8, no. 3, pp. 313–326, 1992.
F. Chaumette and S. Hutchinson, “Visual servo control. ii. advanced approaches [tutorial],” Robotics Automation Magazine, IEEE, vol. 14, no. 1, pp. 109–118, 2007.
S. Hutchinson, G. D. Hager, and P. I. Corke, “A Tutorial on Visual Servo Control,” vol. 12, no. 5, 1996.
M. Gienger, H. Janssen, and C. Goerick, “Task-oriented whole body motion for humanoid robots,” in Humanoid Robots, 2005 5th IEEE-RAS International Conference on, pp. 238–244, 2005.
“Automatic supervisory control of the configuration and behavior of multibody mechanisms,” Systems, Man and Cybernetics, IEEE Transactions on, vol. 7, no. 12, pp. 868–871, 1977.
Aldebaran-robotics, “NAO Technical overview, hardware: Links V4.0.” Retrieved October 25, 2013, from https://community.aldebaran-robotics.com/doc/1-14/%20429%20family/robots/linksrobot.html.
M. A.-R. Marey, Contribution to control modeling in visual servoing, task Redundancy, and joint limits avoidance. PhD thesis, Universite de Rennes 1, 2010.
Acknowledgment
The author would like to thank Prof. M. Braae for his valuable comments.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2015 Springer International Publishing Switzerland
About this paper
Cite this paper
Bombile, M. (2015). Visual Servoing Based Positioning and Object Tracking on Humanoid Robot. In: Elleithy, K., Sobh, T. (eds) New Trends in Networking, Computing, E-learning, Systems Sciences, and Engineering. Lecture Notes in Electrical Engineering, vol 312. Springer, Cham. https://doi.org/10.1007/978-3-319-06764-3_4
Download citation
DOI: https://doi.org/10.1007/978-3-319-06764-3_4
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-06763-6
Online ISBN: 978-3-319-06764-3
eBook Packages: EngineeringEngineering (R0)