Abstract
The probabilistic framework developed in the previous chapter enables a manipulation robot to learn accurate kinematic models of articulated objects. As input, our framework requires a sequence of pose observations of the articulated object. We implemented the perception in the previous chapter using visual markers or by directly recording the end effector trajectory while the robot was manipulating the articulated object. For the daily use in domestic environments, however, both options are not satisfactory: clearly, it is neither desirable to augment all furniture with visual markers nor to guide a robot manually to the handles of all relevant objects.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Sturm, J. (2013). Vision-Based Perception of Articulated Objects. In: Approaches to Probabilistic Model Learning for Mobile Manipulation Robots. Springer Tracts in Advanced Robotics, vol 89. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-37160-8_5
Download citation
DOI: https://doi.org/10.1007/978-3-642-37160-8_5
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-37159-2
Online ISBN: 978-3-642-37160-8
eBook Packages: EngineeringEngineering (R0)