Summary
Humans manage to adapt learned movements very quickly to new situations by generalizing learned behaviors from similar situations. In contrast, robots currently often need to re-learn the complete movement. In this chapter, we propose a method that learns to generalize parametrized motor plans by adapting a small set of global parameters, called meta-parameters. We employ reinforcement learning to learn the required meta-parameters to deal with the current situation, described by states. We introduce an appropriate reinforcement learning algorithm based on a kernelized version of the reward-weighted regression. To show its feasibility, we evaluate this algorithm on a toy example and compare it to several previous approaches. Subsequently, we apply the approach to three robot tasks, i.e., the generalization of throwing movements in darts, of hitting movements in table tennis, and of throwing balls where the tasks are learned on several different real physical robots, i.e., a Barrett WAM, a BioRob, the JST-ICORP/SARCOS CBi and a Kuka KR 6.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Kober, J., Peters, J. (2014). Reinforcement Learning to Adjust Parametrized Motor Primitives to New Situations. In: Learning Motor Skills. Springer Tracts in Advanced Robotics, vol 97. Springer, Cham. https://doi.org/10.1007/978-3-319-03194-1_5
Download citation
DOI: https://doi.org/10.1007/978-3-319-03194-1_5
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-03193-4
Online ISBN: 978-3-319-03194-1
eBook Packages: EngineeringEngineering (R0)