Abstract
In this chapter we introduce the testbed domains used in the remainder of this monograph, where we informally define a domain as a setting for one or more tasks (i.e., MDPs). The primary purpose of this chapter is to provide sufficient background to allow a reader to fully understand the transfer learning experiments in subsequent chapters. In addition to describing each domain, we explain how each can be learned with one or more of the RL methods discussed in the previous chapter. In every domain we will emphasize how some tasks are faster to master than others. In general, experiments in this monograph transfer from relatively simple, quick to learn, source task to a more complex target task. In the target task time scenario, an effective TL algorithm may reduce the target task learning time regardless of task ordering, but only by ordering tasks in order of increasing difficulty will the total training time be reduced.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
Taylor, M.E. (2009). Empirical Domains. In: Transfer in Reinforcement Learning Domains. Studies in Computational Intelligence, vol 216. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-01882-4_4
Download citation
DOI: https://doi.org/10.1007/978-3-642-01882-4_4
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-01881-7
Online ISBN: 978-3-642-01882-4
eBook Packages: EngineeringEngineering (R0)