Abstract
This paper describes an architecture that enables a robot to represent, reason about, and learn affordances. Specifically, Answer Set Prolog is used to represent and reason with incomplete domain knowledge that includes affordances modeled as relations between attributes of the robot and the object(s) in the context of specific actions. The learning of affordance relations from observations obtained through reactive execution or active exploration is formulated as a reinforcement learning problem. A sampling-based approach and decision-tree regression with the underlying relational representation are used to obtain generic affordance relations that are added to the Answer Set Prolog program for subsequent reasoning. The capabilities of this architecture are illustrated and evaluated in the context of a simulated robot assisting humans in an indoor domain.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
We use the terms “robot”, “agent” and “learner” interchangeably in this paper.
References
Driessens, K., Ramon, J.: Relational instance-based regression for relational reinforcement learning. In: International Conference on Machine Learning, pp. 123–130 (2003)
Gabaldon, A.: Activity recognition with intended actions. In: International Joint Conference on Artificial Intelligence (IJCAI), Pasadena, USA, 11–17 July 2009
Gelfond, M., Inclezan, D.: Some properties of system descriptions of \(AL_d\). J. Appl. Non-Class. Logics, Special Issue Equilibr. Logic Answer Set Program. 23(1–2), 105–120 (2013)
Gelfond, M., Kahl, Y.: Knowledge Representation, Reasoning and the Design of Intelligent Agents. Cambridge University Press, New York (2014)
Gibson, J.J.: The Ecological Approach to Visual Perception. Psychology Press (1986)
Gil, Y.: Learning by experimentation: Incremental refinement of incomplete planning domains. In: International Conference on Machine Learning, New Brunswick, USA, pp. 87–95, 10–13 July 1994
Kirk, J., Mininger, A., Laird, J.: Learning task goals interactively with visual demonstrations. Biol. Inspired Cogn. Archit. 18, 1–8 (2016)
Oates, T., Cohen, P.R.: Searching for planning operators with context-dependent and probabilistic effects. In: AAAI Conference on Artificial Intelligence, Portland, USA, 4–8 August 1996
Otero, R.P.: Induction of the effects of actions by monotonic methods. In: International Conference on Inductive Logic Programming, pp. 299–310 (2003)
Ramenzoni, V.C., Davis, T.J., Riley, M.A., Shockley, K.: Perceiving action boundaries: Learning effects in perceiving maximum jumping-reach affordances. Atten. Percept. Psychophys. 72(4), 1110–1119 (2010)
Sahin, E., Cakmak, M., Dogar, M.R., Ugur, E., Ucoluk, G.: To afford or not to afford: A new formalization of affordances towards affordance-based robot control. Adapt. Behav. 15(4), 447–472 (2007)
Sarathy, V., Scheutz, M.: A logic-based computational framework for inferring cognitive affordances. IEEE Trans. Cogn. Devel. Syst. 8(3) (2016)
Shen, W.-M., Simon, H.: Rule creation and rule learning through environmental exploration. In: International Joint Conference on Artificial Intelligence, Detroit, USA, pp. 675–680, 20–25 August 1989
Shu, T., Ryoo, M.S., Zhu, S.-C.: Learning social affordance for human-robot interaction. In: International Joint Conference on Artificial Intelligence (IJCAI), New York, USA, 9–15 July 2016
Sridharan, M., Gelfond, M., Zhang, S., Wyatt, J.: A Refinement-Based Architecture for Knowledge Representation and Reasoning in Robotics. Technical report, April 2017. http://arxiv.org/abs/1508.03891
Sridharan, M., Meadows, B.: A combined architecture for discovering affordances, causal laws, and executability conditions. In: International Conference on Advances in Cognitive Systems (ACS), Troy, USA, 12–14 May 2017
Sridharan, M., Meadows, B., Gomez, R.: ’What can I not do? Towards an architecture for reasoning about and learning affordances. In: International Conference on Automated Planning and Scheduling (ICAPS), Pittsburgh, USA, 18–23 June 2017
Stoffregen, T.A., Yang, C.-M., Giveans, R., Flanagan, M., Bardy, B.G.: Movement in the perception of an affordance for wheelchair locomotion. Ecol. Psychol. 21(1), 1–36 (2009)
Sutton, R.L., Barto, A.G.: Reinforcement Learning: An Introduction. MIT Press, Cambridge (1998)
Tadepalli, P., Givan, R., Driessens, K.: Relational reinforcement learning: an overview. In: Relational Reinforcement Learning Workshop at International Conference on Machine Learning (2004)
Acknowledgments
This work was supported in part by the US Office of Naval Research award N00014-13-1-0766 and Asian Office of Aerospace Research and Development award FA2386-16-1-4071. All conclusions are those of the authors alone.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Sridharan, M., Meadows, B. (2017). Learning Affordances for Assistive Robots. In: Kheddar, A., et al. Social Robotics. ICSR 2017. Lecture Notes in Computer Science(), vol 10652. Springer, Cham. https://doi.org/10.1007/978-3-319-70022-9_1
Download citation
DOI: https://doi.org/10.1007/978-3-319-70022-9_1
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-70021-2
Online ISBN: 978-3-319-70022-9
eBook Packages: Computer ScienceComputer Science (R0)