IEA/AIE 2010: Trends in Applied Intelligent Systems pp 92-101 | Cite as
Down-Up-Down Behavior Generation for Interactive Robots
Abstract
Behavior generation in humans and animals usually employs a combination of bottom-up and top-down patterns. Most available robotic architectures utilize either bottom-up or top-down activation including hybrid architectures. In this paper, we propose a behavior generation mechanism that can seamlessly combine these two strategies. One of the main advantages of the proposed approach is that it can naturally combine both bottom-up and top-down behavior generation mechanisms which can produce more natural behavior. This is achieved by utilizing results from the theory of simulation in neuroscience which tries to model the mechanism used in human infants to develop a theory of mind. The proposed approach was tested in modeling spontaneous gaze control during natural face to face interactions and provided more natural, human-like behavior compared with a state-of-the-art gaze controller that utilized a bottom-up approach.
Keywords
Behavior Generation Interactive Robot Theory Factor Current Role Interaction ProtocolPreview
Unable to display preview. Download preview PDF.
References
- 1.Brooks, R.A.: A robust layered control system for a mobile robot. IEEE Journal of Robotics and Automation 2(1), 14–23 (1986)Google Scholar
- 2.Arkin, R.C., Masahiro Fujita, T.T., Hasegawa, R.: An ethological and emotional basis for human-robot interaction, robotics and autonomous systems. Robotics and Autonomous Systems 42(3-4), 191–201 (2003)MATHCrossRefGoogle Scholar
- 3.Yang, L., Yue, J., Zhang, X.: Hybrid control architecture in bio-mimetic robot, pp. 5699–5703 (June 2008)Google Scholar
- 4.Ulam, P., Arkin, R.: Biasing behavioral activation with intent for an entertainment robot. Intelligent Service Robotics 1(3), 195–209 (2008)CrossRefGoogle Scholar
- 5.Wohlschlager, A., Gattis, M., Bekkering, H.: Action generation and action perception in imitation:an instance of the ideomotor principle. Phil. Trans. R. Soc. Lond. (358), 501–515 (2003)Google Scholar
- 6.Meltzoff, A., Moore, M.: Imitation, memory and the representation of persons. Infant Behav. Dev. (17), 83–99 (1994)Google Scholar
- 7.Mohammad, Y., Nishida, T.: Toward combining autonomy and interactivity for social robots. AI & Society (in press, 2008)Google Scholar
- 8.Breazeal, C.: Socially intelligent robots. Interactions 12(2), 19–22 (2005)CrossRefGoogle Scholar
- 9.Davies, M.: Stone: Mental Simulation. Blackwell Publishers, Oxford (1995)Google Scholar
- 10.Sabbagh, M.A.: Understanding orbitofrontal contributions to theory-of-mind reasoning: Implications for autism. Brain and Cognition (55), 209–219 (2004)Google Scholar
- 11.Oberman, L.M., McCleery, J.P., Ramachandran, V.S., Pineda, J.A.: Eeg evidence for mirror neuron activity during the observation of human and robot actions: Toward an analysis of the human qualities of interactive robots. Neurocomput. 70(13-15), 2194–2203 (2007)CrossRefGoogle Scholar
- 12.Mohammad, Y., Nishida, T.: Learning Interaction Structure using a Hierarchy of Dynamical Systems. Studies in Computational Intelligence. In: Opportunities and Challenges for Next-Generation Applied Intelligence, pp. 253–258. Springer, Heidelberg (2009)CrossRefGoogle Scholar
- 13.Mohammad, Y., Nishida, T.: Controlling gaze with an embodied interactive control architecture. In: Applied Intelligence (2009) (Published online May 30)Google Scholar