Reactive Virtual Creatures for Dexterous Physical Interactions

  • Hironori Mitake
  • Shoichi Hasegawa
  • Makoto Sato
Part of the Lecture Notes in Computer Science book series (LNCS, volume 7060)


Dexterous physical interactions with virtual creatures are important to bring the fun of playing with animals into arts and entertainment. For reality of interaction, virtual creatures need to react to highly varied user inputs in a variety of ways according to physical and psychological laws. We propose constructing virtual creatures using a physical simulator, sensor/attention models, and physical motion controllers. The physical simulator and motion controllers generate highly varied physically real reactions, while sensor/attention models provide psychologically feasible target selection for motion controllers. Having constructed a virtual creature prototype, we realize communicative physical interactions such as guessing and attracting attention by touching it via a haptic device. We have confirmed the prototype’s effectiveness experimentally.


Virtual Creature Attention Sensorimotor System Motion Generation Physical Interaction 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Interactive tokyo 2010 (2010),
  2. 2.
    Abe, Y., Liu, C.K., Popović, Z.: Momentum-based parameterization of dynamic character motion. In: SCA 2004: Proc. of the 2004 ACM SIGGRAPH/Eurographics Symp. on Computer Animation, pp. 173–182. Eurographics Association, Aire-la-Ville (2004)CrossRefGoogle Scholar
  3. 3.
    Brooks, R.A., Stein, L.A.: Building brains for bodies. Auton. Robots 1(1), 7–25 (1994)CrossRefGoogle Scholar
  4. 4.
    Hasegawa, S., Sato, M.: Real-time Rigid Body Simulation for Haptic Interactions Based on Contact Volume of Polygonal Objects. Computer Graphics Forum 23(3), 529–538 (2004), CrossRefGoogle Scholar
  5. 5.
    Hasegawa, S., Toshiaki, I., Hashimoto, N.: Human scale haptic interaction with a reactive virtual human in a realtime physics simulator. In: ACE 2005: Proc. of the 2005 ACM SIGCHI Intl. Conf. on Advances in Computer Entertainment Technology, pp. 149–155. ACM, New York (2005)CrossRefGoogle Scholar
  6. 6.
    Lee, S.P., Badler, J.B., Badler, N.I.: Eyes alive. ACM Trans. Graph. 21(3), 637–644 (2002)Google Scholar
  7. 7.
    Itti, L., Dhavale, N., Pighin, F.: Realistic avatar eye and head animation using a neurobiological model of visual attention. In: Proc. SPIE Intl. Symp. on Optical Science and Technology, vol. 5200, pp. 64–78 (August 2003)Google Scholar
  8. 8.
    Mitsunaga, N., Miyashita, T., Yoshikawa, Y., Ishiguro, H., Kogure, K., Hagita, N.: Robovie-iv: An every day communication robot. Technical report of IEICE. PRMU 105(534), 47–52 (2006),
  9. 9.
    NINTENDO: nintendogs (2005),
  10. 10.
    Peters, C., Sullivan, C.O.: Bottom-up visual attention for virtual human animation. In: Proc. of Computer Animation for Social Agents (2003)Google Scholar
  11. 11.
    Rose, C., Guenter, B., Bodenheimer, B., Cohen, M.F.: Efficient generation of motion transitions using spacetime constraints. In: SIGGRAPH 1996: Proc. of the 23rd Annual Conf. on Computer Graphics and Interactive Techniques, pp. 147–154. ACM Press, New York (1996)CrossRefGoogle Scholar
  12. 12.
    Sato, M., Hirata, Y., Kawarada, H.: Space interface device for artificial reality. Journal of Robotics and Mechatronics 9(3), 177–184 (1997)CrossRefGoogle Scholar
  13. 13.
    SCEE: Eyepet (2009),
  14. 14.
    da Silva, M., Abe, Y., Popović, J.: Interactive simulation of stylized human locomotion. ACM Trans. Graph. 27(3), 1–10 (2008)CrossRefGoogle Scholar
  15. 15.
    Zordan, V.B., Majkowska, A., Chiu, B., Fast, M.: Dynamic response for motion capture animation. ACM Trans. Graph. 24(3), 697–701 (2005)CrossRefGoogle Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • Hironori Mitake
    • 1
  • Shoichi Hasegawa
    • 1
    • 2
  • Makoto Sato
    • 1
  1. 1.P&I Lab.Tokyo Institute of TechnologyKanagawaJapan
  2. 2.PRESTO JSTTokyoJapan

Personalised recommendations