Development of a Tour–Guide Robot Using Dialogue Models and a Cognitive Architecture

Conference paper
Part of the Lecture Notes in Computer Science book series (LNCS, volume 6433)


In this paper, we present the development of a tour–guide robot that conducts a poster session through spoken Spanish. The robot is able to navigate around its environment, visually identify informational posters, and explain sections of the posters that users request via pointing gestures. We specify the task by means of dialogue models. A dialogue model defines conversational situations, expectations and robot actions. Dialogue models are integrated into a novel cognitive architecture that allow us to coordinate both human–robot interaction and robot capabilities in a flexible and simple manner. Our robot also incorporates a confidence score on visual outcomes, the history of the conversation and error prevention strategies. Our initial evaluation of the dialogue structure shows the reliability of the overall approach, and the suitability of our dialogue model and architecture to represent complex human–robot interactions, with promising results.


Dialogue systems service robots human–robot interaction pointing gestures 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Pineda, L.A.: Specification and Interpretation of Multimodal Dialogue Models for Human-Robot Interaction, In: Sidorov, G. (ed.) Artificial Intelligence for Humans: Service Robots and Social Modeling, pp. 33–50. Sociedad Mexicana de Inteligencia Artificial (2008)Google Scholar
  2. 2.
    Pineda, L.A., Meza, I., Salinas, L.: Dialogue Model Specification and Interpretation for Intelligent Multimodal HCI. In: Kuri-Morales, A., Simari, G. (eds.) IBERAMIA 2010. LNCS (LNAI), vol. 6433, pp. 20–29. Springer, Heidelberg (2010)Google Scholar
  3. 3.
    Rascón, C., Avilés, H., Pineda, L.A.: Robotic Orientation towards Speaker in Human-Robot Interaction. In: Kuri-Morales, A., Simari, G. (eds.) IBERAMIA 2010. LNCS (LNAI), vol. 6433, pp. 10–19. Springer, Heidelberg (2010)Google Scholar
  4. 4.
    Tojo, T., Matsusaka, Y., Ishii, T.: A Conversational Robot Utilizing Facial and Body Expressions. In: IEEE International Conference on Systems, Man and Cybernetics, vol. 2, pp. 858–863 (2000)Google Scholar
  5. 5.
    Rogalla, O., Ehrenmann, O., Zoellner, M., Becher, R., Dillmann, R.: Using gesture and speech control for commanding a robot assistant. In: 11th IEEE Workshop on Robot and Human Interactive Communication, pp. 454–459 (2002)Google Scholar
  6. 6.
    Burger, B., Lerasle, F., Ferrane, I., Clodic, A.: Mutual Assistance between Speech and Vision for Human–Robot Interaction. In: IEEE/RSJ International Conference on Intelligent Robotics and Systems, pp. 4011–4016 (2008)Google Scholar
  7. 7.
    Aviles, H., Sucar, E., Vargas, B., Sanchez, J., Corona, E.: Markovito: A Flexible and General Service Robot. In: Liu, D., Wang, L., Tan, K.C. (eds.) Studies in Computational Intelligence, vol. 177, pp. 401–423. Springer, Heidelberg (2009)Google Scholar
  8. 8.
    Toptsis, I., Haasch, A., Hüwel, S., Fritsch, J., Fink, G.A.: Modality Integration and Dialog Management for a Robotic Assistant. In: European Conference on Speech Communication and Technology, pp. 837–840 (2005)Google Scholar
  9. 9.
    Lee, C., Cha, Y.S., Kuc, T.Y.: Implementation of Dialog System for Intelligent Service Robots. In: International Conference on Control, Automation and Systems, pp. 2038–2042 (2008)Google Scholar
  10. 10.
    Stiefelhagen, R., Ekenel, H.K., Fugen, C., Gieselmann, P., Holzapfel, H., Kraft, F., Nickel, K., Voit, M., Waibel, A.: Enabling Multimodal Human–Robot Interaction for the Karlsruhe Humanoid Robot. Trans. on Robotics: Special Issue on Human–Robot Interaction 23(5), 840–851 (2007)CrossRefGoogle Scholar
  11. 11.
    Meza, I., Salinas, S., Venegas, S., Castellanos, H., Chavarria, A., Pineda, L.A.: Specification and Evaluation of a Spanish Conversational System Using Dialogue Models. In: Kuri-Morales, A., Simari, G. (eds.) IBERAMIA 2010. LNCS (LNAI), vol. 6433, pp. 346–355. Springer, Heidelberg (2010)Google Scholar
  12. 12.
    Aviles, H., Meza, I., Aguilar, W., Pineda L.: Integrating Pointing Gestures 2nd International Conference on Agents and Artificial Intelligence, pp. 585–588 (2010)Google Scholar
  13. 13.
    Pineda, L., Castellanos, H., Cuétara, J., Galescu, L., Juárez, J., Llisterri, J., Pérez, P., Villaseñor, L.: The corpus dimex100: Transcription and evaluation. Language Resources and Evaluation (2009)Google Scholar
  14. 14.
    Dutoit, T., Pagel, V., Pierret, N., Bataille, F., Van Der Vrecken, O.: The MBROLA Project: Towards a Set of High-Quality Speech Synthesizers Free of Use for Non–Commercial Purposes. In: 4th International Conference on Spoken Language Processing, vol. 3, pp. 1393–1396 (1996)Google Scholar
  15. 15.
    Walker, M., Litman, D., Kamm, C., Kamm, A.A., Abella, A.: Paradise: A framework for evaluating spoken dialogue agents. In: 35th Annual Meeting of the Association for Computational Linguistics, pp. 271–280 (1997)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2010

Authors and Affiliations

  1. 1.Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas (IIMAS)Universidad Nacional Autónoma de México (UNAM)Mexico

Personalised recommendations