Abstract
The advancement of traffic and computer networks makes the world more and more internationalized and increases the frequency of communications between people who speak different languages and show different nonverbal behaviors. To improve the communication of embodied conversational agent (ECA) systems with their human users, the importance of their capability to cover cultural differences emerged. Various excellent ECA systems are developed and proposed previously, however, the cross-culture communication issues are seldom addressed by researchers. This paper describes a short-term project aiming to explore the possibility of rapidly building multicultural and the multimodal ECA interfaces for a tour guide system by using a generic framework connecting their functional blocks.
Similar content being viewed by others
5. References
J. Cassell, J. Sullivan, S. Prevost, and E. Churchill, eds.,Embodied Conversational Agents. MIT Press, 2000. 41
H. Huang, T. Masuda, A. Cerekovic, K. Tarasenko, I. Pandzic, Y. Nakano, and T. Nishida, “Toward a Universal Platform for Integrating Embodied Conversational Agent Components”, inProceedings of the 10th International Conference Knowledge-Based Intelligent Information and Engineering Systems (KES 2006), (Bournemouth, UK), pp. 220–226, October 2006. 41, 43
The eNTERFACE’06 workshop on multimodal interfaces. http://enterface.tel.fer.hr. 41
OpenAIR protocol. http://www.mindmakers.org/openair/airPage.jsp. 42
NaturalPoint OptiTrack Flex 3. http://www.naturalpoint.com/optitrack/. 43
NEC/Tokin 3D motion sensor. http://www.nec-tokin.com/english/product/3d/. 43
MS SAPI (Microsoft Speech Application Programming Interface). http://www.microsoft.com/speech. 43
Java SAPI (Java Speech Application Programming Interface). http://java.sun.com/products/java-media/speech/. 43
Y. Nakano, M. Okamoto, D. Kawahara, Q. Li, and T. Nishida, “Converting Text into Agent Animations: Assigning Gestures to Text”, inProceedings of The Human Language Technology Conference (HLT-NAACL04), 2004. 43
AIML (Artificial Intelligence Markup Language). http://www.alicebot.org. 43
visage|SDK, Visage player, Visage Technologies. http://www.visagetechnologies.com. 43
Juman. http://nlp.kuee.kyoto-u.ac.jp/nl-resource/juman.html. 43
KNP. http://nlp.kuee.kyoto-u.ac.jp/nl-resource/knp.html. 43
G. Zoric and I. S. Pandzic, “A Real-time Language Independent Lip Synchronization Method Using a Genetic Algorithm”, inProceedings of ICME 2005, July 6–8 2005. 45
ARToolKit. http://artoolkit.sourceforge.net/. 46
S. Kopp, B. Krenn, S. Marsella, A. Marshall, C. Pelachaud, H. Pirker, K. Thorisson, and H. Vilhjalmsson, “Towards a Common Framework for Multimodal Generation: The Markup Language”, inProceedings of IVA2006, August 2006. 47
CUBE-G (Culture-adaptive Behavior Generation for interactions with embodied conversational agents). http://mm-werkstatt.informatik.uni-augsburg.de/projects/cube-g/. 47
F. Landragin, “Visual Perception, Language and Gesture: A Model for their Understanding in Multimodal Dialogue Systems”, inSignal Processing, Elsevier, 2006. 47
Author information
Authors and Affiliations
Rights and permissions
About this article
Cite this article
Huang, HH., Tarasenko, K., Nishida, T. et al. An agent based multicultural tour guide system with nonverbal user interface. J Multimodal User Interfaces 1, 41–48 (2007). https://doi.org/10.1007/BF02884431
Issue Date:
DOI: https://doi.org/10.1007/BF02884431