Abstract
It is difficult for a developer to account for all the surface linguistic forms that users might need in a spoken dialogue computer application. In any specific case users might need additional concepts not pre-programmed by the developer. This chapter presents a method for adapting the vocabulary of a spoken dialogue interface at run-time by end-users. The adaptation is based on expanding existing pre-programmed concept classes by adding new concepts in these classes. This adaptation is classified as a supervised learning method in which users are responsible for indicating the concept class and the semantic representation for the new concepts. This is achieved by providing users with a number of rules and ways in which the new language knowledge can be supplied to the computer. Acquisition of new linguistic knowledge at the surface and semantic levels is done using multiple modalities, including speaking, typing, pointing, touching or image capturing. Language knowledge is updated and stored in a semantic grammar and a semantic database.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bolt, R. A. (1980). Put-That-There: Voice and Gesture at the Graphics Interface. Computer Graphics, 14(3):262–270.
Chomsky, N. (1986). Knowledge of Language: Its Nature, Origin, and Use. New York: Praeger.
Cohen, P. R., Johnston, M., McGee, D., Oviatt, S. L., Clow, J., and Smith, I. (1998). The Efficiency of Multimodal Interaction: A Case Study. In Proceedings of International Conference on Spoken Language Processing (ICSLP), volume 2, pages 249–252, Sydney, Australia.
de Mori, R. (1999). Recognizing and Using Knowledge Structures in Dialog Systems. In Proceedings of the IEEE Workshop on Automatic Speech Recognition and Understanding (ASRU), pages 297–307, Keystone, Colorado, USA.
de Villiers, J. G. and de Villiers, P. A. (1978). Language Acquisition. Cambridge, Massachusetts, and London, England: Harvard University Press.
Dusan, S. and Flanagan, J. (2001). Human Language Acquisition by Computers. In Proceedings of the International Conference on Robotics, Distance Learning and Intelligent Communication Systems, pages 387–392, Malta. WSES/IEEE.
Dusan, S. and Flanagan, J. (2002a). Adaptive Dialog Based upon Multimodal Language Acquisition. In Proceedings of the Fourth International Conference on Multimodal Interfaces (ICMI), pages 135–140, Pittsburgh, PA, USA.
Dusan, S. and Flanagan, J. (2002b). An Adaptive Dialogue System Using Multimodal Language Acquisition. In Proceedings of the International CLASS Workshop: Natural, Intelligent and Effective Interaction in Multimodal Dialogue Systems, pages 72–75, Copenhagen, Denmark.
Gavalda, M. and Waibel, A. (1998). Growing Semantic Grammars. In Proceedings of the 17th International Conference on Computational Linguistics and the 36th Annual Meeting of the Association for Computational Linguistics (COLING-ACL), pages 451–456, Montreal, Quebec, Canada.
Gorin, A. L, Levinson, S. E., Gertner, A., and Goldman, E. (1991). On Adaptive Acquisition of Language. Computer Speech and Language, 5(2):101–132.
Gorin, A. L., Levinson, S. E., and Sankar, A. (1994). An Experiment in Spoken Language Acquisition. IEEE Transaction on Speech and Audio, 2(1):224–240. Part II.
Gorin, A. L., Riccardi, G., and Wright, J. H. (1997). How May I Help You? Speech Communication, 23:113–127.
Henis, E. A, Levinson, S. E., and Gorin, A. L. (1994). Mapping Natural Language and Sensory Information into Manipulatory Actions. In Proceedings of the Yale Workshop on Adaptive and Learning Systems, pages 324–356, Yale University, New Haven.
Miller, L. G. and Gorin, A. L. (1993). Structured Networks for Adaptive Language Acquisition. International Journal of Pattern Recognition and Artificial Intelligence, 7(4):873–898.
Oates, T. (2001). Grounding Knowledge in Sensors: Unsupervised Learning for Language and Planning. PhD thesis, MIT.
Roy, D. K. (1999). Learning Words from Sights and Sounds: A Computational Model. PhD thesis, MIT, Program in Media Arts and Sciences, School of Architecture and Planning.
Sadek, D. and de Mori, R. (1998). Dialogue Systems. In de Mori, R., editor, Spoken Dialogues with Computers, pages 563–582. Academic Press.
Sankar, A. and Gorin, A. L. (1993). Adaptive Language Acquisition in a Multisensory Device. In Mammone, R., editor, Artificial Neural Networks for Speech and Vision, pages 324–356. London: Chapman and Hall.
Young, S. (2002). Talking to Machines (Statistically Speaking). In Proceedings of International Conference on Spoken Language Processing (ICSLP), pages 9–16, Denver, Colorado, USA.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2005 Springer
About this chapter
Cite this chapter
Dusan, S., Flanagan, J. (2005). Adaptive Human-Computer Dialogue. In: van Kuppevelt, J.C.J., Dybkjær, L., Bernsen, N.O. (eds) Advances in Natural Multimodal Dialogue Systems. Text, Speech and Language Technology, vol 30. Springer, Dordrecht. https://doi.org/10.1007/1-4020-3933-6_15
Download citation
DOI: https://doi.org/10.1007/1-4020-3933-6_15
Publisher Name: Springer, Dordrecht
Print ISBN: 978-1-4020-3932-4
Online ISBN: 978-1-4020-3933-1
eBook Packages: Humanities, Social Sciences and LawSocial Sciences (R0)