Context-Aware Multimodal Human–Computer Interaction

  • Siska Fitrianie
  • Zhenke Yang
  • Dragoş Datcu
  • Alin G. Chiţu
  • Léon J. M. Rothkrantz

Abstract

Crisis response and management involve the collaboration of many people. To perform and coordinate their activities, they must rely on detailed and accurate information about the crisis, the environment, and many more factors. To ensure collaboration of emergency services and high-quality care for victims, the ability to supply dynamic and contextually correlated information is necessary. However, current approaches to construct globally consistent views of crises suffer from problems identified in [60]: (a) the setting of events is constantly changing, (b) the information is distributed across geographically distant locations, and (c) the complexity of the crisis management organization makes it difficult and time consuming to collaborate and verify obtained information.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. 1.
    Albacete, P.L., Chang, S., Polese, G., Baker, B.: Iconic language design for people with significant speech and multiple impairments. In: Proc. of the ACM Assets 1994, CA, USA, pp. 23–30. ACM, New York (1994)Google Scholar
  2. 2.
    Alexandersson, J., Becker, T.: The Formal Foundations Underlying Overlay. In: IWCS-5, The Netherlands, pp. 22–36 (2003)Google Scholar
  3. 3.
    Anson, D.K., Moist, P., Przywara, M., Wells, H., Saylor, H., Maxime, H.: The Effects of Word Completion and Word Prediction on Typing Rates using On-Screen Keyboards. In: Proc. of RESNA. RESNA Press, Arlington (2005)Google Scholar
  4. 4.
    Basu, A., Sankar, S., Chakraborty, K., Bhattarcharya, S., Choudhury, M., Patel, R.: Vernacula Educational Communication Tool for the People with Multiple Disabilities. In: Development by Design Conference, Bangalore (2002)Google Scholar
  5. 5.
    Beardon, C.: CD-Icon, an iconic language-based on conceptual dependency. Intelligent Tutoring Media 3(4) (1992)Google Scholar
  6. 6.
    Benjamins, T.: MACSIM: Multi-Agent Crisis Simulator, Interpreter and Monitor. Master thesis, TU Delft (2006)Google Scholar
  7. 7.
    Bevelier, D., Corina, D.P., Neville, H.J.: Brain and Language: a Perspective from Sign Language. Neuron 21, 275–278 (1998)CrossRefGoogle Scholar
  8. 8.
    Bliss, C.K.: The Blissymbols Picture Book. Semantography Press, Sidney (1984)Google Scholar
  9. 9.
    Bosma, W., André, E.: Exploiting emotions to disambiguate dialogue acts. In: Proc. of IUI, Portugal, pp. 85–92 (2004)Google Scholar
  10. 10.
    Bui, T.H., Poel, M., Nijholt, A., Zwiers, J.: A Tractable DDN-POMDP Approach to Affective Dialogue Modeling for General Probabilistic Frame-based Dialogue Systems. In: IJCAI 2007, India (2007)Google Scholar
  11. 11.
    Carpenter, B.: The Logic of Typed Feature Structures. Cambridge University Press, England (1992)MATHCrossRefGoogle Scholar
  12. 12.
    Cassell, J.: Embodied conversational interface agents. Communication ACM 43(4), 70–78 (2000)CrossRefGoogle Scholar
  13. 13.
    Cassell, J., Stocky, T., Bickmore, T., Gao, Y., Nakano, Y., Ryokai, K., Tversky, D., Vaucelle, C., Vilhjálmsson, H.: MACK: Media lab Autonomous Conversational Kiosk. In: Proc. of Imagina 2002, Monte Carlo (2002)Google Scholar
  14. 14.
    Catizone, R., Setzer, A., Wilks, Y.: Multimodal dialogue management in the Comic project. In: Proc. of EACL, Hungary (2003)Google Scholar
  15. 15.
    Champoux, B., Fujisawa, K., Inoue, T., Iwadate, Y.: Transmitting Visual Information: Icons become Words. IEEE - Information Visualization, 244–249 (2000)Google Scholar
  16. 16.
    Chin Jr., G., Stephan, E.G., Gracio, D.K., Kuchar, O.A., Whitney, P.D., Schuchardt, K.L.: Developing Concept-Based User Interfaces for Scientific Computing. Computer 39(9), 26–34 (2006)CrossRefGoogle Scholar
  17. 17.
    Chiţu, A.G., Rothkrantz, L.J.M., Wiggers, P.: Comparison between different feature extraction techniques for audio-visual. Journal on Multimodal User Interfaces 1(1), 7–20 (2007)CrossRefGoogle Scholar
  18. 18.
    Cohn, N.: Visual Syntactic Structures, Towards a Generative Grammar of Visual Language (2003), http://www.emaki.net/essays/
  19. 19.
    Cootes, T.F., Edwards, G.J., Taylor, C.J.: Active Appearance Models. In: European Conf. Computer Vision, pp. 484–498 (1998)Google Scholar
  20. 20.
    Datcu, D., Rothkrantz, L.J.M.: Multimodal workbench for human emotion recognition. In: Software Demo at IEEE CVPR 2007, USA (2007)Google Scholar
  21. 21.
    Datcu, D., Rothkrantz, L.J.M.: The use of Active Appearance Model for facial expression recognition in crisis environments. In: Proc. of ISCRAM 2007, USA, pp. 515–524 (2007)Google Scholar
  22. 22.
    Davis, S., Mermelstein, P.: Comparison of Parametric Representations for Monosyllabic Word Recognition. IEEE ASSP 28, 357–366 (1980)CrossRefGoogle Scholar
  23. 23.
    Dor, R.: The Ear’s Mind, a Computer Model of the Fundamental Mechanisms of the Perception of Sound, Master thesis, TU Delft (2005)Google Scholar
  24. 24.
    Dymon, U.J.: An Analysis of Emergency Map Symbology. Int. Journal of Emergency Management 1(3), 227–237 (2003)CrossRefGoogle Scholar
  25. 25.
    Ekman, P., Friesen, W.F.: Unmasking the Face. Prentice-Hall, Inc., Englewood Cliffs (1975)Google Scholar
  26. 26.
    Farberow, N.L., Frederick, C.J.: Training Manual for Human Service Workers in Major Disasters. Rockville, Maryland - National Institute of Mental Health (1978)Google Scholar
  27. 27.
    Fitrianie, S.: An Icon-based Communication Tool on a PDA, Postgraduate Thesis. TU Eindhoven (2004)Google Scholar
  28. 28.
    Fitrianie, S., Rothkrantz, L.J.M.: Communication in Crisis Situations using Icon Language. In: Proc. of IEEE ICME 2005, The Netherlands, pp. 1370–1373 (2005)Google Scholar
  29. 29.
    Fitrianie, S., Datcu, D., Rothkrantz, L.J.M.: Constructing Knowledge of the World in Crisis Situations using Visual Language. In: Proc of. IEEE SMC 2006, Taiwan, pp. 121–126 (2006)Google Scholar
  30. 30.
    Fitrianie, S., Rothkrantz, L.J.M.: An Adaptive Keyboard with Personalized Language-based Features. In: Matoušek, V., Mautner, P. (eds.) TSD 2007. LNCS (LNAI), vol. 4629, pp. 131–138. Springer, Heidelberg (2007)CrossRefGoogle Scholar
  31. 31.
    Fitrianie, S., Yang, Z., Rothkrantz, L.J.M.: Developing concept-based user interface using icons for reporting observations. In: ISCRAM 2008, USA (2008)Google Scholar
  32. 32.
    Foster, M.E., White, M., Setzer, A., Catizone, R.: Multimodal generation in the COMIC dialogue. In: ACL 2005, USA, pp. 45–48 (2004)Google Scholar
  33. 33.
    Gustafson, J., Bell, L., Beskow, J., Boye, J., Carlson, R., Edlund, J., Granström, B., House, D., Wirén, M.: Adapt - a Multimodal Conversational Dialogue System in an Apartment Domain. In: ICSLP 2000 (2), pp. 134–137 (2000)Google Scholar
  34. 34.
    Housz, T.I.: The elephant’s memory (1994–1996), http://www.khm.de/timot
  35. 35.
    Homeland Security Working Group: Symbology Reference (2003), http://www.fgdc.gov/HSWG/index.html
  36. 36.
    IBM e-business Case Studies: Bullhead City Police Department, http://www.ibm.com/e-business/doc/content/casestudy/47405.html
  37. 37.
    Johanson, B., Fox, A., Winograd, T.: The Interactive Workspaces Project: Experiences with Ubiquitous Computing Rooms. IEEE Pervasive Computing 1(2), 67–74 (2002)CrossRefGoogle Scholar
  38. 38.
    Joshi, A.K., Vijay-Shanker, K.: Compositional Semantics for Lexicalized Tree-Adjoining Grammars. In: Proc. of Computational Semantics, The Netherlands (1999)Google Scholar
  39. 39.
    Kanade, T., Cohn, J.F., Tian, Y.: Comprehensive database for facial expression analysis. In: Proc. of IEEE Automatic Face and Gesture Recognition, France (2000)Google Scholar
  40. 40.
    Karlson, A., Bederson, B., Contreras-Vidal, J.: Understanding Single Handed Use of Handheld Devices. In: Jo, L. (ed.) Handbook of Research on User Interface Design and Evaluation for Mobile Technology (2006) (in press)Google Scholar
  41. 41.
    Keizer, S.: Reasoning under uncertainty in Natural Language Dialogue using Bayesian Networks, Doctoral Dissertation. Twente University (2003)Google Scholar
  42. 42.
    Kjeldskov, J., Kolbe, N.: Interaction Design for Handheld Computers. In: Proc. of APCHI 2002. Science Press, China (2002)Google Scholar
  43. 43.
    Leemans, N.E.M.P.: VIL: A Visual Inter Lingua. Doctoral Dissertation, Worcester Polytechnic Institute, USA (2001)Google Scholar
  44. 44.
    Lucas, B.D., Kanade, T.: An Iterative Image Registration Technique with an Application to Stereo Vision. In: IJCAI 1981, pp. 674–679 (1981)Google Scholar
  45. 45.
    MacKenzie, I.S., Zhang, S.X., Soukoreff, R.W.: Text Entry using Soft Keyboards. Behaviour and Information Technology 18, 235–244 (1999)CrossRefGoogle Scholar
  46. 46.
    Mankoff, J., Abowd, G.D.: Cirrin - A Word-Level Unistroke Keyboard for Pen Input. In: ACM UIST 1998, pp. 213–214 (1998)Google Scholar
  47. 47.
    Maybury, M.: Intelligent User Interfaces: An Introduction. In: Proc. of IUI 1999, pp. 3–4. ACM, New York (1999)CrossRefGoogle Scholar
  48. 48.
    McTear, M.F.: Spoken dialogue technology: enabling the conversational user interface. ACM Computing Surveys 34(1), 90–169 (2002)CrossRefGoogle Scholar
  49. 49.
    Mehrotra, S., Butts, C., Kalashnikov, D., Venkatasubramanian, N., Altintas, K., Hariharan, P., Lee, H., Ma, Y., Myers, A., Wickramasuriya, J., Eguchi, R., Huyck, C.: Camas - a Citizen Awareness System for Crisis Mitigation. In: Proc. of ACM SIGMOD 2004, New York, USA, pp. 955–956 (2004)Google Scholar
  50. 50.
    MESA (Mobile Broadband for Emergency and Safety Applications) project, http://www.projectmesa.org
  51. 51.
    Moore, L.K.: CRS Report for Congress: Public Safety Communication Policy, Confessional Research Service, the Library of Congress (2006)Google Scholar
  52. 52.
    Norman, D.: Things That Make Us Smart. Addison-Wesley Publishing Co., Reading (1993)Google Scholar
  53. 53.
    den Os, E., Boves, L.: Towards Ambient Intelligence: Multimodal Computers that Understand Our Intentions. In: Proc. of eChallenges e-2003 (2003)Google Scholar
  54. 54.
    Otten, J., van Heijningen, B., Lafortune, J.F.: The Virtual Crisis Management center - An ICT implementation to canalize information! In: ISCRAM 2004, Brussels (2004)Google Scholar
  55. 55.
    Oviatt, S., Coulston, R., Lunsford, R.: When Do We Interact Multimodally?: Cognitive Load and Multimodal Communication Patterns. In: Proc. of ICMI 2004, pp. 129–136. ACM, New York (2004)CrossRefGoogle Scholar
  56. 56.
    Perlovsky, L.I.: Emotions, Learning and Control. In: Proc. of International Symposium: Intelligent Control, Intelligent Systems and Semiotics, pp. 131–137 (1999)Google Scholar
  57. 57.
    Pfleger, N., Alexandersson, J., Becker, T.: Scoring Functions for Overlay and their Application in Discourse Processing. In: KONVENS 2002, Germany, pp. 139–146 (2002)Google Scholar
  58. 58.
    Potamiano, A., Ammicht, E., Kuo, H.-K.J.: Dialogue Management the Bell Labs Communicator System. In: Proc. of ICSLP, Beijing, China, vol. 2, pp. 603–606 (2000)Google Scholar
  59. 59.
    Proxem, Proxem Antilope (2008), http://www.proxem.com
  60. 60.
    Ramaswamy, S., Rogers, M., Crockett, A.D., Feaker, D., Carter, M.: WHISPER – Service Integrated Incident Management System. Int. Journal of Intelligent Control and Systems 11(2), 114–123 (2006)Google Scholar
  61. 61.
    Rothkrantz, L.J.M., van Vark, R.J., Peters, A., Andeweg, N.A.: Dialogue control the Alparon system. In: Sojka, P., Kopeček, I., Pala, K. (eds.) TSD 2000. LNCS (LNAI), vol. 1902, pp. 333–338. Springer, Heidelberg (2000)CrossRefGoogle Scholar
  62. 62.
    Schank, R., Abelson, R.: Scripts, Plans, Goals and Understanding. Erlbaum, Hillsdale (1977)MATHGoogle Scholar
  63. 63.
    Museum, S.F.: The Virtual Museum of the City of San Francisco: San Francisco 9-1-1 Dispatch Tapes, October 17 (1989), http://www.sfmuseum.org/1989/sf911.html, http://www.sfmuseum.net/1989/sc911.html
  64. 64.
    Sharma, R., Yeasin, M., Krahnstoever, N., Rauschert, I., Cai, G., Brewer, I., MacEachren, A.M., Sengupta, K.: IEEE, Speech-Gesture Driven Multimodal Interfaces for Crisis Management.  91(9), 1327–1354 (2003)Google Scholar
  65. 65.
    Singhal, A., Rattine-Flaherty, E.: Pencils and Photos as Tools of Communicative Research and Praxis. International Communication Gazette 68(4), 313–330 (2006)CrossRefGoogle Scholar
  66. 66.
    Smith, P.: Toward a Practical Measure of Hypertext Usability. Interacting with Computers 8(4), 365–381 (1996)CrossRefGoogle Scholar
  67. 67.
    Steedman, M., Baldridge, J.: Combinatory Categorial Grammar. In: Borsley, R., Borjars, K. (eds.) Non-Transformational Syntax. Blackwell, Malden (2005)Google Scholar
  68. 68.
    Tatomir, B., Rothkrantz, L.J.M.: Intelligent System for Exploring Dynamic Crisis Environments. In: Van de Walle, B., Turoff, M. (eds.) ISCRAM 2006, Newark, NJ, SA (2006)Google Scholar
  69. 69.
    Vark, R.J.V., de Vreught, J.P.M., Rothkrantz, L.J.M.: Classification of public transport information dialogues using an information-based coding scheme. In: Proc. of ECAI, pp. 55–69 (1996)Google Scholar
  70. 70.
    Viola, P., Jones, M.: Robust Real-time Object Detection. In: 2nd Int. Workshop on Statistical and Computational Theories of Vision-Modeling, Learning, Computing, and Sampling (2001)Google Scholar
  71. 71.
    Wahlster, W.: Dialogue Systems Go Multimodal: The Smartkom experience. In: SmartKom: Foundations of Multimodal Dialogue Systems. Springer, Heidelberg (2006)CrossRefGoogle Scholar
  72. 72.
    Wallace, R.S.: The Elements of AIML Style (2003), http://alicebot.org
  73. 73.
    Ward, D.A., Blackwell, A., MacKay, D.: Dasher - a Data Entry Interface Using Continuous Gesture and Language Models. In: ACM UIST, pp. 129–136. ACM, NY (2000)CrossRefGoogle Scholar
  74. 74.
    WNBC: Exclusive - 911 tapes tell horror of 9/11 (part 1 and 2): Tapes released for first time (2002), http://www.wnbc.com/news/1315646/detail.html
  75. 75.
    Yang, Z., Rothkrantz, L.J.M.: Dynamic Scripting in Crisis Environments. In: HCI International 2007, China, pp. 554–563 (2007)Google Scholar
  76. 76.
    Zhai, S., Kristensson, P.-O.: Shorthand Writing on Stylus Keyboard. In: ACM CHI, pp. 97–104 (2003)Google Scholar
  77. 77.
    Zue, V., Seneff, S., Glass, J., Polifroni, J., Pao, C., Hazen, T., Hetherington, L.: Jupiter - A telephone-based conversational interface for weather information. IEEE Trans. on Speech and Audio Processing 8(1) (2000)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2010

Authors and Affiliations

  • Siska Fitrianie
    • 1
  • Zhenke Yang
    • 1
  • Dragoş Datcu
    • 1
  • Alin G. Chiţu
    • 1
  • Léon J. M. Rothkrantz
    • 1
    • 2
  1. 1.Man Machine InteractionDelft University of TechnologyThe Netherlands
  2. 2.Faculty of Technical SciencesNetherlands Defence AcademyDen HelderThe Netherlands

Personalised recommendations