Skip to main content
Log in

Introduction to the special issue on multimodal corpora for modeling human multimodal behavior

  • Published:
Language Resources and Evaluation Aims and scope Submit manuscript

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

References

  • Almeida, L., Amdal, I., Beires, N., Boualem, M., Boves, L., Os, E., et al. (2002). The MUST Guide to Paris; Implementation and expert evaluation of a multimodal tourist guide to Paris. Multi-modal dialogue in mobile environments, ISCA Tutorial and Research Workshop (IDS’2002), Kloster Irsee, Germany. http://www.isca-speech.org/archive/ids_02.

  • André, E. (2006). Corpus-based approaches to behavior modeling for virtual humans: A critical review, Modeling communication with robots and virtual humans. Workshop of the ZiF: Research Group 2005/2006 “Embodied communication in humans and machines”. Scientific Organization: Ipke Wachsmuth (Bielefeld), Günther Knoblich (Newark).

  • Argyle, M. (2004). Bodily communication (2nd ed.). London and New York: Routledge, Taylor & Francis.

    Google Scholar 

  • Bernsen, N. O., & Dybkjær, L. (2004). Evaluation of spoken multimodal conversation. In Sixth International Conference on Multimodal Interaction (ICMI’2004). New York: Association for Computing Machinery (ACM).

  • Beun, R.-J., & Cremers, A. (2001). Multimodal reference to objects: An empirical approach. In Proceedings of Cooperative Multimodal Communication: Second International Conference (CMC’98). Revised Papers, Tilburg, The Netherlands: Springer-Verlag GmbH.

  • Buisine, S. (2005). Conception et Évaluation d’Agents Conversationnels Multimodaux Bidirectionnels. http://stephanie.buisine.free.fr/, Doctorat de Psychologie Cognitive-Ergonomie, Paris V. 8 avril 2005. Direction J.-C. Martin & J.-C. Sperandio. http://stephanie.buisine.free.fr/.

  • Butterworth, B., & Beattie, G. (1978). Gesture and silence as indicators of planning in speech. In R. N. Campbell & P. Smith (Eds.), Recent advances in the psychology of language: Formal and experimental approaches (pp. 347–360). New York: Plenum.

    Google Scholar 

  • Cassell, J., Bickmore, T., Campbell, L., Vilhjálmsson, H., & Yan, H. (2000). Human conversation as a system framework: Designing embodied conversational agents. In J. Cassell, S. Prevost, & E. Churchill (Eds.), Embodied conversational Agents (pp. 29–63). Cambridge, MA: MIT.

    Google Scholar 

  • Cassell, J., Nakano, Y. I., Bickmore, T. W., Sidner, C. L., & Rich, C. (2001). Annotating and generating posture from discourse structure in embodied conversational agents. In Workshop “Multimodal communication and context in embodied agents”, 5th International Conference on Autonomous Agents, Montreal.

  • Cassell, J., Pelachaud, C., Badler, N., Steedman, M., Achorn, B., Becket, T., et al. (1994). Animated conversation: Rule-based generation of facial expression, gesture and spoken intonation for multiple conversational agents. ACM SIGGRAPH’94. http://www.cs.rutgers.edu/~mdstone/pubs/siggraph94.pdf.

  • Cassell, J., Torres, O., & Prevost, S. (1999). Turn taking vs. discourse structure: How best to model multimodal conversation. In Y. Wilks (Ed.), Machine conversations (pp. 143–154). The Hague: Kluwer.

    Google Scholar 

  • Chen, L., Travis Rose, R. T., Qiao, Y., Kimbara, I., Parrill, F., Welji, H., et al. (2006). Vace multimodal meeting corpus. In Second International Workshop on Machine Learning for Multimodal Interaction-MLMI. Lecture Notes in Computer Science. Berlin: Springer.

  • Cohn, J. F., & Ekman, P. (2005). Measuring facial action. In J. A. Harrigan, R. Rosenthal, & K. Scherer (Eds.), The new handbook of methods in nonverbal behavior research. Oxford University Press.

  • Collier, G. (1985). Emotional expression. Lawrence Erlbaum Associates. http://faculty.uccb.ns.ca/~gcollier/.

  • Dutoit, T., Nigay, L., & Schnaider, M. (2006). In T. Dutoit, L. Nigay, & M. Schnaider (Eds.), Multimodal human–computer interfaces. Elsevier. Journal of Signal Processing. Special Issue on “Multimodal Human–computer Interfaces”, 86(12), 3515–3517. http://dx.doi.org/10.1016/j.sigpro.2006.02.040.

  • Ekman, P. (1999). Basic emotions. In T. Dalgleish & M. J. Power (Eds.), Handbook of cognition & emotion (pp. 301–320). New York: Wiley.

    Chapter  Google Scholar 

  • Ekman, P. (2003). Emotions revealed. Understanding faces and feelings., Weidenfeld & Nicolson. http://emotionsrevealed.com/index.php.

  • Ekman, P., & Friesen, W. V. (1975). Unmasking the face. A guide to recognizing emotions from facial clues. Englewood Cliffs, NJ: Prentice-Hall Inc.

    Google Scholar 

  • Ekman, P., Friesen, W. C., & Hager, J. C. (2002). Facial action coding system. The manual on CD ROM. Research Nexus division of Network Information Research Corporation.

  • Feldman, R. S., & Rim, B. (1991). Fundamentals of nonverbal behavior. Cambridge University Press.

  • Garofolo, J., Laprum, C., Michel, M., Stanford, V., & Tabassi, E. (2004). The NIST Meeting Room Pilot Corpus. Language Resource and Evaluation Conference.

  • Goldin-Meadow, S., Kim, S., & Singer, M. (1999). What the teacher’s hand tell the student mind about math. Journal of Educational Psychology, 91, 720–730. doi:10.1037/0022-0663.91.4.720.

    Article  Google Scholar 

  • Harrigan, J. A., Rosenthal, R., & Scherer, K. (2005). The new handbook of methods in nonverbal behavior research. Oxford University Press.

  • Holzapfel, H., Nickel, K.,& Stiefelhagen, R. (2004). Implementation and evaluation of a constraint-based multimodal fusion system for speech and 3D pointing gestures. ICMI 2004. http://isl.ira.uka.de/fame/publications/FAME-A-WP10-028.pdf.

  • Janin, A., Baron, D., Edwards, J., Ellis, D., Gelbart, D., Morgan, N., et al. (2003). IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP).

  • Johnston, O., & Thomas, F. (1995). The illusion of life: Disney animation, Disney Editions.

  • Kendon, A. (2004). Gesture: Visible action as utterance. Cambridge University Press.

  • Kettebekov, S., Yeasin, M., Krahnstoever, N., & Sharma, R. (2002). Prosody based co-analysis of deictic gestures and speech in Weather Narration Broadcast. Workshop on multimodal resources and multimodal systems evaluation. In Conference on Language Resources and Evaluation (LREC’2002), Las Palmas, Canary Islands, Spain.

  • Kipp, M. (2004). Gesture generation by imitation. From human behavior to computer character animation. Florida, Boca Raton, Dissertation.com. http://www.dfki.de/~kipp/dissertation.html.

  • Kita, S. (2003). Interplay of gaze, hand, torso orientation, and language in pointing. In S. Kita (Ed.), Pointing. Where language, culture, and cognition meet (pp. 307–328). London: Lawrence Erlbaum Associates.

  • Knapp, M. L., & Hall, J. A. (2006). Nonverbal communication in human interaction (6th ed.). Thomson Wadsworth.

  • Kranstedt, A., Kühnlein, P., & Wachsmuth, I. (2004). Deixis in multimodal human–computer interaction. In A. Camurri & G. Volpe (Eds.), Gesture-based communication in human–computer interaction. 5th International Gesture Workshop, GW 2003, Genova, Italy. Springer. LNAI 2915.

  • Krauss, R. M. (1998). Why do we gesture when we speak? Current Directions in Psychological Science, 7, 54–59. http://www.columbia.edu/~rmk7/PDF/CD.pdf. doi:10.1111/1467-8721.ep13175642.

  • Kress, G., Jewitt, C., Ogborn, J., & Tsatsarelis, C. (2001). Multimodal teaching and learning. The rhetorics of the science classroom. Continuum.

  • Loehr, D. (2004). Gesture and intonation. Faculty of the Graduate School of Arts and Sciences of Georgetown University. http://www9.georgetown.edu/faculty/loehrd/pubs_files/Loehr04.pdf.

  • Martin, J. C. (2006). Multimodal human–computer interfaces and individual differences. Annotation, perception, representation and generation of situated multimodal behaviors. Habilitation à diriger des recherches en Informatique. Université Paris XI, 6th December 2006.

  • Martin, J.-C., den Os, E., Kuhnlein, P., Boves, L., Paggio, P., & Catizone, R. (2004). Workshop on multimodal corpora: models of human behaviour for the specification and evaluation of multimodal input and output interfaces. In Association with the 4th International Conference on Language Resources and Evaluation LREC2004. http://www.lrec-conf.org/lrec2004/index.php. Lisbon, Portugal: Centro Cultural de Belem. http://www.limsi.fr/Individu/martin/research/articles/LREC2004-WS-proceedings.pdf.

  • Martin, J.-C., Kuhnlein, P., Paggio, P., Stiefelhagen, R., & Pianesi, F. (2006). Workshop on multimodal corpora: From multimodal behaviour theories to usable models. In Association with the 5th International Conference on Language Resources and Evaluation (LREC2006), Genoa, Italy. http://www.limsi.fr/Individu/martin/tmp/LREC2006/WS-MM/final/proceedings-WS-MultimodalCorpora-v3.pdf.

  • Maybury, M., & Martin, J.-C. (2002). Workshop on multimodal resources and multimodal systems evaluation. In Conference on Language Resources and Evaluation (LREC’2002), Las Palmas, Canary Islands, Spain. http://www.limsi.fr/Individu/martin/research/articles/ws14.pdf.

  • McCowan, I., Carletta, J., Kraaij, W., Ashby, S., Bourban, S., Flynn, M., et al. (2005). The AMI meeting corpus. In Measuring Behavior 2005 Symposium on “Annotating and Measuring Meeting Behavior”.

  • McNeill, D. (1992). Hand and mind —what gestures reveal about thoughts. Chicago, IL: University of Chicago Press.

    Google Scholar 

  • McNeill, D. (2005). Gesture and thought. The University of Chicago Press.

  • McNeill, D., Quek, F., McCullough, K.-E., Duncan, S., Furuyama, N., Bryll, R., et al. (2001). Catchments, prosody and discourse. Gesture, 1(1), 9–33. doi:10.1075/gest.1.1.03mcn.

    Article  Google Scholar 

  • Oviatt, S. L. (2003). Multimodal interfaces. In J. Jacko & A. Sears (Eds.), Human–computer interaction handbook: Fundamentals, evolving technologies and emerging applications (Vol. 14, pp. 286–304). Mahwah, NJ: Lawrence Erlbaum Associates.

  • Pelachaud, C., Braffort, A., Breton, G., Ech Chadai, N., Gibet, S., Martin, J.-C., et al. (2004). AGENTS CONVERSATIONELS: Systèmes d’animation Modélisation des comportements multimodaux Applications: agents pédagogiques et agents signeurs. Action Spécifique du CNRS Humain Virtuel. (Eds.).

  • Pentland, A. (2005). Socially aware computation and communication. IEEE Computer.

  • Piwek, P., & Beun, R. J. (2001). Multimodal referential acts in a dialogue game: From empirical investigations to algorithms. In International Workshop on Information Presentation and Natural Multimodal Dialogue (IPNMD-2001), Verona, Italy. http://www.itri.bton.ac.uk/~Paul.Piwek/#publications.

  • Poggi, I. (1996). Mind markers. In 5th International Pragmatics Conference, Mexico City.

  • Poggi, I. (2003). Mind markers. In M. Rector, I. Poggi, & N. Trigo (Eds.), Gestures. Meaning and use (pp. 119–132). Oporto, Portugal: University Fernando Pessoa Press.

    Google Scholar 

  • Rist, T., André, E., Baldes, S., Gebhard, P., Klesen, M., Kipp, M., et al. (2003). A review of the development of embodied presentation agents and their application fields. In H. Prendinger & M. Ishizuka (Eds.), Life-like characters: Tools, affective functions, and applications (pp. 377–404). Springer.

  • Ruttkay, Z., & Pelachaud, C. (2004). From brows to trust—evaluating embodied conversational agents. Kluwer. http://wwwhome.cs.utwente.nl/~zsofi/KluwerBook.htm.

  • Siegman, A. W., & Feldstein, S. (1985). Multichannel integrations of nonverbal behavior, LEA.

  • Tepper, P., Kopp, S., & Cassell, J. (2004). Content in context: Generating language and iconic gesture without a gestionary. In Workshop on Balanced Perception and Action in ECAs at Automous Agents and Multiagent Systems (AAMAS), New York, NY.

  • van der Sluis, L., & Krahmer, E. (2004). Production experiments for evaluating multimodal generation. In 4th International Conference on Language Resources and Evaluation (LREC’2004).

  • Vinayagamoorthy, V., Gillies, M., Steed, A., Tanguy, E., Pan, X., Loscos, C., et al. (2006). Building expression into virtual characters. In Eurographics Conference State of the Art Reports. http://www.cs.ucl.ac.uk/staff/m.gillies/expressivevirtualcharacters.pdf.

  • Wahlster, W. (2006). SmartKom: Foundations of multimodal dialogue systems. Heidelberg, Germany: Springer.

    Google Scholar 

  • Wegener Knudsen, M., Martin, J.-C., Dybkjær, L., Berman, S., Bernsen, N. O., Choukri, K., et al. (2002a). Survey of NIMM data resources, current and future user profiles, markets and user needs for NIMM resources. ISLE Natural Interactivity and Multimodality. Working Group Deliverable D8.1. http://isle.nis.sdu.dk/reports/wp8/.

  • Wegener Knudsen, M., Martin, J.-C., Dybkjær, L., Machuca Ayuso, M.-J., Bernsen, N. O., Carletta, J., et al. (2002b). Survey of multimodal annotation schemes and best practice. ISLE Natural Interactivity and Multimodality. Working Group Deliverable D9.1. February. http://isle.nis.sdu.dk/reports/wp9/.

Download references

Reviewers

We want to thank the many reviewers who agreed to review manuscripts for this special issue on multimodal corpora. We are very grateful for the hard work they all put into it.

Table 1

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Jean-Claude Martin.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Martin, JC., Paggio, P., Kuehnlein, P. et al. Introduction to the special issue on multimodal corpora for modeling human multimodal behavior. Lang Resources & Evaluation 42, 253–264 (2008). https://doi.org/10.1007/s10579-008-9068-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10579-008-9068-6

Keywords

Navigation