Abstract
Current multi-party video- and audioconferencing systems limit natural communications between participants. People communicate by speech, facial expressions and body gestures. In interactions between three or more people, these communications channels are directed towards particular participants. Spatial proximity and gaze direction are therefore important elements for effective conversational interactions, and yet are largely unsupported in existing conferencing tools. Advanced audioconferencing systems do simulate presence in a shared environment by using ‘virtual humans’ to represent the people taking part in a meeting, but the keyboard and mouse are used to direct conversations to specific people or to change the visual representation to simulate emotion.
This paper describes an experimental implementation of virtual conferencing, which uses machine vision to control a realistic virtual human, with the objective of making ‘virtual meetings’ more like physical ones. The computer vision system provides a more natural interface to the environment, while the realistic representation of users, with appropriate facial gestures and upper body movement, gives more natural visual feedback.
Similar content being viewed by others
References
OnLive Technologies, http://www.OnLive.com/
Sidhu C K and Bowman P: ‘Real people in surreal environments’, BT Technol J, 15,No 4, pp 140–149 (October 1997).
Russ M: ‘Desktop conversations — the future of multimedia conferencing’, BT Technol J, 15,No 4, pp 42–50 (October 1997).
Machin D J: ‘Real-time facial motion analysis for virtual teleconferencing’, IEEE International Conference on Automatic Face and Gesture Recognition (October 1996).
Azarbayejani A, Wren C and Pentland: ‘A real-time 3-D tracking of the human body’, Proceedings of IMAGE'COM 96, Bordeaux, France (also M.I.T. Media Laboratory Perceptual Computing Section Technical Report No. 374) (May 1996).
Moghaddam B and Pentland A: ‘Face recognition using view-based and modular eigenspaces automatic systems for the identification and inspection of humans,’ SPIE, 2277 (July 1994).
Wallin N: ‘Automatic texturing of BT talking head,’ MEng Project report, Dept of Electronics, University of York (1997).
Medit Modelling software, http://www.medit3d.com/
JACK Human Body Modelling Software, Center for Human Modeling and Simulation, University of Pennsylvania, http://www.cis.upenn.edu/~hms/jack.html
‘IRIS Performer’, Silicon Graphics, http://www.sgi.com/Technology/Performer/
Page J H and Breen A P: ‘The Laureate text-to-speech system — architecture and applications’, BT Technol J, 14,No 1, pp 57–67 (July 1992).
MiraLab, University of Geneva, http://miralabwww.unige.ch/
University of California, http://mambo.ucsc.edu/psl/fan.html
French National Audio-Visual Institute, http://www.ina.fr/INA/Recherche/TV/TV.en.html
Ohio State University, http://www.cis.ohio-state.edu/~parent/research/research.html
MPEG4 Synthetic Natural Hybrid Coding, http://www.es.com/mpeg4-snhc/index.html
Whybray M, Morrison G and Mulroy P: ‘Video coding — techniques, standards and applications’, BT Technol J, 15,No 4, pp 86–100 (October 1997).
Ohya J, Kitamura Y, Kishino F, Terashima N, Takemura H and Ishii H: ‘Virtual space teleconferencing: real-time reproduction of 3D human images’, Journal of Visual Communication and Image Representation, 6,No 1 (March 1995).
Author information
Authors and Affiliations
About this article
Cite this article
Mortlock, A.N., Machin, D., McConnell, S. et al. Virtual conferencing. BT Technology Journal 15, 120–129 (1997). https://doi.org/10.1023/A:1018687630541
Issue Date:
DOI: https://doi.org/10.1023/A:1018687630541