Abstract
Tactical audio uses audio feedback to facilitate the precise and accurate positioning of an object with respect to some other object. Existing solutions in pointing and trajectory based Human Computer Interface (HCI) tasks have primarily explored visual feedback, sometimes in rather limited conditions. In this paper we have examined different sonification paradigms for tactical audio to improve the accuracy of pursuit tracking. We have developed a multimodal simulation system as an open environment for evaluation of different sonification and visualization techniques in tactical audio applications. The environment is implemented as an audio-visual scene using the Java3D package. The paper presents the quantitative results of of three pursuit tracking applications using a combination of acoustic and visual guidance and different background conditions. Experiments with 19 participants have shown that acoustic presentation improves the quality of human-computer interaction and reduces the error during pursuit tracking tasks for up to 19%. Moreover, experiments have shown that benefits do not exist in all conditions, indicating the existence of perceptual boundaries of multimodal HCI for different scene complexity and target speeds. Significant benefits of audio modes exist for medium complexity of interaction. User’s questionnaires indicate that users are not aware of quantitative benefits of applied sonification paradigms. We have shown that the most appealing paradigms are not the most effective ones, which necessitates a careful quantitative analysis of proposed multi-modal HCI paradigms.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
E. Jovanov, K. Wagner, V. Radivojevic, D. Starcevic, M. Quinn, D. Karron, “Tactical Audio and Acoustic Rendering in Biomedical Applications”, IEEE Transactions on Information Technology in Biomedicine, Vol. 3, No. 2, June 1999, pp. 109–118.
Wegner K, Karron D, “Surgical Navigation Using Audio Feedback.” In Morgan K S et al (eds): Medicine Meets Virtual Reality: Global Healthcare Grid. IOS Press, Ohmsha, Washington, D.C., pp 450–458, 1997.
E. Jovanov, Z. Obrenovic, D. Starcevic, D. B. Karron, “A Virtual Reality Training System for Tactical Audio Applications”, SouthEastern Simulation Conference SESC’99, Huntsville, Alabama, Oct 1999, pp. 149–154.
Begault D R, 3D Sound for Virtual Reality and Multimedia, Academic Press, Inc., Boston, 1994.
Kramer, G., Ed., Auditory Display, Sonification, Audification and Auditory Interfaces, Addison Wesley, 1994.
Barnard P J, May J, Eds, “Computers, Communication and Usability: Design Issues, research and methods for integrated services,” North Holland Series in Telecommunication, Amsterdam, Elsevier, 1993.
Buxton, W. (1995). Speech, Language & Audition. Chapter 8 in R. M. Baecker, J. Grudin, W. Buxton and S. Greenberg, S. (Eds.) (1995). Readings in Human Computer Interaction: Toward the Year 2000, San Francisco, Morgan Kaufmann Publish
Deatherage, B. H. (1972). Auditory and Other Sensory Forms of Information Presentation. In H. P. Van Cott & R. G. Kinkade (Eds), Human Engineering Guide to Equipment Design (Revised Edition). Washington: U.S. Government Printing Office.
Meera M. Blattner, “Multimodal Integration”, IEEE Multimedia, Winter 1996, pp. 14–24.
Sharon Oviatt, “Ten Myths of Multimodal Interaction”, Comm. of the ACM, Vol. 42, No. 11, Nov 1999, pp. 74–81.
E. Jovanov, D. Starcevic, V. Radivojevic, A. Samardzic, V. Simeunovic, “Perceptualization of biomedical data”, IEEE Engineering in Medicine and Biology Magazine, Vol. 18, No. 1, pp. 50–55, 1999.
Zeljko Obrenovic, Dusan Starcevic, Emil Jovanov, “Experimental Evaluation of Multimodal Human Computer Interface For Tactical Audio Applications”, IEEE International Conference on Multimedia and Expo (ICME) 2002, August 26–29, 2002, Lausanne, Switzerland.
John G. Neuhoff, “Perceiving Acoustic Source Orientation in Three-Dimensional Space”, Proceedings of the 2001 International Conference on Auditory Display, Espoo, Finland, July 29–August 1, 2001, pp. 231–234.
William L. Martens, “Psychophysical Calibration for Controlling the Range of a Virtual Sound Source: Multidimensional Complexity in Spatial Auditory Display”, Proceedings of the 2001 International Conference on Auditory Display, Espoo, Finland, July 29–August 1, 2001, pp. 231–234.
H. Sowizral, K. Rushforth, M. Deering, The Java 3D API Specification, Second Edition, Addison-Wesley Pub Co, Januar 2000, ISBN: 0201710412.
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Obrenovic, Z., Starcevic, D., Jovanov, E. (2003). Toward Optimization of Multimodal User Interfaces for Tactical Audio Applications. In: Carbonell, N., Stephanidis, C. (eds) Universal Access Theoretical Perspectives, Practice, and Experience. UI4ALL 2002. Lecture Notes in Computer Science, vol 2615. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-36572-9_23
Download citation
DOI: https://doi.org/10.1007/3-540-36572-9_23
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-00855-2
Online ISBN: 978-3-540-36572-3
eBook Packages: Springer Book Archive