A Novel Sound Localization Experiment for Mobile Audio Augmented Reality Applications

  • Nick Mariette
Part of the Lecture Notes in Computer Science book series (LNCS, volume 4282)


This paper describes a subjective experiment in progress to study human sound localization using mobile audio augmented reality systems. The experiment also serves to validate a new methodology for studying sound localization where the subject is outdoors and freely mobile, experiencing virtual sound objects corresponding to real visual objects. Subjects indicate the perceived location of a static virtual sound source presented on headphones, by walking to a position where the auditory image coincides with a real visual object. This novel response method accounts for multimodal perception and interaction via self-motion, both ignored by traditional sound localization experiments performed indoors with a seated subject, using minimal visual stimuli. Results for six subjects give a mean localization error of approximately thirteen degrees; significantly lower error for discrete binaural rendering than for ambisonic rendering, and insignificant variation to filter lengths of 64, 128 and 200 samples.


Global Position System Augmented Reality Sound Source Inertial Navigation System Sound Localization 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    Cohen, M., Aoki, S., Koizumi, N.: Augmented Audio Reality: Telepresence/AR Hybrid Acoustic Environments. In: IEEE International Workshop on Robot and Human Communication (1993)Google Scholar
  2. 2.
    Milgram, P., Kishino, F.: A Taxonomy of Mixed Reality Visual Displays. IEICE Transactions on Information Systems, E77–D(12) (1994)Google Scholar
  3. 3.
    Loomis, J.M.: Personal Guidance System for the Visually Impaired Using GPS, GIS, and VR Technologies. In: VR Conference. California State University, Northridge (1993)Google Scholar
  4. 4.
    Holland, S., Morse, D.R., Gedenryd, H.: AudioGPS: Spatial Audio in a Minimal Attention Interface. In: Proceedings of Human Computer Interaction with Mobile Devices (2001)Google Scholar
  5. 5.
    Helyer, N. (1999-2001), Sonic Landscapes Accessed: 22/8/2006Google Scholar
  6. 6.
  7. 7.
    Rozier, J., Karahalios, K., and Donath, J.: Hear & There: An Augmented Reality System of Linked Audio. In: ICAD 2000, Atlanta, Georgia, April 2000 (2000) Google Scholar
  8. 8.
    Warusfel, O.: G.E. LISTEN - Augmenting Everyday Environments through Interactive Soundscapes. In: IEEE VR 2004 (2004)Google Scholar
  9. 9.
    Härmä, A., et al.: Techniques and Applications of Wearable Augmented Reality Audio. In: AES 114TH Convention, Amsterdam, The Netherlands (2003)Google Scholar
  10. 10.
    Zhou, Z., Cheok, A.D., Yang, X., Qiu, Y.: An Experimental Study on the Role of Software Synthesized 3D Sound in Augmented Reality Environments. Interacting with Computers 16, 989 (2004)CrossRefGoogle Scholar
  11. 11.
    Walker, B.N., Lindsay, J.: Auditory Navigation Performance Is Affected by Waypoint Capture Radius. In: ICAD 2004 - The Tenth International Conference on Auditory Display, Sydney, Australia (2004)Google Scholar
  12. 12.
    Loomis, J.M., Klatzky, R.L., Golledge, R.G.: Auditory Distance Perception in Real, Virtual and Mixed Environments. In: Ohta, Y., Tamura, H. (eds.) Mixed Reality: Merging Real and Virtual Worlds, Tokyo, pp. 201–214 (1999)Google Scholar
  13. 13.
    Grantham, D.W., Hornsby, B.W.Y., Erpenbeck, E.A.: Auditory Spatial Resolution in Horizontal, Vertical, andDiagonal Planes. Journal of the Acoustical Society of America 114(2), 1009–1022 (2003)CrossRefGoogle Scholar
  14. 14.
    Strauss, H., Buchholz, J.: Comparison of Virtual Sound Source Positioning with Amplitude Panning and mbisonic Reproduction. The Journal of the Acoustical Society of America 105(2), 934 (1999)CrossRefGoogle Scholar
  15. 15.
    Choe, C.S., Welch, R.B., Gilford, R.M., Juola, J.F.: The “Ventriloquist Effect”: Visual Dominance or Response Bias. Perception & Psychophysics 18, 55–60 (1975)CrossRefGoogle Scholar
  16. 16.
    Larsson, P., Västfjäll, D., Kleiner, M.: Ecological Acoustics and the Multi-Modal erception of Rooms: Real and Unreal Experiences of Auditory-Visual Virtual Environments. In: International Conference on Auditory Display, Espoo, Finland (2001)Google Scholar
  17. 17.
    Brungart, D.S., Simpson, B.D., Kordik, A.J.: The Detectability of Headtracker Latency in Virtual Audio Displays. In: International Conference on Auditory Display, Limerick, Ireland (2005)Google Scholar
  18. 18.
    Point Research, DRM-III Oem Dead Reckoning Module for Personnel Positioning. Fountain Valley, California (2002)Google Scholar
  19. 19.
    Miller, L.E.: Indoor Navigation for First Responders: A Feasibility Study. National Institute of Standards and Technology (2006)Google Scholar
  20. 20.
    Algazi, V.R., Duda, R.O., Thompson, D.M., Avendano, C.: The Cipic Hrtf Database. In: Proc. 2001 IEEE Workshop on Applications of Signal Processing to Audio and Electroacoustics. Mohonk Mountain House, New Paltz (2001)Google Scholar
  21. 21.
    Noisternig, M., Musil, T., Sontacchi, A., Höldrich, R.: A 3D Real Time Rendering Engine for Binaural Sound Reproduction. In: International Conference on Auditory Display, Boston, MA, USA (2003)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2006

Authors and Affiliations

  • Nick Mariette
    • 1
  1. 1.Audio Nomad Group, School of Computer Science and EngineeringUniversity of New South WalesSydneyAustralia

Personalised recommendations