Abstract
In this paper, we compare four different auditory displays in a mobile audio-augmented reality environment (a sound garden). The auditory displays varied in the use of non-speech audio, Earcons, as auditory landmarks and 3D audio spatialization, and the goal was to test the user experience of discovery in a purely exploratory environment that included multiple simultaneous sound sources. We present quantitative and qualitative results from an initial user study conducted in the Municipal Gardens of Funchal, Madeira. Results show that spatial audio together with Earcons allowed users to explore multiple simultaneous sources and had the added benefit of increasing the level of immersion in the experience. In addition, spatial audio encouraged a more exploratory and playful response to the environment. An analysis of the participants’ logged data suggested that the level of immersion can be related to increased instances of stopping and scanning the environment, which can be quantified in terms of walking speed and head movement.
Similar content being viewed by others
Notes
Only data from within the proximity zone were considered, and data points while in the activation zone were excluded as we were only interested in user behaviour while exploring and not once they had reached the activation zone.
Kurtosis is the name of a statistical measure used to describe the distribution of observed data around the mean. A normal distribution has a kurtosis 0, a peaked (tall and skinny) distribution has a positive or high kurtosis, and a flat distribution has a negative or low kurtosis.
References
Shepard M (2006) Tactical Sound Garden [TSG] Toolkit. In: 3rd international workshop on mobile music technology, Brighton, UK
Walker BN, Lindsay J (2006) Navigation performance with a virtual auditory display: effects of beacon sound, capture radius, and practice. Hum Factors 48(2):265–278
Stahl C (2007) The roaring navigator: a group guide for the zoo with shared auditory landmark display. In: Proceedings of MobileHCI 2007. ACM Press, New York, pp 282–386
Blauert J (1997) Spatial hearing: the psychophysics of human sound localization. The MIT Press, Cambridge
Begault DR (1994) 3D sound for virtual reality and multimedia. AP Professional, Boston
Mariette N (2007) From backpack to handheld: the recent trajectory of personal location aware spatial audio. In: Hutchison A (ed) PerthDAC 2007: Proceedings of the 7th digital arts and culture conference. Curtin University of Technology, Perth, pp 233–240
Mynatt E, Back M, Want R, Baer M, Ellis JB (1995) Designing audio aura. In: Proceedings of CHI 1998. ACM Press, New York, pp 566–573
Marentakis GN, Brewster SA (2006) Effects of feedback, mobility and index of difficulty on deictic spatial audio target acquisition in the horizontal plane. In: Proceedings of ACM CHI 2006. ACM Press, New York, pp 359–368
Rozier J, Karahalios K, Donath J (2000) Hear & There: an augmented reality system of linked audio. In: Proceedings of the international conference on auditory display—ICAD 2000
Reid J, Geelhoed E, Hull R, Carter K, Clayton B (2005) Parallel worlds: immersion in location-based experiences. In: Proceedings of CHI 2005, vol 2. ACM Press, New York, pp 1733–1736
Holland S, Morse DR, Gedenryd H (2002) AudioGPS: spatial audio in a minimal attention interface. Pers Ubiquit Comput 6(4):253–259
Cater K, Hull R, O’Hara K, Melamed T, Clayton B (2007) The potential of spatialised audio for location based services on mobile devices: Mediascapes. SAMD: Workshop on Spatialised Audio for Mobile Devices, MobileHCI 2007
McGookin D, Brewster S, Priego P (2009) Audio Bubbles. Employing non-speech audio to support tourist wayfinding. In: Altinsoy ME, Jekosch U, Brewster S (eds) HAID 2009, LNCS, vol 5763, pp 41–50
Magnusson C, Breidegard B, Rassmus-Gröhn K (2009) Soundcrumbs—Hansel and Gretel in the 21st century. In: HAID 2009, LNCS. Springer
Blattner MM, Sumikawa DA, Greenberg RM (1989) Earcons and icons: their structure and common design principles. Hum Comput Interact 4(1):11–44
Gaver WW (1997) Auditory interfaces. In: Helander MG, Landauer TK, Prabhu PV (eds) Handbook of human-computer interaction, 2nd edn. Elsevier, Amsterdam, pp 1003–1041
Lemordant J, Guerraz A (2007) Mobile immersive music. In: Proceedings of the 2007 international computer music conference, ICMC 2007. ICMA, San Francisco, pp 21–24
Etter R, Specht M (2005) Melodious walkabout: implicit navigation with contextualized personal audio contents. Fraunhofer Institute of applied Information Technology, Sankt Augustin
Jones M, Jones S, Bradley G, Warren N, Bainbridge D, Holmes G (2008) ONTRACK: dynamically adapting music playback to support navigation. Pers Ubiquit Comput 12(7):513–525
Strachan S, Eslambolchilar P, Murray-Smith R (2005) gpstunes—controlling navigation via audio feedback. MobileHCI 2005, vol 1. ACM, New York, pp 275–278
Lyons K, Gandy M, Starner T (2000) Guided by Voices: An Audio Augmented Reality System. In: International conference on auditory display (ICAD), Atlanta, pp 57–62
Heller F, Knott T, Weiss M, Borchers J (2009) Multi-user interaction in virtual audio spaces. CHI Extended Abstracts of ACM CHI 2009. ACM Press, pp 4489–4494
Mariette N (2010) Navigation performance effects of render method and head-turn latency in mobile audio augmented reality. CMMR/ICAD 2009, LNCS, vol 5954, pp 239–265
Brungart DS, Simpson BD, Kordik AJ (2005) The detectability of headtracker latency in virtual audio displays. In: Proceedings of the 11th international conference on auditory display (ICAD), Limerick, Ireland, pp 37–42
Vazquez-Alvarez Y, Brewster SA (2010) Designing spatial audio interfaces to support multiple audio streams. In: Proceedings of MobileHCI 2010. ACM Press, New York, pp 253–256
Grossman T, Balakrishnan R (2005) The bubble cursor: enhancing target acquisition by dynamic resizing of the cursor’s activation area. In: Proceedings of CHI 2005. ACM Press, New York, pp 281–290
Acknowledgments
This work was supported by the Ken Browning Travelling Scholarship (University of Glasgow, UK), Nokia and EPSRC research grant EP/F023405 “Gaime”. We would like to express our gratitude to the members of the Madeira-ITI group at Madeira University who participated in this research project.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Vazquez-Alvarez, Y., Oakley, I. & Brewster, S.A. Auditory display design for exploration in mobile audio-augmented reality. Pers Ubiquit Comput 16, 987–999 (2012). https://doi.org/10.1007/s00779-011-0459-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00779-011-0459-0