Navigation Performance Effects of Render Method and Head-Turn Latency in Mobile Audio Augmented Reality
This study assessed participant performance of an outdoor navigation task using a mobile audio augmented reality system. Several quantitative performance measures and one subjective measure were used to compare the perceptual efficacy of Ambisonic and VBAP binaural rendering techniques, and a range of head-turn latencies. The study extends existing indoors research on the effects of head-turn latency for seated listeners.
The pilot experiment found that a source capture radius of 2 meters significantly affected the sole participant’s navigation distance efficiency compared to other radii. The main experiment, using 8 participants, found that render method significantly affected all performance measures except subjective stability rating, while head-turn latency only affected mean track curvature and subjective stability. Results also showed an interaction in which the choice of rendering method mitigated or potentiated the effects of head-turn latency on perceived source stability.
KeywordsMobile audio augmented reality navigation binaural rendering VBAP Ambisonic head-turn latency
Unable to display preview. Download preview PDF.
- 1.Algazi, V.R., Duda, R.O., Thompson, D.M., Avendano, C.: The CIPIC HRTF Database. In: Proc. IEEE Workshop on Applications of Signal Processing to Audio and Electroacoustics, October 21-24, 2001, pp. 99–102. Mohonk Mountain House, New Paltz (2001)Google Scholar
- 2.Benjamin, E., Lee, R., Aficionado, L., Heller, A.: Localization in Horizontal-Only Ambisonic Systems. 131st AES Convention, preprint 6967, 5–8 (2006)Google Scholar
- 4.Mariette, N.: Perceptual Evaluation of Personal, Location Aware Spatial Audio. Ph.D. thesis, School of Computer Science and Engineering, University of New South Wales, [in examination]) (2009)Google Scholar
- 5.Mariette, N.: A Novel Sound Localization Experiment for Mobile Audio Augmented Reality Applications. in: al, Z.P.e. (ed.) 16th Int. Conf. on Artificial Reality and Tele-existence. pp. 132–142, Springer, Berlin/Heidelberg, Hangzhou, China (2006)Google Scholar
- 6.Mariette, N.: Mitigation of binaural front-back confusions by body motion in audio augmented reality. In: Int. Conf. on Auditory Display, Montreal, Canada, June 26-29 (2007)Google Scholar
- 7.McDonald, J.: Data Transformations (2007), http://udel.edu/~mcdonald/stattransform.html
- 8.Miller, J.D., Anderson, M.R., Wenzel, E.M., McClain, B.U.: Latency measurement of a real-time virtual acoustic environment rendering system. In: Int. Conf. on Auditory Display, Boston, MA, USA, 6-9 July (2003)Google Scholar
- 9.Puckette, M.: Pure Data. In: Int. Computer Music Conf, pp. 269–272. Int. Computer Music Association, San Francisco (1996)Google Scholar
- 10.Pulkki, V.: Virtual Sound Source Positioning Using Vector Base Amplitude Panning. J. Audio Eng. Soc. 46(6), 456–466 (1997)Google Scholar
- 11.Strauss, H., Buchholz, J.: Comparison of Virtual Sound Source Positioning with Amplitude Panning and Ambisonic Reproduction. In: The 137th regular Meeting of the Acoustical Society of America (1999)Google Scholar
- 12.Walker, B.N., Lindsay, J.: Auditory Navigation Performance is Affected by Waypoint Capture Radius. In: The 10th Int. Conf. on Auditory Display, Sydney, Australia, July 6-9 (2004)Google Scholar
- 13.Weisstein, E.W.: Curvature (2007), http://mathworld.wolfram.com/Curvature.html
- 14.Wright, M., Freed, A.: Open Sound Control: A New Protocol for Communicating with Sound Synthesizers. In: Int. Computer Music Conf., Thessaloniki, Greece, September 25-30 (1997)Google Scholar