Minimising Pedestrian Navigational Ambiguities Through Geoannotation and Temporal Tagging
The increasing power and ubiquity of smart devices such as mobile phones and PDAs means that a visitor to a city now carries with them a device capable of giving location-specific guiding and routing information. Whilst there have been a number of studies on the use of photographs to supplement text and map-based guiding applications for mobile devices, in this paper we want to propose and give an initial exploratory study of a guiding system that utilises geoannotation to mark photographs. In geoannotation, each photograph is selected from a repository of photographs based on the content and its relevance to the route. The photograph itself is then geoannotated with arrows and other markers on the fly so as to give routing information. Because the photograph in the database will not be taken from the location of the visitor who needs routing information, we need to take care and design cues that are unambiguous. The main contribution of this paper is the discussion of the geoannotation technique, and some informal results from pilot trials on how it helps in addressing certain navigational ambiguities arising in the use of photographs in pedestrian navigation systems.
KeywordsPedestrian Navigation Design Human Factors Ambiguities Geoannotation
Unable to display preview. Download preview PDF.
- 1.A9. A9.com – scanning and mapping cities.(accessed on April 04, 2006), http://maps.a9.com
- 2.Beeharee, A., Steed, A.: Filtering location-based information using visibility. In: Strang, T., Linnhoff-Popien, C. (eds.) LoCA 2005. LNCS, vol. 3479, pp. 306–315. Springer, Heidelberg (2005)Google Scholar
- 3.Beeharee, A.K., Steed, A.: A natural wayfinding exploiting photos in pedestrian navigation systems. In: Mobile HCI, pp. 81–88 (2006)Google Scholar
- 4.Feiner, S., MacIntyre, B., Höllerer, T., Webster, A.: A touring machine: Prototyping 3d mobile augmented reality systems for exploring the urban environment. In: ISWC, pp. 74–81 (1997)Google Scholar
- 5.Freeman, R., Steed, A., Zhou, B.: Rapid scene modelling, registration and specification for mixed reality systems. In: Proceedings of ACM Virtual Reality Software and Technology, pp. 147–150 (November 2005)Google Scholar
- 7.Goodman, J., Gray, P., Khammampad, K., Brewster, S.: Using landmarks to support older people in navigation. In: Brewster, S., Dunlop, M.D. (eds.) Mobile Human-Computer Interaction – MobileHCI 2004. LNCS, vol. 3160, pp. 38–48. Springer, Heidelberg (2004)Google Scholar
- 9.Microsoft. Microsoft live local. (accessed on April 04, 2006), http://local.live.com
- 10.Millonig, A., Schechtner, K.: Developing landmark-based pedestrian navigation systems. In: 8th International IEEE Conference on Intelligent Transport Systems (September 2005)Google Scholar
- 11.Miyazaki, Y., Kamiya, T.: Pedestrian navigation system for mobile phones using panoramic landscape images. In: SAINT 2006. Proceedings of the International Symposium on Applications on Internet, Washington, DC, USA, pp. 102–108. IEEE Computer Society, Los Alamitos (2006)Google Scholar
- 12.TomTom. Tomtom – portable gps navigation for cars. accessed on (September 15, 2005), http://www.tomtom.com
- 13.Verbree, E., Van Anrooij, A.: Interactive navigation through distance added valued panoramic images. In: Maas, H.G., Schneider, D. (eds.) Panoramic Photogrammetry Workshop, Institute of Photogrammetry and Remote Sensing, Dresden, Germany, vol. 34(16), pp. +4 (February 2004)Google Scholar