A Multimodal System for Accessing Driving Directions
The focus of this paper is describing a system that repurposes a web document by a spoken language interface to provide both visual and audio driving directions. The spoken dialog interface is used to obtain the source and destination addresses from a user. The web document retrieved by querying the user on the addresses is parsed to extract the maps and the associated text. Further the system automatically generates two sets of web documents. One of these sets is used to render the maps on a hand-held device and the other set is used for the spoken dialog interface through a traditional phone. The system’s user interface allows navigation both through speech and pen stylus input. The system is built on a PhoneBrowser architecture that allows the user to browse the web by speech control over an ordinary telephone.
- Martin, G.L. “The utility of speech input in user-computer interfaces.” International Journal of Man-Machine Studies, 30:355–375, 1989Google Scholar
- Brown, Michael K., Stephen C. Glinski, Bernard P. Goldman, and Brian C. Schmult. “PhoneBrowser: A Web-Content-Programmable Speech Processing Platform.” Position Paper for The W3C Workshop on Voice Browsers, Cambridge, Massachusetts, October 1998Google Scholar
- C. Schmandt. “Multimedia Nomadic Services on Today’s Hardware.” IEEE Network, September/October 1994.Google Scholar
- Stifelman, Lisa, Barry Arons, Chris Schmandt, and Eric Hulteen, “VoiceNotes: A Speech Interface for a Hand-Held Voice Notetaker.” ACM INTERCHI '93 Conference Proceedings, Amsterdam, The Nether-lands, April 24–29, 1993Google Scholar
- Brown, Michael K. “PhoneBrowser User’s Guide.” Avaya Communication Inc. October 2000Google Scholar
- Morde, Ashutosh. “An Application for Voice Controlled Driving Directions”, Master of Science Thesis, Rutgers-The State University of New Jersey, May 2002.Google Scholar