The VoiceApp System: Speech Technologies to Access the Semantic Web

  • David Griol
  • José Manuel Molina
  • Víctor Corrales
Part of the Lecture Notes in Computer Science book series (LNCS, volume 7023)


Maximizing accessibility is not always the main objective in the design of web applications, specially if it is concerned with facilitating access for disabled people. In this paper we present the VoiceApp multimodal dialog system, which enables to access and browse Internet by means of speech. The system consists of several modules that provide different user experiences on the web. Voice Dictionary allows the multimodal access to the Wikipedia encyclopedia, Voice Pronunciations has been developed to facilitate the learning of new languages by means of games with words and images, whereas Voice Browser provides a fast and effective multimodal interface to the Google web search engine. All the applications in the system can be accessed multimodally using traditional graphic user interfaces such as keyboard and mouse, and/or by means of voice commands. Thus, the results are accessible also for motor-handicapped and visually impaired users and are easier to access by any user in small hand-held devices where graphical interfaces are in some cases difficult to employ.


Dialog Systems Multimodality VoiceXML XHTML+Voice Web Interfaces Speech Interaction 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. 1.
    López-Cózar, R., Araki, M.: Spoken, Multilingual and Multimodal Dialogue Systems. John Wiley & Sons (2005)Google Scholar
  2. 2.
    Beskow, J., Edlund, J., Granström, B., Gustafson, J., Skantze, G., Tobiasson, H.: The MonAMI Reminder: a spoken dialogue system for face-to-face interaction. In: Proc. of Interspeech/ICSLP, pp. 296–299 (2009)Google Scholar
  3. 3.
    Tsai, M.: The VoiceXML dialog system for the e-commerce ordering service. In: Shen, W.-m., Chao, K.-M., Lin, Z., Barthès, J.-P.A., James, A. (eds.) CSCWD 2005. LNCS, vol. 3865, pp. 95–100. Springer, Heidelberg (2006)CrossRefGoogle Scholar
  4. 4.
    Kearns, M., Isbell, C., Singh, S., Litman, D., Howe, J.: CobotDS: A Spoken Dialogue System for Chat. In: Proc. of AAAI 2002, pp. 425–430 (2002)Google Scholar
  5. 5.
    Griol, D., McTear, M.F., Callejas, Z., López-Cózar, R., Ábalos, N., Espejo, G.: A Methodology for Learning Optimal Dialog Strategies. In: Sojka, P., Horák, A., Kopeček, I., Pala, K. (eds.) TSD 2010. LNCS, vol. 6231, pp. 507–514. Springer, Heidelberg (2010)CrossRefGoogle Scholar
  6. 6.
    Stent, A., Stenchikova, S., Marge, M.: Reinforcement learning of dialogue strategies with hierarchical abstract machines. In: Proc. of SLT 2006, pp. 210–213 (2006)Google Scholar
  7. 7.
    Chai, J., Horvath, V., Nicolov, N., Stys, M., Kambhatla, N., Zadrozny, W., Melville, P.: Natural language assistant: A dialog system for online product recommendation. AI Magazine 23, 63–75 (2002)Google Scholar
  8. 8.
    Vesnicer, B., Zibert, J., Dobrisek, S., Pavesic, N., Mihelic, F.: A voice-driven web browser for blind people. In: Proc. of Interspeech/ICSLP, pp. 1301–1304 (2003)Google Scholar
  9. 9.
    Mishra, T., Bangalore, S.: Qme!: a speech-based question-answering system on mobile devices. In: Proc. of HLT 2010, pp. 55–63 (2010)Google Scholar

Copyright information

© Springer-Verlag Berlin Heidelberg 2011

Authors and Affiliations

  • David Griol
    • 1
  • José Manuel Molina
    • 1
  • Víctor Corrales
    • 1
  1. 1.Universidad Carlos III de MadridLeganésSpain

Personalised recommendations