Skip to main content

Towards a Real-Time System for Finding and Reading Signs for Visually Impaired Users

  • Conference paper
Computers Helping People with Special Needs (ICCHP 2012)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 7383))

Included in the following conference series:

Abstract

Printed text is a ubiquitous form of information that is inaccessible to many blind and visually impaired people unless it is represented in a non-visual form such as Braille. OCR (optical character recognition) systems have been used by blind and visually impaired persons for some time to read documents such as books and bills; recently this technology has been packaged in a portable device, such as the smartphone-based kReader Mobile (from K–NFB Reading Technology, Inc.), which allows the user to photograph a document such as a restaurant menu and hear the text read aloud. However, while this kind of OCR system is useful for reading documents at close range (which may still require the user to take a few photographs, waiting a few seconds each time to hear the results, to take one that is correctly centered), it is not intended for signs. (Indeed, the KNFB manual, see knfbreader.com/upgrades_mobile.php , lists “posted signs such as signs on transit vehicles and signs in shop windows” in the “What the Reader Cannot Do” subsection.) Signs provide valuable location-specific information that is useful for wayfinding, but are usually viewed from a distance and are difficult or impossible to find without adequate vision and rapid feedback.

We describe a prototype smartphone system that finds printed text in cluttered scenes, segments out the text from video images acquired by the smartphone for processing by OCR, and reads aloud the text read by OCR using TTS (text-to-speech). Our system detects and reads aloud text from video images, and thereby provides real-time feedback (in contrast with systems such as the kReader Mobile) that helps the user find text with minimal prior knowledge about its location. We have designed a novel audio-tactile user interface that helps the user hold the smartphone level and assists him/her with locating any text of interest and approaching it, if necessary, for a clearer image. Preliminary experiments with two blind users demonstrate the feasibility of the approach, which represents the first real-time sign reading system we are aware of that has been expressly designed for blind and visually impaired users.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Ivanchenko, V., Coughlan, J., Shen, H.: Real-Time Walk Light Detection with a Mobile Phone. In: Miesenberger, K., Klaus, J., Zagler, W., Karshmer, A. (eds.) ICCHP 2010. LNCS, vol. 6180, pp. 229–234. Springer, Heidelberg (2010)

    Chapter  Google Scholar 

  2. Liang, J., Doermann, D., Li, H.: Camera-based analysis of text and documents: a survey. International Journal on Document Analysis and Recognition 7, 83–200 (2005)

    Article  Google Scholar 

  3. Manduchi, R., Kurniawan, S., Bagherinia, H.: Blind Guidance Using Mobile Computer Vision: A Usability Study. In: ACM SIGACCESS Conference on Computers and Accessibility, ASSETS (2010)

    Google Scholar 

  4. Pilu, M., Pollard, S.: A light-weight text image processing method for handheld embedded cameras. In: British Machine Vision Conference (2002)

    Google Scholar 

  5. Sanketi, P., Shen, H., Coughlan, J.: Localizing Blurry and Low-Resolution Text in Natural Images. In: 2011 IEEE Workshop on Applications of Computer Vision (WACV 2011), Kona, Hawaii (January 2011)

    Google Scholar 

  6. Yi, C., Tian, Y.: Assistive Text Reading from Complex Background for Blind Persons. In: Iwamura, M., Shafait, F. (eds.) CBDAR 2011. LNCS, vol. 7139, pp. 15–28. Springer, Heidelberg (2012)

    Chapter  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2012 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Shen, H., Coughlan, J.M. (2012). Towards a Real-Time System for Finding and Reading Signs for Visually Impaired Users. In: Miesenberger, K., Karshmer, A., Penaz, P., Zagler, W. (eds) Computers Helping People with Special Needs. ICCHP 2012. Lecture Notes in Computer Science, vol 7383. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-31534-3_7

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-31534-3_7

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-31533-6

  • Online ISBN: 978-3-642-31534-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics