A method for automatically translating print books into electronic Braille books

Abstract

In this paper, a method for automatically translating scanned images from print books into electronic Braille books is proposed with the objective of reducing the amount of time and cost required for producing Braille books. The proposed method consists of processes for identifying character and image areas in a scanned image, automatically translating characters and images into Braille and tactile graphics, respectively, and positioning Braille and tactile graphics into an electronic Braille page. Experimental results show that the proposed method drastically reduces the time required to translate a print book into an electronic Braille book. Despite the drastic reduction in translation time, the method proposed in this paper does not compromise the ability to recognize information for the visually impaired compared to manually produced Braille books, demonstrating its feasibility in practical applications. Therefore, the proposed method is expected to significantly reduce the time and cost required for producing Braille books, and provide more reading materials for the visually impaired, making significant contributions to enhancing their knowledge and welfare.

This is a preview of subscription content, access via your institution.

References

  1. 1

    Improvement suggestion for ensuring accessibility to knowledge and information of disabled people. ACRC Technical Report. Korea, 2014

  2. 2

    Wong E K, Chen M. A new robust algorithm for video text extraction. Patt Recogn, 2003, 36: 1397–1406

    Article  MATH  Google Scholar 

  3. 3

    Shivakumara P, Phan T Q, Tan C L. A gradient difference based technique for video text detection. In: Proceedings of the 10th International Conference on Document Analysis and Recognition, Barcelona, 2009. 156–160

    Google Scholar 

  4. 4

    Jang D G, Hwang C S. Document image layout analysis using image filters and constrained conditions. KIPS Trans B, 2002, 9: 311–318

    Article  Google Scholar 

  5. 5

    Chun B T, Bae Y I, Kim T Y. Text extraction in videos using topographical feature of characters. In: Proceedings of the IEEE International Conference on Fuzzy System, Seoul, 1999. 1126–1130

    Google Scholar 

  6. 6

    Fu X L, Cai L H, Liu Y, et al. A computational cognition model of perception, memory, and judgment. Sci China Inf Sci, 2014, 57: 032114

    Article  Google Scholar 

  7. 7

    Liu Y, Fu Q, Liu Y, et al. A distributed computational cognitive model for object recognition. Sci China Inf Sci, 2013, 56: 092101

    MathSciNet  Google Scholar 

  8. 8

    Strouthopoulos C, Papamarkos N, Atsalakis A E. Text extraction in complex color document. Patt Recogn, 2002, 35: 1743–1758

    Article  MATH  Google Scholar 

  9. 9

    Chun B T, Song C Y. A method for character segmentation using frequency characteristics and back propagation neural network. J Korea Soc Comput Inf, 2006, 4: 55–60

    Google Scholar 

  10. 10

    Yuan Q, Tan C L. Text extraction from gray scale document images using edge information. In: Proceedings of the 6th International Conference on Document Analysis and Recognition, Seattle, 2001. 302–306

    Google Scholar 

  11. 11

    Park J C. Text region detection of various slope and size of text using the adaptive character-edge map. In: Proceedings of KOCON Conference, 2007. 5–9

    Google Scholar 

  12. 12

    Grover S, Arora K, Mitra S K. Text extraction from document images using edge information. In: Proceedings of the IEEE India Council Conference, Gujarat, 2009. 1–4

    Google Scholar 

  13. 13

    Kim E J. Character area extraction and the character segmentation on the color document. J Korean Inst Intell Syst, 1999, 9: 444–450

    Google Scholar 

  14. 14

    Song Y J, Kim K C, Choi Y W, et al. Text region extraction and text segmentation on camera-captured document style images. In: Proceedings of the 8th International Conference on Document Analysis and Recognition, Seoul, 2005. 172–176

    Google Scholar 

  15. 15

    Kim J S, Kim S H. Three-level color clustering algorithm for binarizing scene text images. KIPS Trans B, 2005, 12: 1–8

    Article  Google Scholar 

  16. 16

    Jung K C, Han J H. Hybrid approach to efficient text extraction in complex color images. Patt Recogn Lett, 2004, 25: 679–699

    Article  Google Scholar 

  17. 17

    Jung K C. Hybrid approach of texture and connected component methods for text extraction in complex images. IEEK, 2004, 41: 175–186

    Google Scholar 

  18. 18

    Huang T J, Tian Y H, Li J, et al. Salient region detection and segmentation for general object recognition and image understanding. Sci China Inf Sci, 2011, 54: 2461–2470

    MathSciNet  Article  Google Scholar 

  19. 19

    Zhou L, Hu D W, Zhou Z T. Scene recognition combining structural and textural features. Sci China Inf Sci, 2013, 56: 078106

    MathSciNet  Google Scholar 

  20. 20

    Smith R. An overview of the Tesseract OCR Engine. In: Proceedings of the 12th International Conference on Document Analysis and Recognition. Washington DC: IEEE, 2007. 629–633

    Google Scholar 

  21. 21

    Ministry of Culture, Sports and Tourism. Korea Braille Regulations, MCST Regulation 2006-39. Korea, 2006

  22. 22

    Way T P, Barner K E. Automatic visual to tactile translation-part II: evaluation of the TACTile image creation system. IEEE Trans Rehabil Eng, 1997, 5: 95–105

    Article  Google Scholar 

  23. 23

    Rotard M, Knödler S, Ertl T. A tactile web browser for the visually disabled. In: Proceedings of the 16th ACM Conference on Hypertext and Hypermedia. New York: ACM, 2005. 15–22

    Google Scholar 

  24. 24

    Ladner R E, Ivory M Y, Rao R, et al. Automating tactile graphics translation. In: Proceedings of the 7th International ACM SIGACCESS Conference on Computers and Accessibility. New York: ACM, 2005. 150–157

    Google Scholar 

  25. 25

    Petit G, Dufresne A, Levesque V, et al. Refreshable tactile graphics applied to schoolbook illustrations for students with visual impairment. In: Proceedings of the 10th International ACM SIGACCESS Conference on Computers and Accessibility. New York: ACM, 2008. 89–96

    Google Scholar 

  26. 26

    Chen J J, Nagaya R, Takagi N. An extraction method of solid line graph elements in mathematical graphs for automating translation of tactile graphics. In: Proceedings of the 13th IEEE International Symposium on Advanced Intelligent Systems. Washington DC: IEEE, 2012. 422–427

    Google Scholar 

  27. 27

    Takagi N, Chen J J. A broken line classification method of mathematical graphs for automating translation into scalable vector graphic. In: Proceedings of the IEEE 43rd International Symposium on Multiple-Valued Logic. Washington DC: IEEE, 2013. 71–76

    Google Scholar 

  28. 28

    Chen J J, Takagi N. A pattern recognition method for automating tactile graphics translation from hand-drawn maps. In: Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics. Washington DC: IEEE, 2013. 4173–4178

    Google Scholar 

  29. 29

    Chen Q, Zhao L, Lu J, et al. Modified two-dimensional Otsu image segmentation algorithm and fast realization. IET Image Process, 2012, 6: 426–433

    MathSciNet  Article  Google Scholar 

  30. 30

    Deng Y N, Kenney C, Moore M S, et al. Peer group filtering and perceptual color image quantization. In: Proceedings of the IEEE International Symposium on Circuits and Systems. Washington DC: IEEE, 1999. 21–24

    Google Scholar 

  31. 31

    Gong Y. Advancing content-based image retrieval by exploiting image color and region features. Multimedia Syst, 1999, 7: 449–457

    Article  Google Scholar 

  32. 32

    Shi J, Tomasi C. Good features to track. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Washington DC: IEEE, 1994. 593–600

    Google Scholar 

Download references

Author information

Affiliations

Authors

Corresponding author

Correspondence to Jinsoo Cho.

Rights and permissions

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Park, T., Jung, J. & Cho, J. A method for automatically translating print books into electronic Braille books. Sci. China Inf. Sci. 59, 072101 (2016). https://doi.org/10.1007/s11432-016-5575-z

Download citation

Keywords

  • Braille
  • tactile graphic
  • electronic Braille book
  • automatic translation
  • the visually impaired