Abstract
In order to design AR (Augmented Reality) glasses for the deaf, this paper studies the virtual display interface of AR glasses. It is found that 89.8% of the deaf have willingness to use AR glasses based on questionnaires of 1900 deaf people. An experiment with 10 deaf subjects and 10 hearing subjects was done to calculate times of completing specific tasks. The results show that: (1) 80% of the deaf subjects have 3 to 5 zones in the interface with significant differences on cognitive efficiency, compared with the 6 to 7 zones from 80% of the hearing subjects. (2) The heat maps of the cognitive efficiency distribution shows that there is an obvious difference on the distribution patterns of regional sensitivity between the two groups. (3) The significant difference on cognitive efficiency of the deaf shown in different display zones is less than that of hearing people.
References
World Health Organization. https://www.who.int/zh/news-room/fact-sheets/detail/deafness-and-hearing-loss
Mi, K.: A review of the research on communication mode of hearing impaired. Modern Spec. Educ. 08, 72–76 (2016)
McCormack, A., Fortnum, H.: Why do people fitted with hearing aids not wear them? Int. J. Audiol. 52(5), 360–368 (2013)
Schweizer, H.: Smart glasses: technology and applications Student report Ubiquitous computing seminar FS (2014)
Sony. Sony Access Glasses. http://goo.gl/0DKFoQ
Jain, D., Findlater, L., et al.: Head-mounted display visualizations to support sound awareness for the deaf and hard of hearing. The CHI, 241–250 (2015)
Levine, A., Billawa, S., Bridge, L., et al.: FMRI correlates of visual motion processing in hearing and deaf adults. J. Vis. 14(10), 297 (2014)
Bavelier, D., Dye, M.W.C., Hauser, P.C.: Do deaf individuals see better? Trends Cognit. Sci. 10(11), 512–518 (2006)
Lu, X., Pas, E.I.: Socio-demographics. activity participation and travel behavior. Department of Civil and Environmental Engineering, pp. 3–4
Timmermans, H., et al.: Analyzing space-time behaviour: new approaches to old problems. Progress Hum. Geogr. 26(2), 175–190 (2002)
Shao, J.: Research on the information encoding method of helmet mounted display system interface based on visual perception theory. Southeast Univ. 77, 113–118 (2016)
Google Inc. https://developers.google.com/glass/design/style
Ruch Theodore, C., Fulton, J.F.: Medical physiology and biophysics (1960)
Ding, Y.: Ergonomics. Beijing Institute of Technology Press (04), 39–41 (2011)
Matthews, T., Carter, S., et al.: Scribe4Me: Evaluating a Mobile Sound Transcription Tool for the Deaf. Technical Report No. UCB/EECS 9, 159–176 (2006)
Matthews, T., Fong, J., et al.: Evaluating non-speech sound visualizations for the deaf. Behav. Inf. Technol. 25(2), 333–351 (2007)
Stevens, C.: Neuroplasticity as a double-edged sword: deaf enhancements and dyslexic deficits in motion processing. J. Cognit. Neurosci. 18(5), 701–714 (2006)
Loke, W.H., Song, S.: Central and peripheral visual processing in hearing and non hearing individuals. Bull. Psychon. Soc. 29(5), 437–440 (1991)
Acknowledgments
This research is supported by the Fundamental Research Funds for Central Universities (2017ZX013).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Li, Z., Mo, X., Shi, C., Jiang, S., Jiang, L. (2020). The Research of Visual Characteristics for the Deaf Wearing AR Glasses. In: Ahram, T. (eds) Advances in Human Factors in Wearable Technologies and Game Design. AHFE 2019. Advances in Intelligent Systems and Computing, vol 973. Springer, Cham. https://doi.org/10.1007/978-3-030-20476-1_16
Download citation
DOI: https://doi.org/10.1007/978-3-030-20476-1_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-20475-4
Online ISBN: 978-3-030-20476-1
eBook Packages: EngineeringEngineering (R0)