Abstract
A myriad of infotainment systems has found its applications in the automobile industry with the burgeoning demand for user comfort and interaction. However, operating such infotainment systems entertain secondary tasks to be carried out at the expense of the primary task of driving. This can increase the cognitive load on the driver and has the potential to keep road safety at stake. This paper presents an intelligent interactive head-up display (HUD) on the windscreen of the driver that does not require them to take eyes off the road while undertaking secondary tasks like playing music, operating vent controls, watching navigation map, and so on. The interactive HUD allows the user to navigate and make selections using eye gaze. The HUD also incorporates provisions to estimate driver’s cognitive load and distraction level. User studies show that the system improves driving performance in terms of mean deviation from lane in an ISO 26022 lane changing task compared to touch screen system and participants can undertake ISO 9241 pointing tasks in less than 2 s on average inside a car.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Chang W, Hwang W, Ji YG (2011) Haptic seat interfaces for driver information and warning systems. Int J Hum Comput Interact 27(12):1119–1132. https://doi.org/10.1080/10447318.2011.555321
Ohn-Bar E, Trivedi M (2014) Hand gesture recognition in real-time for automotive interfaces: a multimodal vision-based approach and evaluations. IEEE Trans Intell Trans Syst
Feld M, Meixner G, Mahr A, Seissler M, Kalyanasundaram B (2013) Generating a personalized UI for the car: a user-adaptive rendering architecture. In: UMAP 2013, LNCS, vol 7899, pp 344–346
Normark CJ (2015) Design and evaluation of a touch-based personalizable in-vehicle user interface. Int J Hum Comput Interact
May KR, Walker BN, Gable TM (2014) A multimodal air gesture interface for in vehicle menu navigation. In: Adjunct proceedings of ACM automotive UI 2014
Weinberg G, Knowles A, Langer P (2014) BullsEye: an au automotive touch interface that’s always on target. In: Adjunct proceedings of ACM automotive UI 2014
Zhai S, Morimoto C, Ihde S (1999) Manual and gaze input cascaded (MAGIC) pointing. In: ACM SIGCHI conference on human factors in computing system (CHI)
Ashdown M, Oka K, Sato Y (2005) Combining head tracking and mouse in-put for a GUI on multiple monitors. In: CHI Late breaking result
Zhang Y, Bulling A, Gellersen H (2013) SideWays: a gaze interface for spontaneous interaction with situated displays. In: ACM SIGCHI conference on human factors in computing system (CHI)
Fu Y, Huang TS (2007) hMouse: head tracking driven virtual computer mouse. In: IEEE workshop on applications of computer vision
CameraMouse. http://www.cameramouse.com. Accessed 22 Sept 2018
Fejtova M et al (2009) Hands-free interaction with a computer and other technologies. Univers Access Inf Soc 8
Lutteroth C, Penkar M, Weber G (2015) Gaze vs. Mouse: a fast and accurate gaze-only click alternative. In: Proceedings of ACM UIST 2015
Biswas P, Langdon P (2013) A new interaction technique involving eye gaze tracker and scanning systems. In: ACM eye tracking South Africa (ETSA)
Mondragon CK, Bleacher B (2015) Eye tracking control of vehicle entertainment systems, Patent No. WO2013036632
Poitschke T, Laquai F, Stamboliev S, Rigoll G (2011) Gaze-based interaction on multiple displays in an automotive environment. IEEE Int Conf Syst Man Cybern (SMC), 543–548. ISSN: 1062-922X
Biswas P, Prabhakar G (2018) Detecting drivers’ cognitive load from saccadic intrusion. Trans Res Part F Traffic Psychol Behav 54 (Elsevier)
Biswas P (2016) Exploring the use of eye gaze controlled interfaces in automotive environments. Springer. ISBN: 978-3-319-40708-1
Gamgoz N (2000) Effects of hue, saturation, and brightness on attention and preference. Bilkent University
Campbell JL, Brown JL, Graving JS, Richard CM, Lichty MG, Bacon LP, Sanquist T (2018) Human factors design guidance for level 2 and level 3 automated driving concepts (Report No. DOT HS 812 555). National Highway Traffic Safety Administration, Washington, DC
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Ramakrishnan, A., Madan, M., Prabhakar, G., Deshmukh, S., Biswas, P. (2020). Eye Gaze Controlled Head-up Display. In: Fong, S., Dey, N., Joshi, A. (eds) ICT Analysis and Applications. Lecture Notes in Networks and Systems, vol 93. Springer, Singapore. https://doi.org/10.1007/978-981-15-0630-7_46
Download citation
DOI: https://doi.org/10.1007/978-981-15-0630-7_46
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-15-0629-1
Online ISBN: 978-981-15-0630-7
eBook Packages: EngineeringEngineering (R0)