Abstract
This chapter presents research activities of UTDrive: the smart vehicle project, at the Center for Robust Speech Systems, University of Texas at Dallas. The objectives of the UTDrive project are to collect and research rich multi-modal data recorded in actual car environments for analyzing and modeling driver behavior. The models of driver behavior under normal and distracted driving conditions can be used to create improved in-vehicle human–machine interactive systems and reduce vehicle accidents on the road. The UTDrive corpus consists of audio, video, brake/gas pedal pressure, head distance, GPS information (e.g., position, velocity), and CAN-bus information (e.g., steering-wheel angle, brake position, throttle position, and vehicle speed). Here, we describe our in-vehicle data collection framework, data collection protocol, dialog and secondary task demands, data analysis, and preliminary experimental results. Finally, we discuss our proposed multi-layer data transcription procedure for in-vehicle data collection and future research directions.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
H. Abut, H. Erdogan, A. Ercil et al., “Data collection with ‘UYANIK’: too much pain; but gains are coming,” Chapter 3 in this book.
A. Baron, P. Green, “Safety and usability of speech interfaces for in-vehicle tasks while driving: a brief literature review,” UMTRI-2006-5: The University of Michigan, Transportation Research Institute, Ann Arbor, February 2006, pp.1–8.
C. Carter, R. Graham, “Experimental comparison of manual and voice controls for the operation of in-vehicle systems,” in Proceedings of the IEA2000/HFES2000 Congress, Santa Monica, CA.
T. Chen, “Audio-visual speech processing,” IEEE Signal Process Mag., vol. 18, no. 1, pp. 9–21, 2001.
H. Erdogan, A. Ercil, H.K. Ekenel, S.Y. Bilgin, I. Eden, M. Kirisci, H. Abut, “Multimodal person recognition for vehicular applications,” N.C. Oza et al. (Eds.) MCS-2005, Monterey, CA, LNCS-3541, pp. 366–375, Jun. 2005.
C. Forlines, B. Schmidt-Nielsen, B. Raj, P. Wittenburg, P. Wolf, “Comparison between spoken queries and menu-based interfaces for in-car digital music selection,” TR2005-020, Mitsubishi Electric Research Laboratories, Cambridge, MA.
J.H.L. Hansen, D. Cairns, “ICARUS: a source generator based real-time system for speech recognition in noise, stress, and Lombard effect,” Speech Commun., vol. 16, no. 4, pp. 391–422, July 1995.
J.H.L. Hansen, “Analysis and compensation of speech under stress and noise for environmental robustness in speech recognition,” Speech Commun.: Special Issue on Speech Under Stress, vol. 20, no. 2, pp. 151–170, Nov. 1996.
T.B. Hughes, H.S. Kim, J.H. DiBiase, H.F. Silverman, “Performance of an HMM speech recognizer using a real-time tracking microphone array as input,” IEEE Trans. Speech Audio Process., vol. 7, no. 3, pp. 346–349, 1999.
E. Lombard, “Le signe de l'elevation de la voix,” Ann. Maladies Oreille Larynx, Nez, Pharynx, vol. 37, pp. 101–119, 1911.
C. Miyajima, Y. Nishiwaki, K. Ozawa, T. Wakita, K. Itou, K. Takeda, F. Itakura, “Analysis and modeling of personality in driving behavior and its application to driver identification,” Proc. IEEE, vol. 95, no. 2, pp. 427–437, Feb. 2007.
C. Miyajima, T. Kusakawa, T. Nishino, N. Kitaoka, K. Itou, K. Takeda, “On-going data collection for driving behavior signal,” Chapter 4 in this book.
A. Pentland, A. Liu, “Modeling and prediction of human behavior,” Neural Comput., vol. 11, pp. 229–242, 1999.
M. Pettitt, G. Burnett, A. Stevens, “Defining driver distraction,” World Congress on Intelligent Transport Systems, San Francisco, Nov. 2005.
D. Salvucci, E.R. Boer, A. Liu, “Toward an integrated model of driver behavior in a cognitive architecture,” Transport. Res. Rec., no. 1779, pp. 9–16, 2001.
A. Wahab, T.C. Keong, H. Abut, K. Takeda, “Driver recognition system using FNN and statistical methods,” H. Abut, J.H.L. Hansen, K. Takeda (Eds.) Chapter 3 in Advances for in-vehicle and mobile systems, Springer, New York, 2007.
X.X. Zhang and J.H.L. Hansen, “CSA-BF: a constrained switched adaptive beamformer for speech enhancement and recognition in real car environment,” IEEE Trans. Speech Audio Proc., vol. 11, no. 6, pp. 733–745, Nov. 2004.
X.X. Zhang, K. Takeda, J.H.L. Hansen, T. Maeno, “Audio-visual speaker localization for car navigation systems,” in INTERSPEECH-2004, Jeju Island, Korea, 2004.
Acknowledgments
This work is supported by NEDO, Japan and University of Texas at Dallas under project EMMITT. The authors would like to thank all the student assistants for their great contributions during this project.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer Science+Business Media, LLC
About this chapter
Cite this chapter
Angkititrakul, P. et al. (2009). UTDrive: The Smart Vehicle Project. In: Takeda, K., Erdogan, H., Hansen, J.H.L., Abut, H. (eds) In-Vehicle Corpus and Signal Processing for Driver Behavior. Springer, Boston, MA. https://doi.org/10.1007/978-0-387-79582-9_5
Download citation
DOI: https://doi.org/10.1007/978-0-387-79582-9_5
Published:
Publisher Name: Springer, Boston, MA
Print ISBN: 978-0-387-79581-2
Online ISBN: 978-0-387-79582-9
eBook Packages: EngineeringEngineering (R0)