Abstract
The present investigation examines the correlations of emotional user states and performance in command and control environments. The aim is to gain insights into whether and how the integration of the emotional state into a diagnostic component of an adaptive human-machine system can take place. While positive valence is expected to be associated with high performance (Hypothesis 1), negative valence is assumed to be associated with low performance (Hypothesis 2). In a laboratory experiment, a command and control task in the domain of anti-air warfare was performed. Emotional valence was assessed with a software for recognition of emotional face expressions (Emotient FACET). To measure performance, we assessed performance decrements that occurred whenever a subtask was not accomplished in time. Data from 22 participants were used for the analysis (45% female, 17–52 years, M = 30.96, SD = 9.76). Regression analyses were conducted to test the hypotheses. Contrary to expectations, there were no correlations between emotional valence and performance at the group level. Even though empirical results failed to support the hypotheses, significant correlations between positive valence and performance were found for 36% of the subjects, 45% for negative valence and performance. These results indicate that individual models are necessary for the analysis of emotional user state. We discuss practical implications and suggest improvements to the paradigm.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bainbridge, L.: Ironies of automation. Automatica 19(6), 775–779 (1983)
Parasuraman, R., Manzey, D.: Complacency and bias in human use of automation: an attentional integration. Hum. Factors 52(3), 381–410 (2010)
Åkerstedt, T., Mollard, R., Samel, A., Simons, M., Spencer, M.: The role of EU FTL legislation in reducing cumulative fatigue in civil aviation. Presented at the ETSC Meeting (2003)
Young, M.S., Stanton, N.A.: Malleable attentional resources theory: a new explanation for the effects of mental underload on performance. Hum. Factors: J. Hum. Factors Ergon. Soc. 44(3), 365–375 (2002)
Fuchs, S., et al.: Anwendungsorientierte Realisierung adaptiver Mensch-Maschine-Interaktion für Sicherheitskritische Systeme (ARAMIS). Final project report (grant no. E/E4BX/HA031/CF215). Fraunhofer FKIE, Wachtberg (2019)
Eyben, F., et al.: Emotion on the road-necessity, acceptance, and feasibility of affective computing in the car. Adv. Hum.-Comput. Interact. 2010(1768), 1–17 (2010)
Jones, C.M., Jonsson, I.-M.: Automatic recognition of affective cues in the speech of car drivers to allow appropriate responses. In: Proceedings of the 17th Australian Conference on Computer-Human Interaction, pp. 1–10. Citizens Online: Considerations for Today and the Future, Canberra (2005)
Groeger, J.A.: Understanding Driving: Applying Cognitive Psychology to a Complex Everyday Task. Taylor & Francis Ltd., Hove (2000)
Kontogiannis, T.: Stress and operator decision making in coping with emergencies. Int. J. Hum.-Comput. Stud. 45(1), 75–104 (1996)
Madhavan, P., Wiegmann, D.A., Lacson, F.C.: Automation failures on tasks easily performed by operators undermine trust in automated aids. Hum. Factors 48(2), 241–256 (2006)
Vassiliou, M.S., Alberts, D.S., Agre, J.R.: C2 Re-Envisioned: The Future of the Enterprise. CRC Press, Boca Raton (2014)
Littlewort, G., et al.: The computer expression recognition toolbox (CERT). In: Proceedings of the 2011 International Conference on Automatic Face & Gesture Recognition and Workshops, pp. 298–305, (2011)
Lemay Center for Doctrine: Key considerations of command and control. Annex 3-30 - Command and Control, pp. 6–10 (2014)
Vassiliou, M.S., Alberts, D.S., Agre, J.R., Alberts, D.S., Agre, J.R.: C2 Re-Envisioned: The Future of the Enterprise. CRC Press, Boca Raton (2014)
Wohl, J.G.: Force management decision requirements for air force tactical command and control. IEEE Trans. Syst. Man Cybern. 11(9), 618–639 (1981)
Gaillard, A.: Concentration, stress and performance. In: Hancock, P.A., Szalma, J.L. (eds.) Performance Under Stress, pp. 59–65. Ashgate Publishing, Burlington (2008)
Panganiban, A.: Effects of anxiety on change detection in a command and control task. Master of Science thesis, University of Cincinnati, Cincinnati (2011)
Stecklov, G., Goldstein, J.: Terror attacks influence driving behavior in Israel. Proc. Nat. Acad. Sci. (PNAS) 101(40), 14551–14556 (2004)
Zimasa, T., Jamson, S., Henson, B.: Are happy drivers safer drivers? Evidence from hazard response times and eye tracking data. Transp. Res. Part F: Traffic Psychol. Behav. 46(1), 14–23 (2017)
Wells-Parker, E., et al.: An exploratory study of the relationship between road rage and crash experience in a representative sample of US drivers. Accid. Anal. Prev. 34(3), 271–278 (2002)
Deffenbacher, J.L., Oetting, E.R., Lynch, R.S.: Development of a driving anger scale. Psychol. Rep. 74(1), 83–91 (1994)
iMotions: iMotions facial expressions analysis pocket guide (2016)
James, L., Nahl, D.: Road Rage and Aggressive Driving: Steering Clear of Highway Warfare. Prometheus Books, Amherst (2000)
Cohn, J.F., Schmidt, K., Gross, R., Ekman, P.: Individual differences in facial expression: Stability over time, relation to self-reported emotion, and ability to inform person identification. In: Proceedings of the 4th IEEE International Conference on Multimodal Interfaces, pp. 491–496 (2002)
Hudlicka, E.: Affect-adaptive interaction in behavioral health technologies. In: Proceedings of the 22nd International Conference on Human-Computer Interaction in Copenhagen, in press (2020)
Taggart, R.W., Dressler, M., Kumar, P., Khan, S., Coppola, J.F.: Determining emotions via facial expression analysis software. Behav. Res. Methods 50(4), 1446–1460 (2016)
Fridkin, K.L., Gershon, S.A., Courey, J., LaPlant, K.: Gender differences in emotional reactions to the first 2016 presidential debate. Polit. Behav., 1–31 (2019). https://doi.org/10.1007/s11109-019-09546-9
Iqbal, F.: Understanding user interaction in a video game by using eye tracking and facial expressions analysis. Unpublished Master’s thesis, University of Tampere, Tampere, Finnland (2015)
Takahashi, K., Yamada, K., Nakano, T., Yamamoto, S.: Method of detecting concentration on cellular phone call from facial expression change by image processing. In: 2005 IEEE International Conference on Systems, Man and Cybernetics, vol. 4, pp. 3444–3448 (2005)
Cai, H., Lin, Y.: Modeling of operators’ emotion and task performance in a virtual driving environment. Int. J. Hum.-Comput. Stud. 69(9), 571–586 (2011)
McDuff, D., Kaliouby, R., Kassam, K. Picard, R.: Affect valence inference from facial action unit spectrograms. In 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition - Workshops, pp. 17–24 (2010)
Russell, J.: A circumplex model of affect. J. Personal. Soc. Psychol. 39, 1161–1178 (1980)
Posner, J., Russell, J.A., Peterson, B.S.: The circumplex model of affect: an integrative approach to affective neuroscience, cognitive development, and psychopathology. Dev. Psychopathol. 17(3), 715–734 (2005)
Appelhans, B.M., Luecken, L.J.: Heart rate variability as an index of regulated emotional responding. Rev. Gen. Psychol. 10(3), 229–240 (2006)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Schmitz-Hübsch, A., Fuchs, S. (2020). Challenges and Prospects of Emotional State Diagnosis in Command and Control Environments. In: Schmorrow, D.D., Fidopiastis, C.M. (eds) Augmented Cognition. Theoretical and Technological Approaches. HCII 2020. Lecture Notes in Computer Science(), vol 12196. Springer, Cham. https://doi.org/10.1007/978-3-030-50353-6_5
Download citation
DOI: https://doi.org/10.1007/978-3-030-50353-6_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50352-9
Online ISBN: 978-3-030-50353-6
eBook Packages: Computer ScienceComputer Science (R0)