Abstract
The examination of neural resource allocation during complex action sequence execution is critical to understanding human behavior. While physical systems are usually used for such assessment, virtual/remote systems offer other approaches with potential benefits such as remote training/evaluation. Here we describe a virtual environment (VLEARN) operated via the internet that has been developed to study the cognitive-motor mechanisms underlying the execution of goal-oriented action sequences in remote and laboratory settings. This study aimed to i) examine the feasibility of evaluating human cognitive-motor behavior when individuals operate VLEARN to complete various tasks; and ii) assess VLEARN by comparing its usability and the resulting performance, mental workload, and mental/physical fatigue during virtual and physical task execution. Results revealed that our approach allowed human cognitive-motor behavior assessment as the tasks completed physically and virtually via VLEARN had similar success rates. Also, there was a relationship between the complexity of the virtual control systems and the dependency on those to complete tasks. Namely, relative to controls with more functionalities, when VLEARN enabled simpler controls, above average usability and similar levels of cognitive-motor performance for both physical and virtual task execution were observed. Thus, a simplification of some aspects of the VLEARN control interface should enhance its usability. Our approach is promising for examining human cognitive-motor behavior and informing multiple applications (e.g., telehealth, remote training).
A. A. Shaver and N. Peri—both are co-first authors.
J. Purtilo and R. J. Gentili—both are co-last authors.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Notes
- 1.
- 2.
The 2 min time limit for the physical trials was set from prior work which clearly established that it was largely enough for task completion and thus did not bias the present study.
References
Wickens, C.D.: Multiple resources and mental workload. Hum. Factors 50(3), 449–455 (2008)
Young, M.S., Brookhuis, K.A., Wickens, C.D., Hancock, P.A.: State of science: mental workload in ergonomics. Ergonomics 58(1), 1–17 (2015)
Shaw, E.P., et al.: Measurement of attentional reserve and mental effort for cognitive workload assessment under various task demands during dual-task walking. Biol. Psychol. 134, 39–51 (2018)
Shuggi, I.M., et al.: Motor Performance, mental workload and self-efficacy dynamics during learning of reaching movements throughout multiple practice sessions. Neuroscience 423, 232–248 (2019)
Welsh, M.C., Huizinga, M.: Tower of Hanoi disk-transfer task: influences of strategy knowledge and learning on performance. Learn. Individ. Differ. 15(4), 283–298 (2005)
Wulf, G., Shea, C.H.: Principles derived from the study of simple skills do not generalize to complex skill learning. Psychon. Bull. Rev. 9(2), 185–211 (2002)
Hardy, D.J., Wright, M.J.: Assessing workload in neuropsychology: an illustration with the Tower of Hanoi test. J. Clin. Exp. Neuropsychol. 40(10), 1022–1029 (2018)
Hauge, T.C., et al.: A novel application of Levenshtein distance for assessment of high-level motor planning underlying performance during learning of complex motor sequences. J. Mot. Learn. Dev. 8(1), 67–86 (2019)
Radüntz, T.: The effect of planning, strategy learning, and working memory capacity on mental workload. Sci. Rep. 10(1), 7096 (2020)
Vakil, E., Lev-Ran Galon, C.: Baseline performance and learning rate of conceptual and perceptual skill-learning tasks: the effect of moderate to severe traumatic brain injury. J. Clin. Exp. Neuropsychol. 36(5), 447–454 (2014)
Milla, K., Bakhshipour, E., Bodt, B., Getchell, N.: Does movement matter? Prefrontal cortex activity during 2D vs. 3D performance of the Tower of Hanoi puzzle. Front. Hum. Neurosci. 13, 156 (2019)
Steunebrink, B., Wang, P., Goertzel, B. (eds.): AGI -2016. LNCS (LNAI), vol. 9782. Springer, Cham (2016). https://doi.org/10.1007/978-3-319-41649-6
Katz, G., Huang, D.W., Hauge, T., Gentili, R., Reggia, J.: A novel parsimonious cause-effect reasoning algorithm for robot imitation and plan recognition. IEEE Trans. Cognit. Dev. Syst. PP(99), 1–17 (2017)
Hauge, T.C., Katz, G.E., Davis, G.P., Huang, D.W., Reggia, J.A., Gentili, R.J.: High-level motor planning assessment during performance of complex action sequences in humans and a humanoid robot. Int. J. Soc. Robot 13, 981–998 (2021)
Shaver, A., Shuggi, I., Katz, G., Davis, G., Reggia, J., Gentili, R.: Effects of practicing structured and unstructured complex motor sequences on performance and mental workload. In: North American Society for the Psychology of Sport and Physical Activity Virtual Conference, Journal Sport Exercise Psychology, vol. 42, no. S1, pp. S56–S56. Human Kinetic Publisher Inc (2020)
Huang, D.W., Katz, G.E., Langsfeld, J.D., Gentili, R.J., Reggia, J.A.: A virtual demonstrator environment for robot imitation learning. In: IEEE International Conference on Technologies for Practical Robot Applications (TePRA), Woburn, MA, USA, pp. 1–6 (2015)
Huang, Di-Wei., Katz, G., Langsfeld, J., Oh, H., Gentili, R., Reggia, J.: An object-centric paradigm for robot programming by demonstration. In: Schmorrow, D.D., Fidopiastis, C.M. (eds.) AC 2015. LNCS (LNAI), vol. 9183, pp. 745–756. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-20816-9_71
Bangor, A., Kortum, P.T., Miller, J.T.: An empirical evaluation of the system usability scale. Intl. Int. J. Hum.-Comput. Int. 24(6), 574–594 (2008)
Kortum, P.T., Bangor, A.: Usability ratings for everyday products measured with the system usability scale. Int. J. Hum.-Comput. Int. 29(2), 67–76 (2013)
Bangor, A., Kortum, P., Miller, J.: Determining what individual SUS scores mean: adding an adjective rating scale. J. Usability Stud. 4(3), 114–123 (2009)
Hart, S.G., Staveland, L.E.: Development of NASA-TLX (Task Load Index): results of empirical and theoretical research. Adv. Psychol. 52, 139–183 (1988)
Childs, A.: Qualtrics, Provo, UT, USA (2020). https://www.qualtrics.com
Sauro, J., Lewis, J.R.: Quantifying the User Experience: Practical Statistics for User Re-search, 2nd edn. Morgan Kaufmann, Cambridge (2016)
Barnum, C.N.: Usability Testing Essentials, 2nd edn. Morgan Kaufmann, Cambridge (2021)
Levenshtein, V.I.: Binary codes capable of correcting deletions, insertions, and reversals. Soviet physics doklady 10(8), 707–710 (1966)
Wagner, R.A., Fischer, M.J.: The string-to-string correction problem. J. ACM 21(1), 168–173 (1974)
Jaquess, K.J., et al.: Empirical evidence for the relationship between cognitive workload and attentional reserve. Int. J. Psychophysiol. 121, 46–55 (2017)
Mahon, A., Bendžiūtė, S., Hesse, C., Hunt, A.R.: Shared attention for action selection and action monitoring in goal-directed reaching. Psychol. Res. 84(2), 313–326 (2018). https://doi.org/10.1007/s00426-018-1064-x
Acknowledgment
This work was supported by The Office of Naval Research (N00014–19-1–2044).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Shaver, A.A. et al. (2022). Assessment of a Novel Virtual Environment for Examining Human Cognitive-Motor Performance During Execution of Action Sequences. In: Schmorrow, D.D., Fidopiastis, C.M. (eds) Augmented Cognition. HCII 2022. Lecture Notes in Computer Science(), vol 13310. Springer, Cham. https://doi.org/10.1007/978-3-031-05457-0_28
Download citation
DOI: https://doi.org/10.1007/978-3-031-05457-0_28
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-05456-3
Online ISBN: 978-3-031-05457-0
eBook Packages: Computer ScienceComputer Science (R0)