Abstract
Computer-Based training (CBT) is a widely used strategy in interactive learning and skills training to provide the trainee with information while in training. Extended Reality (XR) is a set of technologies used in the fourth industrial revolution for this purpose. However, an important XR limitation is the cognitive overload in trainees due to continuous changes in attention. The effects of the changes in attention on cognitive load have been studied in environments such as printed material and desktop PC’s. However, such effects are not yet fully known on XR platforms. This study’s aim was to identify strategies to reduce the changes in attention in instructional/educational materials, which use extended reality to present information to trainees. The presented information can guide the use of combined strategies to reduce the cognitive overload generated by XR display platforms. Therefore, an extensive literature review was carried out. From a set of 1946 articles, 53 studies were selected. The selected studies evaluated the cognitive load in instructional materials that used XR as a visualization platform. The results showed three strategy groups: those associated with the spatial integration of information, those associated with the use of visual features and those associated with the content segmentation. Effects of this strategies in the user performance and user cognitive load are discussed.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Extended Reality (XR) is considered a set of technologies including virtual reality (VR), mixed reality (MR) and augmented reality (AR), ALL part of the fourth industrial revolution and are widely used in learning and training environments using the computer-based paradigm. For example, augmented reality is widely used to allow combining objects (real and virtual) and providing spatially positioned information in training environments [1, 2]. According to several studies, the main utility of AR is to provide procedural and support information during training [3,4,5,6,7]. According to Lebel [7] y Gavish [4], the information provided improves trainee performance and it´s a fundamental component while training since it provides initial support for the learning and automation phases [8]. The information is delivered to the trainee using a Graphical User Interface (GUI), through small units called "Displays" [9]. However, according to Evans, [10] y Acampora [11] AR advantages in training could be reduced by cognitive overload generated by the poor disposition of information. An inadequate disposition of information overloads the trainees working memory and reduces cognitive resources available for learning [12, 13]. As established by the cognitive load theory (CLT) [14], working memory overload can have an external source and increases the cognitive resources consumption due to the attention changes required to search for specific information within the virtual environment [15]. Therefore, Ayres and Sweller [16, 17], state that the Split attention principle is important to avoid attention changes. This principle states that, trainees should avoid dividing visual attention among multiple sources of information to ensure a proper learning process. Split attention of instruction occurs when trainees are required to divide their visual attention and mentally integrate information from separate (spatial or temporal) sources. Because each information source is essential to understanding the displayed instructional material, the external cognitive load is likely to be increased. Applying the principle has challenges: the presentation format selection [16, 18] and the positioning of information in allotted virtual spaces [19, 20]. On the other hand, other principles such as signaling [21], and coding, are applied to drive attention during learning, but the information found is segmented or dispersed. Despite the wide use of extended reality, few studies have evaluated its impact on cognitive load during multimedia learning [22]. Akçayır [23] mentions the lack of data on cognitive load factors in said studies associated with ER technologies and the existing ones, show contradictory results. Similarly, Ens [24] mentions the need to study the effect on cognitive load and trainee performance, when multiple information spaces are used in augmented reality. Also, Garcia-San Juan [25] states that the effect of separation and presentation of information on multiple displays on the trainee performance is unknown. Finally, Rashid [26] shows the need to study split attention and information distribution (in the virtual environment), to reduce changes in the trainee’s visual attention. For this reason, the work aim was to review the literature to identify strategies oriented to avoid changes in attention while training with multimedia and XR technologies. Attention changes increase cognitive load by this, in this study several techniques used to deal with changes in attention in Augmented reality, Virtual Reality and Mixed Reality were reported.
In this review, three groups of strategies were identified: Regarding to spatial integration, regarding to visual attributes uses and regarding to content segmentation. Effects of spatial integration formats and contiguity effect condition was identified. Empirical studies were identified that relate visual attributes such as color and signaling forms, with negative effects on the trainee performance, the attention changes and distraction [1, 2, 27, 28]. Content segmentation techniques used to reduce cognitive load and improve the user performance in training and learning material was identified. Finally, results are discussed and the few studies that evaluate cognitive load due to the graphical user interface configuration in extended reality technologies are evidenced. The results of this study may serve as a guide for graphical user interfaces design for: instructional, learning and training materials with extended reality technologies, which reduce attention changes and cognitive overload in trainees. Despite this, the authors consider it necessary to carry out more in-depth studies that relate the properties of the user interface with the effect on changes in attention and cognitive load.
2 Theorethical framework
2.1 Training with interactive technologies
In recent years, extended reality technologies have been emerging within the fourth industrial revolution as tools for simulation-based training (SBT) [29]. The impact identified is relevant, especially in industrial and medical applications [30]. According to Shaywitz [31], a simulation is "a recreation of reality aspects in a setting or environment". Interactive simulators are used as an effective tool for KSA (Knowledge, Skills, and Attitude) training, incorporating several levels of realism: physical, functional, or psychological [31]. Knowledge is obtained by an information acquisition and subsequent learning; Skill is developed through repetition and regular practice; finally, Attitude, relates the knowledge integration and the perform the task ability. SBT is used to create training environments related to high-risk situations [32] without compromising the trainee integrity and safety. Additionally, SBT can represent out or the ordinary and unknown environments with 24/7 availability for practice [31]. However, as Vandewaetere states, the training effectiveness using SBT depends on an adequate instructional design that includes authentic and real learning tasks in a simulated environment [9].
2.2 Extended reality in learning and training
Extended reality is an umbrella term that covers VR, AR and MR. XR tech modifies the user interface to immerse the user in the virtual environment (VR) or augments the user´s surrounding (AR), or both of those (MR). Virtual reality (VR) has made a breakthrough and is now used in a wide range of applications. It allows a user to be completely immersed and involved, making them feel "present"[33]. VR has been widely used for training applications due to the ability to design specific, complex and safe environments at low cost. In fact, training using virtual reality can achieve performance levels similar to real training environments [34], and in some cases, performance has improved [35, 36]. For instance, in educational, industrial and medical applications, VR has shown wide efficacy as a training tool. In medical applications, such as surgical equipment training in the operating room [37], González concluded that the automation of procedural steps was the main contribution. Additionally, the trainee can remember the most important steps thanks to the possibility of repeating the procedure as many times as necessary. Siu [38] proposed a VR-based system to predict the acquisition of surgical skills. VR improved the skill level in basic surgical tasks such as: Peg transfer and Needle passing tasks, thanks to the adaptive VR incorporation. The main achievements were improvement in the ability to interactively assess learning and skills impairment, optimize skills re-learning and surgical skills practice. Sattar [39] evaluated VR in abdominal surgery training using laparoscopy with 87 medical students, compared to video and text-based learning. Sattar found that motivation and knowledge transfer was significantly higher for VR training compared to other methods. Also, it was identified that VR fostered active learning. Mickiewicz [40] evaluated the VR impact on otolaryngology resident’s training. VR training allowed the participants to significantly improve virtual antromastoidectomy performance and resulted in a shortening surgery duration and number of mistakes made. Also, VR training provides a structured, safe and supportive environment to familiarize oneself with complex anatomy and to practice surgical skills. Ho [41] used virtual reality for training in medical device’s assembly, finding superiority in VR training compared to traditional training. Ho found that incorporating AI with VR achieved higher effectiveness scores, increased confidence level and shorter training time compared to traditional training. Besides, VR has been used to evaluate the competencies of residents in orthopedics [42], as a remote training tool in the Covid-19 pandemic [43], introducing curriculum changes to incorporate VR [42]. In other applications, Pila [44] uses VR in training systems for inspection and verification tasks on production plants. VR increased the realism level, immersion and interaction between the components, the environment and the user. Kalkan [45] implemented VR to improve the performance in complex assembly tasks. Results revelated that VR-based training reduced training time up to 25% per subject and increased performance up to 27%. Lee [46] developed a VR application for procedural tasks training in factories, resulting in performance improvement. However, according to Le, multiple design considerations need to be taken into account in designing VR (GUI) interfaces for factory training (Fig. 1).
VR Application designed for operating room (OR) equipment use training [37]
Another XR technology is Augmented Reality. AR is an interactive technology that combines digital and physical information through different technological devices interacting with the surroundings/environment in real-time [47], or as De la Torre mentions [48], “it is a technology that allows user interaction with the physical and real-world that surrounds it". AR combines computer-generated 3D objects and text superimposed on real images and video, in real-time, as seen in Fig. 2. Therefore, AR is a technology that facilitates the integration of real elements with added virtual ones, to create a new communicative scenography [49]. Santos [50] states three features that make AR an interactive technology to create learning experiences: Added information in the real world improve perception by superimposing real objects and virtual text, this reduces the cognitive load due to reduced working memory usage; Contextual visualization improves spatial integration by providing relevant clues in the real environment that help the trainee to build his/her knowledge base; finally, the simultaneous use of vision and touch senses to present information improves integration based on body movement.
Augmented reality. Everything you see is real, with an additional layer of superimposed data in your field of view. O´Conell. [51]
Often, AR is used in initial phases of training with remote assistance for maintenance tasks. For instance, AR is widely used in medical procedures to increase accuracy [52] and several applications used AR in surgical skills training [53, 54]. Martín-Gutierrez [55] used AR to train engineering students in spatial skills. The experimental group improved their spatial skills significantly compared to the control group. Similarly, Lee [56] shows AR applications for military vehicle and equipment maintenance training. In manufacturing industry, it is used for vehicle assembly line training, electronic devices assembly [57] and industrial plants inspection. Gavish [58] compared AR with other maintenance and assembly training methods. In the evaluation, operators who used AR had significantly better performance in the task-time compared to those who did not. Westerfield [27], combined AR with an intelligent tutoring system for computer motherboard assembly training, and the results showed an increase in user scores and performance. In medical applications, AR has been used in surgical procedure training. Sanne [59], analyzed four AR simulators for laparoscopic surgery training and shows the advantages for training basic skills such as manual coordination, grip and traction, as well as for advanced skills, such as clip application, suture placement, dissections, among other. Likewise, the importance of AR to provide information during the procedure is evidenced, mainly in the early stages of training. Finally, Chowriappa [60], demonstrated how AR improves skills acquisition with minimum demand for cognitive load in the training of Uretrovesical Anastomosis—UVA. This study emphasizes that AR allows the anatomical and scenario variability, didactic education and user´s adaptation, in an integrated way along with the true practical acquisition of surgical skills.
Mixed reality is a technology in which virtual objects and other digital information are superimposed over the real-time view of the physical objects providing a composite view with guides for carrying out a task with or on physical objects. For instance, Kobayashi [62] implemented a mixed reality prototype to transfer knowledge from experts’ surgeons to novice students. Using a mixed reality interface, a set of cues were projected to guide “blind insertion” invasive procedural training. Margarido [61] designed a mixed reality simulator to train pinpoint insertion of intravenous needles. Using an OST, the mixed system provides a visual guide to the insertion task that helps novice students performing the central venous catheterization procedure. The MR system development evidenced a significative difference in task time and number of errors made by novice participants. Brunzini [63] applied mixed reality to train lumbar puncture. A training scenario was developed for novice students to practice. Results confirmed the great potential of extended reality as a support tool for future medical procedures, both for medical training and when performing real life procedures. Other mixed reality applications have been used in the industrial field such as on-the-job training. Sautter [64] shows potential for effective industrial sector training with regard to key aspects of constructivist learning. In addition, MR should be applied regarding the specific learning tasks, for instance, integrating workers in new planning processes right from the beginning. In maintenance tasks, Su [65] proposes an auxiliary equipment maintenance system based on Mixed Reality technology. The system drives the maintenance procedure guidance and equipment training phase, where users only need to follow the steps set by the system. MR can help repairment personnel check and carry out maintenance operations without a maintenance manual, especially in case of difficult complex industrial equipment maintenance as well as in its potential difficult circumstances.
2.3 Working memory and cognitive load
The cognitive factors (or cognitive ergonomics) studied by Human Factors, comprise, on one hand, communication and teamwork, and on the other hand, perception, work memory, reasoning and motor skills [66]. Long-term memory is the place where knowledge is stored, likewise, short-term memory or working memory is responsible for processing the information required during task execution. Or as Baddeley [67] defines it, Working Memory is a dedicated system that maintains and stores information in the short term, in periods between 15 and 30 s, and is the basis of human thought processes related to the conscious activity that a person develops. Based on the limited processing capacity of working memory, Sweller proposes a theory of cognitive load [68, 69] and later Mayer proposes the cognitive theory of multimedia learning [70, 71]. These theories raise premises related to the use of limited cognitive resources and the limited capacity of a learner when dealing with new information. The cognitive load that a learner presents can be a) Intrinsic, it is the load inherent to the task complexity; b) Extrinsic, is the load that saturates, pollutes, and affects working memory, generally caused by spatially misplaced instructional materials or user interfaces whose representations hinder cognitive processes; c) Germane is responsible for contributing to learning. The three types of cognitive load are complementary, so it is usually sought to reduce the Extraneous cognitive load, free up working memory resources, and achieve more efficient learning [72]. Therefore, the main strategy proposed by Sweller, to deal with cognitive overload, is the reduction of extraneous cognitive processing. For example, in Information Search and Matching [17, 68]. Based on the evidence, Sweller and Chandler [68, 73], have proposed some principles for the reduction of extraneous cognitive load: coherence, redundancy, signaling, temporal contiguity, and spatial contiguity. As Jeroen [3] explains, the Redundancy Principle indicates that the presentation of redundant information typically has a negative impact on learning. However, learners have to find out that the information from different sources is actually redundant, which is a cognitively demanding process that does not contribute to meaningful learning; the Signaling Principle or attention-focusing principle indicates that learning may be improved if the learner’s attention is focused on the critical aspects of the learning task or the presented information. It reduces the need for visual search and so frees up cognitive resources that may then be devoted to schema construction and automation, with positive effects on transfer test performance; the Temporal Contiguity principle or temporal split-attention principle indicates that learning from mutually referring information sources is facilitated if these sources are not separated from each other in time, i.e., if they are presented simultaneously, and the Spatial Contiguity or split-attention principle, refers to the finding that higher transfer test performance is reached when mutually referring information sources are physically integrated with each other in space. Finally, the spatial contiguity and signaling principles is a special interest for the spatial integration of information and attention changes, due to the ability to relate objects in a 3D space and associate them in a spatial context.
2.4 Split attention and signaling principles
According to Ayres and Sweller [16], the split attention principle establishes that, for instructional material design, it´s important to prevent trainees from dividing their attention among multiple sources of information, to ensure mental integration. Due to the need to integrate multiple sources of information, the extraneous cognitive load increases having a negative impact on learning. This effect is observed when trainees who study with integrated information formats outperform those who study with the same information but in non-integrated formats. In Kalyuga [74], there is evidence that spatial integration of information sources reduces attention shifts without the need for additional relationships. This increases effective working memory, reducing the cognitive load, and increasing test performance. Likewise, the author mentions the need to be careful to avoid information redundancy that could eliminate any positive effect.
Mayer [71], identifies some conditions where the spatial contiguity effect is stronger. a. Trainees with low prior knowledge: b. Spatial integrated formats are more effective than separate formats for low-knowledge trainees, but not for high-knowledge trainees [75]; c. Text and non-redundant images, it is applied when multiple sources of information would be unintelligible in isolation; d. Complex Lessons are applied when the material is complex, but may not apply when the material is so simple that the separate design does not overload the cognitive system; e. Interactive formats, the principle of spatial contiguity can be strengthened when trainees create integrated presentations by moving text to relevant parts of a graphic. In the same way when the material has high levels of interactivity. In a Meta-analysis, Schroeder [76], studied some variables like: instructional medium, image type, element interactivity, learning experience, redundancy of learning material, domain, education level, world region, testing type, duration of intervention, and incentives. Finally, the analysis found some correlated effects: learning experience has a small effect, redundancy of learning material has medium effect, educational level has big effect. Pouw [77] studies in two experiments the split-attention effect. Results indicate that increased cognitive load demands due to spatial separation of information is a viable underlying mechanism for the split-attention effect. However, spatial separation is likely not the only, nor a sufficient, condition for the “split-attention effect” to occur.
According to Van Gog [78], The signaling principle, also known as the cueing principle, refers to the finding that people learn more deeply from a multimedia message when cues are added that guide attention to the relevant elements of the material or highlight the organization of the essential material. In multimedia learning, this means that the attention allocation of novices, who lack prior knowledge of a task, may rely more on the characteristics of the stimulus material. Some visual features are used to implement signaling principle, for instance, text-based cues, Pictured-based Cues, Coding cues, Cues based on animation. Several techniques are used to signaling: Color, Blinking and movement. Grogorick [79], Moon [80], Jin [81], Bernhard [82], Grogorick [83], Jamet [84], Jarodzka [85], used color coding to guide attention towards relevant information. Finally, Grogorick [83] and El-Nasr [86] used element movement to guide attention. Position changes and short paths were used.
3 Materials and methods
3.1 Procedure
3.1.1 Review question
The methodology proposed by Kitchenham [87] and Petersen [88] was used to conduct the review. This study’s aim was to review the literature to identify strategies oriented to avoid changes in spatial attention or reduce its effect as well as how these strategies are implemented within the graphical user interface to provide procedural information to trainees. Therefore, the following review question (RQ) was posed:
RQ: Which strategies are applied to reduce spatial attention changes and cognitive load in learning environments that uses Extended Reality to provide information to trainees?
3.1.2 Research context
PICOC (Population, Intervention, Comparison, and Outcomes) was used to identify keywords and create search strings to answer the review question:
-
Population: Journals papers, book chapters, conference proceedings and doctoral thesis.
-
Intervention: User experimental studies that evaluated: performance (time, number of errors), cognitive load and/or mental effort.
-
Comparison: comparison but without doing any intervention
-
Results: Set of studies that evaluates user interfaces in XR technologies, performance (times, number of errors), cognitive load and/or mental effort.
-
Context / level of coverage: Articles in English and Spanish published since 1990.
3.1.3 Keywords
The identified keywords are shown in Table 1. A list of 12 search terms was obtained, grouped into four categories: a. Related to User's activity; b. Related to cognitive effect; c. Related to XR Technology; d. Related to user interface. Additionally, synonym terms were included. Finally, a search string was built using the Boolean connectors AND, OR and NOT.
3.1.4 Scientific databases
The strings were used to search in four multidisciplinary databases: EBSCO, Web of Science, SCOPUS, with a time frame between 1990 and 2022 (Table 2).
3.1.5 Inclusion and exclusion criteria
According to Pettersen [88] suggest, To guarantee the study identification process the first author generated a validation set of studies after the second author completed the search and inclusion/exclusion process (Fig. 3).
Mixed reality System development by Margarido [61]
Inclusion criteria applied:
-
Journal Articles, Conference Proceedings, and Doctoral Thesis.
-
Studies in Spanish or English language
-
Studies that evaluated cognitive load, mental load, mental effort, performance or learning.
-
Studies that evaluated instructional material oriented to learning and training.
-
Studies that used as display devices: desktop PC’s, mobile devices, High definition displays, Headset displays, for Augmented, Mixed Reality, Virtual Reality and Multimedia environments.
-
Studies on demographics ranging between 18 and 50 years old.
Exclusion criteria applied:
-
Non-experimental studies
-
Reviews and studies without users
-
Studies about cultural applications
-
Studies including participants with cognitive disabilities, low vision, vision problems
-
Studies with underage participants or over 50 years old.
With the set of 53 chosen primary studies, a quality assessment was carried out using the following questions:
-
Is the study’s aim clearly identifiable as an assessment that measures users' cognitive load or performance?
-
Do the study findings relate aspects, such as spatial contiguity or visual factors, with measures of users' cognitive load or performance?
-
Are the experimental studies clearly defined?
3.1.6 Studies selection
From database searches, 13,762 articles were obtained. Once the analysis by title was made, the following results were found (see Fig. 4):
EBSCO: A total of 1893 results were obtained, of which, once the title was reviewed, 492 were obtained, excluding 1401 because they did not meet the inclusion criteria. There were 49 articles selected from the abstract reading phase, also excluding 443 articles because they did not have experimental studies. When a detailed analysis was performed, 39 articles were obtained, excluding 10 for not providing relevant information for this research. Finally, 27 articles were obtained with valuable information, after exclude 12 studies by do not answer the quality assessment questions.
WoS: A total of 6385 articles were obtained, of which, after performing a title analysis, we obtained 975, excluding 5410 articles for not meeting the inclusion criteria. There were 57 articles retrieved from the abstract reading phase, excluding 918 for not having experimental studies. When performing a detailed analysis, 36 articles were obtained, excluding 21 for not providing relevant information for the research. Finally, there were 11 articles selected with valuable information for the study after exclude 25 studies by quality assessment questions.
Scopus: A total of 5484 articles were obtained, once a title analysis was performed, 479 were obtained, excluding 5005 because they did not meet the inclusion criteria. There were 49 articles selected from reading the abstracts, excluding 430 articles because they did not have experimental studies. When performing a detailed analysis, 35 articles were obtained, excluding 14 for not providing relevant information for the research. Finally, 15 articles were obtained with valuable information for the study after exclude 20 studies by quality assessment questions.
3.1.7 Classification and analysis
The strategies identified were grouped according to the factor type to avoid split attention. Afterward, the documents belonging to each topic were counted. For example, for the spatial integration strategy, the distance was identified as the main property. For the addressing strategies, characteristics related to coding, signaling, and segmentation were identified (Fig. 5).
3.2 Evaluation validity
3.2.1 Descriptive validity
Descriptive validity is the measure by which findings are described accurately and objectively. To mitigate factors affecting descriptive validity, a data collection template was designed to support the record. The template guides the data extraction process objectively. Therefore, descriptive validity is considered to be consistent, and the factors are objectively controlled.
3.2.2 Theoretical validity
The theoretical validity is determined by the reviewer's ability to capture what is intended to be captured. In addition, confounding factors such as bias and reviewer selection play an important role.
Identification and Sampling: Studies may have been inadvertently omitted. To mitigate this, the search was supplemented with a retrospective sampling after reading the full text. Biases associated with reviewers may appear during selection and data extraction. Study selection was made by an individual author. To reduce these factors that compromise validity, study identification was assessed by creating a set of reference articles. This was done by the first author, through snowball sampling. A small set of new studies was obtained during this phase, indicating that the overall conclusions of this review will not change. There are other potential compromising factors, such as activities describing the study that may be omitted, misinterpreted, or incompletely reported. This may affect the classification of activities. As a control action, the first author checked the extraction.
Data extraction and classification: During this phase, the reviewer's bias is also a threat factor. In this phase, it is considered useful for one person to extract the information and another to review the extraction. To reduce this threat factor, the first reviewer evaluated all extractions performed by the professional. Although this step involves human judgment, the threat factor can be eliminated.
3.2.3 Interpretative validity
Interpretive validity is achieved when the conclusions are drawn are in accordance with the data, and therefore, adjust and reaffirm the conclusion validity. A threat factor in data interpretation is researcher bias. However, the authors have experience in similar reviews, which reduces the threat factor related to data interpretation.
3.2.4 Repeatability
Repeatability requires detailed information about the research process. For this reason, the process followed is described in detail, as well as measures to reduce the threat factors to validity. In a direct way, repeatability was favored by using the control measures adopted.
4 Results
4.1 Publications overview
4.1.1 Publications per Year
The first study was published by Kalyuga[74] in 1999. Between 1999 and 2022, a trend analysis of the average number of publications grouped into six-year intervals is conducted. 0.5 per year were published in the first period (1999–2004), 1.6 per year in the second period (2005–2010), 2.1 per year in the third period (2011–2016), and 4.5 per year in the fourth period (2017–2022). There is a definite upward trend in the number of publications, especially over the most recent period (2017–2022), when there has been a more than doubling of publications.
4.1.2 Studies aims/objectives
Studies objectives identified were grouped in three groups, see Fig. 6:
-
Spatial Integration Techniques: Comparative studies about instructional materials with physical integration formats 25% (n = 13). Integration by proximity [89]–[92], active [93] and non-integrated integration[94]–[97], to avoid attention changes was identified.
-
Comparative evaluation: A set of studies 34% (n = 18) was as an objective the comparative evaluation between XR materials vs. paper-based instructions [58, 98, 99]. The evaluations were centered on cognitive load and user performance when attention changes occurred.
-
Visual Features assessment: Several studies 42% (n = 22) using visual features in instructional material to deal with attention changes and reduce cognitive load [92, 95, 100, 101]. Visual features as color, shape and signaling were identified.
4.1.3 Field of instruction
The studies identified were aimed at the evaluations mentioned in Sect. 4.1.2. These evaluations were conducted in various fields of instruction. They were grouped into three main fields, see Fig. 7:
-
Engineering Education: A set of studies (n = 22), equivalent to 42%, focus on engineering education. For instance, in higher education [96, 102, 103], electrical and mechanical engineering education[74], among others. The studies’ aim was to evaluate instruction material features in user cognitive load.
-
Basic Sciences: (n = 17) studies equivalent to 32%, were applied in the educational area of basic sciences such as Biology [104], physics, chemistry [105], mathematics [75, 101] and health. For these studies the main objective was to define guidelines in the use of visual features for XR instructional material.
-
Industrial Learning-Training: Finally, (n = 14) studies corresponding to 26% focused on learning, training and industrial assistance[106], mainly in assembly tasks [94], maintenance [92, 107] and reduction of errors during task execution [100].
4.1.4 Type of activity
The instructional material designed in the studies had several purposes as educational activities. The type of activities identified in the studies was:
-
Learning: learning was identified as has the greatest number of studies (n = 37) equivalents to 70%. In other words, the material was used in a preliminary stage of training, during the presentation of information and knowledge learning.
-
Training: 30% (n = 16) studies used the material in a later stage of training, in applications for military training [108], devices assembly[107], improving athlete performance [97] and in surgical procedure performance[89]. For example, Kotranza [89] studied real-time feedback for breast surgery, improving physicians' efficiency and psychomotor skills.
4.1.5 Technology
Several XR technologies were used in the studies reviewed. Augmented Reality, Virtual Reality and Mixed reality were used as educational tools. Next, the XR tech was used as a follow:
-
Augmented reality: This tech was the most used XR technology with educational purposes. 55% of studies (n = 29) used Augmented Reality to deliver information to students and trainees, for instance, in engineering education and basic sciences[109]. AR in work stations [94, 110], class rooms [101], plant inspection and daily activities [111, 112].
-
Virtual Reality: This technology was used in (n = 13) 25% studies for maintenance training and task assembly, scenario simulations and skill training. Mostly in work stations [58, 95] to evaluate the user performance using VR.
-
Mixed reality: In (n = 5) 9% some studies MR was used for real object, tools or machine operation training. This tech was used because virtual objects can be manipulated in the scenario [103, 113].
-
Multimedia: (n = 6) 11% of studies used Multimedia instruction to compare with XR technologies, mainly to evaluate spatial integrated formats [114, 115].
4.1.6 Visualization devices
Several devices were used in the studies experiments:
-
HMD: Head-Mounted Displays or Headsets was the most used with 57% of studies (n = 30). Generally, in Mixed reality and Virtual Reality applications because of the ability to have the hands freed to execute a learning or training activity. Several devices were used in the experiments. AR devices as Epson-Moverio [97, 113, 116, 117], Google Glass [107, 112] and AR glass prototypes [91, 98, 101, 118], VR devices as Oculus Gear VR [37], Oculus Rift [101] and HTC[117], Mixed Reality Devices as Hololens2 [113, 119].
-
High Displays or AR Projections. These devices were used in Augmented Reality and Virtual Reality 2D in (n = 12) studies 23%. For instance, [120,121,122] used high size displays to project AR content, for instance in Cooper [122], was used to visualize a car activity.
-
Computer Monitor: (n = 5) 9% projects use AR in workstation using camera and desktop, e.g. [94, 96, 99], mostly to conduct assembly task training.
-
Tablets: In maintenance activities (n = 4) 8% studies was use AR with tablets [100, 123], and learning content activities [109].
-
Smartphones: AR is very common in smartphones. However, only a few studies (n = 2) 4% use AR in smartphone [92, 96].
4.1.7 Device experience level
The reviewed studies conducted assessment with participants whose level of experience in using these types of devices could be established as follows:
-
Beginner: In (n = 33) studies, equivalent to 62%, subjects whose level of experience was Beginner participated. This was mainly observed with HMD devices for AR [97, 107, 112, 113] and VR [58, 95] in learning and training activities.
-
Intermediate: (n = 9) 17% of the studies had participants with an intermediate level of experience. These studies mainly assessed those who used Smartphones [92, 96] or tablets [100, 109, 123].
-
Expert: Five (n = 5) studies with expert participants comparatively evaluated the experience of novices Vs experts or were studies that only sought to capture the experience of the participant.
-
N/R: Three (n = 6) 11% studies did not report the level of experience in using the device.
4.1.8 User knowledge level
The studies reviewed focused on learning and training activities. Its objective was to extend knowledge level and improve participants’ performance. Most of the studies assessed the level of knowledge of the participants in the tests. The reported level was as follows:
-
Low: In 64% (n = 34) of studies were conducted with participants who reported low or no knowledge in the area of learning/training [58, 103, 107].
-
Intermediate: In 11% (n = 6) the level of knowledge reported was intermediate. Mainly, in studies that seek to increase the performance of activities using XR.
-
High: In (n = 8) studies, equivalent to 15%, the level of knowledge reported was expert. These studies focused on comparing the level of performance once the technology was used for training [89, 100, 123].
-
N/R: (n = 5) studies did not report the users’ knowledge level.
However, studies such as Choi [96] and Fujimoto [98], where they used groups of subjects combining the level of knowledge and experience in the use of the visualization platform, stand out.
4.1.9 Graphical user interface features
The studies reviewed applied several visual features to avoid user attention changes. 58% of studies used features based on closeness between elements, while 17% did not use these features. Two formats related to proximity were identified: integrated format and active integration format. 17% of studies identified comparatively evaluated participant performance using the integrated and non-integrated formats. Finally, 40% of studies that applied physical integration complemented with other features. Task-time, accuracy and number of errors were used as performance measurement; and Learning, retention and knowledge transfer were used as learning outcome measurement.
Most studies focused on comparing factors that prevent attention changes, evaluating the user's performance and learning outcomes in several visualization platforms while performing learning or training tasks. In 17% of these studies, non-integrated formats were compared with integrated formats. In 28% XR technology (AR/VR/MR/Multimedia) was compared with printed material for instruction and training. In Table 3, a general description is shown. Another group of features identified are related to visual features such as color, shape, or symbols which are used to drive attention. The following point (4.2.1) describes how the features were used to avoid attention shifts in training tasks that use multimedia XR technologies.
4.1.10 Regarding spatial proximity
Three groups were identified related to spatial proximity between information source and spatial objects: a) Non-integrated, split, or distant format. Corresponds to formats whose design is spatially distant, that is, with a spatial separation that avoids spatial integration. b) Integrated format, defined by Schroeder [76], as formats whose information sources are close to each other. c) Active Integration Formats are used when trainees receive information in a non-integrated format and must perform the integration themselves by interacting with the content.
-
Non-integrated formats: Printed instructional materials are a kind of non-integrated format, as it requires a constant shift of attention between the guidance provided by the material and the actual setting where the task is performed. We identified (n = 20) 38% studies that used this type of format, 21% (n = 11) used only non-integrated format. In AR, 19% (n = 10) of the 53 studies reviewed use this format, 2% (n = 1) gets cognitive load reduction, 11% (n = 6)) better task-time, 2% (n = 1) better accuracy and 2% (n = 1) better overall performance and 2% (n = 1) did not report reduction in cognitive load or performance. Of this 21% (n = 11) that used non-integrated format, only two studies were evaluated regarding cognitive load. Azimi [113] and Kruijff [101] reported low cognitive load in trainees with low knowledge level. However, Wang [107] gets opposite results and reported high cognitive load. Complementally, in the use of non-integrated format, 28% (n = 15) of studies used codification features and 21% (n = 11) used signaling ones. In VR, 11% (n = 6) used Non-Integrated format. In Kalyuga [74] and Seufert [105], which comprise 6% (n = 3) of studies reviewed, reported low cognitive load, but in contrast, Rao [95] reported high cognitive load. In regard to performance, 4% (n = 2) reported short task-time completion and 9% (n = 5) showed better overall performance. In Multimedia 6% (n = 3) of studies used this format, 4% (n = 2) reported better learning outcomes and there were no reports on cognitive load or performance. Finally, in 19% (n = 10) of studies, apart from the non-integrated format use they were conducted also using codification or signaling features.
-
Integrated Formats: 72% (n = 38) of the studies selected used integrated formats, see Fig. 8. 58% (n = 31) of all studies reviewed used only integrated format, 32% (n = 17) used integrated format along with visual features and 19% (n = 10) compared Integrated format with non-integrated formats. In AR 38% (n = 20) of all 53 studies used this format and 19% (n = 10) found a reduction in cognitive load, 21% (n = 11) shorter task-time completion, 7% (n = 4) better learning outcome and 4%(n = 2)) reported better learning outcomes. In Multimedia, 19% (n = 10) applied integrated formats, 11% (n = 6) reported better learning outcome and 4% (n = 2) shorter task-time completion. In VR, 13% (n = 5) of studies use integrated format, 2% (n = 1) found a cognitive load reduction and 7% (n = 4) shorter task-time completion. In MR, 6% (n = 3) of studies used integrated format, 2% (n = 1) reported cognitive load reduction and 4% (n = 2) better performance.
Active integration format: Active integration requires trainees to perform an action to move content to an integrated position. Bodemer [93] compared active integration and static integrated formats where the trainee gets up to 75% better performance when active integration is used. Also, it was identified that active integration showed significant performance improvements compared to the non-integrated formats. In another study, Bodemer [141], reports the difference between 2D formats (Non-integrated, integrated and actively integrated), and identified the usefulness of active integration, especially in instructional materials with high difficulty and complexity.
4.1.11 Effect conditions identified
Some additional conditions to proximity effect were identified: Presentation format, complexity task and device Field of view. The influences of representation formats were identified: text-text (T-T), text-graphic (T- G), graphic-text (G-T). Trainees who used text-text (T-T) and text-graphic (T-G) representation formats had better performance results, in contrast to graphic-text formats, who had lower performance results. Finally, some studies propose a design guideline based on task-type.
In the selected studies, the integrated format demonstrated its advantages compared to the non-integrated format. The results obtained improvements in performance, task time, precision, efficiency and, to a lesser extent, learning. Likewise, in the eyetracking analysis, the integrated format generated more integrative saccades. This explains that the integrated format encourages the use of a processing pattern, and the user process the information as a single unit, while when they are not integrated, they process them as two different units (Fig. 9).
Fujimoto Study [98]. On the left, Integration by proximity: the information is displayed close to the object. On the center: information is displayed in a random position with respect to the object. On the Right: Non-integrated, where Information is displayed in a position far from the object
4.1.12 Regarding visual features
Several visual features were used to reduce cognitive load and improve performance. In Fig. 10, a summary of visual features is shown.
4.1.12.1 Codification
The main use of codification was to provide correlated information. Therefore: color 40% (n = 21), shape 19% (n = 10), size 2% (n = 1), and positioning 2% (n = 1) were used as codification features. 44% (n = 23) of all studies used at least one codification characteristic, as shown in Table 3. The most frequent use of color was to apply the same color tone for content with correlated information source. A strategy aimed at reducing information pinpoint as well as cognitive load. Although, 55% (n = 29) of studies did not use any codification feature. In AR, 23% (n = 12) of all studies used codification, 4% (n = 2) reported a cognitive load reduction, 13% (n = 7) reported shorter task-time completion, 6% (n = 3) better performance results and 2% (n = 1) better learning outcome. In VR, 11% (n = 6) use codification features, 2% (n = 1) reported reduction and 2% (n = 1) no reduction in cognitive load, 4% (n = 2) better accuracy results and 4% (n = 2) a decrease in the number of errors. However, 4% (n = 2) reported the longest task-time completion. In MR, 2% (n = 1) reported a low cognitive load, 2% (n = 1) shorter task-time completion and 2% (n = 1) better learning outcome. Finally, in Multimedia, 6% (n = 3) reported better learning outcome but no performance result was reported, see Fig. 10. Shape was used as codification feature always complementing color codification in 17% of studies, 11% (n = 6) in AR, 4% (n = 2) in VR and 2%(n = 1) in MR. Generally, Color as the visual attribute mostly used.
In Rao [95] used color and shape to indicate the instruction location that the trainee should follow; this use was associated with lower mental effort in trainees. Kim [142] showed that color coding improves information comprehension. Related to time-task completion, in Sirakaya, there was no evidence of shorter task-time completion associated with color use [68]. Therefore, Sirakaya [92] mentions the need to limit the number of colors used, since the simultaneous use of many colors could eliminate any positive effects [92] and recommends a maximum use of 5 colors simultaneously. In Kruijff [92, 101] the color preference on labels was identified, its influence on text readability and its correlation with the background. Blue was the most preferred by test participants with OST (optical see-through) devices, gray and green were the preferred for VST (video see-through) devices. The impact of color use was evidenced by eye-tracking analysis, where it was observed an improvement on retention and learning results. In studies by Ozcelik [126] and Choi [96], it was observed that the average duration of ocular fixation indicated deeper processing. Thus, trainees who used color-coded format, paid more attention to the color-coded information in contrast to trainees who received un-coded information. It was also evident in Herrema [94], that the use of the same color to code the instruction (in text) and signaling (using some figure), its positioning in the workspace where the action must be performed improved task-time completion. In the Rao [95], Choi [96] and DeCrescenzio [91] studies, the color was used to code the text describing the task objective, 3D models objects and visual signals associated with the final objective. In other studies, such as Fujimoto [98] and Gavish [143], color coding was used, which complemented with graphic elements, such as boxes, indicators, icons and arrows [89, 97] also allowed for an improvement and shortening of tasks.
4.1.12.2 Signaling
Van Gog [78] defines signaling as the set of signals or clues used to guide attention to the relevant elements of the instructional material. Signaling has been used in several ways to reduce cognitive load and better performance as well, see Fig. 8. In 28% (n = 15) of studies, direct signaling techniques were used. The most used characteristics were, labeling with 6% (n = 3) and boxes with 9% (n = 5). Also, 21% (n = 11) of studies used symbolic signaling techniques to indicate the target position required, 17% (n = 9) used arrows and 2% (n = 1) used a dot. Finally, 21% (n = 11) studies used other target positioning signals such as transparencies, flashing dots, highlighting tool, and animation. The most used characteristic was transparency (with 13%), used to highlight areas or elements of interest where attention is desired. Also, flashes, highlights or blinks, used color or other features. Kruijff [101] evaluates vertical, horizontal, and rotational movements getting cognitive load reduction and drive the user’s attention (Fig. 11).
Direct Signaling: As a common strategy, primitive geometric figures such as squares, rectangles, and circles were used to frame the object of interest[113]. Direct signaling was the most used feature in AR Material, where 23% (n = 12) studies reported their used, 13% (n = 7) cognitive load reduction, 13% (n = 7) better performance results and 2%(n = 1) better learning outcomes. In VR and MX, only 2% (n = 1) of studies report better performance results and 2% (n = 1) better learning outcomes, see Fig. 12.
In Herrema[94], Wachsmuth [52, 90], DeCrescenzio [91], Kishishita [75], Wang[107], used boxes to limit objects and guide their identification within the visual field. However, some studies used primitive shapes with segmented contours as in Park [103] and Izard [37], where object detection achieved better performances results. Bodemer [93] used labeling to direct attention within inner and outer field of vision. Bell [144] used to connect lines to establish a visual link between the object of interest and its label. In Tatzgern[145] and Madsen [109], it states that, for a better effectiveness labels must be perpendicular to user view, the pinpoint located in its center, the connection line should not occlude the object but as short as possible to maintain a compact display. The use of labels for comparison tasks should reduce the use of connecting elements (preferably lines) and not reduce the field of view. For search tasks, the field of view should be expanded using connecting lines with greater contours and evident [104]. The labeling was also accompanied by subtle animation. In Kruijff [101] lateral, vertical and circular animations were given to labels. Circular animation was mostly identified in VST (video see-through) devices. For VST (video see-through) devices, blinking obtained better results. The combination of direct signaling with animations was also observed in this study. Biocca[146] uses a guide based on a flexible tunnel of frames that begins at the user's head and ends at the object location and orientation, while showing an animation that progressively reduces the frame size till it reaches said object. Circles were also used in Baudisch [124] and Wang [107], to surround target objects out of field of view with rings that reach its limits indicating the presence of more objects. This technique was used to indicate location and infer the distance to target based on the ring arc size. A variation of this technique is proposed by Renner [117], where circles in waves form propagate towards the target, like the concentric circles that appear when throwing a stone in the lake, only with a reverse direction. These feature combinations show error levels similar to techniques that use arrows. In addition, there was a significant reduction in task-time completion. However, these techniques have limitations with target positioning since its representation is difficult outside the user's field of view.
Symbol Signaling: Some studies use symbols to guide users’ attention. Symbol signaling was the most used feature in AR, where 9% (n = 5) of studies reported cognitive load reduction, 11% (n = 6) reported better user performance and 2% (n = 1) better learning outcomes. In VR, only 2% (n = 1) of studies reported cognitive load reduction and 4% (n = 2) better user performance. In MX, only 2% (n = 1) reported better learning outcome. In Multimedia none of the 53 studies used symbols to drive users’ attention, as seen in Fig. 13.
For instance, Hořejší [99] used arrows, Biocca used omnidirectional attention tunnels [125, 147] and Baudisch [124] used Halos. In Renner [117] three techniques were compared to locate objects outside the field of view (like arrows, waves, and radar bleeping). Attention diversion using arrows showed better results in task-time completion, precision actioning, information readability, and technique usefulness. A technique variant is to place an arrow on the screen outline pointing towards the target location. As a complementary feature, animated 3D arrows were used in Radkowski [110], Choi [96] y DeCrescenzio [91], to indicate turns and point towards parts to be assembled. This combined use reduced the task-time completion as well as the number of errors made. Another identified combination was predictive signals, where the principle of predictability is applied to indicate the location where the visual stimulus will appear. Using arrow prediction, it was possible to increase task performance and considered highly suitable for AR environments, according to Volmer[131], due to mental effort reduction results.
4.1.13 Regarding content segmentation
4.1.13.1 Segmentation
In some studies, content required to execute a task exceeds the available field of view space and could overload the trainee, see Fig. 14. In 21% (n = 11) of studies, content segmentation was divided into smaller parts to avoid or reduce attention changes. In 9% (n = 5) of studies content was segmented with step-by-step instructions. Other studies used the discovery, proposed by Gavish [58], to keep content hidden, but easily found by clicking on an icon.
In AR, 9% (n = 5) of studies reported better user performance when using segmentation techniques, a cognitive load reduction in 4% (n = 2) and better learning outcomes in 2% (n = 1) of them. In VR, out of all 53 studies only 6% (n = 3) of them reported better user performance and 2% (n = 1) reported cognitive load reduction, any of them reported better learning outcomes. In Multimedia, only 2% (n = 1) of studies reported cognitive load reduction and 2% (n = 1) better learning outcomes. Finally, in MX no study reported segmentation techniques.
Segmentation techniques was the most used in AR, for example, Rijnsburger [112], proposes that labels contain a five word maximum to avoid mental overload. Progressive disclosure is also a segmentation technique used in VR environments, Herrema [94] applied it in interfaces overloaded with information to reduce mental effort. Discovery has been used in projects like SKILLS [58], Hanley [148] proposes this concept, which consists of placing an icon to be interacted with to temporarily display information. Liarokapis [121] and Gavish [58] uses labels to link information to objects in AR environments, however, some difficulties when handling detailed descriptions, due to the descriptive text length. Thus, it was divided into smaller sections controlled by the user through a scroll bar. Finally, another segmentation variant is information extraction. Gabbard proposes that irrelevant information be eliminated and that correlated information be grouped together [118]. Rijnsburger [112] states another variant as information segments through personalization, especially for devices with reduced field of view as well as using labels with a 5 word maximum.
4.1.13.2 Hyperlinking
This technique is based on relational hyperlink use where links use visual indicators to relate content and allow the trainee to control playback. Furthermore, in some measure, this technique was found to be more effective than an integrated format. For example, Seufert [105], evaluated learning material in molecular biology under this concept. Relevant words, images, or formulas were accompanied by hyperlinks. When the trainee clicks any of them, it points to the corresponding element in the presentation. The results showed a reduction in trainee cognitive load. Gavish [58], also implemented a similar technique to show details of an assembly procedure instructions done step by step.
5 Discussion and conclusion
In the review process, it was identified that only a few studies focused on cognitive load evaluation. As Ens [24] suggests, AR designers need cognitive effects information to design more effective learning and training environments. The most recent studies are focused on user performance while learning and training, but only a few studies are focused on cognitive effects. XR technologies are most used for learning than for training. In contrast to Gopher [149] state that AR is useful as a training technology strategy due to shorter task-time completion achieved by trainees. AR is the most used technology, mainly using Headsets as display devices. According to Zhang [150], being hands-free it allows to manipulate objects and tools.
XR technologies are used by novice and entry-level users, thus, as Gutierrez and Agati [55, 151] states, XR technologies are useful in learning/training preliminary phases for entry-level knowledge trainees.
6 Regarding spatial proximity
Three spatial formats were identified: Non-integrated, integrated and Active Integration format. The Non-Integrated format allowed for trainees to achieve the best results by using AR technology, mainly, better user performance (task-time completion) and Cognitive load reduction. In the review, it was found that non-integrated format represented better learning outcomes when used in Multimedia technology. Generally, Non-Integrated format achieved a small cognitive load reduction in all XR technologies and better user performance but not better learning outcomes. In contrast, as Gloria Mark [152] states: trainees often compensate spatial separation between information sources by executing tasks faster. This justifies the users’ higher performance but also induces a higher overall cognitive load, higher workload, stress, high level frustration and more pressed for time. Due to these harsh training conditions, it is possible that the learning environment was inadequate for knowledge transfer and retention.
Several studies are based on close proximity stimuli causing associations between said stimuli, as mentioned by Ozceli [126]. Kalyuga [74] supports this thesis, adding that the distance between integrated formats should be the minimum possible. However, the distance or separation measurement unit is not specified. In the review process, no study evidenced the establishment of any rule relating the separation distance with any of cognitive load measures. In this sense, Ogueta [153] specifies this separation variable in angles measured in degrees, it is also identified a proportional correlation between distance, task-time completion, and number of errors made. In other words, the greater the separation distance, the longer was the time for task completion, and higher number of errors made. In addition, in XR technologies the visual separation uses a 3D coordinate system. This implies that separation could not be on an angle, it should be in spatial metrics as inches or centimeters since it must be coordinated with workspace real-world measurements. Additionally, the coordinate system implies movement, therefore it could cause depth perception issues, as Ping [154] describes. To solve this limitation, Diaz [155], proposed guidelines to design user perception techniques and take advantage of users' visual abilities to better blend the physical and virtual world.
The Integrated format was widely used in AR, reporting reduction in cognitive load, better user performance and better learning outcomes. However, in more recent studies, the results related to the reduction of cognitive load are not conclusive [122, 139, 140]. Possibly due to a better understanding and use of technology. However, the best results in learning outcomes were achieved in multimedia technology. XR Integrated format achieved positive results in relation to cognitive load reduction and lower mental effort. This could be due to dual information source processing, according to Fujimoto [98]. In integrated formats, information sources are processed as a single information source, i.e. when the object and source information are close to each other, the user processes all the information at the same time reducing the mental effort. Despite this result, on some studies as Radkowsky[106] and Sirakaya [92], the task-time completion advantage is little or non-significative. This result could be due to the little experience with this devices as Seufert mentions; or also entry-level knowledge and a small trainee sample as Sirakaya [92] reports. Regarding learning outcomes, as Holsanova [114] and Mayer[156] states, spatial contiguity is decisive to facilitate learning. However, in XR technologies we found more prominent results in Cognitive load reduction and user performance than learning outcomes. In learning outcome, the results could be influenced by test design. Seufert and Sirakaya used ad-hoc test, and as Makransky [115] mentions, the subjective measures can provide different information to test the theoretical mechanisms involved in multimedia learning. In XR technologies, Non-integrated and Integrated formats frequently were accompanied by visual features (codification and signaling), complementing the effect to obtain better learning or performance results.
As Mayer [71] states, spatial contiguity principle has a strong effect when the following conditions are evidenced: Low knowledge level, Non-Redundant text and imagens, complex learning sessions or interactive formats. In the review it was not possible to identify these conditions in all 47 studies. However, we can stablish that trainee knowledge level influences task-time completion results when a spatially integrated format is used [75, 101]. Likewise, it was observed that performance increases for trainees whose previous knowledge on the subject is high. However, it showed that in most studies the participants had a low level of experience and low knowledge level. The foregoing shows that although good performance results can be obtained, these could be significantly improved if tests are carried out with participants with high prior subject knowledge [130]. Similarly, the effectiveness results could be justified, as mentioned by Ginns [157] and Polvi [100] influenced by the complexity and type of task, such as information search tasks and information comparison tasks.
Another condition identified was the representation format. The user's working memory behavior is not the same when different representation formats are evaluated. Generally, representation formats are grouped into symbolic (text and other symbols) and pictorial (images and other representations). Likewise, in this type of spatial integration format, the integrations between text-text and text-graphic representation formats showed better trainee behavior with respect to learning performance and comprehension. However, the studies that showed improvements in learning performance did not show significant differences in task-time completion or mental effort [93]. This is because symbol representations are easy to interpret or does not require experience for interpretation [158]. However, pictorial representations such as graphics require prior knowledge and visual spatial ability, to be adequately interpreted, which implies higher working memory processing [133]. Additionally, the representation format processing could be influenced by the its complexity, that is, low complexity representations such as 2D graphics require greater visual spatial ability and processing than the high complexity representations such as 3D models and photographs [157]. Finally, there were differences observed between visualization platforms. For the AR platform with Displays, an increase in the task-time completion was observed, which may show the presence of high cognitive load. For HMD, more precision was observed in the identification of information, but not task-time completion improvements. Finally, in training activities, the integrated format showed better user performance. However, for learning material is difficult to provide concrete design guidelines because when trying to use such guidelines mismatched results were obtained [159] and also due to integrated format results in learning not being conclusive, as Beege [159] reports. In the same way, Cammerat [133] states that knowledge transfer and retention results are mixed, therefore, proposes more studies be performed regarding spatial proximity and learning outcomes. Studies that have shown mixed results of the principle of spatial contiguity, such as Cammerat [133] and Beege [159], use objective evaluation such as eye-tracking and were done on multimedia technology. However, in this review, there was no evidence of any study that evaluated in depth the spatial contiguity principle with XR technologies for learning or instructional material design. Although some multimedia principles are replicable in 3D technologies as mentioned by Lacoche [160], the use conditions of multimedia technologies are different, due to the workspace size and visual aspects involved.
7 Regarding visual attributes
The most used codification technique was color codification. The content of all related information sources uses the same color tone. This technique was the most used in AR technology, participants achieve better performance and a significant cognitive load reduction. This result is possible due to the easy mapping principle, which establishes the ease required to identify objects in a workspace, reported by Tatzgern [145]. In the same way, studies showed improvements in retention and knowledge transfer. These results were due to the efficiency in locating complementary information sources, and the ability to attract user attention to objects with highlighted information. The preference for blue in AR (OST) devices may originate because this color coincides with the cone sensitivity in the peripheral visual field. This indicates that the device display influences the color selection and may be related to the amount of light present in the projections of each device. According to Ozcelik [126] a 5 color maximum is recommended to be used simultaneously. These results are consistent with the Sirakaya research [92], where an adverse effect was noticed due to confusion by a high number of colors used simultaneously. Likewise, Kalyuga [74] also agrees when reporting that using too many colors simultaneously can impose a significant load on working memory and eliminate any positive effect of color codification. Finally, Kalyuga [74] proposed color codification as an alternative to spatial integration to reduce attention changes, especially in computer-based instructional material with few space available. Kaylyuga proposition is based on an eye-tracking study, when working memory processing using color coding format is similar to spatial integration format. However, some limitations such as the number of colors available to use simultaneously [126]; possible perception changes of color in HMD devices caused by lighting and the device's projection technology, according to Livingston [161]; and the colored element position, since color perception worked best the closer it is to central vision focal point and also requires the user’s gaze tracking to detect the best element positioning.
Direct signaling technique was the most used in multimedia technology. Primitive shapes, labels and subtle animations were used where overall results achieved better learning outcome. However, in AR better user performance was achieved. The shape and label uses are related to device Field of view, i.e. locating objects of interest inside or outside the field of view being possible due to task-type requirements. In other words, the label elements selection depends on task-type as mentioned by Polys [104]. As well, to Madsen [109], search tasks needed for visual exploration in the work field is proportional to visual field size. In other words, for large visual field devices, more visual exploration is required and therefore a greater cognitive load is demanded. However, according to Tatzgern [109] for comparison tasks, the visual field should be limited, and although there are continuous changes in attention focus, the location is already known and does not require additional spatial searches, which forces higher cognitive processing. Finally, for user performance training purposes the field of view should be constrained not widen, identified by Ren[162] where he correlates FOV size and performance.
Likewise, the effectiveness of vertical, horizontal and label rotation animation was established to drive attention. Therefore, the use of subtle animation in elements has some advantages. Mainly because it worked best at the visual field edge, while color and shape worked best the closer to central vision focal point it is [75]. This condition does not require user gaze tracking to identify the best element positioning and could be used in large field of view devices. Another advantage is represented by the use of animation of a single object at a time, this could be used to guide procedures and focus attention on the main object.
The signaling based on symbols preferred used arrows as elements for task-time completion and number of errors made reduction. Despite using rings or arcs to guide attention, arrows were the most used. This is possible due to the decrease in the processing required to interpret information represented by symbols. Herrema states when symbols are used they can be quickly interpreted by the user, since recognizing a symbol reduces the processing required for recognition[94]. However, it represents a significant limitation since known symbols must be used to avoid misinterpretation. Additionally, symbols must indicate the action, for example, the arrow indicates direction of movement or displacement [91]. This may represent a limitation related to the type of symbols that can be used, since they must meet recognition conditions.
8 Regarding content segmentation
Instructional and training materials are becoming more comprehensive and complex. Due to this, the amount of content is greater. Therefore, content segmentation was identified as a technique to avoid cognitive overload in trainees. Despite being used in few studies, effects were identified in increasing the performance and reducing cognitive load. However, segmentation techniques represent a weakness evidenced by Hanley [148], since the required instructions are divided into several spaces, generating a violation of the spatial contiguity principle. However, under restrictions such as a 5-word maximum, adverse effects could be avoided. This is consistent with the principle of conciseness mentioned in ISO 9241–12[163], which suggests that users should not be overloaded with extraneous information. Also consistent with the results of Florax [102]that indicate that learning can be improved by segmentation to a greater extent and to a lesser extent with labeling.
References
Dickey, R.M., Srikishen, N., Lipshultz, L.I., Spiess, P.E., Carrion, R.E., Hakky, T.S.: Augmented reality assisted surgery: A urologic training tool. Asian J. Androl. 18(5), 732–734 (2016). https://doi.org/10.4103/1008-682X.166436
Barsom, E.Z., Graafland, M., Schijven, M.P.: Systematic review on the effectiveness of augmented reality applications in medical training. Surg. Endosc. 30(10), 4174–4183 (2016). https://doi.org/10.1007/s00464-016-4800-6
van Merriënboer, J.J.G., Kester, L.: The four-component instructional design model: Multimedia principles in environments for complex learning. In: Mayer, R. (ed.) The Cambridge Handbook of Multimedia Learning, pp. 104–148. Cambridge University Press, Cambridge (2014)
Gavish, N., Gutierrez, T., Webel, S., Rodriguez, J.: Design guidelines for the development of virtual reality and augmented reality training systems for maintenance and assembly tasks. Int. Conf. Ski. 29, 1–4 (2011). https://doi.org/10.1051/bioconf/201101000
Ong, N.T., Bowcock, A., Hodges, N.J.: Manipulations to the timing and type of instructions to examine motor skill performance under pressure. Front. Psychol. 1, 196 (2010). https://doi.org/10.3389/fpsyg.2010.00196
Anderson, R., Campbell, M.J.: Accelerating skill acquisition in rowing using self-based observational learning and expert modelling during performance. Int. J. Sports Sci. Coach. 10(2–3), 425–437 (2015). https://doi.org/10.1260/1747-9541.10.2-3.425
LeBel, M.E., Haverstock, J., Cristancho, S., van Eimeren, L., Buckingham, G.: Observational learning during simulation-based training in arthroscopy: is it useful to novices? J. Surg. Educ. (2017). https://doi.org/10.1016/j.jsurg.2017.06.005
Spruit, E.N., Band, G.P.H., Hamming, J.F., Ridderinkhof, K.R.: Optimal training design for procedural motor skills: a review and application to laparoscopic surgery. Psychol. Res. 78(6), 878–891 (2014). https://doi.org/10.1007/s00426-013-0525-5
Vandewaetere, M., Manhaeve, D., Aertgeerts, B., Clarebout, G., Van Merriënboer, J.J.G., Roex, A.: 4C/ID in medical education: How to design an educational program based on whole-task learning: AMEE Guide No. 93. Med. Teach. 37(1), 4–20 (2015). https://doi.org/10.3109/0142159X.2014.928407
Evans, G., Miller, J., Iglesias Pena, M., MacAllister, A., Winer, E.: Evaluating the microsoft hololens through an augmented reality assembly application, 10197, 101970V (2017), doi: https://doi.org/10.1117/12.2262626.
Acampora, G., Cook, D.J., Rashidi, P., Vasilakos, A.V.: A survey on ambient intelligence in healthcare. Proc. IEEE 101(12), 2470–2494 (2013). https://doi.org/10.1109/JPROC.2013.2262913
Luis Alejandro Andrade-Lotero: Teoría de la carga cognitiva, diseño multimedia y aprendizaje: un estado del arte, Magis. Rev. Int. Investig. en Educ., 5, 75–92 (2012)
Bertrand, J., Bhargava, A., Madathil, K. C., Gramopadhye, A., Babu, S. V.: The effects of presentation method and simulation fidelity on psychomotor education in a bimanual metrology training simulation, 2017 IEEE Symp. 3D User Interfaces, 3DUI 2017 - Proc., pp. 59–68 (2017) https://doi.org/10.1109/3DUI.2017.7893318.
Tabbers, H. K., Martens, R. L., Van Merrie, J. J. G.: Multimedia instructions and cognitive load theory : Effects of modality and cueing, 71–81 (2004)
Cheng, T., Lu, Y., Yang, C. (2015) Using the multi-display teaching system to lower cognitive load, 18, 128–140 (2015)
Ayres, P., Sweller, J.: The split-attention principle in multimedia learning, In: The Cambridge Handbook of Multimedia Learning, Second Edition, pp. 206–226 (2014)
Sweller, J.: Measuring cognitive load, pp. 1–2 (2018). https://doi.org/10.1007/s40037-017-0395-4.
Kalyuga, S., Chandler, P., Sweller, J.: From the archive: Managing split-attention and redundancy in multimedia instruction by S. Kalyuga, P. Chandler, & J. Sweller (1999). Applied Cognitive Psychology, 13, 351-371 with commentary,” Appl. Cogn. Psychol., 25(S1), S123–S144 (2011).https://doi.org/10.1002/acp.1773
Ens, B., Hincapié-Ramos, J. D., Irani, P.: Ethereal planes: A design framework for 2D information spaces in 3D mixed reality environments, In: SUI 2014 - Proceedings of the 2nd ACM Symposium on Spatial User Interaction, pp. 2–12 (2014) doi: https://doi.org/10.1145/2659766.2659769.
Ens, B.: Spatial Analytics Interfaces, (2016) [Online]. Available: http://hci.cs.umanitoba.ca/assets/publication_files/Barrett_Ens_Final_Version.pdf.
Renkl, A., Scheiter, K.: Studying visual displays: how to instructionally support learning. Educ. Psychol. Rev. 29(3), 599–621 (2017). https://doi.org/10.1007/s10648-015-9340-4
Lai, A., Chen, C., Lee, G.: An augmented reality-based learning approach to enhancing students science reading performances from the perspective of the cognitive load theory, (2018). https://doi.org/10.1111/bjet.12716.
Akçayır, M., Akçayır, G.: Advantages and challenges associated with augmented reality for education: A systematic review of the literature. Educ. Res. Rev. 20, 1–11 (2017). https://doi.org/10.1016/j.edurev.2016.11.002
Ens, B., Irani, P.: Spatial analytic interfaces: Spatial user interfaces for in situ visual analytics. IEEE Comput. Graph. Appl. 37(2), 66–79 (2017). https://doi.org/10.1109/MCG.2016.38
Garcia-Sanjuan, F., Jaen, J., Nacher, V.: Toward a general conceptualization of multi-display environments. Front. ICT 3(September), 1–15 (2016). https://doi.org/10.3389/fict.2016.00020
Rashid, U., Nacenta, M. A., Quigley, A.: The cost of display switching: A comparison of mobile, large display and hybrid UI configurations, Proc. Int. Work. Conf. Adv. Vis. Interfaces, pp. 99–106 (2012) doi: https://doi.org/10.1145/2254556.2254577.
Westerfield, G., Mitrovic, A., Billinghurst, M.: Intelligent augmented reality training for motherboard assembly. Int. J. Artif. Intell. Educ. 25(1), 157–172 (2015). https://doi.org/10.1007/s40593-014-0032-x
Vergara, T., Pedraza, G.: User Interface model-based for basic- skills training in medical applications, pp. 1–6
Patterson, F., Zibarras, L., Kerrin, M., Lopes, S., Price, R.: Development of competency models for assessors and simulators in high-stakes selection processes. Med. Teach. 36(12), 1082–1085 (2014). https://doi.org/10.3109/0142159X.2014.930112
Hamstra, S.J., Brydges, R., Hatala, R., Zendejas, B., Cook, D.A.: Reconsidering fidelity in simulation-based training. Acad. Med. 89(3), 387–392 (2014). https://doi.org/10.1097/ACM.0000000000000130
Shaywitz, D. A., Martin, J. B., Ausiello, D. A.: Approaches to training, 9343(8) (2013)
Scerbo, M.W., Anderson, B.L.: Medical simulation, Handb. Hum. Factors Ergon. Heal. Care Patient Saf. (2nd Ed., no. February, pp. 557–572 (2011)
Luong, T., Argelaguet, F., Martin, N., Lecuyer, A.: Introducing mental workload assessment for the design of virtual reality training scenarios, In: Proceedings of 2020 IEEE Conference Virtual Reality 3D User Interfaces, VR 2020, pp. 662–671 (2020). https://doi.org/10.1109/VR46266.2020.1581086856229.
Wallet, G., Sauzéon, H., Rodrigues, J., N’Kaoua, B. : Transfer of spatial knowledge from a virtual environment to reality: Impact of route complexity and subject’s strategy on the exploration mode, J. Virtual Real. Broadcast., 6(4), 572–574 (2009). [Online]. Available: http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.163.2542&rep=rep1&type=pdf.
Bowman, D.A., McMahan, R.P.: Virtual reality: How much immersion is enough? Computer (Long. Beach. Calif) 40(7), 36–43 (2007). https://doi.org/10.1109/MC.2007.257
Nagendran, M., Gurusamy, K.S., Aggarwal, R., Loizidou, M., Davidson, B.R.: Virtual reality training for surgical trainees in laparoscopic surgery. Cochrane Database Syst. Rev. 8, 2013 (2013). https://doi.org/10.1002/14651858.CD006575.pub3
Izard, S.G., Juanes, J.A., García Peñalvo, F.J., Estella, J.M.G., Ledesma, M.J.S., Ruisoto, P.: Virtual reality as an educational and training tool for medicine. J. Med. Syst. 42(3), 50 (2018). https://doi.org/10.1007/s10916-018-0900-2
Siu, K.C., Best, B.J., Kim, J.W., Oleynikov, D., Ritter, F.E.: Adaptive virtual reality training to optimize military medical skills acquisition and retention. Mil. Med. 181(5), 214–220 (2016). https://doi.org/10.7205/MILMED-D-15-00164
Sattar, M.U., Palaniappan, S., Lokman, A., Hassan, A., Shah, N., Riaz, Z.: Effects of virtual reality training on medical students’ learning motivation and competency. Pakistan J. Med. Sci. 35(3), 852–857 (2019). https://doi.org/10.12669/pjms.35.3.44
Mickiewicz, P., Gawęcki, W., Gawłowska, M.B., Talar, M., Węgrzyniak, M., Wierzbicka, M.: The assessment of virtual reality training in antromastoidectomy simulation. Virtual Real. (2021). https://doi.org/10.1007/s10055-021-00516-3
Ho, N., Wong, P.M., Chua, M., Chui, C.K.: Virtual reality training for assembly of hybrid medical devices. Multimed. Tools Appl. 77(23), 30651–30682 (2018). https://doi.org/10.1007/s11042-018-6216-x
Vaughan, N., Dubey, V. N., Wainwright, T. W., Middleton, R. G.: Can virtual-reality simulators assess experience and skill level of orthopaedic surgeons? pp. 105–108 (2015)
Cannito, L., et al.: Health anxiety and attentional bias toward virus-related stimuli during the COVID-19 pandemic. Sci. Rep. 10(1), 1–12 (2020). https://doi.org/10.1038/s41598-020-73599-8
Pila, B., Alcoser, E., Pruna, E., Escobar, I.: Inspection and verification training system of production lines in automated processes, through virtual environments, In: Augmented Reality, Virtual Reality, and Computer Graphics, pp. 603–620 (2021).
Kalkan, Ö.K., Karabulut, Ş, Höke, G.: Effect of virtual reality-based training on complex industrial assembly task performance. Arab. J. Sci. Eng. (2021). https://doi.org/10.1007/s13369-021-06138-w
Le, K.-D., Azhar, S., Lindh, D., Ziobro, D.: VRQUEST: designing and evaluating a virtual reality system for factory training, In: Human-Computer-Interaction -- INTERACT 2021, pp. 300–305 (2021)
Techtrends and March: How AR Works in Education and Training, 56(2), (2012) Accessed: Mar. 10, 2018. [Online]. Available: https://link.springer.com/content/pdf/10.1007%2Fs11528-012-0559-3.pdf.
Cabero Almenara, J., Barroso Osuna, J.: Posibilidades educativas de la Realidad Aumentada. J. New Approaches Educ. Res. 6(1), 44–50 (2016). https://doi.org/10.7821/naer.2016.1.140
Fischer, M.: What is augmented reality? (2015)
Santos, M.E.C., Chen, A., Taketomi, T., Yamamoto, G., Miyazaki, J., Kato, H.: Augmented reality learning experiences: Survey of prototype design and evaluation. IEEE Trans. Learn. Technol. 7(1), 38–56 (2014). https://doi.org/10.1109/TLT.2013.37
O’Connell, K.: Designing for Mixed Reality (2016)
Bautista, L.E.R., Maradei, F.G., Pedraza, G.F.: Augmented reality user interaction to computer assisted orthopedic surgery system, (2018). https://doi.org/10.1145/3293578.3293590
Wang, X., Ong, S.K., Nee, A.Y.C.: A comprehensive survey of augmented reality assembly research. Adv. Manuf. 4(1), 1–22 (2016). https://doi.org/10.1007/s40436-015-0131-4
Wildman, J. L.: Using simulation-based training to enhance, 8(4), 559–573 (2009).
Martín-Gutiérrez, J., Contero, M., Alcañiz, M.: Augmented reality to training spatial skills. Procedia Computer Science 77, 33–39 (2015). https://doi.org/10.1016/j.procs.2015.12.356
Lee, B. K.: Augmented reality in education and training, pp. 13–21 (2012)
Gopher, D.: Skill training in Multimodal virtual environments, 41(SUPPL.1). CRC Press, Tayolor & Francis Group, Boca Raton, Florida (2012)
Gavish, N., et al.: Evaluating virtual reality and augmented reality training for industrial maintenance and assembly tasks. Interact. Learn. Environ. 23(6), 778–798 (2015). https://doi.org/10.1080/10494820.2013.815221
Botden, S. M. B. I., Jakimowicz, Æ. J. J.: What is going on in augmented reality simulation in laparoscopic surgery?, pp. 1693–1700 (2009). https://doi.org/10.1007/s00464-008-0144-1
Chowriappa, A. et al.: Augmented-reality-based skills training for robot-assisted urethrovesical anastomosis : a multi-institutional randomised controlled trial, pp. 336–345 (2015). https://doi.org/10.1111/bju.12704
Mendes, H. C. M., Costa, C. I. A. B., da Silva, N. A., Leite, F. P., Esteves, A., Lopes, D. S.: PIÑATA: Pinpoint insertion of intravenous needles via augmented reality training assistance, Comput. Med. Imaging Graph., 82 (2020). https://doi.org/10.1016/j.compmedimag.2020.101731
Kobayashi, L., Zhang, X.C., Collins, S.A., Karim, N., Merck, D.L.: Exploratory application of augmented reality/mixed reality devices for acute care procedure training. West. J. Emerg. Med. 19(1), 158–164 (2018). https://doi.org/10.5811/westjem.2017.10.35026
Brunzini, A., Papetti, A., Germani, M., Adrario, E.: Mixed reality in medical simulation: A comprehensive design methodology. Proc. Des. Soc. 1(AUGUST), 2107–2116 (2021). https://doi.org/10.1017/pds.2021.472
Sautter, B., Daling, L.: Mixed reality supported learning for industrial on-the-job training. SSRN Electron. J. (2021). https://doi.org/10.2139/ssrn.3864189
Su, X., Cheng, Z., Luo, B.: An auxiliary industrial equipment maintenance system using mixed reality, In: 2021 IEEE 8th International Conference on Industrial Engineering and Applications ICIEA 2021, pp. 503–508 (2021) https://doi.org/10.1109/ICIEA52957.2021.9436694
HFES - Human Factors and Ergonomics Society: What is Human Factors and Ergonomics? (2021). https://www.hfes.org/About-HFES/What-is-Human-Factors-and-Ergonomics#:~:text=Ergonomics (or human factors) is, being and overall system performance. (accessed Feb. 02, 2021).
Baddeley, A.: Working memory: Looking back and looking forward. Nat. Rev. Neurosci. 4(10), 829–839 (2003). https://doi.org/10.1038/nrn1201
Sweller, J., Chandler, P.: Evidence for cognitive load theory evidence for cognitive load theory, January 2015, pp. 37–41 (2009). https://doi.org/10.1207/s1532690xci0804
Sweller, J.: Element interactivity and intrinsic, extraneous, and germane cognitive load. Educ. Psychol. Rev. 22(2), 123–138 (2010). https://doi.org/10.1007/s10648-010-9128-5
Mayer, R. E.: Multimedia instruction, In: Handbook of Research on Educational Communications and Technology: Fourth Edition, Springer New York, New York, NY, pp. 385–399 (2014)
Mayer, R. E.: Principles for reducing extraneous processing in multimedia learning : coherence, signaling, redundancy, spatial contiguity, and temporal contiguity principles, In: The Cambridge Handbook of Multimedia Learning, pp. 183–200 (2012)
Paas, F., Tuovinen, J., Tabbers, H., Van Gerven, P.W.M.: Cognitive load measurement as a means to advance cognitive load theory. Educ. Psychol. 1520(38), 43–52 (2010). https://doi.org/10.1207/S15326985EP3801
Sweller, J., Chandler, P., Sweller, J., Chandler, P.: Why some material is difficult to learn why some material is difficult to learn, October 2014, pp. 37–41 (2009). https://doi.org/10.1207/s1532690xci1203
Kalyuga, S., Chandler, P., Sweller, J.: Managing split-attention and redundancy in multimedia instruction. Appl. Cogn. Psychol. 25(SUPPL. 1), 123–144 (1999). https://doi.org/10.1002/acp.1773
Kishishita, N., Kiyokawa, K., Orlosky, J., Mashita, T., Takemura, H., Kruijff, E.: Analysing the effects of a wide field of view augmented reality display on search performance in divided attention tasks, ISMAR 2014 IEEE International Symposium on Mixed and Augmented Reality-Science and Technology 2014, Proceedings, pp. 177–186 (2014). https://doi.org/10.1109/ISMAR.2014.6948425
Schroeder, N.L., Cenkci, A.T.: Spatial contiguity and spatial split-attention effects in multimedia learning environments: a meta-analysis. Educ. Psychol. Rev. 30(3), 679–701 (2018). https://doi.org/10.1007/s10648-018-9435-9
Pouw, W., Rop, G., de Koning, B., Paas, F.: The cognitive basis for the split-attention effect. J. Exp. Psychol. Gen. (2019). https://doi.org/10.1037/xge0000578
van Gog, T.: The signaling (or cueing) principle in multimedia learning, In: The Cambridge Handbook of Multimedia Learning, Second Edition, pp. 263–278 (2014).
Grogorick, S., Albuquerque, G., Tauscher, J.P., Magnor, M.: Comparison of unobtrusive visual guidance methods in an immersive dome environment. ACM Trans. Appl. Percept. (2018). https://doi.org/10.1145/3238303
Moon, J., Ryu, J.: The effects of social and cognitive cues on learning comprehension, eye-gaze pattern, and cognitive load in video instruction. J. Comput. High. Educ. 33(1), 39–63 (2021). https://doi.org/10.1007/s12528-020-09255-x
Jin, S.H.: Visual design guidelines for improving learning from dynamic and interactive digital text. Comput. Educ. 63, 248–258 (2013). https://doi.org/10.1016/j.compedu.2012.12.010
Bernhard, M., Stavrakis, E., Hecher, M., Wimmer, M.: Gaze-to-object mapping during visual search in 3D virtual environments. ACM Trans. Appl. Percept. 11(3), 1–17 (2014). https://doi.org/10.1145/2644812
Grogorick, S., Stengel, M., Eisemann, E., Magnor, M.: Subtle gaze guidance for immersive environments. Proc. - SAP 2017, ACM Symp. Appl. Percept. (2017). https://doi.org/10.1145/31198813119890
Jamet, E.: An eye-tracking study of cueing effects in multimedia learning. Comput. Human Behav. 32, 47–53 (2014). https://doi.org/10.1016/j.chb.2013.11.013
Jarodzka, H., Janssen, N., Kirschner, P.A., Erkens, G.: Avoiding split attention in computer-based testing: Is neglecting additional information facilitative? Br. J. Educ. Technol. 46(4), 803–817 (2015). https://doi.org/10.1111/bjet.12174
El-Nasr, M.S., Vasilakos, A., Rao, C., Zupko, J.: Dynamic intelligent lighting for directing visual attention in interactive 3-D scenes. IEEE Trans. Comput. Intell. AI Games 1(2), 145–153 (2009). https://doi.org/10.1109/TCIAIG.2009.2024532
Kitchenham, B., et al.: Systematic literature reviews in software engineering – A tertiary study. Inf. Softw. Technol. 52(8), 792–805 (2010). https://doi.org/10.1016/j.infsof.2010.03.006
Petersen, K., Vakkalanka, S., Kuzniarz, L.: Guidelines for conducting systematic mapping studies in software engineering: An update. Inf. Softw. Technol. 64, 1–18 (2015). https://doi.org/10.1016/j.infsof.2015.03.007
Kotranza, A., Lind, D.S., Lok, B.: Real-time evaluation and visualization of learner performance in a mixed-reality environment for clinical breast examination. IEEE Trans. Vis. Comput. Graph. 18(7), 1101–1114 (2012). https://doi.org/10.1109/TVCG.2011.132
Wachsmuth, S., Wrede, S., Hanheide, M.: Coordinating interactive vision behaviors for cognitive assistance. Comput. Vis. Image Underst. 108(1–2), 135–149 (2007). https://doi.org/10.1016/j.cviu.2006.10.018
De Crescenzio, F., Fantini, M., Persiani, F., Di Stefano, L., Azzari, P., Salti, S.: Augmented reality for aircraft maintenance training and operations support. IEEE Comput. Graph. Appl. 31(1), 96–101 (2011). https://doi.org/10.1109/MCG.2011.4
Sirakaya, M., Cakmak, E.K.: Effects of augmented reality on student achievement and self-efficacy in vocational education and training. Int. J. Res. Vocat. Educ. Train. 5(1), 1–18 (2018). https://doi.org/10.13152/IJRVET.5.1.1
Bodemer, D., Ploetzner, R., Feuerlein, I., Spada, H.: The active integration of information during learning with dynamic and interactive visualisations. Learn. Instr. 14(3), 325–341 (2004). https://doi.org/10.1016/j.learninstruc.2004.06.006
Herrema, J. S.: Evaluation of graphical user interfaces for augmented reality based manual assembly support (2013)
Rao, A.K., Pramod, B.S., Chandra, S., Dutt, V.: Influence of Indirect Vision and Virtual Reality Training Under Varying Manned/Unmanned Interfaces in a Complex Search-and-Shoot Simulation, vol. 780. Springer International Publishing, Berlin (2019)
Choi, S.H., Kim, M., Lee, J.Y.: Situation-dependent remote AR collaborations: Image-based collaboration using a 3D perspective map and live video-based collaboration with a synchronized VR mode. Comput. Ind. 101(May), 51–66 (2018). https://doi.org/10.1016/j.compind.2018.06.006
Kim, Y., Hong, S., Kim, G.J.: Augmented reality-based remote coaching for fast-paced physical task. Virtual Real. 22(1), 25–36 (2018). https://doi.org/10.1007/s10055-017-0315-2
Fujimoto, Y., Yamamoto, G., Miyazaki, J., Kato, H.: Relation between location of information displayed by Augmented Reality and user’s memorization, In: ACM International Conference Proceeding Series. pp. 0–7 (2012). https://doi.org/10.1145/2160125.2160132
Hořejší, P.: Augmented reality system for virtual training of parts assembly. Procedia Eng. 100(January), 699–706 (2015). https://doi.org/10.1016/j.proeng.2015.01.422
Polvi, J., et al.: Handheld guides in inspection tasks: augmented reality versus picture. IEEE Trans. Vis. Comput. Graph. 24(7), 2118–2128 (2018). https://doi.org/10.1109/TVCG.2017.2709746
Kruijff, E., Orlosky, J., Kishishita, N., Trepkowski, C., Kiyokawa, K.: The influence of label design on search performance and noticeability in wide field of view augmented reality displays. IEEE Trans. Vis. Comput. Graph. 25(9), 2821–2837 (2019). https://doi.org/10.1109/TVCG.2018.2854737
Florax, M., Ploetzner, R.: What contributes to the split-attention effect? The role of text segmentation, picture labelling, and spatial proximity. Learn. Instr. 20(3), 216–224 (2010). https://doi.org/10.1016/j.learninstruc.2009.02.021
Park, K.B., Kim, M., Choi, S.H., Lee, J.Y.: Deep learning-based smart task assistance in wearable augmented reality. Robot. Comput. Integr. Manuf. 63(December 2018), 101887 (2020). https://doi.org/10.1016/j.rcim.2019.101887
Polys, N.F., Bowman, D.A., North, C.: The role of Depth and Gestalt cues in information-rich virtual environments. Int. J. Hum. Comput. Stud. 69(1–2), 30–51 (2011). https://doi.org/10.1016/j.ijhcs.2010.05.007
Seufert, T., Bru, R.: Cognitive load and the format of instructional aids for coherence formation, 331, pp. 321–331 (2006). https://doi.org/10.1002/acp.1248
Radkowski, R.: Investigation of visual features for augmented reality assembly assistance. In: Shumaker, R., Lackey, S. (eds.) HCI International 2015, vol. 9179, pp. 488–498. Springer International Publishing, Cham (2015)
Wang, C.H., Tsai, N.H., Lu, J.M., Wang, M.J.J.: Usability evaluation of an instructional application based on Google Glass for mobile phone disassembly tasks. Appl. Ergon. 77(January), 58–69 (2019). https://doi.org/10.1016/j.apergo.2019.01.007
Jerome, C.J., Witmer, B., Mouloua, M.: Attention orienting in augmented reality environments: effects of multimodal cues. Proc. Hum. Factors Ergon. Soc. Annu. Meet. 50(17), 2114–2118 (2006). https://doi.org/10.1177/154193120605001785
Madsen, J.B., Tatzqern, M., Madsen, C.B., Schmalstieg, D., Kalkofen, D.: Temporal coherence strategies for augmented reality labeling. IEEE Trans. Vis. Comput. Graph. 22(4), 1415–1423 (2016). https://doi.org/10.1109/TVCG.2016.2518318
Radkowski, R., Herrema, J., Oliver, J.: Augmented reality-based manual assembly support with visual features for different degrees of difficulty. Int. J. Hum. Comput. Interact. 31(5), 337–349 (2015). https://doi.org/10.1080/10447318.2014.994194
Lei, X., Tsai, Y.-L., Rau, P.-L. P.: Effect of layout on user performance and subjective evaluation in an augmented-reality environment, In: Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), vol. 11576 LNCS, pp. 376–385 (2019).
Rijnsburger, W., Kratz, S.: Personalized presentation annotations using optical HMDs. Multimed. Tools Appl. 76(4), 5607–5629 (2016). https://doi.org/10.1007/s11042-016-4064-0
Azimi et al.: Can mixed-reality improve the training of medical procedures?, In: Proceedings under Annual International Conference of the IEEE Engineering in Medicine and Biology Society EMBS, vol. 2018-July, pp. 4065–4068 (2018). https://doi.org/10.1109/EMBC.2018.8513387
Holsanova, J., Holmberg, N., Holmqvist, K.: Reading information graphics: The role of spatial contiguity and dual attentional guidance. Appl. Cogn. Psychol. 23(9), 1215–1226 (2009). https://doi.org/10.1002/acp.1525
Makransky, G., Terkildsen, T.S., Mayer, R.E.: Role of subjective and objective measures of cognitive processing during learning in explaining the spatial contiguity effect. Learn. Instr. 61(December 2018), 23–34 (2019). https://doi.org/10.1016/j.learninstruc.2018.12.001
Eiberger, A., Kristensson, P. O., Mayr, S., Kranz, M., Grubert, J.: Effects of depth layer switching between an optical see-through head-mounted display and a body-proximate display, In: Symposium on Spatial User Interaction, pp. 1–9 (Oct. 2019). https://doi.org/10.1145/3357251.3357588
Renner, P., Pfeiffer, T.: Attention guiding techniques using peripheral vision and eye tracking for feedback in augmented-reality-based assistance systems, In: 2017 IEEE Symposium on 3D User Interfaces (3DUI)-Proceedings, pp. 186–194 (2017). https://doi.org/10.1109/3DUI.2017.7893338
Gabbard, J.L., Mehra, D.G., Swan, J.E.: Effects of AR display context switching and focal distance switching on human performance. IEEE Trans. Vis. Comput. Graph. 25(6), 2228–2241 (2019). https://doi.org/10.1109/TVCG.2018.2832633
Bork, F., Schnelzer, C., Eck, U., Navab, N.: Towards efficient visual guidance in limited field-of-view head-mounted displays. IEEE Trans. Vis. Comput. Graph. 24(11), 2983–2992 (2018). https://doi.org/10.1109/TVCG.2018.2868584
Arbeláez, J.C., Viganò, R., Osorio-Gómez, G.: Haptic Augmented Reality (HapticAR) for assembly guidance. Int. J. Interact. Des. Manuf. 13(2), 673–687 (2019). https://doi.org/10.1007/s12008-019-00532-3
Liarokapis, F., White, M., Lister, P.: Augmented reality interface toolkit, In: Proceedings. Eighth International Conference on Information Visualisation, 2004. IV 2004., vol. 8, pp. 761–767 (2004). https://doi.org/10.1109/IV.2004.1320227
Cooper, N., Millela, F., Cant, I., White, M.D., Meyer, G.: Transfer of training - Virtual reality training with augmented multisensory cues improves user experience during training and task performance in the real world. PLoS One 16(3), 1–22 (2021). https://doi.org/10.1371/journal.pone.0248225
Webel, S., Bockholt, U., Engelke, T., Gavish, N., Olbrich, M., Preusche, C.: An augmented reality training platform for assembly and maintenance skills. Rob. Auton. Syst. 61(4), 398–403 (2013). https://doi.org/10.1016/j.robot.2012.09.013
Baudisch, P., Rosenholtz, R.: Halo: a technique for visualizing off-screen objects, In: Proceedings of the Conference on Human Factors in Computing Systems - CHI ’03, (5), p. 481 (2003). https://doi.org/10.1145/642611.642695
Biocca, F., Tang, A., Owen, C., Fan, X.: The omnidirectional attention funnel: a dynamic 3D cursor for mobile augmented reality systems, In: Proceedings of the 39th Annual Hawaii International Conference on System Sciences (HICSS’06), 1(C), 22c-22c (2006). https://doi.org/10.1109/HICSS.2006.476
Ozcelik, E., Karakus, T., Kursun, E., Cagiltay, K.: An eye-tracking study of how color coding affects multimedia learning. Comput. Educ. 53(2), 445–453 (2009). https://doi.org/10.1016/j.compedu.2009.03.002
Fecteau, J.H., Korjoukov, I., Roelfsema, P.R.: Location and color biases have different influences on selective attention. Vision Res. 49(9), 996–1005 (2009). https://doi.org/10.1016/j.visres.2009.03.013
Cierniak, G., Scheiter, K., Gerjets, P.: Explaining the split-attention effect: Is the reduction of extraneous cognitive load accompanied by an increase in germane cognitive load? Comput. Human Behav. 25(2), 315–324 (2009). https://doi.org/10.1016/j.chb.2008.12.020
Schmidt-weigand, F., Kohnert, A., Glowalla, U.: Using route and survey information to generate cognitive maps. Appl. Cogn. Psychol. 23(2), 223–235 (2009). https://doi.org/10.1002/acp
Kim, S., Rehder, B.: How prior knowledge affects selective attention during category learning: An eyetracking study. Mem. Cognit. 39(4), 649–665 (2011). https://doi.org/10.3758/s13421-010-0050-3
Volmer, B., et al.: A comparison of predictive spatial augmented reality cues for procedural tasks. IEEE Trans. Vis. Comput. Graph. 24(11), 2846–2856 (2018). https://doi.org/10.1109/TVCG.2018.2868587
Condino, S., Carbone, M., Piazza, R., Ferrari, M., Ferrari, V.: Perceptual limits of optical see-through visors for augmented reality guidance of manual tasks. IEEE Trans. Biomed. Eng. 67(2), 411–419 (2019). https://doi.org/10.1109/TBME.2019.2914517
Cammeraat, S., Rop, G., de Koning, B.B.: The influence of spatial distance and signaling on the split-attention effect. Comput. Human Behav (2020). https://doi.org/10.1016/j.chb.2019.106203
Bang, S., Lee, H., Woo, W.: Effects of Augmented Content’s Placement and Size on User’s Search Experience in Extended Displays, pp. 184–188 (2020). https://doi.org/10.1109/ismar-adjunct51615.2020.00056
Markov-Vetter, D., Luboschik, M., Islam, A.T., Gauger, P., Staadt, O.: The effect of spatial reference on visual attention and workload during viewpoint guidance in augmented reality. Proc. - SUI 2020 ACM Symp. Spat. User Interact (2020). https://doi.org/10.1145/33859593418449
Lee, H., Jung, J., Lee, H.K., Yang, H.S.: Discipline vs guidance: comparison of visual engagement approaches in immersive virtual environments. Multimed. Tools Appl. 80(20), 31239–31261 (2021). https://doi.org/10.1007/s11042-020-10267-z
Vogt, A., Albus, P., Seufert, T.: Learning in virtual reality: bridging the motivation gap by adding annotations. Front. Psychol. (2021). https://doi.org/10.3389/fpsyg.2021.645032
Harada, Y., Ohyama, J.: Quantitative evaluation of visual guidance effects for 360-degree directions. Virtual Real. 26(2), 759–770 (2022). https://doi.org/10.1007/s10055-021-00574-7
Thees, M., et al.: Augmented reality for presenting real-time data during students’ laboratory work: comparing a head-mounted display with a separate display. Front. Psychol. 13(March), 1–16 (2022). https://doi.org/10.3389/fpsyg.2022.804742
Krüger, J.M., Bodemer, D.: Application and investigation of multimedia design principles in augmented reality learning environments. Inf. (2022). https://doi.org/10.3390/info13020074
Bodemer, D., Faust, U.: External and mental referencing of multiple representations. Comput. Human Behav. 22(1), 27–42 (2006). https://doi.org/10.1016/j.chb.2005.01.005
Kim, K., Carlis, J.V., Keefe, D.F.: Comparison techniques utilized in spatial 3D and 4D data visualizations: A survey and future directions. Comput. Graph. 67, 138–147 (2017). https://doi.org/10.1016/J.CAG.2017.05.005
Gavish, N., Shelef, M.: Evaluating two modes of observational learning in cognitive-spatial task training, pp. 222–231 (2013)
Bell, B., Feiner, S., Höllerer, T.: View management for virtual and augmented reality, In: Proceedings of the 14th Annual ACM Symposium on User Interface Software and Technology, vol. 3, no. 2, pp. 101–110 (2001) [Online]. Available: http://portal.acm.org/citation.cfm?id=502363.
Tatzgern, M., Kalkofen, D., Grasset, R., Schmalstieg, D.: Hedgehog labeling: View management techniques for external labels in 3D space, Proc. - IEEE Virtual Real., pp. 27–32 (2014). https://doi.org/10.1109/VR.2014.6802046
Biocca, F., Tang, A., Owen, C., Xiao, F.: Attention funnel: Omnidirectional 3D cursor for mobile augmented reality platforms. Conf Hum. Factors Comput. Syst. - Proc. 2, 1115–1122 (2006)
Biocca, F., Owen, C., Tang, A., Bohil, C.: Attention issues in spatial information systems: Directing mobile users’ visual attention using augmented reality. J. Manag. Inf. Syst. 23(4), 163–184 (2007). https://doi.org/10.2753/MIS0742-1222230408
Hanley, G.L.: e-Learning and the Science of Instruction, vol. 18, no. 1. John Wiley & Sons, Inc., Hoboken, NJ, USA (2016)
Gopher, D., Krupenia, S., Gavish, N.: Skill training in multimodal virtual environments, In: Advances in Cognitive Ergonomics, pp. 883–892 (2016)
Zhang, H. et al.: Augmented Reality in Education A New Technology for Teaching and Learning. (2020)
Agati, S. S., Bauer, R. D., Hounsell, M. D. S., Paterno, A. S.: Augmented reality for manual assembly in industry 4.0: gathering guidelines, Proceedings of the 2020 22nd Symposium Virtual Augment. Reality, SVR 2020, no. November, pp. 179–188 (2020). https://doi.org/10.1109/SVR51698.2020.00039
Mark, G., Gudith, D., Klocke, U.: The cost of interrupted work: More speed and stress, In: Conference on Human Factors in Computing Systems - Proceedings, pp. 107–110 (2008). https://doi.org/10.1145/1357054.1357072
García Ogueta, M.: Atención y sus cambios en el espacio visual. Cognitiva 3(2), 205–236 (1991)
Ping, J., Liu, Y., Weng, D.: Comparison in depth perception between virtual reality and augmented reality systems, In: 26th IEEE Conference on Virtual Reality and 3D User Interfaces, VR 2019 - Proceedings, pp. 1124–1125 (2019). https://doi.org/10.1109/VR.2019.8798174
Diaz, C., Walker, M., Szafir, D. A., Szafir, D.: Designing for depth perceptions in augmented reality, In: Proceedings of 2017 IEEE International Symposium on Mixed and Augmented, ISMAR 2017, pp. 111–122 (2017). https://doi.org/10.1109/ISMAR.2017.28
Mayer, R.E.: Using multimedia for e-learning. J. Comput. Assist. Learn. 33(5), 403–423 (2017). https://doi.org/10.1111/jcal.12197
Ginns, P.: Meta-analysis of the modality effect. Learn. Instr. 15(4), 313–331 (2005). https://doi.org/10.1016/j.learninstruc.2005.07.001
Schüler, A.: The integration of information in a digital, multi-modal learning environment. Learn. Instr. 59(July), 76–87 (2019). https://doi.org/10.1016/j.learninstruc.2017.12.005
Beege, M., Wirzberger, M., Nebel, S., Schneider, S., Schmidt, N., Rey, G.D.: Spatial continuity effect vs. spatial contiguity failure. revising the effects of spatial proximity between related and unrelated representations. Front. Educ. 4(August), 1–20 (2019). https://doi.org/10.3389/feduc.2019.00086
Lacoche, J., Duval, T., Arnaldi, B., Maisel, E., Royan, J.: 3DPlasticToolkit: Plasticity for 3D user interfaces, In: Virtual Reality and Augmented Reality. Lecture Notes in Computer Science., 2019, pp. 62–83.
Livingston, M.a.: Human Factors in Augmented Reality Environments (2013).
Ren, D., Goldschwendt, T., Chang, Y., Hollerer, T.: Evaluating wide-field-of-view augmented reality with mixed reality simulation, Proc. - IEEE Virtual Real., vol. 2016-July, pp. 93–102 (2016). https://doi.org/10.1109/VR.2016.7504692
ISO: Ergonomics of human-system interaction — Part 12. ISO 9241–12:2018. Internacional Organization for Standardization, 2018, [Online]. Available: https://www.iso.org/obp/ui/#iso:std:iso:9241:-11:ed-2:v1:en.
Funding
Open Access funding provided by Colombia Consortium.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Bautista, L., Maradei, F. & Pedraza, G. Strategies to reduce visual attention changes while learning and training in extended reality environments. Int J Interact Des Manuf 17, 17–43 (2023). https://doi.org/10.1007/s12008-022-01092-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12008-022-01092-9