Abstract
Background
In minimally invasive surgery (MIS), trainees need to learn how to interpret the operative field displayed on the laparoscopic screen. Experts currently guide trainees mainly verbally during laparoscopic procedures. A newly developed telestration system with augmented reality (iSurgeon) allows the instructor to display hand gestures in real-time on the laparoscopic screen in augmented reality to provide visual expert guidance (telestration). This study analysed the effect of telestration guided instructions on gaze behaviour during MIS training.
Methods
In a randomized-controlled crossover study, 40 MIS naive medical students performed 8 laparoscopic tasks with telestration or with verbal instructions only. Pupil Core eye-tracking glasses were used to capture the instructor’s and trainees’ gazes. Gaze behaviour measures for tasks 1–7 were gaze latency, gaze convergence and collaborative gaze convergence. Performance measures included the number of errors in tasks 1–7 and trainee’s ratings in structured and standardized performance scores in task 8 (ex vivo porcine laparoscopic cholecystectomy).
Results
There was a significant improvement 1–7 on gaze latency [F(1,39) = 762.5, p < 0.01, ηp2 = 0.95], gaze convergence [F(1,39) = 482.8, p < 0.01, ηp2 = 0.93] and collaborative gaze convergence [F(1,39) = 408.4, p < 0.01, ηp2 = 0.91] upon instruction with iSurgeon. The number of errors was significantly lower in tasks 1–7 (0.18 ± 0.56 vs. 1.94 ± 1.80, p < 0.01) and the score ratings for laparoscopic cholecystectomy were significantly higher with telestration (global OSATS: 29 ± 2.5 vs. 25 ± 5.5, p < 0.01; task-specific OSATS: 60 ± 3 vs. 50 ± 6, p < 0.01).
Conclusions
Telestration with augmented reality successfully improved surgical performance. The trainee’s gaze behaviour was improved by reducing the time from instruction to fixation on targets and leading to a higher convergence of the instructor’s and the trainee’s gazes. Also, the convergence of trainee’s gaze and target areas increased with telestration. This confirms augmented reality-based telestration works by means of gaze guidance in MIS and could be used to improve training outcomes.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Minimally invasive surgery (MIS) has become the gold standard in many procedures in intraabdominal surgery throughout the last thirty years [1,2,3]. MIS brings faster postoperative recovery, reduction in postoperative pain, lower rate of surgical site infections and shorter hospital stays [4]. However, MIS also harbours some technical challenges which may result in longer operative times, mostly due to prolonged learning curves [5]. The indirect camera view, the lack of haptic feedback and difficult instrument handling due to the fulcrum and pivoting effects are possible explanations for this phenomenon [6, 7].
To compensate for this prolonged learning curve structured and extensive training of minimally invasive skills is essential. Various box trainers, Virtual or Augmented Reality training systems can significantly decrease the time needed to learn not only basic tasks such as suturing and knot tying but also surgical procedures such as cholecystectomies [8,9,10]. However, clear and understandable instruction at the training site is still needed to maximize the learning experience of the trainee [11].
To improve the communication between the instructor and the trainee, some innovative programmes and add-ons have been developed lately [12]. Augmented Reality (AR) systems blend artificial images on the MIS screen of the real situs, helping the trainer to point out important structures using arrows, points and sketching over the MIS image [12,13,14]. A newly developed AR tool called iSurgeon allows using the instructor’s own hand to project gestures in real-time onto the laparoscopic screen and provide visual guidance. Pointing at the target structures or demonstration of the correct execution of the movement becomes effortless and very concise [15,16,17]. Previous studies have shown that using the system for instruction can lead to time savings and fewer mistakes while performing basic tasks [16, 18, 19].
These results suggest that the changed gaze behaviour of the trainee might play a crucial role in learning and conducting new tasks and it is boosted by AR telestration [12, 18, 19]. An increasing number of studies observing the gaze behaviour of surgeons in laparoscopic surgery and training have shown significant differences in the gaze behaviour of experienced surgeons and surgical novices. [20,21,22]. AR telestration was able to improve the fixation concentration, gaze latency, gaze convergence and collaborative gaze convergence [12, 18, 19]. Therefore, the aim of the present study was to analyze the effects of AR telestration on the laparoscopic performance of laparoscopically naïve trainees with regard to eye gaze behaviour.
Materials and methods
Study design
The study was designed as a randomized-controlled, crossover trial with 40 laparoscopically naive medical students recruited through the medical faculty of the University of Heidelberg as part of a clinical elective course. The participants were randomized into two groups (AR telestration and no AR telestration) with an allocation ratio of 1:1 and switched groups after completing all tasks (Fig. 1). A pragmatic sample size was chosen based on the results of previous studies [16].
The participation was on a voluntary basis and all participants signed an informed consent form and filled in the demographic data and surgical experience. The study was performed in the Training Center for Minimally Invasive Surgery at the Department for General, Visceral, and Transplant Surgery at Heidelberg University Hospital, Germany between September 2020 and February 2021. The study was approved by the local ethics committee at Heidelberg University (S-436/2018).
Laparoscopic tasks
All participants underwent a basic laparoscopic consisting of 6 tasks from the basic module (Task 3–8) on a Virtual Reality (VR) Trainer (LAP Mentor III, 3D Systems, Rock Hill, USA) and two PEG transfers and threading rubber bands through multiple eyelets in a box trainer. Following that, eight different laparoscopic tasks were performed: (1) PEG-Transfer (2) circle marking (3) needle parkour (4) Grabbing and transferring silicone loops (5) Unravelling small intestine convolute (6) blood vessel ligation (7) felt cloth exposition (8) cholecystectomy in a cadaveric porcine liver (Supplementary Table 1; Supplementary Fig. 1). The tasks were selected based on previous studies to investigate the effects of AR instruction in diverse environments [16, 23]. Tasks 1–7 were considered basic tasks and were performed twice. Task 8 was considered an advanced task and was performed only once. All tasks were performed in a Szabo–Berci–Sackier box trainer using Karl Storz laparoscopy instruments and a standard laparoscopy tower (KARL STORZ GmbH & Co. KG, Tuttlingen, Germany).
Instruction mode
The verbal instructions were standardised for all participants and were delivered in a simple and standardized way. For the instruction with AR, the iSurgeon system [15,16,17] was used in addition to verbal instructions. Apart from that additional usage of the iSurgeon to telestrate hand gestures, there were no differences in the training between both instruction modalities. The iSurgeon system used an RGB-D camera (colour resolution: 1920 × 1080 pixels, 30 fps, depth resolution: 512 × 424 pixels, 30 fps) to detect the hands of the instructor and project them over the image on the laparoscopic screen. To operate it the instructor moved his hand beneath the sensor. Instructor and participant stood next to each other with enough space to not interfere with their movements. The instructions were provided by the same instructor throughout the study. Before the study, the instructor was introduced to the use of the iSurgeon and practiced it as well as the standardized verbal instructions until proficiency. The functional principle of the iSurgeon and the meaning of the verbal instructions were explained to the participants before the tasks were carried out.
Eye tracking system
Pupil Core eye-tracking glasses (Pupil labs GmbH, Berlin, Germany) were used for the detection of both instructor’s and trainee’s eye movement (Fig. 2). The gazes were recorded using Pupil Capture software Version 1.23–4 (Pupil labs GmbH, Berlin, Germany). Further, the world cameras detected the environment in front of the instructor and the participant—in particular, the screen with the laparoscopic image. Overall, the gaze position, fixations, audio and the laparoscopic screen as a surface were recorded in a synchronized manner. The collected data were annotated and analysed with the Pupil Player software Version 1.23–4 (Pupil labs GmbH, Berlin, Germany) [24].
Gaze metrics and performance assessment
The following gaze parameters were collected:
-
(i)
Gaze latency. i.e. the time from instruction until the first fixation of the trainee’s gaze on a target area [19]
-
(ii)
Gaze convergence, i.e. the absement, namely the integral of the gaze displacement, between the instruction and the first fixation of the trainee on the target area [18]
-
(iii)
Collaborative gaze convergence, i.e. the absement of the instructor’s and trainee’s gaze over the time needed to merge over a target area [19]
For tasks 1–7, the performance was measured by the number of errors and time needed to complete the tasks. For the cholecystectomy, the global and task-specific Objective Structured Assessments of Technical Skills (OSATS) and the time needed to complete the tasks were used to measure the outcomes [25, 26]. The subjective workload of the participant was measured by a modified NASA Task Load Index (NASA-TLX) [27]. In addition, the blink rate was analysed as another marker for cognitive workload [28].
Statistical analysis
Python 3 (Version 3.9.7, Python Software Foundation, Delaware, USA) with Pandas (Version 1.3.4, [29]) was used for data manipulation and Pingouin (Version 0.5.2, [30]) for statistical analysis. Within-group comparison was done with the Wilcoxon rank-sum test (tasks 1–7) and between-group comparison with the Mann–Whitney U test (task 8). These data were reported as medians and interquartile ratios. The effects of modality and task type to each metric were assessed with two-way repeated measures ANOVA and reported as F-values, p-values and ηp2-values.
Results
Population characteristics
A total of 42 participants were screened. There were two dropouts due to personal reasons. The total number included in the study was 40. The study group characteristics can be found in Table 1. There was no relevant difference found in the population characteristics between both groups.
Eye gaze outcomes
Significant differences in all eye gaze parameters were observed in tasks 1–7 between both groups (iSurgeon vs. verbal). Gaze latency was significantly lower with AR, reaching approximately 0.21 ± 0.19 s for iSurgeon and 2.04 ± 1.51 s for the verbal only group (F(1,39) = 762.5, p < 0.01, ηp2 = 0.95). Gaze convergence in iSurgeon group was also significantly lower at 0.02 ± 0.04 pixels*sec compared to 0.55 ± 0.49 pixels*sec in the verbal only group (F(1,39) = 482.8, p < 0.01, ηp2 = 0.93). Collaborative gaze convergence also improved upon instruction with AR, reaching 0.05 ± 0.06 pixels*sec in the iSurgeon group and 0.56 ± 0.44 pixels*sec in the verbal only group (F(1,39) = 408.4, p < 0.01, ηp2 = 0.91). (Fig. 3).
The heatmaps of gaze fixations indicated that trainees focused on the target structures more precisely and had a higher overlap with the instructor’s gaze upon instruction with iSurgeon (Fig. 4).
Performance and workload outcomes
Using the AR system for instruction resulted in a lower number of errors (tasks 1–7; 0.18 ± 0.56 vs. 1.94 ± 1.80, F(1,39) = 433.5, p < 0.01, ηp2 = 0.92) and faster completion time (tasks 1–7; 118 ± 73 vs. 148 ± 81.5 s, F(1,39) = 97.7, p < 0.01, ηp2 = 0.71) compared to verbal instruction only (Fig. 5). No significant difference in task duration was shown for task 8 (4781 ± 1257 vs. 5.024 ± 1447 s, F(1) < 0.01, p = 0.98, ηp2 < 0.01). Also, the score ratings for laparoscopic cholecystectomy (task 8) were significantly higher in the iSurgeon group, resulting in higher mean global OSATS of 29 ± 2.5 vs. 25 ± 5.5, (p < 0.01) and mean task-specific OSATS of 60 ± 3 vs. 50 ± 6 (p < 0.01, Fig. 5).
Instruction with the AR system resulted in significantly lower cognitive workload in the NASA-TLX in tasks 1–7 (50 ± 21 vs. 56 ± 22, p < 0.01), whereas it was not significantly lower during laparoscopic cholecystectomy (28 ± 14 vs. 37 ± 15, p = 0.12) (Fig. 6). Throughout the basic tasks (tasks 1–7) the blink rate was lower upon instruction with iSurgeon (1.8 ± 5.2 vs. 3.4 ± 6.2, F(1,39) = 39.62, p < 0.01, ηp2 = 0.50) while no significant differences were found in task 8 (4.7 ± 9.1 vs. 5.0 ± 7.9, F(1) = 0.05, p = 0.83, ηp2 < 0.01, Fig. 6).
Discussion
In the present study, telestration with augmented reality using the iSurgeon system successfully improved not only the trainee’s gaze behaviour but also resulted in improved laparoscopic performance. Additionally, the cognitive workload was reduced in most of the tasks.
The Trainee’s gaze behaviour was improved by reducing the gaze latency and improving the gaze convergence and the collaborative gaze convergence in the surgical tasks in the present study using the augmented reality telestration system as compared to only using verbal guidance in laparoscopic surgical training. The lower gaze latency with the iSurgeon indicates a lower time from instruction to fixation on targets [18]. This could have helped the trainee to identify the targets faster. The faster target identification may have led to the higher gaze convergence, which indicates a higher convergence of the participants’ gaze and the target areas [18]. The higher collaborative gaze convergence indicates that the gaze of trainer and trainee converged more during the task [19]. One important purpose of telestration systems like the iSurgeon is to help a novice to see what an expert sees [15,16,17]. The higher collaborative gaze convergence indicates that this purpose of the iSurgeon was achieved in the study. In the last years, an increasing number of studies observing the gaze behaviour of surgeons in MIS and surgical training have been published [12, 18,19,20,21,22]. Several studies showed that more experienced surgeons have a higher fixation rate (median (IQR) 1.86 (0.3) vs 0.96 (0.3); p = 0.006) and a higher dwell time (median (IQR) 792 (159) vs 469 (109) s; p = 0.028) than novices while performing more complex laparoscopic procedures, e.g. inguinal hernia repair [20]. Further, it has been shown that experts look predominantly at target structures (r = 0.655, p < 0.05) and less at non-essential structures than novices (r = − 0.619, p < 0.05) [21] and disengage their gaze faster from the previous subtask (mean(SD) expert surgeons: − 1(93) ms, intermediate surgeons: − 189(160) ms, novices: − 296(179) ms) while performing cholecystectomies [22]. However, novices’ gazes became more focused and less scattered upon instruction with AR pointers [12]. Further, the time between instruction and the gaze on a target (gaze latency) decreased by 48% and the convergence of gaze and target over time (gaze convergence) improved by 33% while being instructed with AR compared to verbal instruction [18, 19]. But only gaze latency, gaze convergence and collaborative gaze convergence seemed to correlate with a better performance during training, especially with lower error rates and less time needed to complete basic laparoscopic tasks [18, 19]. The results in the present study matched with previous studies analysing gaze latency, gaze convergence and collaborative gaze convergence, as it also showed an improvement with telestration. This suggests that AR-based telestration works by means of guiding the trainees’ gazes and is thus a promising tool to improve gaze behaviour in laparoscopic training.
The surgical performance of the participants in the present study was improved with the iSurgeon through lower error rates, higher performance scores and reduced times in most of the tasks. The error rate was reduced in all the tasks which indicates that AR telestration led to fewer misconceptions and improved the quality of the tasks. Also, the achieved points in the global and task-specific OSATS scores indicate an improved quality as compared to only verbal guidance. The higher global OSATS score indicates an improved general surgical performance during task 8. The higher task-specific OSATS score indicates a better performance of the cholecystectomy for the use of telestration with AR [31, 32]. A reason for the improved quality of the performances in the tasks might have been improved communication through the iSurgeon [16]. Improving communication not only in training but also in the real operating theatre is essential. Because multiple studies have shown that unclear communication and misunderstandings can lead to intraoperative mistakes and cause complications [33, 34]. Therefore, it is essential to try to improve intraoperative communication—one possibility for that could be telestration with AR [16]. The time efficiency was improved in the basic tasks (tasks 1–7) in the present study with AR when comparing them combined but showed no difference in task 8. The improvement of the time efficiency in the basic tasks could have been achieved through faster recognition of the target structures by the participants. A reason why the time was not reduced in the cholecystectomy (task 8) could have been the complexity of the task. Cholecystectomy is a rather complex task for laparoscopically naïve medical students and this could explain the similar task times, albeit better performance in the AR group compared to the control. The clearer instructions with AR at key steps may have led to a better quality of the tasks by preventing mistakes. Analysing the task duration is still important because transferred to the operating theatre longer operation times can increase the risk of complications like wound infections, thromboses and cardiopulmonary failure [35, 36]. The results of this study indicate that instructing trainees in laparoscopic training with AR can improve the quality of their performance and reduce the task time in most of the tasks. Also, the performance in more complex procedures like cholecystectomy was influenced by AR through improving the performance compared to verbal instructions. Hence, AR telestration may improve the quality and efficiency of surgical training. Translation to improved clinical outcomes will have to be shown in future studies.
The cognitive workload was improved with AR compared to verbal instructions by lowering the NASA-TLX and the blink rate in most of the tasks. Both, the score in the NASA-TLX and the blink rate were lower during instruction with AR in tasks 1–7 combined but showed no significant difference in task 8. The lack of workload reduction in task 8 seems to be in concordance with the lacking time difference between both instruction modalities in task 8. The reason for that might also have been that the task was too complex for the inexperienced participants causing the iSurgeon to not have a significant effect on the already very high workload by the task itself. However, the lower score in the NASA-TLX in the basic tasks indicates a reduced subjective workload [27] The lower blink rate in the basic tasks indicates a reduced objectively measured workload in those tasks [28, 37, 38]. Measuring the workload during training and in the operating theatre is an established way to analyse surgical performances [39] [40]. Decreasing the workload of surgeons is especially important in the real operating theatre. It has been shown that an elevated workload can cause higher error rates and worse intraoperative performances of surgeons [41]. In this study, AR telestration was able to reduce both the subjective and objective workload of the participants in most of the tasks. Therefore, AR telestration may be a way to reduce workload during surgical training.
The study has some limitations. The data annotation of the eye gaze data could not be blinded due to technical reasons, i.e. the projected iSurgeon hand of the instructor was visible throughout the analysis process. To limit a possible detection bias caused by the unblinded data annotation annotating with strict and objective standards and through only having one person who annotated all the data were performed. Yet, the blink rate could not be influenced by the annotator and was determined directly from the recorded and exported data. Another limitation of the study is the transferability into the real operating theatre. It is unclear if the results observed in the tasks in the box trainer and the cadaveric cholecystectomy would also be observed during real surgical procedures. However, the gaze behaviour was analysed during multiple and different tasks to mimic as many different aspects of surgical procedures as possible. Additionally, the participants were not surgeons but laparoscopically naïve medical students. This may reduce the transferability of the results on surgeons but has the advantage of having a homogenous study population.
In the present study, telestration with AR led to an improved surgical performance in training. AR improved the gaze behaviour by reducing the time from instruction to target fixation and improving the convergence of the participants’ gazes with target areas as well as with the instructor’s gaze. Additionally, the quality and time efficiency was improved and the workload was reduced in most of the tasks using telestration with AR with the iSurgeon system. These results suggest that AR systems may be effective training tools by adjusting trainee’s gaze behaviour to instructor’s, which may result in better overall performances and trainee’s comfort while performing tasks. Hence, it may be a promising tool to improve surgical training. If these results could transfer to the operating theatre, intraoperative communication and gaze behaviour could be improved, which may lead to fewer misunderstandings, errors, complications and reduce intraoperative time. Future studies are needed to assess implementation possibilities into clinical practice and the effects of AR telestration in different training situations as well as in the operating theatre.
References
Ohtani H et al (2012) Meta-analysis of the results of randomized controlled trials that compared laparoscopic and open surgery for acute appendicitis. J Gastrointest Surg 16(10):1929–1939
Buia A, Stockhausen F, Hanisch E (2015) Laparoscopic surgery: a qualified systematic review. World J Methodol 5(4):238–254
Wehrtmann FS et al (2020) Learning curves of laparoscopic Roux-en-Y gastric bypass and sleeve gastrectomy in bariatric surgery: a systematic review and introduction of a standardization. Obes Surg 30(2):640–656
Mehta HB et al (2018) Outcomes of laparoscopic colectomy in younger and older patients: an analysis of nationwide readmission database. J Laparoendosc Adv Surg Tech A 28(4):370–378
Harrysson IJ et al (2014) Systematic review of learning curves for minimally invasive abdominal surgery: a review of the methodology of data collection, depiction of outcomes, and statistical analysis. Ann Surg 260(1):37–45
Nickel F, Kowalewski KF, Müller-Stich BP (2015) Risk awareness and training for prevention of complications in minimally invasive surgery. Chirurg 86(12):1121–1127
Gallagher AG et al (1998) An ergonomic analysis of the fulcrum effect in the acquisition of endoscopic skills. Endoscopy 30(7):617–620
Nickel F et al (2013) Virtual reality does not meet expectations in a pilot study on multimodal laparoscopic surgery training. World J Surg 37(5):965–973
Nickel F et al (2014) One or two trainees per workplace in a structured multimodality training curriculum for laparoscopic surgery? Study protocol for a randomized controlled trial-DRKS00004675. Trials 15:137
Kowalewski KF et al (2019) Sensor-based machine learning for workflow detection and as key to detect expert level in laparoscopic suturing and knot-tying. Surg Endosc 33:3732–3740
Rivera C, Cao CGL (2005) Toward improved communication in laparoscopic surgery: accounting for multiple frames of reference and mental rotations. Proc Hum Factors Ergon Soc 49(11): 999–1003
Feng Y et al (2019) Virtual pointer for gaze guidance in laparoscopic surgery. Surg Endosc 34:3533–3539
Sparwasser PM et al (2018) Augmented reality and virtual reality in the operating theatre status quo und quo vadis. Aktuelle Urol 49(6):500–508
Jarc AM et al (2017) Proctors exploit three-dimensional ghost tools during clinical-like training scenarios: a preliminary study. World J Urol 35(6):957–965
Nickel F, Cizmic A, Chand M (2021) Telestration and augmented reality in minimally invasive surgery: an invaluable tool in the age of COVID-19 for remote proctoring and telementoring. JAMA Surg 157:169–170
Wild C et al (2022) Telestration with augmented reality for visual presentation of intraoperative target structures in minimally invasive surgery: a randomized controlled study. Surg Endosc 36:7453–7461
Müller LR et al (2022) Robust hand tracking for surgical telestration. Int J Comput Assist Radiol Surg 17:1477–1486
Chetwood AS et al (2012) Collaborative eye tracking: a potential training tool in laparoscopic surgery. Surg Endosc 26(7):2003–2009
Kwok KW et al (2012) Collaborative gaze channelling for improved cooperation during robotic assisted surgery. Ann Biomed Eng 40(10):2156–2167
Tien T et al (2015) Differences in gaze behaviour of expert and junior surgeons performing open inguinal hernia repair. Surg Endosc 29(2):405–413
Evans-Harvey K et al (2020) Comparison of surgeon gaze behaviour against objective skill assessment in laparoscopic cholecystectomy-a prospective cohort study. Int J Surg 82:149–155
Zheng B et al (2021) Action-related eye measures to assess surgical expertise. BJS Open 5(5):zrab068
Nickel F et al (2017) Face validity of the pulsatile organ perfusion trainer for laparoscopic cholecystectomy. Surg Endosc 31(2):714–722
Kassner M, Patera W, Bulling A (2014) Pupil: an open source platform for pervasive eye tracking and mobile gaze-based interaction. In Proceedings of the 2014 ACM International Joint Conference on Pervasive and Ubiquitous Computing: Adjunct Publication. 2014, Association for Computing Machinery: Seattle, Washington. p. 1151–1160
Martin JA et al (1997) Objective structured assessment of technical skill (OSATS) for surgical residents. Br J Surg 84(2):273–278
Nickel F et al (2016) Direct observation versus endoscopic video recording-based rating with the objective structured assessment of technical skills for training of laparoscopic cholecystectomy. Eur Surg Res 57(1–2):1–9
Hart S, Staveland L (1988) Development of NASA-TLX (task load index): results of empirical and theoretical research. Adv Psychol 52:139–183
Richstone L et al (2010) Eye metrics as an objective assessment of surgical skill. Ann Surg 252(1):177–182
McKinney W (2011) pandas: a foundational python library for data analysis and statistics. Python High Perform Sci Comput 14:1–9
Vallat R (2018) Pingouin: statistics in Python. J Open Source Softw 3:1026
Kowalewski KF et al (2018) One or two trainees per workplace for laparoscopic surgery training courses: results from a randomized controlled trial. Surg Endosc 33:1523–1531
Kowalewski KF et al (2018) LapTrain: multi-modality training curriculum for laparoscopic cholecystectomy-results of a randomized controlled trial. Surg Endosc 32(9):3830–3838
Sevdalis N et al (2012) Quantitative analysis of intraoperative communication in open and laparoscopic surgery. Surg Endosc 26(10):2931–2938
Manser T (2009) Teamwork and patient safety in dynamic domains of healthcare: a review of the literature. Acta Anaesthesiol Scand 53(2):143–151
Cheng H et al (2017) Prolonged operative duration increases risk of surgical site infections: a systematic review. Surg Infect (Larchmt) 18(6):722–735
Daley BJ et al (2015) How slow is too slow? Correlation of operative time to complications: an analysis from the Tennessee surgical quality collaborative. J Am Coll Surg 220(4):550–558
Erridge S et al (2018) Comparison of gaze behaviour of trainee and experienced surgeons during laparoscopic gastric bypass. Br J Surg 105(3):287–294
Gao J et al (2019) Subjective and objective quantification of the effect of distraction on physician’s workload and performance during simulated laparoscopic surgery. Med Sci Monit 25:3127–3132
Ruiz-Rabelo JF et al (2015) Validation of the NASA-TLX score in ongoing assessment of mental workload during a laparoscopic learning curve in bariatric surgery. Obes Surg 25(12):2451–2456
Bell SW et al (2022) The national aeronautics and space administration-task load index: NASA-TLX: evaluation of its use in surgery. ANZ J Surg. https://doi.org/10.1111/ans.17830
Yurko YY et al (2010) Higher mental workload is associated with poorer laparoscopic performance as measured by the NASA-TLX tool. Simul Healthc 5(5):267–271
Funding
Open Access funding enabled and organized by Projekt DEAL. The development of the iSurgeon system was supported by Bundesministerium für Wirtschaft und Energie (Federal Ministry of Economics and Energy) BMWi EXIST programme and the European Social Fund (Grant number 03EFJBW167).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Disclosures
Felix Nickel reports receiving travel support and support for training courses by Johnson and Johnson, Medtronic, Intuitive Surgical, Cambridge Medical Robotics, and Karl Storz as well as consultancy fees by Karl Storz. All other authors report no financial ties or conflicts of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
464_2022_9859_MOESM1_ESM.jpg
Supplementary file1 (JPG 141 kb)—Supplementary Fig. 1 Instructor’s hand is displayed on the laparoscopic screen during the laparoscopic task. A PEG transfer, B Marking Circles, C Needle Parkour, D Grabbing and transferring vessel loops E Unravelling silicone small intestine, F Suture ligations of blood vessel in a silicone model, G Picking up felt cloth, H Cholecystectomy in a cadaveric poricine liver model
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Felinska, E.A., Fuchs, T.E., Kogkas, A. et al. Telestration with augmented reality improves surgical performance through gaze guidance. Surg Endosc 37, 3557–3566 (2023). https://doi.org/10.1007/s00464-022-09859-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00464-022-09859-7