Abstract
Stereopsis may be an advantage of robotic surgery. Perceived robotic ergonomic advantages in visualisation include better exposure, three-dimensional vision, surgeon camera control, and line of sight screen location. Other ergonomic factors relating to visualisation include stereo-acuity, vergence–accommodation mismatch, visual–perception mismatch, visual–vestibular mismatch, visuospatial ability, visual fatigue, and visual feedback to compensate for lack of haptic feedback. Visual fatigue symptoms may be related to dry eye or accommodative/binocular vision stress. Digital eye strain can be measured by questionnaires and objective tests. Management options include treatment of dry eye, correction of refractive error, and management of accommodation and vergence anomalies. Experienced robotic surgeons can use visual cues like tissue deformation and surgical tool information as surrogates for haptic feedback.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Visualisation is one of the two facets of sensory ergonomics in surgery. There are important ergonomic factors to consider with regards to visualisation in robotic surgery. Most are advantageous, but some are detrimental [1]. Conventional laparoscopic surgery has several drawbacks that can impair complete visualisation of the operative field, which is presented as a monocular view on a two-dimensional (2D) video display. The surgeon relies on an assistant to control the camera and must contend with the loss of peripheral vision and potential inadvertent tissue injury. The surgeon also loses depth perception and instead relies on visual cues, such as light and shade, relative size, and motion parallax to give a sense of depth [2]. However, indirect depth cues can be degraded in situations of limited light such as operating in the pelvis or in situations of limited contrast such as within dark blood-stained tissue [3]. In addition, the disparity between the location of the display monitor and the actual operative field (where we look versus where we operate) can result in visual and mental fatigue.
Robotic surgery ameliorates these issues by providing better exposure, immersive three-dimensional (3D) vision, surgeon camera control, and line of sight screen location. There are other visualisation ergonomic factors, some of which may be detrimental, to consider when assessing the difference between laparoscopic and robotic systems. This includes stereo-acuity, vergence–accommodation mismatch, visual–perception mismatch, visual–vestibular mismatch, visuospatial ability, visual fatigue, and visual feedback to compensate for lack of haptic feedback. Some of the detrimental visualisation ergonomic factors can be overcome with experience and adaptation by the brain sensory system.
Exposure
Visualisation exposure can be enhanced with robotic surgery because of better illumination, magnification, and better viewing angles. Optimal surgical illumination should offer sufficient ambient light with the ability to focus on the surgeon’s operative field even at difficult angles; without shadow, glare, and artifact [4]. Light emanating from the laparoscope provides more focussed illumination compared with overhead lights during open surgery. Degradation of the light cable used in laparoscopic surgery can reduce the intensity of light. Improved visualisation can occur using the robot camera magnification technology to position the camera further away from a smoke or steam generating dissection site (e.g., during pelvic surgery). Use of magnification together with motion scaling can improve vision and accuracy [5]. The port hopping feature of the da Vinci Xi robot (currently the most recognised and widely used robot-assisted system) allows re-positioning of the camera to facilitate optimal multi-quadrant visualisation during surgery [6].
Three-dimensional vision
3D visualisation provided by the robotic console can reduce ocular strain, task performance time, and error rates [4] Falk et al. found that peak velocity was reduced (p = 0.01) and the deceleration phase of motion was prolonged (P < 0.05) in a study of 15 surgeons performing six tasks with 2D compared with 3D vision [7]. 3D laparoscopy has been shown to shorten operative times and improve precision when compared to 2D laparoscopy [8]. A meta-analysis comparing 3D laparoscopic system with 2D laparoscopy in the treatment of different urological conditions found shorter operative time (p < 0.0001) and lower estimated blood loss (p = 0.001) for radical prostatectomy surgery [9]. However, 3D vision has not been shown to be superior to two-dimensional vision for all laparoscopic surgery [9, 10].
The modern robotic console may provide a better stereopsis experience for surgeons compared to the traditional 3D images created by shuttering or polarizing images [11]. The da Vinci robotic system uses two separate optical trains with two independent three-chip cameras and display systems to present slightly different images to each eye, mimicking binocular vision. The optical system minimises geometric distortion across the field of view, enabling stereo fusion of images even near the edges.
Surgeon camera control
Robotic surgery returns the control of the camera to the operating surgeon on a stable platform, resulting in a steadier view of the operative field with better centering, horizon adjustment, zoom control, and instrument visualisation. In a simulation study of 84 medical students, surgeon performance was significantly influenced by the quality of the laparoscopic camera navigator [12]. The use of robotic laparoscopic holder has been shown to provide consistent control of the view of the operative field [13]. The quality of camera navigation has been shown to be better with robotic versus conventional laparoscopic rectal cancer surgery [14].
Line of sight screen location
The da Vinci console returns the screen location back in line with the hands of the surgeon, as occurs naturally in open surgery. This view of the surgeon’s hands performing a task is intuitive and familiar. Studies looking at the optimal position of image display in endoscopic surgery found that task performance improves when the image display is close to the hands [15]. Hanna et al. found endoscopic knot tying was better (P < 0.01) and faster (p < 0.01) with hand-level “gaze-down” viewing than eye-level viewing in a laboratory study of ten endoscopic surgeons [10]. In addition, the downgaze viewing results in a smaller ocular surface for exposure to the effects of tear film evaporation compared with the horizontal gaze when viewing the screen during laparoscopic surgery. [16]
Stereo-acuity
Stereopsis is a binocular sensation of relative depth caused by horizontal disparity of retinal images from both eyes which are then cortically integrated [17]. Stereo-acuity is a measure of the smallest depth that an individual can detect with binocular disparity [3]. Stereo-acuity is variable among individuals and declines with age [18]. In a study of 66 surgeons, reduced or no stereopsis was found in 2–14% depending on the test used [19]. Worse stereopsis with age might be due to insufficient correction of refractive error [20]. Hence, the visualisation advantage of 3D robotic surgery over the conventional 2D surgery may not be significant for some surgeons.
Stereopsis can be measured globally by random dot stereo tests and locally by contour stereo tests [18]. Images of stereo figures are embedded in a background of random points with random dot stereo tests and testing requires wearing anaglyph red and green glasses. Images are separated and presented separately to each eye with contour stereo tests and testing requires wearing cross-polarizing filtered glasses. Contour stereo tests have the disadvantage of monocular cues. High-quality stereo vision of the surgeon has been shown to offer no advantage when using the 2D display but associated with significantly reduced operating time and error when using the 3D display [18]. Gietzelt et al. found that surgical experience could partly compensate for a poorer stereo-acuity [18].
Vergence–accommodation mismatch
With normal visual activity and 2D laparoscopy, the eyes automatically accommodate and converge to the same distance on a specific object or on the display screen [3]. Vergence and accommodation are neurally coupled, with accommodative changes evoking changes in vergence (accommodative vergence), and vergence changes evoking changes in accommodation (vergence accommodation) [21]. Accommodative vergence can be detected only during monocular fixation, because the need for binocular fusion is the stimulus to convergence. Fincham and Walton confirmed the evidence of vergence accommodation by measuring accommodation both during condition of constant light vergence and changed convergence, as well as condition of constant convergence and changed light vergence [21]. Vergence–accommodation mismatch (VAM) can occur with 3D vision associated with closed robotic console systems, because the accommodation distance is fixed from the surgeon’s eyes to the display screen and the vergence distance depends on the simulated distance.
This VAM may be influenced in different ways with laparoscopic and robotic 3D vision [10, 22, 23]. The changes in convergence and accommodation (and mismatch) at the typical viewing distance of 50 cm with laparoscopy and open robotic console systems are much smaller than the head-mounted displays of closed robotic console systems. On the other hand, in contrast with robotic surgery, rapid movement of the camera in laparoscopic surgery exaggerates 3D display vergence–accommodation mismatch [24].
To see objects clearly on 3D displays requires the viewer to counteract the neural coupling between vergence and accommodation. Visual discomfort and fatigue can occur because of this vergence–accommodation conflict. Optimising focus cues (accommodation and blur in the retinal image) can improve stereo-acuity, reduce distortions, and reduce viewer fatigue and discomfort [25].
Visual–perception mismatch
Visual perception is related to how visual information is processed by the brain. Perceptual learning can lead to performance improvement through practice or training [26]. Visual–perception mismatch (VPM) occurs with laparoscopic and robotic surgery but can be reduced with increasing experience. With laparoscopic surgery, the fulcrum point created by the trocar at the abdominal wall results in counterintuitive movements of the hand and the laparoscopic instrument, creating a visualisation–perception disconnect [27]. The surgeon console–robot interface of robotic systems removes the fulcrum effect through software technology. The robotic system can translate the surgeon’s movements exactly resulting in surgeon hand movements replicated by movement of the instrument tip in the same direction. However, the technological advances of motion scaling and filtering of surgeon tremor available on the robotic system can lead to a discrepancy between vision and surgeon hand movements.
With the robotic system, the use of the clutch control (which allows greater movement of the robotic arms from the smaller confines of the surgeon console) can potentially introduce a visual–perceptual mismatch that can add an additional cognitive load [28]. Abiri et al. found significantly more peg drops (P = 0.011) and longer time to completion with a peg transfer task (P < 0.001) in 45 novice surgeons operating with a single-axis misalignment [28]. With robotic surgery, use of the clutch can lead to a brain sensory misalignment of the location of surgeon hands at the console and the location of the robotic arms as visualised through the console binoculars. However, preference by experienced surgeons to maintain their hands in an ergonomic neutral position may allow them to unconsciously overcome this VPM.
Visual–vestibular mismatch
Visual–vestibular mismatch (VVM) can occur when a person is overusing their visual cues for determining their position in space, when it might be more appropriate to use the information from the inner ear vestibular organs (which sense orientation, movement, and balance). This can occur during laparoscopic surgery if the assistant is not keeping the laparoscope steady and the surgeon may interpret visual motion as self-motion. With robotic surgery, the camera is held more steadily and VVM may be less of a problem.
Visual and vestibular signals are the primary sources of sensory information for self-motion. Conflict among these signals is adaptable but can result in vertigo, postural instability, and simulator sickness [29]. There is a trade-off between visual–vestibular integration and conflict detection that is mediated by eye movements [30]. Compared with laparoscopic surgery, closed robotic console surgery results in less head movement with resting of the forehead on the head rest and the eyes on the binoculars. This results in more eye movements and less retinal motion when the visual target is redirected. Therefore, use of the robotic visualisation system can improve conflict detection at the cost of impaired integration performance. [30]
Visuospatial ability
Visuospatial ability (VSA) is correlated with surgical skill, especially in less-experienced surgeons at the beginning of their learning curve. VSA may be somewhat innate but can often be improved with training. VSA can be tested with the Perceptual Ability Test (PAT) which tests a person’s ability to interpret 2D shapes in a 3D context. The PAT has been validated and shown to correlate with technical performance [31]. Better VSA has been shown to correlate with improved performance by novice surgeons on a robotic surgery simulator [32]. High PAT performers had a higher simulation score (49% higher, p = 0.005) and completed the exercise faster (36% faster, p = 0.009) than the low PAT performers. Higher initial VSA may result in greater improvement in performance and time with training especially for complex tasks [33].
Visual fatigue
Longer robotic operating times also put the surgeon at risk of visual fatigue from constant viewing of the video display through the binoculars, leading to computer vision syndrome, which includes symptoms such as blurred vision, dry eyes, eyestrain (asthenopia), and headache [34, 35]. There are two main causes of visual fatigue: external symptoms relating to dry eye and internal symptoms relating to accommodative or binocular vision stress [36]. These symptoms have been previously studied in office workers and computer gamers. A study found that more than 4 h of computer gaming was associated with both ocular and physical discomfort [35]. In a survey of 432 robotic surgeon, eye symptoms were found to be more common with longer years practicing robotic surgery [23]. The causes of visual discomfort when viewing stereo displays include cross-talk between the two eye images, misalignment of the images, head position, vergence–accommodation mismatch, visibility of flicker or motion artifacts, and visual–vestibular conflicts [24, 37].
Digital eye strain may be measured subjectively with questionnaires or objectively with critical flicker-fusion frequency, blinking and squinting characteristics, accommodative function, and pupil characteristics [36, 38, 39]. Management approaches include correction of refractive error, management of dry eye (downgaze, lubricating eye drops, omega-3 fatty acids supplementation, and blink efficiency training), regular screen breaks (20/20/20 rule of 20 s break every 20 min to view a distant object at 20 feet), consideration of vergence and accommodation problems, and use of blue-light filtering lenses [36].
Visual feedback
Experienced robotic surgeons can use visual cues like tissue deformation as surrogates for force and haptic feedback (which is not available with the da Vinci robotic system). Meccariello et al. performed a study on 25 surgeons assessing their ability to recognise the thickness of custom-made membranes without use of haptic feedback [40]. They found a significant difference in performance/average score between experts and non-experts (p < 0.05). In addition, 67% of expert surgeons correctly identified the location of a metal clip in the correct quadrant of a membrane compared to 37% of non-experts. Restoration of three-dimensional vision may be important to compensate for loss of force feedback, with experienced surgeons able to feel a pseudo-tactile sensation [41, 42]. Vision-based approaches to predict force can also use surgical tool information, such as the tool-tip trajectory, velocity, and grasper status [43]. Combining information arising from different sensory modalities is essential to interact effectively with the environment. Cross-modal interactions between different sensory modalities may be automatic and unconscious. Lunghi et al. used continuous flash suppression to investigate whether haptic signals can interact with visual signals outside of visual awareness. They found that touch can accelerate the rise to consciousness of a suppressed visual stimulus but must be congruent, with matching in either spatial frequency or orientation [44]. The perception of haptic feedback may be explained by the converse relationship of this visual–haptic congruency.
Open robotic console systems
New robotic surgical systems have been developed and are now available for use since the expiration of patents held by Intuitive Surgical (makers of the da Vinci robotic system). The Cambridge Medical Robotics Versius, the Medtronic Hugo, and the TransEnterux Surgical Senhance robotic systems all provide an open console system which require the use of 3D polarising glasses [45,46,47]. The open console design enables better verbal and nonverbal communication between the surgical team members. The visualisation ergonomic considerations for the closed da Vinci system would differ for the open console systems but has not been compared directly. The visualisation advantages of the open console system include improved camera control, less reliance on visual cues for haptic feedback, and potentially less VAM and VVM. The Senhance surgical system offers the added advantage of camera control by eye movements through an infrared eye-tracking system [45]. Both the Versius and Senhance robotic systems offer haptic feedback from the handles. The visualisation disadvantages of the open console system relate to the need to use polarised glasses like for 3D laparoscopic surgery, and include decrease brightness of the operating field, need for horizontal gaze, less stereopsis, and more visual fatigue [45]. In addition, visualisation with the Hugo system is complicated by its two monitors: a 3D-HD screen with head tracking system and a right sided touchscreen surgeon interactive display which can select visual filters and rotate the camera [48].
Conclusion
Understanding the ergonomics of visualisation during robotic surgery is important. Stereopsis requires visual cortex fusion of two slightly different images from each of the eyes. Optimal visualisation occurs when VAM and cross-modality conflict relating to the sensations of proprioception and balance are compensated. Visual fatigue from viewing stereoscopic images is more common when the images are outside the depth of focus and can be caused by cross-talk, misalignment, head position, VAM, and VVM. Management approaches include correction of refractive error, use of lubricating eye drops, blink efficiency training, and regular screen breaks.
Data availability statement
Not applicable.
References
Wong SW, Ang ZH, Yang PF, Crowe P (2022) Robotic colorectal surgery and ergonomics. J Robot Surg 16(2):241–246. https://doi.org/10.1007/s11701-021-01240-5. (Epub 2021 Apr 22 PMID: 33886064)
Stylopoulos N, Rattner D (2003) Robotics and ergonomics. Surg Clin North Am 83(6):1321–1337
Sakata S, Grove PM, Stevenson ARL (2016) Effect of 3-dimensional vision on surgeons using the da vinci robot for laparoscopy: more than meets the eye. JAMA Surg 151(9):793–794. https://doi.org/10.1001/jamasurg.2016.0412
Curlin J, Herman CK (2020) Current State of Surgical Lighting. Surg J (N Y) 6(2):e87–e97. https://doi.org/10.1055/s-0040-1710529. (PMID:32577527;PMCID:PMC7305019)
Cassilly R, Diodato MD, Bottros M, Damiano RJ Jr (2004) Optimizing motion scaling and magnification in robotic surgery. Surgery 136(2):291–294. https://doi.org/10.1016/j.surg.2004.05.002. (PMID: 15300193)
Darwiche F, Swain S, Kallingal G, Punnen S, Manoharan M, Parekh DJ, Gonzalgo ML (2015) Operative technique and early experience for robotic-assisted laparoscopic nephroureterectomy (RALNU) using da Vinci Xi. Springerplus 27(4):298. https://doi.org/10.1186/s40064-015-1076-6. (PMID:26140262;PMCID:PMC4483175)
Falk V, Mintz D, Grünenfelder J et al (2001) Influence of three-dimensional vision on surgical telemanipulator performance. Surg Endosc 15(11):1282–1288
Sørensen SM, Savran MM, Konge L, Bjerrum F (2016) Three-dimensional versus two-dimensional vision in laparoscopy: a systematic review. Surg Endosc 30(1):11–23. https://doi.org/10.1007/s00464-015-4189-7. (Epub 2015 Apr 4 PMID: 25840896)
Dirie NI, Wang Q, Wang S (2018) Two-dimensional versus three-dimensional laparoscopic systems in urology: a systematic review and meta-analysis. J Endourol 32(9):781–790. https://doi.org/10.1089/end.2018.0411. (PMID:29969912;PMCID:PMC6156697)
Hanna GB, Shimi SM, Cuschieri A (1998) Randomised study of influence of two-dimensional versus three-dimensional imaging on performance of laparoscopic cholecystectomy. Lancet 351(9098):248–251
Wang T, Zheng B (2019) 3D presentation in surgery: a review of technology and adverse effects. J Robot Surg 13(3):363–370
Huettl F, Huber T, Duwe M, Lang H, Paschold M, Kneist W (2020) Higher quality camera navigation improves the surgeon’s performance: evidence from a pre-clinical study. J Minim Access Surg. 16(4):355–359. https://doi.org/10.4103/jmas.JMAS_143_19. (PMID: 31793451; PMCID: PMC7597869)
Kavoussi LR, Moore RG, Adams JB, Partin AW (1995) Comparison of robotic versus human laparoscopic camera control. J Urol 154(6):2134–2136
Kim JS, Piozzi GN, Kwak JM, Kim J, Kim T, Choo J, Yang G, Lee TH, Baek SJ, Kim J, Kim SH (2022) Quality of laparoscopic camera navigation in robot-assisted versus conventional laparoscopic surgery for rectal cancer: an analysis of surgical videos through a video processing computer software. Int J Med Robot. 18(4):e2393. https://doi.org/10.1002/rcs.2393
Hanna GB, Shimi SM, Cuschieri A (1998) Task performance in endoscopic surgery is influenced by location of the image display. Ann Surg 227(4):481–484
Rosenfield M (2011) Computer vision syndrome: a review of ocular causes and potential treatments. Ophthalmic Physiol Opt 31(5):502–515. https://doi.org/10.1111/j.1475-1313.2011.00834.x. (Epub 2011 Apr 12 PMID: 21480937)
O’Connor AR, Tidbury LP (2018) Stereopsis: are we assessing it in enough depth? Clin Exp Optom 101(4):485–494
Gietzelt C, Datta R, Busshoff J, Bruns T, Wahba R, Hedergott A (2022) The influence of stereoscopic vision on surgical performance in minimal invasive surgery-a substudy of the IDOSP-Study (Influence of 3D- vs. 4 K-display systems on surgical performance in minimal invasive surgery). Langenbecks Arch Surg 407(7):3069–3078. https://doi.org/10.1007/s00423-022-02608-3. (Epub 2022 Jul 22. PMID: 35869334; PMCID: PMC9640404)
Biddle M, Hamid S, Ali N (2014) An evaluation of stereoacuity (3D vision) in practising surgeons across a range of surgical specialities. Surgeon 12(1):7–10. https://doi.org/10.1016/j.surge.2013.05.002. (Epub 2013 Jun 10 PMID: 23764432)
Tuna MB, Kilavuzoglu AE, Mourmouris P, Argun OB, Doganca T, Obek C, Ozisik O, Kural AR (2020) Impact of refractive errors on da vinci si robotic system. JSLS. 24(3):e2020.00031. https://doi.org/10.4293/JSLS.2020.00031. (PMID: 32831541; PMCID: PMC7434396)
Fincham EF, Walton J (1957) The reciprocal actions of accommodation and convergence. J Physiol 137(3):488–508. https://doi.org/10.1113/jphysiol.1957.sp005829. (PMID:13463783;PMCID:PMC1363021)
Zhou J, Xu HJ, Liang CZ et al (2015) A comparative study of distinct ocular symptoms after performing laparoscopic surgical tasks using a three-dimensional surgical imaging system and a conventional two-dimensional surgical imaging system. J Endourol 29(7):816–820
Lee GI, Lee MR, Green I et al (2017) Surgeons’ physical discomfort and symptoms during robotic surgery: a comprehensive ergonomic survey study. Surg Endosc 31(4):1697–1706
Banks MS, Kim J, Shibata T (2013) Insight into vergence–accommodation mismatch. Proc SPIE Int Soc Opt Eng. https://doi.org/10.1117/12.2019866. (PMID: 24244832; PMCID: PMC3824256)
Hoffman DM, Girshick AR, Akeley K, Banks MS (2008) Vergence-accommodation conflicts hinder visual performance and cause visual fatigue. J Vis 8(3):33.1–30. https://doi.org/10.1167/8.3.33. (PMID: 18484839; PMCID: PMC2879326)
Lu ZL, Dosher BA (2022) Current directions in visual perceptual learning. Nat Rev Psychol. 1:654–668. https://doi.org/10.1038/s44159-022-00107-2
Schneeberger EW, Michler RE (2001) An overview of the intuitive system: the surgeon’s perspective. Oper Tech Thorac Cardiovasc Surg 6(3):170–176. https://doi.org/10.1053/otct.2001.26962
Abiri A, Tao A, LaRocca M et al (2017) Visual-perceptual mismatch in robotic surgery. Surg Endosc 31(8):3271–3278
Nishiike S, Okazaki S, Watanabe H, Akizuki H, Imai T, Uno A, Kitahara T, Horii A, Takeda N, Inohara H (2013) The effect of visual-vestibulosomatosensory conflict induced by virtual reality on postural stability in humans. J Med Invest 60(3–4):236–239. https://doi.org/10.2152/jmi.60.236. (PMID: 24190041)
Garzorz IT, MacNeilage PR (2017) Visual-vestibular conflict detection depends on fixation. Curr Biol 27(18):2856-2861.e4. https://doi.org/10.1016/j.cub.2017.08.011. (Epub 2017 Sep 7 PMID: 28889981)
Kramer GA, Kubiak AT, Smith RM (1989) Construct and predictive validities of the perceptual ability test. J Dent Educ 53(2):119–125 (PMID: 2915079)
Suozzi BA, O’Sullivan DM, Finnegan KT, Steinberg AC (2013) Can visuospatial ability predict performance and learning curves on a robotic surgery simulator? Female Pelvic Med Reconstr Surg 19(4):214–218. https://doi.org/10.1097/SPV.0b013e318298b364. (PMID: 23797520)
Roch PJ, Rangnick HM, Brzoska JA, Benner L, Kowalewski KF, Müller PC, Kenngott HG, Müller-Stich BP, Nickel F (2018) Impact of visual-spatial ability on laparoscopic camera navigation training. Surg Endosc 32(3):1174–1183. https://doi.org/10.1007/s00464-017-5789-1. (Epub 2017 Aug 24 PMID: 28840317)
Alhusuny A, Cook M, Khalil A, Xie Y, Johnston V (2020) Neck/shoulder problems and visual disturbances among surgeons: a scoping review. J Surg Res 247:413–428. https://doi.org/10.1016/j.jss.2019.09.064. (Epub 2019 Oct 31 PMID: 31679800)
Lee JW, Cho HG, Moon BY et al (2019) Effects of prolonged continuous computer gaming on physical and ocular symptoms and binocular vision functions in young healthy individuals. PeerJ 7:e7050. https://doi.org/10.7717/peerj.7050
Sheppard AL, Wolffsohn JS (2018) Digital eye strain: prevalence, measurement and amelioration. BMJ Open Ophthalmol. 3(1):e000146. https://doi.org/10.1136/bmjophth-2018-000146. (PMID: 29963645; PMCID: PMC6020759)
Nam KW, Park J, Kim IY, Kim KG (2012) Application of stereo-imaging technology to medical field. Healthc Inform Res. 18(3):158–163. https://doi.org/10.4258/hir.2012.18.3.158. (Epub 2012 Sep 30. PMID: 23115737; PMCID: PMC3483472)
Kim T, Lee EC (2020) Experimental verification of objective visual fatigue measurement based on accurate pupil detection of infrared eye image and multi-feature analysis. Sensors (Basel) 20(17):4814. https://doi.org/10.3390/s20174814. (PMID:32858920;PMCID:PMC7506756)
Chi CF, Lin FT (1998) A comparison of seven visual fatigue assessment techniques in three data-acquisition VDT tasks. Hum Factors 40(4):577–590. https://doi.org/10.1518/001872098779649247. (PMID: 9974230)
Meccariello G, Faedi F, AlGhamdi S et al (2016) An experimental study about haptic feedback in robotic surgery: may visual feedback substitute tactile feedback? J Robot Surg 10(1):57–61
van der Meijden OA, Schijven MP (2009) The value of haptic feedback in conventional and robot-assisted minimal invasive surgery and virtual reality training: a current review. Surg Endosc 23(6):1180–1190
Hagen ME, Meehan JJ, Inan I, Morel P (2008) Visual clues act as a substitute for haptic feedback in robotic surgery. Surg Endosc 22(6):1505–1508. https://doi.org/10.1007/s00464-007-9683-0. (Epub 2007 Dec 11 PMID: 18071811)
Jung WJ, Kwak KS, Lim SC (2020) Vision-based suture tensile force estimation in robotic surgery. Sensors (Basel) 21(1):110. https://doi.org/10.3390/s21010110. (PMID:33375388;PMCID:PMC7796030)
Lunghi C, Lo Verde L, Alais D (2017) Touch accelerates visual awareness. Iperception 8(1):2041669516686986. https://doi.org/10.1177/2041669516686986. (PMID:28210486;PMCID:PMC5298565)
Koukourikis P, Rha KH (2021) Robotic surgical systems in urology: what is currently available? Investig Clin Urol 62(1):14–22. https://doi.org/10.4111/icu.20200387. (PMID:33381927;PMCID:PMC7801159)
Mayor N, Coppola AS, Challacombe B (2022) Past, present and future of surgical robotics. Trends Urol Men Health 13:7–10. https://doi.org/10.1002/tre.834
Hughes T, Rai B, Madaan S, Chedgy E, Somani B (2023) The availability, cost, limitations, learning curve and future of robotic systems in urology and prostate cancer surgery. J Clin Med 12(6):2268. https://doi.org/10.3390/jcm12062268
Gueli Alletti S, Chiantera V, Arcuri G, Gioè A, Oliva R, Monterossi G, Fanfani F, Fagotti A, Scambia G (2022) Introducing the new surgical robot HUGO™ RAS: system description and docking settings for gynecological surgery. Front Oncol 9(12):898060. https://doi.org/10.3389/fonc.2022.898060. (PMID: 35756633; PMCID: PMC9218341)
Funding
Open Access funding enabled and organized by CAUL and its Member Institutions. The authors declare that no funds, grants, or other supports were received during the preparation of this manuscript.
Author information
Authors and Affiliations
Contributions
All authors contributed to the study conception and design. The first draft of the manuscript was written by SWW and all authors commented on previous versions of the manuscript. All authors read and approved the final manuscript.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no relevant financial or non-financial interests to disclose.
Ethical approval
This article does not contain any studies with human or animal subjects performed by any of the authors.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Wong, S.W., Crowe, P. Visualisation ergonomics and robotic surgery. J Robotic Surg 17, 1873–1878 (2023). https://doi.org/10.1007/s11701-023-01618-7
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11701-023-01618-7