FormalPara Key Summary Points

Simulator-based training can enhance student education without additional burden on classmates or patients.

A significant effect of simulator-based training was already proven for direct and indirect ophthalmoscopy but the effect of fundus biomicroscopy training has not yet been studied.

The aim of this study was to evaluate the effectiveness of simulator-based fundus biomicroscopy training with the slit lamp.

Training on the Eyesi Slit Lamp Simulator significantly improved students’ examination techniques in fundus biomicroscopy and their ability to locate pathological fundus lesions.

Further studies should be conducted to evaluate whether longer training periods result in superior technical skills and findings, as well as a significant improvement in correct clinical diagnosis.

Introduction

Ophthalmoscopy has a long history in ophthalmology [1, 2]. In contrast to direct ophthalmoloscopy [3], indirect ophthalmoscopy [4], including biomicroscopic funduscopy with the slit lamp [5, 6], offers a binocular stereoscopic fundus evaluation with a wide field of view [7]. The different examination options should also be offered to students, not only to increase their interest in this special field of ophthalmology but also to consolidate their knowledge of ophthalmological diagnoses (e.g., fundus hypertonicus, diabetic retinopathy) with relevance to almost all other medical subspecialties [8].

This is crucial, as previous research demonstrated that medical students and practicing physicians experience a lack of confidence in their ophthalmoscopic abilities due to inadequate training during medical school [9,10,11].

A well-developed medical specialist training is necessary in order to have highly qualified physicians for the future [12]. Nevertheless, nowadays, a high workload with shortage of time and staff in clinical settings often makes it more difficult to introduce medical students or residents to the various examination methods in a structured and well-supervised way [13]. Many students feel that they have not received adequate training in the past in regards to conducting physical examinations, making a diagnosis, and establishing the correct management of various diseases [14].

Additionally, clinical medicine is increasingly focused on patient safety and quality and patients may be worried that students and residents are training on them [15]. Especially in the early education stage ethical concerns about using patients for training purposes arise. Physical examination of pears is an alternative method to improve students' skills, but usually only healthy eyes can be inspected and willingness to be examined is low [16].

Therefore, simulator-based training can offer a useful complement to reduce barriers to medical education without extra strain on classmates or even patients. The impact and educational utility of simulator-based medical education have grown in the past 40 years and are likely to increase in the future [17]. Lee et al. published a systematic review and analyzed studies with different training modalities: virtual reality, wet lab, dry-lab models, and e-learning [18]. Models with the strongest validity evidence were the Eyesi Surgical, Eyesi Direct Ophthalmoscope, and Eye Surgical Skills Assessment Test. Regarding the Eyesi training simulators (Haag-Streit Simulation, Mannheim, Germany), a significant effect in various studies was already proven for both surgical [19,20,21] and diagnostic skills in direct and indirect ophthalmoscopy [22,23,24,25]. In 2022, the Eyesi Slit Lamp Simulator was market launched. The simulator is integrated into a BQ 900 slit-lamp model from Haag-Streit Diagnostics and therefore offers all the functions of a real slit lamp: anterior chamber examination, gonioscopy, and fundus biomicroscopy (Fig. 1).

Fig. 1
figure 1

Students can practice posterior segment examinations using a 90D ophthalmoscope lens with the Eyesi Slit Lamp simulator. To visualize the retina, they must insert the 90D lens into the virtual slit beam and then slowly move it towards the model eye until they see a focused image through the slit-lamp microscope

After demonstrating a significant training effect for the Eyesi Slit Lamp simulator in anterior chamber examination [26], we sought to evaluate the effectiveness of simulator-based fundus biomicroscopy training. The study design and subject group (students) remained unchanged to ensure equal conditions for both studies. However, we included an additional OSCE observer to assess interrater reliability.

Methods

This prospective randomized controlled trial was conducted in accordance with the Declaration of Helsinki. The Ethics Committee of the Landesärztekammer Hessen reviewed the study (2022-3167-AF) and determined that ethical approval is not required. All study participants provided informed written consent. Additionally, consent for publication of clinical pictures was given by the patient volunteers.

In total, 243 medical students in their eighth semester at Saarland University Medical Center were invited to participate in the study via e-mail. Eligibility was restricted to those who had previously completed a comprehensive 1-week ophthalmology clerkship at the same institution, consisting of lectures and practical sessions on ophthalmologic examination and imaging methods, as well as ophthalmologic diseases and emergencies. Thirty positive responses were obtained. Block randomization was used. Students were assigned to two treatment groups: the traditional training group and the simulator-based training group with a 1:1 allocation ratio. A block size of six students was set. Out of 30 students, two students withdrew and 28 met the inclusion criteria and were included in the study (Supplemental Fig. S1). Written informed consent was obtained from all participants prior to enrollment.

Five 60-min courses for 2–3 students were held, both for the simulator-based training group and for the traditional training group. Both training groups were guided respectively by one ophthalmology-certified faculty instructor.

The students were given the opportunity to complete the alternative training after the OSCE examination, in which all participated.

Traditional Training Structure

The traditional training began with learning essential examination techniques to perform a systematic fundus examination. This included adjusting the refraction on the eyepieces and the interpupillary distance, positioning the patient correctly at the slit lamp, adjusting the light to the patient's eye, optimally localizing the 90D lens between patient's eye and slit, performing lateral and vertical translation, adapting of slit width and slit length and the interpretation of the inverted fundus image. In the next step, the essential fundus structures had to be localized and described. These structures include the macula, optic disc, superior and inferior temporal vascular arcade, upper and lower nasal vascular arcade and the equatorial region/outer periphery (vortex veins). Determining the structures was achieved through mutual examinations. Therefore, at least one eye of the students had to be dilated. For the last part of the training, the faculty instructor presented clinical images relevant to the students' curriculum and discussed fundus pathologies, including glaucoma, papilledema, dry AMD, wet AMD, diabetic retinopathy (background and high risk), toxoplasmosis, branch vein occlusion and hypertensive retinopathy.

Simulator-Based Training Structure

The simulator-based course started with abstract tasks to learn the different examination techniques. Therefore, each separate slit-lamp function was set as a separate task using a gamified teaching method: Students had to find and identify common items found in a virtual ophthalmologist’s office. Once a level was completed, the acquired function was retained for subsequent tasks. This step-by-step approach allowed students to become familiar with more and more slit-lamp functions as they progressed through the course. After completing all selected abstract tasks for device handling, the students had to examine and diagnose virtual patients, suffering from the same diseases with associated fundus pathologies, which were shown as clinical pictures to the traditional group (Fig. 2). Each virtual patient was presented with a brief medical history, current symptoms, as well as their visual acuity and intraocular pressure values.

Fig. 2
figure 2

a Glaucomatous optic nerve (left) taken during the simulator course compared to a real fundus biomicroscopy image (right). b Toxoplasmosis scar (left) taken during the simulator course compared to a real fundus biomicroscopy image (right). c Non-proliferative diabetic retinopathy (left) taken during the simulator course compared to a real fundus biomicroscopy image (right). d Dry age-related macular degeneration (left) taken during the simulator course compared to a real fundus biomicroscopy image (right)

OSCE

An OSCE examination for fundus biomicroscopy was developed to assess the skills of both training groups. We used the OSCE form for direct ophthalmoscopy developed by Boden et al. [22]. This form was previously adapted for slit-lamp examination in our prior study analyzing the impact of anterior chamber examination using the Eyesi Slit Lamp [23]. For this study, the form was extended to include fundus biomicroscopy (Table 1).

Table 1 Prospective rubrics to guide the two masked assessors in assigning scores to the medical students’ OSCE performance in fundus biomicroscopy

The following rubrics were created to guide the masked observer in assigning scores for the student’s fundus biomicroscopy skills: preparation for fundus examination (max six points), finding relevant structures of the fundus (max five points), describing the structures found (max eight points), correct diagnosis of the patient volunteer’s disease (max three points), explanation of ophthalmic procedures and appropriate application such as pupil dilation (max six points), and the correct diagnosis of five clinical images randomly chosen from nine images (max five points). For this purpose, images of the same disease patterns were used in the simulator and traditional course, but varied fundus photographs were taken for the OSCE. Scores per task were assigned depending on the difficulty level and importance for the fundus biomicroscopy examination (Table 1). All the rubrics were added up to calculate the total score (max 33 points) for the overall performance.

The students’ OSCE performance was evaluated by two independent observers. The first masked observer (observer 1) was a resident of the Department of Ophthalmology of the Saarland University Medical Center in his fourth year.

The second masked observer (observer 2) was a professor of ophthalmology with over 20 years of experience in medical student training, who volunteered to assess the OSCE. The examination room was on a separate floor to ensure that they had no contact with students during the program. Ten volunteers were recruited from the inpatients for the OSCE. To eliminate any potential exchange of diagnostic information among students, each subgroup examined a separate patient.

Post-assessment Survey

After attending the OSCE and both courses, each student filled out a post-assessment survey which was designed following the principles of survey tool development from the Medical Didactics course of the Goethe University Frankfurt. The course supervisor ensured that each student responded only once. The following categories were collected and evaluated: usefulness of the fundus biomicroscopy simulator course, increase in knowledge, effectiveness of abstract tasks, training of fundus biomicroscopy techniques, training of necessary techniques in the context of pathologies, three-dimensional localization of pathologies, efficacy of multimedia learning of pathologies, efficacy of independent examination training, preparation for a real fundus biomicroscopy examination, ability to recognize diseases with fundus biomicroscopy after simulator training, proportion of medical training on simulators. The students were asked to assign point values to each question. The questionnaire responses were scored on a scale from 1 to 7. A score of 1 represented low importance, whereas 7 represented very high importance.

Additionally, a self-assessment form was distributed to the students to compare the effectiveness of the simulator versus the traditional course in regards to gaining curriculum-relevant skills such as applying fundus biomicroscopy techniques, examining a healthy eye, recognizing a pathology, and assessing pathology location. Open-ended queries explored the benefits of fundus biomicroscopy simulator training and traditional training, as well as proposals for improvement.

Data and Statistical Analysis

The medical student volunteers received written and oral information about the study and were informed of their right to withdraw their participation at any time. The study data were only accessible to the authors and confidentiality was maintained by pseudonymizing the materials of the OSCE examination and survey forms. The data were analyzed using Excel, IBM SPSS Statistics v.28, and BiAS for Windows v. 11.12. A Fisher’s exact test was performed to investigate the homogeneity of the two training groups.

The primary outcome measures were the exam grades of both the simulator and traditional groups. Initially, the scores were tested for normal distribution through the Shapiro–Wilk test. As the scores were not normally distributed, the statistical analysis relied on the Wilcoxon–Mann–Whitney U-exact test to identify the differences between the objective OSCE scores of the simulator and traditional group [27, 28]. The data were described using median (interquartile range [IQR]).

Additionally, the Rosenthal's effect size was calculated, with values of 0.1 indicating a small effect, 0.3 indicating a medium effect, and 0.5 indicating a large effect [29]. The Spearman correlation coefficient was computed between the masked observers for interrater reliability (IRR), using a two-sided test with Edgeworth approximation. Secondary outcomes included students’ subjective assessment of the efficacy of the traditional versus the simulator course. Binomial tests were used for comparison. Post-assessment surveys were analyzed descriptively, with frequencies presented in contingency tables. A significance level of p < 0.05 was assumed for all tests.

Results

A total of 28 medical students in their eighth semester were enrolled in the study and randomized to simulator-based (n = 14) or traditional (n = 14) training.

Out of 28 students, two were unable to complete the entire course, including the OSCE, and one had to be excluded from the statistical analysis due to a significant language barrier. Twenty-five students [simulator group (n = 14), traditional trained group (n = 11)] remained for statistical analysis.

Regarding the characteristics of the simulator versus the traditional trained group (Table 2), there was no statistically significant difference in prior fundus biomicroscopy experience (p = 0.18), previous experience with the subject “ophthalmology” (p = 0.34) or presence of own eye diseases (p = 0.66). Furthermore, the planned subspecialty for medical residency is listed in Table 2.

Table 2 Characteristics of 25 medical student participants in a randomized controlled trial comparing simulator-based vs. traditional techniques for learning fundus biomicroscopy

The overall performance (total score) of the medical students in the fundus biomicroscopy OSCE was statistically ranked significantly higher in the simulator group (27.0 [5.25]/28.0 [3.0] vs. 20.0 [7.5]/16.0 [10.0]) by both observers with an interrater reliability of IRR < 0.001 and a significance level of p = 0.003 for observer 1 and p < 0.001 for observer 2. The highest overall performance score (33.0/33.0) was assigned by both observers to the same student, a student in the simulator group. Table 3 summarizes the scores assigned to the students in each group.

Table 3 Statistical analysis of grades assigned by the two masked ophthalmologists during the fundus biomicroscopy OSCE comparing the simulator and traditionally trained group

The scores given to students trained on simulators were consistently higher than those given to students trained traditionally. Interrater reliability was verified as significant for all subtasks (≤ 0.002) except subtask 1 “preparation for fundus examination” (p = 0.12). For this subtask, no significant difference could be found by observer 1 (5.5 [2.0] vs. 5.0 [2.0], p = 0.85), in contrast to observer 2 (5.0 [1.0] vs. 3.0 [2.0], p < 0.001).

For subtask 2 “Finding of relevant fundus structures” statistical significance was narrowly missed for observer 1 (3.5 [1.86] vs. 2.0 [1.0], p = 0.06), but was already reached for observer 2 (4.0 [0.75] vs. 1.0 [1.5], p < 0.001), IRR = 0.002.

Both observers scored the simulator group significantly higher (p = 0.009/p < 0.001) for the subtask 3 “correct description of the structures found” (5.0 [1.75]/6.0 [1.75] vs. 4.0 [2.0]/3.0 [3.0]), IRR < 0.001.

For subtask 4 “accurate diagnosis of the examined patient” a statistical significance (p = 0.02) was determined for observer 1 (3.0 [0.75] vs. 0.0 [1.75]), but narrowly missed for observer 2 with p = 0.05 (3.0 [0.75] vs. 1.0 [2.0]), IRR < 0.001.

Regarding subtask 5 “comments on students’ examination approach”, there was no significant difference found for the scores assigned by observer 1 (6.0 [0.0] vs. 6.0 [2.0], p = 0.13), contrary to observer 2 (6.0 [0.0] vs. 4.0 [2.0], p = 0.001), IRR < 0.001.

No significant difference between the simulator and traditional group was seen for subtask 6 “correct recognition of diseases on five clinical fundus pictures” (4.25 [0.88]/5.0 [1.0] vs. 4.0 [1.0]/4.0 [1.5]) for both observers with an interrater reliability of p < 0.001.

The students’ self-assessment forms showed similar subjective results. Students felt significantly better trained in the use of biomicroscopy and its practical background (p = 0.04), in the detection of fundus pathologies (p < 0.001) and in the correct localization of pathologies using the biomicroscopy simulator (p < 0.001).

Learning how to perform an examination of a healthy eye showed no statistical difference between the two learning methods (p = 0.42).

All 25 students returned their questionnaires. Point values for each question were scored on a scale from 1 to 7 (from 1 = “does not apply at all” to 7 “fully applies”). Both groups, the simulator group, which was trained on the slit lamp simulator prior to taking the OSCE exam, and the traditional group, which received simulator-based slit lamp training after taking the OSCE exam, reported a primarily positive experience with the slit lamp simulator course and would recommend it to others (7.0 [0.0]). They noticed an increase in knowledge over the training period (7.0 [0.0]), the abstract tasks helped them to get a better feeling for the procedure (7.0 [0.0]), they felt able to understand the biomicroscopy techniques in general (7.0 [1.0]) as well as in the context of pathologies (7.0 [1.0]). They gained a better sense of the location of the pathologies (7.0 [0.0]) and their typical manifestation form (7.0 [0.0]). As a result, they felt more prepared to identify diseases using fundus biomicroscopy after the simulator training (7.0 [1.0]). The interactive multi-media learning approach supported them to memorize the different fundus pathologies (7.0 [0.0]). The majority would like to have more simulated medical training during their medical studies (7.0 [0.0]). The detailed results can be found in Table 4.

Table 4 Results of the post-assessment survey of 25 medical students regarding efficacy of the fundus biomicroscopy simulator coursea

In order to be adequately prepared to perform fundus biomicroscopy on real patients (6.0 [1.0]), they required more simulator training time as stated in the open field questions.

Based on the open field responses, other major topics among the stated advantages of simulator training based on open field responses were the ability to learn to visualize simulated fundus pathologies and not just inspect healthy eyes during the mutual examinations (14 [56.0%] of 25 responses), as well as the ability to train unrestrictedly without burdening patients or fellow students (12 [48.0%] of 25 responses). Additionally, eight of 25 [32.0%] students reported that they could improve their proficiency in using the slit lamp and 90D lens more quickly by receiving direct feedback from the simulator on their settings and the fundus structures found.

The primary benefit of examining real patients during slit-lamp training, as reported by 20 out of 25 respondents [80.0%], was the opportunity to interact with patients, which helped to alleviate inhibitions and fears associated with real patient contact. The students were only able to practice positioning the patient's head in front of the real slit lamp, as the mannequin's head is already attached to the simulator. This was reported by five out of 25 respondents (20%).

Discussion

Various publications have shown the added value of simulator training compared to e-learning or labs [13, 18, 30,31,32,33]. In a virtual reality environment, a non-existing environment is simulated and an augmented reality involves haptic elements from the real world to enhance the experience [34]. On a computerized literature search using PubMed, we found previous studies that investigated the effect of training on the Eyesi Direct and Indirect Ophthalmoscope, but there was no study that dealt with the effect of fundus biomicroscopy training with the Slit Lamp Simulator, launched in 2022.

For direct ophthalmoscopy training, studies have shown that examination technique improves with the use of a simulator as a supplemental learning tool.

Boden et al. could show that the simulator-trained group achieved significantly higher scores compared to the traditionally trained group [22]. They randomized 34 medical students during their ophthalmological internship for a 45-min classical and simulator training following a 5-min introduction. The traditionally trained group achieved a learning success rate of 78% on the subsequent OSCE, while the simulator-based group achieved a higher score of 91% with less variation in all sub-disciplines. These results are comparable to our study; here the success rate of the traditionally trained group was 62% for observer 1 and 52% for observer 2. Meanwhile the simulator-based group achieved a significantly higher score of 80% for observer 1 and 85% for observer 2. Overall, the average total score was slightly lower for both groups. This could be attributed to the more challenging examination technique of indirect fundus biomicroscopy in comparison to direct funduscopy.

Boden et al. also underlined the advantages of patient- and instructor-independent training possibilities, i.e., a reduction of light exposure for patients and test subjects, as well as a standardized and controlled mediation of physiological and pathological findings.

In Howell and associates’ study, medical student volunteers underwent a longer training period with the direct ophthalmoscope simulator [24]. After a 1-h didactic instruction course, 17 students were randomized to an additional hour of training on a direct ophthalmoscope simulator, and another 16 students to a supervised practice examining classmates. After a 1-week independent student practice using the assigned training methods, masked ophthalmologist observers assessed the students’ ophthalmoscopy skills. They identified better performance among students in the simulator-trained group regarding three measures: longer independent practice time, technique (ability to handle a direct ophthalmoscope and conduct an examination) with p = 0.034, and, based on the results of one patient examination, the ability to locate fundus lesions accurately (p = 0.013). An improved practical technique for fundus biomicroscopy following simulation training could also be shown in the present study. The simulator trained group was able to better visualize fundus structures (observer 1 p = 0.06; observer 2 p < 0.001) and to describe typical medical findings (observer 1 p = 0.009; observer 2 p < 0.001). Training time was predetermined in our study, but in the open field questions, students also requested longer training times for fundus biomicroscopy with the Eyesi Slit Lamp simulator.

Longer training periods could also be necessary to demonstrate not only superior technical skills and findings, as mentioned above, but also a significant improvement regarding the correct clinical diagnosis.

Learning practical skills is much easier for direct ophthalmoscopy than indirect ophthalmoscopy. In a previous study using the Eyesi Direct and Eyesi Indirect simulator, we demonstrated improved management of simulated clinical cases after a 4-h training period for direct and indirect ophthalmoscopy. However, statistical significance was only observed for direct ophthalmoscopy [23]. This study found similar results with significantly better technical skills and medical findings, but no statistical difference in diagnosing diseases. The limited duration of the course may have contributed to this issue. A more extended training period, lasting several weeks and providing opportunities to practice and repeat the identification of various pathologies, could enhance students’ ability to identify correct and differential diagnoses.

The technical skills of the simulator group were significantly better compared to the traditional group. This may be attributed to the direct objective performance assessment provided by the simulator regarding the students’ settings and fundus structures found, as well as the design of the tiers, which are methodically built on one another. Training units with the Eyesi Slit Lamp’s comprehensive and modular curriculum to examine and diagnose simulated fundus pathologies (other than training on healthy eyes of fellow students) and the ability to learn unrestrictedly without burdening patients or fellow students (e.g., by dilating the pupil) were also seen as advantages by the students, as stated in their open field responses.

The strengths of our study include its randomized design, comparing fundus biomicroscopy simulator training to traditional training allowing equal training times for both groups. All attendees completed a 1-week ophthalmology clerkship first to ensure equal prerequisites. An objective assessment of acquired skills was performed by two masked observers to prove interrater reliability. A significant interrater reliability could be shown for all tasks except for “preparation for fundus examination”, including rather soft evaluation criteria such as the correct positioning of the patient, the slit light, and the 90D lens.

Patient volunteers were used for the OSCE in order to verify the transferability of examining simulated patients on real patients. In addition to the objective results, the study also evaluated the subjective assessment of the effectiveness of simulated fundus biomicroscopy training. The subjective assessment corresponded to the objective results. Students reported feeling better prepared for everyday clinical practice and requested more simulation training in the future.

The limitations of this study include the relatively small number of student volunteers and the limited training time. Therefore, statistically significant improvements in practical techniques and localization of fundus lesions were observed for the simulator group compared to the traditional group. However, no significant improvement was observed in diagnosing patients. The students reported subjectively that they were able to learn pathologies more effectively using the simulation slit lamp. It is recommended that further studies investigate the potential benefits of longer training periods, e.g., over weeks or months. The scoring system of our study was adapted from the direct ophthalmoscopy OSCE template of Boden et al. [22]. However, it still uses a non-validated scoring system. The validity evidence for the slit-lamp simulator has to be evaluated carefully along existing formal validation frameworks [35]. So far the Eyesi Indirect Ophthalmoscope Simulator [36] and Eyesi Surgical Simulator [37,38,39] have been verified for construct and face validity in former studies.

Overall, simulation tools can enhance the possibilities of student training. They can serve as an alternative to real patients. As an advantage, trainees can make mistakes and learn from them without the fear of harming the patient [31]. The students can train systematically in a self-guided manner according to their own learning speed and can subjectively develop a significantly higher level of confidence after the simulation training [40], too. Based on our experience in this and previous studies [22, 23, 26, 41], a lecture on simulator training was included in the 1-week ophthalmology clerkship for students at the Saarland University, Germany, and simulator training courses are offered on a voluntary basis 1 day after the clinical clerkship, which are well attended.

Conclusions

In summary, our results demonstrate that training on the Eyesi Slit Lamp Simulator significantly enhances practical skills in fundus biomicroscopy and thereby the ability to locate fundus lesions accurately. Further studies are needed to verify whether an extended training period of weeks or months would result in improved diagnosis. Additionally, students also subjectively reported a significant improvement in practical skills.

The Eyesi Slit Lamp Simulator can be a valuable supplement to traditional training methods. Students can independently train fundus biomicroscopy without the need for suitable patients or an all-time-present instructor to get better prepared for everyday clinical practice.