Abstract
Objectives
Digital breast tomosynthesis (DBT) can improve diagnostic accuracy compared to 2D mammography, but DBT reporting is time-consuming and potentially more fatiguing. Changes in diagnostic accuracy and subjective and objective fatigue were evaluated over a DBT reporting session, and the impact of taking a reporting break was assessed.
Materials and methods
Forty-five National Health Service (NHS) mammography readers from 6 hospitals read a cancer-enriched set of 40 DBT cases whilst eye tracked in this prospective cohort study, from December 2020 to April 2022. Eye-blink metrics were assessed as objective fatigue measures. Twenty-one readers had a reporting break, 24 did not. Subjective fatigue questionnaires were completed before and after the session. Diagnostic accuracy and subjective and objective fatigue measures were compared between the cohorts using parametric and non-parametric significance testing.
Results
Readers had on average 10 years post-training breast screening experience and took just under 2 h (105.8 min) to report all cases. Readers without a break reported greater levels of subjective fatigue (44% vs. 33%, p = 0.04), which related to greater objective fatigue: an increased average blink duration (296 ms vs. 286 ms, p < 0.001) and a reduced eye-opening velocity (76 mm/s vs. 82 mm/s, p < 0.001). Objective fatigue increased as the trial progressed for the no break cohort only (ps < 0.001). No difference was identified in diagnostic accuracy between the groups (accuracy: 87% vs. 87%, p = 0.92).
Conclusions
Implementing a break during a 2-h DBT reporting session resulted in lower levels of subjective and objective fatigue. Breaks did not impact diagnostic accuracy, which may be related to the extensive experience of the readers.
Clinical relevance statement
DBT is being incorporated into many mammography screening programmes. Recognising that reporting breaks are required when reading large volumes of DBT studies ensures this can be factored in when setting up reading sessions.
Trial registration
Clinical trials registration number: NCT03733106
Key Points
• Use of digital breast tomosynthesis (DBT) in breast screening programmes can cause significant reader fatigue.
• The effectiveness of incorporating reading breaks into DBT reporting sessions, to reduce mammography reader fatigue, was investigated using eye tracking.
• Integrating breaks into DBT reporting sessions reduced reader fatigue; however, diagnostic accuracy was unaffected.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Adopting digital breast tomosynthesis (DBT) as a standard of care in breast screening programmes could improve patient outcomes and clinical workflow due to its reported increase in cancer detection rate and reduced recall rate in high recall environments, compared to 2D digital mammography alone [1,2,3,4,5]. DBT facilitates cancer detection by offering greater power to resolve overlapping layers of breast tissue, reducing the likelihood of missing obscured lesions or recalling artefactual findings. As a consequence, DBT images present greater image content to search and interpret compared to 2D digital mammography, increasing the read time and cognitive cost to the clinician [6]. In the context of breast screening, reading large volumes of DBT images could induce more severe fatigue in mammography readers compared to 2D studies, which has the potential to compromise diagnostic accuracy over prolonged screening sessions.
Previous studies in radiology have demonstrated the negative effects of reader fatigue on diagnostic accuracy and case interpretation efficiency [7,8,9]. However, these studies have primarily focused on specialties and imaging modalities other than DBT. Furthermore, these studies often compare radiologists’ performance in two different reporting sessions when fatigue levels would be expected to be very different, for instance, comparing a reporting session before starting a work shift to one after finishing a shift and comparing reporting in day shifts to overnight shifts [8, 9].
The aim of this prospective cohort study was to evaluate the changes in diagnostic accuracy and subjective and objective fatigue over a DBT reporting session, and how taking a break in reporting can affect these parameters. We hypothesised that implementing breaks within a DBT session would lead to lower levels of fatigue and reduced error rates. Identifying the point of fatigue onset in DBT reporting via blink characteristics could help to inform standards of DBT reporting session duration to limit reader fatigue and its negative impacts on patient outcomes, as breast screening programmes transition to this new modality.
Materials and methods
Participants and inclusion criteria
This prospective cohort study was conducted as a sub-study within the UK PROSPECTS Trial (ClinicalTrials.gov Identifier: NCT03733106) which has London–Dulwich Research Ethics Committee approval and all study participants provided written consent. The PROSPECTS Trial is a prospective randomised trial of DBT plus standard 2D digital mammography or synthetic 2D mammography (S2D) compared to standard 2D digital mammography in breast cancer screening.
Forty-five mammography readers from 6 National Health Service Breast Screening Programme centres participated from December 2020 to April 2022. All readers from centres participating in the PROSPECTS Trial were invited to take part, and consenting readers were consecutively recruited at each centre. Participants were NHS Breast Screening Programme mammography readers including board-certified consultant radiologists, radiographers (consultant radiographers and advanced practitioners, who are technologists with Master’s level training in mammographic interpretation) and breast clinicians (doctors who work in the field of breast care, but are not radiologists). All screening mammograms in the NHS breast screening programme are independently double read; and all participating mammography readers interpreted a minimum of 5000 mammograms per year, with a minimum of 1500 screening mammograms as the first reader. All readers had received prior training in DBT interpretation. Participants received continuing professional development (CPD) points and certification for their participation.
Eye tracking data from 30 of the 45 participants included in the present investigation were analysed previously [10, 11]. These studies investigated the use of eye-blink behaviour as fatigue and cognitive markers in DBT reporting; but 21 of these 30 participants had reporting breaks in their reading sessions—a previous limitation [10, 11]. In the present study, data has been collected from a further 15 participants who were not permitted a reporting break, enabling a comparison between participants depending on whether a break was allowed.
DBT case set
Participants independently read 40 anonymised cases possessing both 2D digital mammography and DBT images. Cases were chosen by an expert breast radiologist with more than 20 years’ experience, J.J., providing a variety of difficulty. There was also a variety of case pathology and finding types (Table 1). Participants were blinded to the proportions of each pathology type in the test set. Cases were presented to each participant in a random order. Case images were viewed on a Hologic SecurView workstation (Hologic Inc.) with a 4200 × 2800 pixel, mammography-approved BARCO monitor (BARCO Ltd.). Up to 4 views (left and right breast, MLO and CC) from a single case could be reviewed. Cases opened with 2D digital mammography images by default. DBT mode could be toggled on and off, reflecting real clinical practice (note that participants preferentially read the cases in DBT mode). The hanging protocols for each case could be changed by the participant, and all image manipulation tools were available to allow participants to simulate real-life reading.
Eye tracking equipment
Three non-intrusive eye tracking cameras (SmartEyePro, SmartEye AB) were mounted to the clinical monitor to record eye-blink data (60 Hz sampling rate). Equipment was set up in each participant’s natural reading environment, at their NHS screening centre (Fig. 1).
Procedure
Before the experiment, participants completed a demographics survey to account for confounding variables and read two practice cases to ensure familiarity with the procedure and image viewing software (these were not recorded). Participants then examined each case in their own time and verbally reported each breast as normal or benign (return to screen) or indeterminate, suspicious or highly suspicious (recall). Participants indicated the location of any abnormality on the images, which were recorded using the PERFORMS online reporting software by a supervising experimenter [12, 13]. PERFORMS (Personal Performance in Mammographic Screening) is an international external quality assurance scheme for mammography readers; further details on PERFORMS can be found elsewhere [14].
For the first cohort of eye tracking trials (n = 20; 3 screening centres), participants had reporting breaks every 40 min. The duration of the break was measured, but we did not record what the reader did during this period. For the second cohort (n = 24; the remaining 3 centres), participants were not permitted to take a break. One extra participant from the second batch needed to take a break at 40 min, therefore was categorised as having a break (n = 1). Participants completed a subjective fatigue survey before and after the reporting session, where participants rated their fatigue levels on a percentage scale from 0 (not fatigued) to 100 (extremely fatigued).
Blink data processing, quality filtering and exclusion criteria
Studies in a variety of psychological settings have reported increased blink rates in the fatigued state, as well as changes in individual blink dynamics, including longer blink durations in the fatigued state [15,16,17,18,19]. Previous investigation during DBT reporting concluded that blink frequency was an unreliable measure of fatigue in this context, and hence, only characteristics of the blink events (including blink duration and peak eye-opening velocity Fig. 2) were analysed here [10, 11].
Eye-blink data were analysed using a blink detection algorithm developed in-house [20, 21]. Blink data were automatically subjected to quality assessment as part of the algorithm; cases that did not pass the quality filter, due to missing and noisy data (resulting from eye obstruction and calibration issues), were excluded (Fig. 3).
Statistical analysis
All DBT cases had a known outcome. Malignant and benign cases were confirmed by biopsy, and normal cases had a normal 3-year follow-up mammogram. For each participant, each case was classified as true positive (TP), true negative (TN), false positive (FP) or false negative (FN) compared to pathology.
Normality was tested for using the Kolmogorov-Smirnov test. Significance testing was calculated for the differences between the break and no break cohorts. A chi-square test of independence was performed to examine the relation between categorical variables. The Mann-Whitney U test and independent samples t-test were used to check for significance for non-parametric and parametric continuous variables respectively. Kruskal-Wallis tests were performed to investigate the change in non-parametric blink metrics over the course of the DBT reporting session. The α-level for statistical significance was set at .05 for all analyses. Statistical calculations were performed using Python 3.8.3 (Python Software Foundation) by GP and AT. Data generated or analysed during the study are available from the corresponding author by request.
Results
Participant characteristics
We initially included 45 participants who reported 40 DBT images, yielding 1800 cases (Fig. 3). Following quality filtering, the blink data associated with 190 cases were excluded due to poor-quality eye tracking data. Three participants were excluded since all of their associated cases were excluded. Forty-two participants with 1610 cases remained in the blink analysis. All 45 participants remained in the diagnostic accuracy analysis. Forty participants were included in the subjective fatigue analysis; 5 were excluded due to an incomplete fatigue survey.
Demographic, training and session duration information for the participants are illustrated in Table 2. Most readers were radiology consultants (69%, n = 31 of 45) followed by advanced radiographic practitioners (16%, n = 7 of 45). We found no evidence of a difference in job roles between the break and no break cohorts (p = 0.54). Similarly, there was no significant difference between the cohorts in the frequencies of corrective lenses, day of the week and time of session (p = 0.37, p = 0.11, p = 0.66, respectively). The break cohort consisted mainly of readers from earlier sites, whereas the no break cohort were mainly from later sites (p < 0.001)—reflecting the change in methodology mentioned previously.
Both cohorts were well matched in terms of experience. We found no evidence of a difference in the years in post, years reading DBT and number of DBT cases read per year (p = 0.43, p = 0.44, p = 0.86, respectively). Participants had a sound baseline experience in radiology illustrated by a median of 10 years in their post (IQR=12 years). The median DBT experience was 5 years (IQR = 5 years), suggesting participants were not novices in DBT interpretation.
Participants on average took just under 2 h to complete reading all 40 DBT cases. Session duration (excluding break durations) was similar between the two cohorts (109.9 min vs. 102.0 min, p = 0.38). The median duration of a break was 7.6 min (IQR = 9 min).
Subjective fatigue
Of the participants who completed the fatigue survey, those who had a break (n = 16) reported significantly lower levels of subjective fatigue difference after the reading sessions compared to those who did not have breaks (n = 24) (mean, SD: 33% ± 22 vs. 44% ± 17, respectively, p = 0.04; Fig. 4). We found no evidence of a difference in the starting levels of fatigue between the break and no break cohorts (mean, SD: 32% ± 0.2 vs. 24% ± 0.2, respectively, p = 0.19).
Objective fatigue (blink metrics)
Over the whole trial, participants in the no break cohort exhibited a greater blink duration (296 ms vs 286 ms, p < 0.001), and a reduced peak eye-opening velocity (POV) (76 mm/s vs 82 mm/s, p < 0.001), compared to the break cohort. Additionally, changes in blink metrics in both cohorts were noted over the time course of the reporting session, where the session was split by case chronology (Fig. 5). During the interpretation of the first 10 cases, the blink metrics were similar between the two cohorts (blink duration: 285 ms vs 282 ms, p = 0.14; POV: 81 mm/s vs 82 mm/s, p = 0.19). However, during the second, third and last 10 cases, the blink duration was greater in the no break cohort (p = 0.02, p = 0.01 and p = 0.003, respectively), and the POV was reduced in the no break cohort (p = 0.02, p < 0.001 and p < 0.001, respectively) compared to the break cohort.
Using Kruskal-Wallis tests to compare blink metrics in each cohort over the time course revealed no evidence of a difference between case order and blink duration or POV in the break cohort (p = 0.09 and p = 0.88, respectively). However, significant changes were noted in the no break cohort (p < 0.001 and p < 0.001). Post hoc pairwise tests in the no break cohort highlighted significant changes in the blink metrics after reporting the first 10 cases compared to later cases (Supplementary Tables 1 and 2).
Diagnostic accuracy
All participants (n = 45) had a median sensitivity of 94.7% (IQR = 10.5%), a mean specificity of 85.1% (SD = 7.5%) and a mean accuracy of 87.1% (SD = 5.4%). There was no evidence of a difference in the three diagnostic accuracy measures investigated between the break and the no break groups (Table 3). Additionally, the diagnostic accuracy measures were similar for both groups when the reading session was split by case chronology (Table 3). Although not significant, a 10% reduction was observed for sensitivity in the no break cohort when comparing the performance in the first 20 cases to the second 20 cases (100% vs. 90%, p = 0.09), whereas the sensitivities in the break cohort were matched when the session was divided in this way (92.3% vs. 92.3%, p = 0.27).
Discussion
Digital breast tomosynthesis (DBT) has the potential to transform screening programmes; however, fatigue and its potential negative impacts on diagnostic accuracy need to be considered. In our study, two cohorts of mammography readers read a cancer-enriched set of 40 DBT cases, with and without breaks. Those without a break reported greater levels of subjective fatigue post reporting session (44% vs. 33%, p = 0.04) which was related to a greater blink duration and reduced peak eye-opening velocity (POV), compared to those who had breaks (blink duration: 296 ms vs. 286 ms, p < 0.001; POV: 76 mm/s vs 82 mm/s, p < 0.001). Furthermore, an increase in the blink duration and a reduction in the POV were noted as the trial progressed for the no break cohort (p < 0.001 and p < 0.001). There was no evidence of a difference in diagnostic accuracy between the cohorts (p = 0.92) or over time within either cohort (p = 0.73 and p = 0.60).
Due to the nature of the study, the first 20 participants were from specific screening centres, whom all had breaks. However, since both cohorts were equally matched in potential confounding factors, we do not expect the differences in centre location as a cause of significant findings. Notably, the readers in this study had extensive radiology experience (including DBT). Ten years was the average time in post, with half of that time reading tomosynthesis cases. This is generalisable to the current screening radiologists in the NHS [22].
Several studies in a diverse number of study environments have demonstrated that blink duration and peak eye-opening velocity are reliable markers of fatigue, noting a positive correlation between blink duration and fatigue, and a negative correlation between POV and fatigue [15,16,17,18, 23]. This was reflected in our study where blink duration was greater, and POV reduced in the no break cohort (p < 0.001 and p < 0.001). Furthermore, blink metrics were similar between both cohorts at the start of the trial (for the first 10 cases) and then increased through the trial in the no break cohort (p < 0.001 and p < 0.001), whereas the blink metrics were more stable through the trial in the break cohort. The significant changes in the blink metrics over time were most notable between the first and second 10 cases for the no break group. These findings suggest that the fatigue level of the no break participants increased considerably after reporting the first 10 cases and increased more gradually through the remainder of the session. Conversely, the fatigue level of the participants in the break cohort was more consistent through the trial, potentially related to the presence of reading breaks in their reporting sessions. Interestingly, reporting breaks were only relatively short, lasting on average 7.6 min, yet still seemed to have a marked effect on the subjective and objective fatigue measures.
We observed no evidence of difference in diagnostic accuracy between the two cohorts. This may be related to the extensive experience level of the study participants. Bernstein et al [24] recently explored the effect of time of day on DBT interpretation and reported that radiologists with 5 or fewer years of post-training experience exhibited increased recall and false positives with every consecutive hour of DBT reading (with increasing fatigue). However, there was no increased recall or false positives for radiologists with more than 5 years of experience [24]. Krupinski et al [8] investigated the performance of radiology residents and consultants in lung CT nodule detection before starting and after finishing a work shift. Fatigue measures were greater for all participants in the later session, and receiver operating characteristic analyses showed that resident performance reduced in the later session, but consultant performance actually improved from the earlier to later session [8]. These results suggest that experienced radiologists are more resistant to the negative impacts of fatigue than relative novices, and potentially a higher threshold of fatigue is required to elicit a meaningful reduction in diagnostic accuracy for these clinicians [8, 24].
Study limitations should be acknowledged. The test set only contained a relatively small number of cases enriched with challenging cancers and so is not representative of typical screening populations; consequently, reader behaviour may not be generalisable to real-world reporting. Additionally, diagnostic accuracy metrics may have been artificially high due to the Hawthorne effect [25, 26]. In the screening population, only a small number of cases are true positives, and so in a loaded malignant case set, recall behaviour may be exaggerated. Artificially high recall rates may have blunted any potential difference in diagnostic accuracy. Future studies should also include junior mammography readers. These participants will constitute a large proportion of future readers utilising DBT routinely for screening. Therefore, it would be beneficial to understand how fatigue impacts their reporting. Finally, although participants provided subjective fatigue levels using a percentage scale, a validated fatigue questionnaire could have been implemented [27].
In conclusion, a break during a 2-h DBT reporting session resulted in lower levels of subjective fatigue. Blink metrics, recognised as objective fatigue measures, demonstrated a significant increase in fatigue for participants that were not permitted breaks compared to those who were and were seen to increase significantly for participants without reporting breaks as the trial progressed. Implementing breaks did not significantly impact diagnostic accuracy in this study; however, this may be related to the experienced sample of radiologists, case mix and number of cases in the reporting session. With the potential serious, but preventable harm related to fatigue, and the growing uptake of DBT into screening programmes, it is vital to understand how fatigue manifests in mammography readers reporting with this modality. Information from these studies can help to inform clinical guidelines and standards on the optimal length of time or number of cases per reading session before onset of fatigue.
Abbreviations
- DBT:
-
Digital breast tomosynthesis
- IQR:
-
Interquartile range
- NHS:
-
National Health Service
- SD:
-
Standard deviation
References
Conant EF, Barlow WE, Herschorn SD, Weaver DL, Beaber EF (2019) Tosteson ANA, et al Association of digital breast tomosynthesis vs digital mammography with cancer detection and recall rates by age and breast density. JAMA Oncol 5:635. https://doi.org/10.1001/JAMAONCOL.2018.7078
Johnson K, Lång K, Ikeda DM, Åkesson A, Andersson I, Zackrisson S (2021) Interval breast cancer rates and tumor characteristics in the prospective population-based Malmö breast tomosynthesis screening trial. Radiology 299:559–567. https://doi.org/10.1148/RADIOL.2021204106
Heindel W, Weigel S, Gerß J et al (2022) Digital breast tomosynthesis plus synthesised mammography versus digital screening mammography for the detection of invasive breast cancer (TOSYMA): a multicentre, open-label, randomised, controlled, superiority trial. Lancet Oncol 23:601–611. https://doi.org/10.1016/S1470-2045(22)00194-2
Durand MA, Friedewald SM, Plecha DM et al (2021) False-negative rates of breast cancer screening with and without digital breast tomosynthesis. Radiol 298:296–305. https://doi.org/10.1148/RADIOL.2020202858
Houssami N, Zackrisson S, Blazek K et al (2021) Meta-analysis of prospective studies evaluating breast cancer detection and interval cancer rates for digital breast tomosynthesis versus mammography population screening. Eur J Cancer 148:14–23. https://doi.org/10.1016/J.EJCA.2021.01.035
Dang PA, Freer PE, Humphrey KL, Halpern EF, Rafferty EA (2014) Addition of tomosynthesis to conventional digital mammography: effect on image interpretation time of screening examinations. Radiology 270:49–56. https://doi.org/10.1148/RADIOL.13130765
Waite S, Kolla S, Jeudy J et al (2017) Tired in the reading room: the influence of fatigue in radiology. J Am Coll Radiol 14:191–197. https://doi.org/10.1016/J.JACR.2016.10.009
Krupinski EA, Berbaum KS, Caldwell RT, Schartz KM, Madsen MT, Kramer DJ (2012) Do long radiology workdays impact nodule detection in dynamic CT interpretation? J Am Coll Radiol 9:191. https://doi.org/10.1016/J.JACR.2011.11.013
Hanna TN, Zygmont ME, Peterson R et al (2018) The effects of fatigue from overnight shifts on radiology search patterns and diagnostic performance. J Am Coll Radiol 15:1709–1716. https://doi.org/10.1016/J.JACR.2017.12.019
Chen Y, Sudin E, Partridge G et al (2023) Measuring reader fatigue in the interpretation of screening digital breast tomosynthesis. Br J Radiol 96:1143. https://doi.org/10.1259/bjr.20220629
Partridge G, Phillips P, Darker I, Chen Y (2022) Investigating reading strategies and eye behaviours associated with high diagnostic performance when reading digital breast tomosynthesis (DBT) images. Proc. SPIE 12035, Medical Imaging 2022: Image Perception, Observer Performance, and Technology Assessment, 1203508; https://doi.org/10.1117/12.2611388
PERFORMS. iPERFORMS Web Site. n.d. https://iperforms.com/performs/ (accessed December 7, 2021).
Chen Y, Gale A (2018) Performance assessment using standardized data sets: the PERFORMS scheme in breast screening and other domains. In: Samei E, Krupinski EA, editors. The handbook of medical image perception and techniques, Cambridge University Press, p. 328–42. https://doi.org/10.1017/9781108163781.022.
Gale A, Chen Y (2020) A review of the PERFORMS scheme in breast screening. Br J Radiol;93. https://doi.org/10.1259/BJR.20190908.
Yamada Y, Kobayashi M (2018) Detecting mental fatigue from eye-tracking data gathered while watching video: evaluation in younger and older adults. Artif Intell Med 91:39–48. https://doi.org/10.1016/J.ARTMED.2018.06.005
Li J, Li H, Wang H, Umer W, Fu H, Xing X (2019) Evaluating the impact of mental fatigue on construction equipment operators’ ability to detect hazards using wearable eye-tracking technology. Autom Constr 105:102835. https://doi.org/10.1016/J.AUTCON.2019.102835
Schleicher R, Galley N, Briest S, Galley L (2008) Blinks and saccades as indicators of fatigue in sleepiness warnings: looking tired? Ergonomics 51:982–1010. https://doi.org/10.1080/00140130701817062
Martins R, Carvalho JM (2015) Eye blinking as an indicator of fatigue and mental load—a systematic review. Occupational Safety and Hygiene III - Selected Extended and Revised Contributions from the International Symposium on Safety and Hygiene:231–5. https://doi.org/10.1201/B18042-48.
Maffei A, Angrilli A (2018) Spontaneous eye blink rate: an index of dopaminergic component of sustained attention and fatigue. Int J Psychophysiol 123:58–63. https://doi.org/10.1016/J.IJPSYCHO.2017.11.009
Baccour MH, Driewer F, Kasneci E, Rosenstiel W (2019) Camera-based eye blink detection algorithm for assessing driver drowsiness. IEEE Intelligent Vehicles Symposium, Proceedings 2019; 987–93. https://doi.org/10.1109/IVS.2019.8813871.
Searjeant M, Phillips P, Roy D, Gale A, Chen Y (2021) Blink identification with eye tracking: a software processing program. Med Imaging 2021: Image Percep Observ Perform Technol Assess 11599:171–182. https://doi.org/10.1117/12.2580967
Chen Y, James JJ, Michalopoulou E, Darker IT, Jenkins J (2022) Performance of radiologists and radiographers in double reading mammograms: the UK National Health Service Breast Screening Program. Radiology https://doi.org/10.1148/RADIOL.212951
Cori JM, Anderson C, Shekari Soleimanloo S, Jackson ML, Howard ME (2019) Narrative review: do spontaneous eye blink parameters provide a useful assessment of state drowsiness? Sleep Med Rev 45:95–104. https://doi.org/10.1016/J.SMRV.2019.03.004
Bernstein MH, Baird GL, Lourenco AP (2022) Digital breast tomosynthesis and digital mammography recall and false-positive rates by time of day and reader experience. Radiology https://doi.org/10.1148/RADIOL.210318
Evans KK, Birdwell RL, Wolfe JM (2013) If you don’t find it often, you often don’t find it: why some cancers are missed in breast cancer screening. PLoS One 8:e64366. https://doi.org/10.1371/JOURNAL.PONE.0064366
Gur D, Bandos AI, Cohen CS et al (2008) The “laboratory” effect: comparing radiologists’ performance and variability during prospective clinical and laboratory mammography interpretations. Radiology https://doi.org/10.1148/radiol.2491072025
Kaida K, Takahashi M, Åkerstedt T et al (2006) Validation of the Karolinska Sleepiness Scale against performance and EEG variables. Clin Neurophysiol 117:1574–1581. https://doi.org/10.1016/J.CLINPH.2006.03.011
Acknowledgements
We would like to thank all members of the PROSPECTS trial and PROSPECTS site leads and PIs who helped organise data collection. Finally, we would like to thank all the participants who voluntarily took part in the eye tracking experiment.
Funding
We acknowledge the funding and the continued support by Hologic and King’s College Hospital.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Guarantor
The scientific guarantor of this publication is Yan Chen.
Conflict of interest
The authors of this manuscript declare no relationships with any companies, whose products or services may be related to the subject matter of the article.
Statistics and biometry
No complex statistical methods were necessary for this paper.
Informed consent
Written informed consent was obtained from all subjects (participating radiologists/radiographers) in this study.
Ethical approval
This study has undergone independent external review as part of the King’s Research and Development protocol, and has been approved by the London – Dulwich Research Ethics Committee.
Study subjects or cohorts overlap
The eye tracking data from 30 of the 45 participants included in the present investigation were analysed previously [1, 2]. These studies investigated the use of eye-blink behaviour as fatigue and cognitive markers in DBT reporting; but 21 of these 30 participants had reporting breaks in their reading sessions—a previous limitation [1, 2]. In the present study, data has been collected from a further 15 participants who were not permitted a reporting break, enabling a comparison between participants depending on whether a break was allowed.
[1]Chen Y, Sudin E, Partridge G, Taib A, Darker I, Phillips P, et al Measuring reader fatigue in the interpretation of screening digital breast tomosynthesis. British Journal of Radiology 2023 Jan 12;96(1143):20220629. doi: 10.1259/bjr.20220629.
[2]Partridge G, Phillips P, Darker I, Chen Y. Investigating reading strategies and eye behaviours associated with high diagnostic performance when reading digital breast tomosynthesis (DBT) images. Medical Imaging 2022: Image Perception, Observer Performance, and Technology Assessment 2022;12035:36–47. https://doi.org/10.1117/12.2611388.
Methodology
-
prospective
-
experimental
-
multicentre study
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
George John William Partridge and Adnan Gani Taib are joint first authors.
Supplementary Information
ESM 1
(PDF 170 kb)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Partridge, G.J.W., Taib, A.G., Phillips, P. et al. Take a break: should breaks be enforced during digital breast tomosynthesis reading sessions?. Eur Radiol 34, 1388–1398 (2024). https://doi.org/10.1007/s00330-023-10086-4
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00330-023-10086-4