Abstract
Purpose
Our objective was to develop and evaluate a Generic Integrated Objective Structured Assessment Tool (GIOSAT) to integrate Medical Expert and intrinsic (non-medical expert) CanMEDS competencies with non-technical skills for crisis simulation.
Methods
An assessment tool was designed and piloted using two pediatric anesthesia scenarios (laryngospasm and hyperkalemia). Following revision of the tool, we used previously recorded videos of anesthesia residents (n = 50) who managed one of two intraoperative advanced cardiac life support (ACLS) scenarios (ventricular tachycardia or ventricular fibrillation). Four independent trained raters, blinded to the residents’ level of training, analyzed the video recordings using the GIOSAT scale. Inter-rater reliability was calculated using intraclass correlations (ICCs) for single raters (single measure) and the average of the four raters (average measure), and construct validity was investigated by correlating GIOSAT scores with postgraduate year of residency (PGY).
Results
Total GIOSAT scores for the ACLS scenarios had single measure ICCs of 0.62 and average measure ICCs of 0.85. Inter-rater reliability was substantial for both Medical Expert and intrinsic competencies (single measure ICCs 0.69 and 0.62, respectively; average measure ICCs 0.90 and 0.82, respectively). We found significant correlations between PGY level and total GIOSAT score (r = 0.36; P = 0.011) and between PGY level and Medical Expert competencies (r = 0.42; P = 0.003); however, correlations were not found between PGY level and intrinsic CanMEDS competencies (r = 0.24; P = 0.09).
Conclusion
Inter-rater reliability of the total GIOSAT scores using four trained raters was substantial. Significant correlation between PGY and (i) total GIOSAT score and (ii) Medical Expert competencies supports construct validity. Evidence of validity was not obtained for intrinsic CanMEDS competencies.
Résumé
Objectif
Notre objectif était de mettre au point et d’évaluer un Outil d’évaluation structuré objectif intégré générique (GIOSAT - Generic Integrated Objective Structured Assessment Tool) afin d’intégrer les compétences CanMEDS d’expert médical et intrinsèques (autres que expert médical) à des habiletés non techniques pour la simulation de crise.
Méthode
Un outil d’évaluation a été conçu et mis au banc d’essai à l’aide de deux scénarios d’anesthésie pédiatrique (laryngospasme et hyperkaliémie). Après révision de l’outil, nous nous sommes servis de clips vidéo enregistrés auparavant de résidents en anesthésie (n = 50) qui avaient réussi un de deux scénarios d’ACLS (soins intensifs post-réanimation cardiaque) peropératoires (tachycardie ventriculaire ou fibrillation ventriculaire). Quatre évaluateurs formés indépendants, ne connaissant pas le niveau de formation des résidents, ont analysé les enregistrements vidéo à l’aide de l’échelle GIOSAT. La fiabilité inter-évaluateur a été calculée à l’aide de corrélations intraclasse (CIC) pour évaluateurs uniques (mesure unique) et de la moyenne des quatre évaluateurs (mesure moyenne), et la validité conceptuelle a été examinée en corrélant les scores de GIOSAT à l’année de résidence.
Résultats
Les scores GIOSAT totaux pour les scénarios d’ACLS ont eu des CIC de 0,62 en mesure unique et des CIC de 0,85 en mesure moyenne. La fiabilité inter-évaluateur était substantielle aussi bien pour les compétences d’expert médical que pour les compétences intrinsèques (CIC en mesure unique 0,69 et 0,62, respectivement; CIC en mesure moyenne de 0,90 et 0,82, respectivement). Nous avons observé des corrélations significatives entre l’année de résidence et le score total sur l’échelle de GIOSAT (r = 0,36; P = 0,011) et entre l’année de résidence et les compétences d’expert médical (r = 0,42; P = 0,003); toutefois, aucune corrélation n’a été observée entre l’année de résidence et les compétences CanMEDS intrinsèques (r = 0,24; P = 0,09).
Conclusion
La fiabilité inter-évaluateur des scores totaux sur l’échelle de GIOSAT en faisant appel à quatre évaluateurs formés était substantielle. Une corrélation significative entre l’année de résidence et (i) le score total GIOSAT et (ii) les compétences d’expert médical soutient la validité conceptuelle de notre outil. Aucune donnée probante de validité n’a été obtenue pour les compétences CanMEDS intrinsèques.
Similar content being viewed by others
Current theory in medical education emphasizes competencies that extend beyond medical knowledge and procedural skills.1 In Canada, residency training and assessment are organized according to the CanMEDS competency framework which incorporates seven competencies: the intrinsic CanMEDS roles of Communicator, Collaborator, Health Advocate, Manager, Scholar, and Professional are positioned around the Medical Expert competency which occupies the central role.2 Primary causes of adverse events in health care are more often related to communication and teamwork than they are to knowledge deficit or lack of technical skills (TSs).3-6 High-reliability industries have referred to the skills required to manage critical situations as crisis resource management or non-technical skills (NTSs).7,8 These NTSs are increasingly identified in health care and constitute a perspective on competency that parallels and overlaps the intrinsic competencies of the CanMEDS framework.
Controversy exists in the literature regarding the taxonomy, nomenclature, and definitions of the terms used for managing critical situations in health care. For high-reliability organizations, NTSs have been defined as “the cognitive, social, and personal resource skills that complement TSs and contribute to safe and efficient task performance”.9 Seven skills have been well described: situation awareness, decision-making, communication, teamwork, leadership, managing stress, and coping with fatigue. The adoption of industry concepts, such as TSs and NTSs, is problematic and has recently been questioned because of inaccuracy.10 In contrast with a dichotomous approach, the CanMEDS framework offers a more nuanced view that reflects the complex interaction and overlap of the various generic core competencies for physicians. Nevertheless, the competencies in the CanMEDS framework have not been designed specifically for crisis management.
The Medical Expert competency is traditionally assessed using written, oral, and objective structured clinical examinations (OSCEs).11 Communication and interviewing skills are also assessed using OSCEs.12 The intrinsic CanMEDS competencies as well as NTSs are generally assessed in the workplace training environment and are under-represented in formal assessments. Since it is challenging to assess crisis management skills using traditional assessment tools, simulation has been proposed as an alternative assessment tool to evaluate both technical and non-technical skills.13 This proposal has stimulated the development of various assessment tools to measure the performance of NTSs during clinical crisis simulation. These scales do not explicitly mention CanMEDS, but they overlap with the Communicator, Collaborator, and Manager competencies.8,14 The focus of other evaluation tools is almost solely on the Medical Expert role, neglecting intrinsic competencies and the NTSs discourse.15-17 In our view, CanMEDS has the requisites to serve as the conceptual framework to facilitate translation and integration of the industry concepts of TSs and NTSs to medical and health care education.
The objective of this study was to integrate CanMEDS intrinsic and Medical Expert competencies with NTSs into a Generic Integrated Objective Structured Assessment Tool (GIOSAT) that is capable of assessing residents’ performance during clinical crisis simulation and is appropriate for different specialties, scenarios, and environments. We investigated the reliability and construct validity of the GIOSAT using simulation scenarios with anesthesia residents.
Methods
Development of the tool
Using an approach similar to that of other researchers,18 we developed the GIOSAT following the steps shown in Fig 1.19 A group of five physicians (V.M.N., A.N., P.M., D.D., H.W.) from a variety of clinical backgrounds in anesthesia, intensive care, and obstetrics and with experience in postgraduate medical education defined the purpose of the assessment tool.
Two physicians and a senior investigator (V.M.N., H.W., and S.H.) performed a literature search in English and French using the internet search engines, PubMed, Medline, and Ovid, for the period from January 2000 to December 2010. Search terms included: assessment, assessment tools, simulation, crisis resource management, non-technical skills, anesthesia, critical care, surgery, emergency medicine, medicine, and pediatrics. We included articles that tested an assessment tool during crisis simulation and investigated construct validity as part of the study. We hand-searched reference lists from relevant papers and also included resulting key articles. The original search produced 828 publications, and the abstracts of these articles were screened, leaving 86 articles for full text review. Eighteen publications were finally selected. The retained articles were classified according to their main focus as NTSs,8,14,20-24 Medical Expert competency,16,17,25-30 or both.13,15,31 The assessment tools were classified as checklists and global rating scales (GRS) (Fig. 2), and items in each assessment tool were classified according to the evidence of construct validity.32 Assessment tool content was mapped to corresponding CanMEDS competencies and rated as 0 = not applicable or 1 = applicable (table available as Electronic Supplementary Material).
We used a modified Delphi process33 to design an assessment tool based on the existing literature. A draft of the assessment tool was sent to the original group, and after two iterations of the first version, the GIOSAT was developed. We designed a GRS of four points (1 = poor, 2 = marginal, 3 = acceptable, 4 = good) with free text for comments divided into two sections: Medical Expert competency with 12 items and intrinsic CanMEDS competencies with eight items.
Pilot study
After obtaining Ethics Board approval at the Children’s Hospital of Eastern Ontario Research Institute (Dec 3, 2008), we performed a pilot study (Dec 2008 to May 2009) with two video recorded pediatric anesthesia scenarios (laryngospasm, and hyperkalemia) based on the perioperative cardiac arrest and closed claims studies in pediatric anesthesia.6,34 We invited anesthesia residents (postgraduate year (PGY) 3-5) to participate in the study and obtained informed consent. Ten residents participated in the laryngospasm scenario and 14 participated in the hyperkalemia scenario.
Four trained raters independently assessed the residents’ video performance using the GIOSAT. Inter-rater intraclass correlation (ICC) was used to analyze reliability categorized according to Landis and Koch.35 The demographics of residents participating in the pilot study are shown in Table 1. The hyperkalemia scenario had moderate to substantial ICCs in all Medical Expert competencies. The Communicator, Collaborator, and Manager competencies had substantial ICCs (0.69-0.77), but there was poor reliability for the Professional competency (ICC = 0.06). Health Advocate and Scholar roles were not identified in these scenarios and were left blank by raters (Table 2). We did not undertake a formal validity analysis for this pilot study due to the small number of participants.
Validation study
The pilot study underwent critical review by an expert group of medical education researchers (D.B., S.B., V.N., and S.H.). The results of this process were used to redesign the tool, and a second study was performed to evaluate the new version. The revised version of the GIOSAT is divided into two sections: Medical Expert competencies with eight items and intrinsic competencies with six items. Each item has abbreviated descriptors and is scored with a GRS (1 = very poor to 6 = very good). Summed Medical Expert items, intrinsic items, and total GIOSAT scores are 8-48, 6-36, and 14-84, respectively (Appendix; available as Electronic Supplementary Material).
In this second study, we used video recordings from other previously reported research36 in order to examine the reliability and validity of the newly designed GIOSAT scale. Research Ethics Board approval was obtained from St. Michael’s Hospital, Toronto for the original study (October 3, 2008) comparing debriefing techniques with a pre-test/post-test design. Only pre-test video recordings were used in our analysis to avoid any bias. The sample size was calculated a priori for the original study. We used the method recommended by Cohen for an intraclass correlation of 0.7 using four raters for a power of 0.8 and found 35 participants.37 We decided to include all pre-test video recordings (n = 50). An amendment was approved by the same Research Ethics Board to use previously collected video records for our study (February 3, 2010). Written informed consent was previously obtained from the subjects.
Fifty anesthesia residents with different levels of training (PGY 2-5) were randomized to perform in one of two intraoperative advanced cardiac life support (ACLS) scenarios lasting five minutes: ventricular fibrillation due to hyperkalemia or pulseless ventricular tachycardia secondary to myocardial infarction.36 Four independent raters (three anesthesiologists and one emergency physician) from the University of Ottawa were trained to use the GIOSAT.38 Raters, blinded to subjects’ identity and PGY level of training, independently scored all residents’ performances using GIOSAT (October- November 2010). To examine the relative difficulty of the two scenarios, a comparison of the GIOSAT scores and competency scores was analyzed with Student’s t test.
Investigation of reliability
The ICC was determined with a consistency definition using a two-way random model for both single measures (individual rater) and average measures (the average of the four raters’ scores) for total GIOSAT scores, summed Medical Expert scores, summed intrinsic scores, and individual item scores.
Investigation of construct validity
Our primary outcome was the correlation between total GIOSAT scores and PGY level, as measured using Spearman’s correlation coefficient (Rho). Our secondary outcomes were correlations between PGY level and summed scores for Medical Expert items, summed scores for intrinsic items, and individual items. We hypothesized that residents’ performances would improve with level of training.
Comparison of performance in the two scenarios
The potential confounding effect of gender and type of scenario on the GIOSAT score was analyzed with a two-way analysis of variance. A Z-test was performed to compare ICCs between scenarios corrected for multiple testing using Holm’s method.39 P values < 0.05 were considered significant. We used SPSS® version 18 (SPSS, INC. 2010, Chicago, IL, USA) for the statistical analysis.
Results
Results of the validation study
The revised GIOSAT was used in the second study. Residents’ distribution according to scenario, gender, and level of training is shown in Table 3. The PGY level was similar in both scenarios, and there was an apparent imbalance in gender distribution between scenarios (Fisher’s exact test P = 0.045). No significant differences in GIOSAT scores were found between scenarios: Medical Expert scores P = 0.40; intrinsic scores P = 0.56; and total scores P = 0.54 (Student’s t test) (Table 4). Scholar was scored as not applicable (N/A) by raters in 69% of the ratings (139/200 ratings), and for that reason, it was excluded from the analysis.
Investigation of reliability
Inter-rater intraclass correlations with pooled results of both scenarios are shown in Fig 3. The ICCs were substantial for single measure summed scores of Medical Expert competencies (0.69), intrinsic competencies (0.62), and total GIOSAT scores (0.62). The single measure ICCs for individual Medical Expert items were moderate to substantial (0.43-0.69), except for examine the patient and equipment (0.29). The single measure ICCs for individual intrinsic items were fair to moderate (0.36-0.60). The average measure ICCs were substantial to almost perfect for individual items (0.61-0.90) and almost perfect for summed Medical Expert items, summed intrinsic items, and total scores (0.87-0.90).
Investigation of construct validity
For our primary outcome, we found a significant correlation between PGY and total GIOSAT scores (r = 0.36; P < 0.011) (Fig. 4). For our secondary outcomes, we found a significant correlation between PGY and summed Medical Expert competencies (r = 0.42; P < 0.003), but not for summed intrinsic competencies (r = 0.24; P = 0.09).
Comparison of performance in the two scenarios
The results of the analysis of variance showed that residents’ GIOSAT scores were not significantly influenced by scenario type or residents’ gender. We found significant differences between scenarios for single measure ICCs in two of the Medical Expert competencies, diagnosis and differentials and dealing with changing situations (P = 0.03 and 0.006, respectively), and in one intrinsic competency, Collaborator (P = 0.006). No differences between scenarios were found in average measure ICCs.
Discussion
We developed a global rating scale “GIOSAT” that integrates NTSs concepts with the CanMEDS competencies for generic assessment during crisis simulation. A pilot study was used to refine the GIOSAT through an iterative process. We found evidence of substantial reliability (single measures) for the refined tool using four raters for Medical Expert, intrinsic, and total scores. There was evidence of construct validity for the Medical Expert and total scores, but we found no evidence of construct validity for the intrinsic scores. The Medical Expert section of GIOSAT has good psychometric properties and could be used for summative assessment during simulated crises. Although the metrics for reliability and validity for the total score is acceptable, the intrinsic competencies section is problematic, which may raise concerns for total scores as well.
Non-technical skills have been described as being difficult to define10,40 and assess.8,14,20 The reliability of GIOSAT is comparable with previous studies of NTSs that have shown fair to substantial reliability. As would be expected, NTSs scales have greater reliability when their items are summed, as we have found with the GIOSAT. Our finding of poor reliability and construct validity of some of the intrinsic competencies not included in the NTSs discourse is in keeping with the literature on professionalism which is not often taught formally (i.e., part of the hidden curriculum) and is difficult to assess.41,42
Our study has a number of limitations. The data are based on a re-analysis of scenarios not designed to identify certain competencies, such as Professional, Health Advocate, and Scholar. It may also be that the ACLS scenarios were too short to identify all CanMEDS competencies fully. Increasing the number of scenarios and changing scenario design may improve construct validity of the intrinsic section. It has also been shown that reliability improves with increased testing time, for instance, it has been shown in OSCEs that several hours of testing and ten or more cases are required for high-stakes examinations.12
A further limitation is the possibility that intrinsic competencies are underrepresented in GIOSAT and the Medical Expert role is overrepresented. Future iterations of the GIOSAT tool may emphasize the intrinsic competencies either in terms of the number of items or the extensiveness of the descriptors. According to current concepts, reliability and validity are not properties of the instrument but properties of the instrument’s scores and interpretations.43 The same instrument used in a different setting or with different subjects may produce different results; thus, our results may not necessarily be generalizable to other populations.
The GIOSAT Medical Expert competencies section relating to psychometric properties is appropriate to be used for summative assessment. The intrinsic competencies section and, by implication, the scale as a whole are not yet appropriate for summative assessment.
The aim of future research will be to identify the number of raters, scenarios, and examinees necessary to establish reliability and generalizability. The development of scenarios specifically designed to challenge specific domains (Professional, Health Advocate and Scholar: PHAS Roles) may be required for the appropriate use of GIOSAT for summative assessment.
References
The Royal College of Physicians and Surgeons of Canada. CanMEDS 2005 Framework. CanMEDS: better standards, better physicians, better care. Available from URL: http://www.rcpsc.medical.org (accessed November 2012).
Sherbino J, Frank JR, Flynn L, Snell L. “Intrinsic roles” rather than “armour”: renaming the “non-medical expert roles” of the CanMEDS framework to match their intent. Adv Health Sci Educ Theory Pract 2011; 16: 695-7.
Kohn LT, Corrigan JM, Donaldson MS. To Err is Human: Building a Safer Health Care System. Washington, DC: National Academy Press; 1999 .
Canadian Institute for Health Information. Health Care in Canada, 2004. Available from URL: www.cihi.ca (accessed November 2012).
McIlvaine WB. Human error and its impact on anesthesiology. Semin Anesth 2006; 25: 172-9.
Jimenez N, Posner KL, Cheney FW, Caplan RA, Lee LA, Domino KB. An update on pediatric anesthesia liability: a closed claims analysis. Anesth Analg 2007; 104: 147-53.
Gaba DM, Howard SK, Fish KJ, Smith BE, Sowb YA. Simulation-based training in anesthesia crisis resource management (ACRM): a decade of experience. Simulation & Gaming 2001; 32: 175-93.
Fletcher G, Flin R, McGeorge P, Glavin R, Maran N, Patey R. Anesthetists’ Non-Technical Skills (ANTS): evaluation of a behavioral marker system. Br J Anaesth 2003; 90: 580-8.
Flin R, O’Connor P, Crichton M. Safety at the Sharp End - A Guide to Non-Technical Skills. Farnham: England, Ashgate Publishing Limited; 2008: 1-16.
Nestel D, Walker K, Simon R, Aggarwal R, Andreatta P. Nontechnical skills: an inaccurate and unhelpful descriptor? Simul Healthc 2011; 6: 2-3.
Bandiera G, Sherbino J, Frank JR. The CanMEDS Assessment Tools Handbook. An Introductory Guide to Assessment Methods for the CanMEDS Competencies. Ottawa: The Royal College of Physicians and Surgeons of Canada; 2006 .
Hawkins RE, Boulet JR. Direct observation: standardized patients. In: Holmboe ES, Hawkins RE, editors. Practical Guide to the Evaluation of Clinical Competence. Philadelphia: Mosby Elsevier; 2008. p. 102-18.
Savoldelli GL, Nail VN, Joo HS, et al. Evaluation of patient simulator performance as an adjunct to the oral examination for senior anesthesia residents. Anesthesiology 2006; 104: 475-81.
Kim J, Neilipovitz D, Cardinal P, Chiu M, Clinch J. A pilot study using high-fidelity simulation to formally evaluate performance in the resuscitation of critically ill patients: The University of Ottawa Critical Care Medicine, High Fidelity Simulation, and Crisis Resource Management I Study. Crit Care Med 2006; 34: 2167-74.
Brett-Fleegler MB, Vinci RJ, Weiner DL, Harris SK, Shih MC, Kleinman ME. A simulator-based tool that assesses pediatric resident resuscitation competency. Pediatrics 2008; 121: e597-603.
Morgan PJ, Cleave-Hogg D, DeSousa S, Tarshis J. High-fidelity patient simulation: validation of performance checklists. Br J Anaesth 2004; 92: 388-92.
Murray DJ, Boulet JR, Kras JF, Woodhouse JA, Cox T, McAllister JD. Acute care skills in anesthesia practice: a simulation-based resident performance assessment. Anesthesiology 2004; 101: 1084-95.
Ringsted C, Hodges B, Scherpbier A. ‘The research compass’: an introduction to research in medical education: AMEE Guide No. 56. Med Teach 2011; 33: 695-709.
Hamstra SJ. Key note address: the focus on competencies and individual learner assessment as emerging themes in medical education research. Acad Emerg Med 2012; 19: 1336-43.
Graham J, Hocking G, Giles E. Anaesthesia non-technical skills: can anaesthetists be trained to reliably use this behavioural marker system in 1 day? Br J Anaesth 2010; 104: 440-5.
Yule S, Flin R, Maran N, Rowley D, Youngson G, Paterson-Brown S. Surgeons’ non-technical skills in the operating room: reliability testing of the NOTSS behavior rating system. World J Surg 2008; 32: 548-56.
Jankouskas T, Bush MC, Murray B, et al. Crisis resource management: evaluating outcomes of a multidisciplinary team. Simul Healthc 2007; 2: 96-101.
Guise JM, Deering SH, Kanki BG, et al. Validation of a tool to measure and promote clinical teamwork. Simul Healthc 2008; 3: 217-23.
Malec JF, Torsher LC, Dunn WF, et al. The Mayo high performance teamwork scale: reliability and validity for evaluating key crew resource management skills. Simul Healthc 2007; 2: 4-10.
Boulet JR, Murray D, Kras J, Woodhouse J, McAllister J, Ziv A. Reliability and validity of a simulation-based acute care skills assessment for medical students and residents. Anesthesiology 2003; 99: 1270-80.
Chopra V, Gesink BJ, de Jong J, Bovill JG, Spierdijk J, Brand R. Does training on an anaesthesia simulator lead to improvement in performance? Br J Anaesth 1994; 73: 293-7.
Scavone BM, Sproviero MT, McCarthy RJ, et al. Development of an objective scoring system for measurement of resident performance on the human patient simulator. Anesthesiology 2006; 105: 260-6.
Devitt JH, Kurrek MM, Cohen MM, Cleave-Hogg D. The validity of performance assessments using simulation. Anesthesiology 2001; 95: 36-42.
Murray DJ, Boulet JR, Avidan M, et al. Performance of residents and anesthesiologists in a simulation-based skill assessment. Anesthesiology 2007; 107: 705-13.
Murray DJ, Boulet JR, Kras JF, McAllister JD, Cox TE. A simulation-based acute skills performance assessment for anesthesia training. Anesth Analg 2005; 101: 1127-34.
Gaba DM, Howard SK, Flanagan B, Smith BE, Fish KJ, Botney R. Assessment of clinical performance during simulated crisis using both technical and behavioral ratings. Anesthesiology 1998; 89: 8-18.
Kane M, Crooks T, Cohen A. Validating measures of performance. Educational Measurement: Issues and Practice 1999; 18: 5-17.
Jones J, Hunter D. Consensus methods for medical and health services research. BMJ 1995; 311: 376-80.
Bhananker SM, Ramamoorthy C, Geiduscheck JM, et al. Anesthesia-related cardiac arrest in children: update from the Pediatric Perioperative Cardiac Arrest Registry. Anesth Analg 2007; 105: 344-50.
Landis JR, Koch GG. The measurement of observer agreement for categorical data. Biometrics 1977; 33: 159-74.
Boet S, Bould MD, Bruppacher HR, Desjardins F, Chandra DB, Naik VN. Looking in the mirror: self-debriefing versus instructor debriefing for simulated crises. Crit Care Med 2011; 39: 1377-81.
Cohen J. Statistical Power Analysis for the Behavioral Sciences, Second Edition. Hillsdale, NJ: Lawrence Erlbaum Associates, Inc., Publishers; 1988.
American Heart Association. Part 4: Advanced life support. Circulation 2005; 112: III25-54.
Holm S. A simple sequentially rejective multiple test procedure. Scand J Stat 1979; 6: 65-70.
Glavin RJ. Skills, training and education. Simul Healthc 2011; 6: 4-7.
Lynch DC, Surdyk PM, Eiser AR. Assessing professionalism: a review of the literature. Med Teach 2004; 26: 366-73.
Bahaziq W, Crosby E. Physician professional behaviour affects outcomes: a framework for teaching professionalism during anesthesia residency. Can J Anesth 2011; 58: 1039-50.
Cook DA, Beckman TJ. Current concepts in validity and reliability for psychometric instruments: theory and application. Am J Med 2006; 119: 166.e7-16.
Acknowledgements
The authors sincerely thank Kevin Thomas and Drs Michelle Chiu, Simone Crooks, George Dumitrascu, and Robert Elliott for facilitating the pilot study. We also thank the residents of the University of Toronto and the University of Ottawa for their participation in the study.
Funding
This study was supported in part by an education grant from the Postgraduate Office and funds from the Department of Anesthesiology, University of Ottawa.
Competing interests
None declared.
Author information
Authors and Affiliations
Corresponding author
Additional information
Author contributions
Victor M. Neira was the principal investigator and corresponding author. He was involved in data extraction and analysis, pilot study design, coordination and communication with the other investigators, and validation of the study. He contributed to the presentation, writing, editing, and correction of all drafts and the final paper. Stanley J. Hamstra was the senior investigator and guarantor. He was involved in the critical review and supervision in all steps of the investigation, including the development of the rating scale, pilot study, validation study, development of drafts and editions, and final paper presentation. Victor M. Neira and Hilary Writer were involved in the literature review. Hilary Writer was involved in the literature search and contributed to the design of the first version of the scale. Dylan M. Bould, Sylvain Boet, and Viren Naik were involved in the critical review of the pilot study results, the design of the second version of the scale, acquisition of data, and analysis and interpretation of the validation study. Amy Nakajima, Philipp Mossdorf, and Dermot Doherty were responsible for the conception and design of the first revision of the scale. Amy Nakajima, Philipp Mossdorf, Amy Roeske, Devin Sydor, Stephen Noseworthy, and Dermot Doherty were involved in the acquisition of the data (rater) and analysis and interpretation of the pilot study. Dylan M. Bould, Sylvain Boet, Viren Naik, Amy Nakajima, Philipp Mossdorf, Amy Roeske, Devin Sydor, Stephen Noseworthy, Dermot Doherty, and Hilary Writer were involved in critical revision of the paper content. Nicholas Barrowman was involved in the statistic analysis and the interpretation of the pilot and validation studies. He also contributed graphics for the final paper.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Rights and permissions
About this article
Cite this article
Neira, V.M., Bould, M.D., Nakajima, A. et al. “GIOSAT”: a tool to assess CanMEDS competencies during simulated crises. Can J Anesth/J Can Anesth 60, 280–289 (2013). https://doi.org/10.1007/s12630-012-9871-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12630-012-9871-9