Abstract
The Objective Structured Clinical Exam (OSCE) is a widely used method of assessment in medical education. Rater cognition has become an important area of inquiry in the medical education assessment literature generally, and in the OSCE literature specifically, because of concerns about potential compromises of validity. In this study, a novel approach to mixed methods that combined Ordinal Logistic Hierarchical Linear Modeling and cognitive interviews was used to gain insights about what examiners were thinking during an OSCE. This study is based on data from the 2010 to 2014 administrations of the Clinician Assessment for Practice Program OSCE for International Medical Graduates (IMGs) in Nova Scotia. An IMG is a physician trained outside of Canada who was a licensed practitioner in a different country. The quantitative data were examined alongside four follow-up cognitive interviews of examiners conducted after the 2014 administration. The quantitative results show that competencies of (1) Investigation and Management and (2) Counseling were highly predictive of the Overall Global score. These competencies were also described in the cognitive interviews as the most salient parts of OSCE. Examiners also found Communication Skills and Professional Behavior to be relevant but the quantitative results revealed these to be less predictive of the Overall Global score. The interviews also reveal that there is a tacit sequence by which IMGs are expected to proceed in an OSCE, starting with more basic competencies such as History Taking and building up to Investigation Management and Counseling. The combined results confirm that a hidden pattern exists with respect to how examiners rate candidates. This study has potential implications for research into rater cognition, and the design and scoring of practice-ready OSCEs.
This is a preview of subscription content, access via your institution.



References
Ambady, N., & Rosenthal, R. (1992). Thin slices of expressive behavior as predictors of interpersonal consequences: A meta-analysis. Psychological Bulletin, 111(2), 256–274. doi:10.1037/0033-2909.111.2.256.
Bejar, I. I. (2012). Rater cognition: Implications for validity. Educational Measurement: Issues and Practice, 31(3), 2–9. doi:10.1111/j.1745-3992.2012.00238.x.
Berendonk, C., Stalmeijer, R. E., & Schuwirth, L. W. T. (2012). Expertise in performance assessment: assessors’ perspectives. Advances in Health Science Education, 18, 559–571. doi:10.1007/s10459-012-9392-x.
Beretvas, S. N., & Kamata, A. (2005). The multilevel measurement model: Introduction to the special issue. Journal of Applied Measurement, 6(3), 247–254.
Bobko, P., Roth, P. L., & Buster, M. A. (2007). The usefulness of unit weights in creating composite scores: A literature review, application to content validity, and meta-analysis. Organizational Research Methods, 10(4), 689–709. doi:10.1177/1094428106294734.
Boulet, J. R., Cooper, R. A., Seeling, S. S., Norcini, J. J., & McKinley, D. W. (2009). U.S. citizens who obtain their medical degrees abroad: An overview, 1992–2006. Health Affairs, 28(1), 226–233. doi:10.1377/hlthaff.28.1.226.
Boursicot, K. A. M., & Burdick, W. P. (2014). Structured assessments of clinical competence. In T. Swanwick (Ed.), Understanding medical education: Evidence, theory and practice (2nd ed., pp. 293–304). New York: Wiley.
Brennan, R. L. (2001). An essay on the history and future of reliability from the perspective of replications. Journal of Educational Measurement, 38(4), 295–317. doi:10.1111/j.1745-3984.2001.tb01129.x.
Canadian Institute for Health Information. (2009, August). International Medical Graduates in Canada: 1972 to 2007 Executive Summary. Retrieved February 1, 2015 from http://secure.cihi.ca/free_products/img_1972-2007_aib_e.pdf.
Corp, I. B. M. (2012). IBM SPSS Statistics for Windows, Version 21.0. Armonk, NY: IBM Corp.
Cox, M., Irby, D. M., & Epstein, R. M. (2007). Assessment in medical education. New England Journal of Medicine, 356(4), 387–396. doi:10.1056/NEJMra054784.
CRAN. (2015). R 3.1.3 “Smooth Sidewalk”. http://cran.r-project.org/.
Creswell, J. W., Klassen, A. C., Plano Clark, V. L., & Smith, K. C. (2011, August) for the Office of Behavioral and Social Sciences Research. Best practices for mixed methods research in the health sciences. National Institutes of Health. Retrieved August 1, 2015 from http://obssr.od.nih.gov/mixed_methods_research/pdf/Best_Practices_for_Mixed_Methods_Research.pdf.
Crisp, V. (2012). An investigation of rater cognition in the assessment of projects. Educational Measurement: Issues and Practice, 31(3), 10–20. doi:10.1111/j.1745-3992.2012.00239.x.
Cronbach, L. J., Gleser, G. C., Nanda, H., & Rajaratnam, N. (1972). The dependability of behavioral measurements: Theory of generalizability for scores and profiles. New York: Wiley.
Cronbach, L. J., & Meehl, P. E. (1955). Construct validity in psychological tests. Psychological Bulletin, 52(4), 281–302.
Douglas, S., & Selinker, L. (1992). Analyzing oral proficiency test performance in general and specific purpose contexts. System, 20(3), 317–328. doi:10.1016/0346-251x(92)90043-3.
Eckes, T. (2012). Operational rater types in writing assessment: Linking rater cognition to rater behavior. Language Assessment Quarterly, 9, 270–292. doi:10.1080/15434303.2011.649381.
Epstein, R. M., & Hundert, E. M. (2002). Defining and assessing professional competence. Jama, 287(2), 226–235.
Fuller, R., Homer, M., & Pell, G. (2013). Longitudinal interrelationships of OSCE station level analyses, quality improvement and overall reliability. Medical Teacher, 35, 515–517. doi:10.3109/0142159X.2013.775415.
Gingerich, A., & Eva, K. W. (2011). Rater-based assessments as social judgments: Rethinking the etiology of rater errors. Academic Medicine, 86, S1–S7. doi:10.1097/ACM.0b013e31822a6cf8.
Gingerich, A., Kogan, J., Yeates, P., Govaerts, M., & Holmboe, E. (2014a). Seeing the “black box” differently: Assessor cognition from three research perspectives. Medical Education, 48, 1055–1068. doi:10.1111/medu.12546.
Gingerich, A., van der Vleuten, C. P. M., & Eva, K. W. (2014b). More consensus than idiosyncrasy: Categorizing social judgments to examine variability in Mini-CEX ratings. Academic Medicine, 89, 1510–1519. doi:10.1097/ACM.0000000000000486.
Goldstein, H. (1986). Multilevel mixed linear model analysis using iterative generalized least squares. Biometrika, 73(1), 43–56. doi:10.1093/biomet/73.1.43.
Hodges, B., & McIlroy, J. H. (2003). Analytic global OSCE ratings are sensitive to level of training. Medical Education, 37, 1012–1016.
Hodges, B., Regehr, G., McNaughton, N., Tiberius, R., & Hanson, M. (1999). OSCE checklists do not capture increasing levels of expertise. Academic Medicine, 74, 1129–1134.
Joe, J. N., Harmes, J. C., & Hickerson, C. A. (2011). Using verbal reports to explore rater perceptual processes in scoring: A mixed methods application to oral communication assessment. Assessment in Education: Principles, Policy & Practice, 18, 239–258. doi:10.1080/0969594X.2011.577408.
Johnston, J. L., Lundy, G., McCullough, M., & Gormley, G. J. (2013). The view from over there: Reframing the OSCE through the experience of standardised patient raters. Medical Education, 47(9), 899–909. doi:10.1111/medu.12243.
Kamata, A. (2001). Item analysis by the hierarchical generalized linear model. Journal of Educational Measurement, 38(1), 79–93. doi:10.1111/j.1745-3984.2001.tb01117.x.
Kamata, A., Bauer, D. J., & Miyazaki, Y. (2008). Multilevel measurement modeling. In A. A. O’Connell & D. B. McCoach (Eds.), Multilevel modeling of educational data (pp. 345–390). Charlotte, NC: Information Age Publishing.
Kane, M. T. (1992). The assessment of professional competence. Evaluation and the Health Professions, 15(2), 163–182.
Kane, M. T. (2013). Validation as a pragmatic, scientific activity. Journal of Educational Measurement, 50(1), 115–122. doi:10.1111/jedm.12007.
Kane, M. T., & Bejar, I. I. (2014). Cognitive frameworks for assessment, teaching, and learning: A validity perspective. Psicología Educativa, 20(2), 117–123. doi:10.1016/j.pse.2014.11.006.
Kelley, T. L. (1927). Interpretation of educational measurements. New York: World Book Co. Retrieved February 1, 2014 from http://hdl.handle.net/2027/mdp.39015001994071.
Khan, K. Z., Gaunt, K., Ramachandran, S., & Pushkar, P. (2013). The objective structured clinical examination (OSCE): AMEE Guide No. 81. Part II: Organisation & Administration. Medical Teacher, 35(9), e1447–e1463. doi:10.3109/0142159X.2013.818635.
Kishor, N. (1990). The effect of cognitive complexity on halo in performance judgment. Journal of Personnel Evaluation in Education, 3, 377–386.
Kishor, N. (1995). The effect of implicit theories on raters’ inference in performance judgment: Consequences for the validity of student ratings of instruction. Research in Higher Education, 36(2), 177–195. doi:10.1007/BF02207787.
Kogan, J. R., Conforti, L., Bernabeo, E., Iobst, W., & Holmboe, E. (2011). Opening the black box of clinical skills assessment via observation: A conceptual model. Medical Education, 45(10), 1048–1060. doi:10.1111/j.1365-2923.2011.04025.x.
Liao, S. C., Hunt, E. A., & Chen, W. (2010). Comparison between inter-rater reliability and inter-rater agreement in performance assessment. Annals of the Academy of Medicine, Singapore, 39(8), 613–618.
Linacre, J. M., & Wright, B. D. (2002). Construction of measures from many-facet data. Journal of Applied Measurement, 3(4), 486–512.
MacLellan, A.-M., Brailovsky, C., Rainsberry, P., Bowmer, I., & Desrochers, M. (2010). Examination outcomes for international medical graduates pursuing or completing family medicine residency training in Quebec. Canadian Family Physician, 56(9), 912–918.
Maudsley, R. (2008). Assessment of international medical graduates and their integration into family practice: The clinical assessment for practice program. Academic Medicine, 83, 309–315.
Medical Council of Canada. (2013, November). Guidelines for the development of objective structured clinical examination (OSCE) cases. Retrieved February 1, 2015, from http://mcc.ca/wp-content/uploads/osce-booklet-2014.pdf.
Messick, S. (1975). The standard problem: Meaning and values in measurement and evaluation. American Psychologist, 30(10), 955–966. doi:10.1037/0003-066X.30.10.955.
Messick, S. (1994). The interplay of evidence and consequences in the validation of performance assessments. Educational Researcher, 23(2), 13–23. doi:10.3102/0013189X023002013.
Miles, M. B., Huberman, A. M., & Saldana, J. (2014). Qualitative data analysis: A methods sourcebook (3rd ed.). Thousand Oaks: Sage.
Mislevy, R. J. (1993). Foundations of a new test theory. In N. Frederikson, R. Mislevy, & I. I. Bejar (Eds.), Test theory for a new generation of tests (pp. 19–49). Hilllsdale, NJ: Lawrence Erlbaum Associates.
Newble, D. (2004). Techniques for measuring clinical competence: Objective structured clinical examinations. Medical Education, 38(2), 199–203. doi:10.1046/j.1365-2923.2004.01755.x.
Norcini, J. J., Boulet, J. R., Opalek, A., & Dauphinee, W. D. (2014). The relationship between licensing examination performance and the outcomes of care by international medical school graduates. Academic Medicine, 89, 1157–1162. doi:10.1097/ACM.0000000000000310.
Osborne, J. W. (2000). Advantages of hierarchical linear modeling. Practical Assessment, Research & Evaluation, 7(1). Retrieved February 6, 2015 from http://PAREonline.net/getvn.asp?v=7&n=1.
Page, G., Bordage, G., & Allen, T. (1995). Developing key-feature problems and examinations to assess clinical decision-making skills. Academic Medicine, 70(3), 194.
Raudenbush, S., & Bryk, A. S. (1986). A hierarchical model for studying school effects. Sociology of Education, 59(1), 1–17. doi:10.2307/2112482.
Raudenbush, S. W., Bryk, A. S., & Congdon, R. (2004). HLM 6 for Windows. Skokie, IL: Scientific Software International, Inc.
Regehr, G., Eva, K., Ginsburg, S., Halwani, Y., & Sidhu, R. (2011). Assessment in postgraduate medical education: Trends and issues in assessment in the workplace (Members of the FMEC PG consortium). Retrieved February 1, 2015 from https://www.afmc.ca/pdf/fmec/13_Regehr_Assessment.pdf.
Regehr, G., MacRae, H., Reznick, R. K., & Szalay, D. (1998). Comparing the psychometric properties of checklists and global rating scales for assessing performance on an OSCE-format examination. Academic Medicine, 73(9), 993–997.
Sandilands, D. D., Gotzmann, A., Roy, M., Zumbo, B. D., & de Champlain, A. (2014). Weighting checklist items and station components on a large-scale OSCE: Is it worth the effort? Medical Teacher, 36(7), 585–590. doi:10.3109/0142159X.2014.899687.
Shepard, L. A. (1997). The centrality of test use and consequences for test validity. Educational Measurement: Issues and Practice, 16(2), 5–24. doi:10.1111/j.1745-3992.1997.tb00585.x.
ten Cate, O., Snell, L., & Carraccio, C. (2010). Medical competence: The interplay between individual ability and the health care environment. Medical Teacher, 32(8), 669–675. doi:10.3109/0142159X.2010.500897.
Toops, H. A. (1927). The selection of graduate assistants. Personnel Journal (Pre-1986), 6, 457–472.
van der Vleuten, C. P. M. (1996). The assessment of professional competence: Developments, research and practical implications. Advances in Health Science Education, 1(1), 41–67. doi:10.1007/BF00596229.
van der Vleuten, C. P. M., & Schuwirth, L. W. T. (2005). Assessing professional competence: From methods to programmes. Medical Education, 39(3), 309–317. doi:10.1111/j.1365-2929.2005.02094.x.
Walsh, A., Banner, S., Schabort, I., Armson, H., Bowmer, M. I., & Granata, B. (2011). International Medical Graduates—Current issues (Members of the FMEC PG consortium). Retrieved February 1, 2015 from http://www.afmc.ca/pdf/fmec/05_Walsh_IMG%20Current%20Issues.pdf.
Wickham, H., & Chang, W. (2015). Ggplot2: An implementation of the grammar of graphics, Version 1.0.1. http://cran.r-project.org/web/packages/ggplot2/index.html.
Williams, R. G., Klamen, D. A., & McGaghie, W. C. (2003). Cognitive, social and environmental sources of bias in clinical performance ratings. Teaching and Learning in Medicine, 15(4), 270–292. doi:10.1207/S15328015TLM1504_11.
Willis, G. B. (2005). Cognitive interviewing: A tool for improving questionnaire design. Thousand Oaks: Sage.
Wolfe, E. W. (2004). Identifying rater effects using latent trait models. Psychology Science, 46(1), 35–51.
Wolfe, E. W. (2006). Uncovering rater’s cognitive processing and focus using think-aloud protocols. Journal of Writing Assessment., 2(1), 37–56. http://www.journalofwritingassessment.org/archives/2-1.4.pdf.
Wong, G. Y., & Mason, W. M. (1985). The hierarchical logistic regression model for multilevel analysis. Journal of the American Statistical Association, 80(391), 513–524. doi:10.2307/2288464.
Wood, T. J. (2014). Exploring the role of first impressions in rater-based assessments. Advances in Health Science Education, 19, 409–427. doi:10.1007/s10459-013-9453-9.
Acknowledgments
The authors wish to acknowledge the support of the College of Physicians & Surgeons of Nova Scotia in conducting the research study. The authors wish to acknowledge Susan Elgie, Dr. Lorelei Lingard, and Dr. Tomoko Arimura for their support in reviewing this manuscript.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
The Ordinal Logistic Hierarchical Linear Modeling (OLHLM) is slightly more complex as the outcome variable is an ordered categories and not continuous data. OLHLM is very similar to Logistic Hierarchical Linear Modeling, where the outcome is binary, and we estimating likelihood of receiving very good or poor ratings. Since there are four categories (poor, borderline, satisfactory, very good) of probability in our case three functions (1–3) are used and propensity of being in each of the categories is based on the change in log odds. The model below, represents a three level OLHLM with 4 category outcomes, 3 of the 4 categories are estimated, as it assumed that the 4th category is 1—the probability of being in the other three.
Station level model
Candidate level
Year level
Its important that when reading an HLM model to look at the different levels. In our case since there are no person characteristics at level 2 and no year characteristics at level 3, we are estimating the error variance at each of those levels and the competency coefficients. As such the:
\( \phi_{ijm(1)}^{\prime } \) is the probability that person j in year m scores a 1; \( \phi_{ijm(2)}^{\prime } \) is the probability that person j in year m scores a 1 or 2; \( \phi_{ijm(3)}^{\prime } \) is the probability that person j in year m scores a 1, 2 or 3; \( \pi_{ojm} \) is the intercept term; \( \pi_{kjm} \) is the coefficient for the kth competency for person j in year m; \( X_{kjm} \) is the kth competency score for person j in year m; \( \delta_{(2)} \) is the threshold value between category 2 and 1; \( \delta_{(3)} \) is the threshold value between category 3 and 2; \( r_{0jm} \) is the random component of \( \pi_{0jm} \); \( \beta_{00m} \) is an effect of the reference competency in year m; \( \beta_{k0m} \) is an effect of the kth competency in year m; \( u_{00m} \) is the random component of \( \beta_{00m} \); \( \gamma_{000} \) is the overall effects of competencies; \( \gamma_{k00} \) is the coefficient for competency k.
The HLM program uses the above model and provides estimates for each of the values above. However it does not provide the probability estimates of a candidate belonging to the poor, borderline, satisfactory or very good category at each station. These probability estimates need to be calculated. The following formulas are used to calculate the probability of being in each of the categories.
\( \phi '_{ijm\left( 1 \right)} \) denotes the probability of being in the poor category, when the \( X_{kjm} \) (i.e. competency scores) are at the average (i.e. set to zero). To calculate \( \phi '_{ijm\left( 4 \right)} \) the probability of being in the very good category we subtract 1 from the probability of being in the satisfactory, borderline, or very good categories.
Rights and permissions
About this article
Cite this article
Chahine, S., Holmes, B. & Kowalewski, Z. In the minds of OSCE examiners: uncovering hidden assumptions. Adv in Health Sci Educ 21, 609–625 (2016). https://doi.org/10.1007/s10459-015-9655-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10459-015-9655-4