Abstract
Objective Structured Clinical Examinations (OSCEs) have become ubiquitous as a form of assessment in medical education but involve substantial resource demands and considerable local variation. A detailed understanding of the processes by which OSCEs are designed and administered could improve feasibility and sustainability. This exploration of OSCE design is informed by Practice Theory, which suggests assessment design processes are dynamic, social and situated activities. The overall purpose is to provide insights that inform on-the-ground OSCE administration. Fifteen interviews were conducted with OSCE academics and administrators from three medical schools in Australia, the United Kingdom and Canada. Drawing from post-qualitative inquiry, Schatzki’s Practice Theory was used both as a sensibility and as an analytic framework. OSCE design was characterised by planning activities, administration activities, negotiation activities and bureaucratic activities; it involves significant and resource-intensive effort in negotiation and coordination. There was considerable local variation but at the same time activities were remarkably consonant across national boundaries. There was a tension between general understandings such as reliability and validity that underpin the OSCE and the improvisational practices associated with design and administration. Our findings highlighted the role of blueprints as a key coordinating artefact but with too many rules and procedures prompting cycles of bureaucracy and complexity. Emphasising coordination rather than standardisation might ease workloads, support adaptation to local environments and prevent an overly reductive approach to this assessment format.
Similar content being viewed by others
References
Bearman, M., & Ajjawi, R. (2018a). Actor-network theory and the OSCE: formulating a new research agenda for a post-psychometric era. Advances in Health Sciences Education, 23(5), 1037–1049. https://doi.org/10.1007/s10459-017-9797-7
Bearman, M., & Ajjawi, R. (2018b). From ‘seeing through’to ‘seeing with’: assessment criteria and the myths of transparency. Frontiers in Education,. https://doi.org/10.3389/feduc.2018.00096.
Bearman, M., Dawson, P., Bennett, S., Hall, M., Molloy, E., et al. (2017). How university teachers design assessments: A cross-disciplinary study. Higher Education, 74(1), 49–64. https://doi.org/10.1007/s10734-016-0027-7
Bevan, J., Russell, B., & Marshall, B. (2019). A new approach to OSCE preparation - PrOSCEs. BMC Medical Education, 19(1), 126. https://doi.org/10.1186/s12909-019-1571-5.
Boud, D., Dawson, P., Bearman, M., Bennett, S., Joughin, G., et al. (2018). Reframing assessment research: through a practice perspective. Studies in Higher Education, 43(7), 1107–1118. https://doi.org/10.1080/03075079.2016.1202913.
Brown, C., Ross, S., Cleland, J., & Walsh, K. (2015). Money makes the (medical assessment) world go round: The cost of components of a summative final year Objective Structured Clinical Examination (OSCE). Medical Teacher, 37(7), 653–659. https://doi.org/10.3109/0142159X.2015.1033389.
Brown, J. S., & Duguid, P. (2001). Knowledge and organization: A social-practice perspective. Organization Science, 12(2), 198–213. https://doi.org/10.1287/orsc.12.2.198.10116
Catchpole, K., & Russ, S. (2015). The problem with checklists. BMJ Quality & Safety, 24(9), 545–549. https://doi.org/10.1136/bmjqs-2015-004431
Criscione-Schreiber, L. (2020). Turning objective structured clinical examinations into reality. Rheumatic Disease Clinics, 46(1), 21–35. https://doi.org/10.1016/j.rdc.2019.09.010
Fenwick, T., & Landri, P. (2012). Materialities, textures and pedagogies: socio-material assemblages in education. Pedagogy, Culture & Society, 20(1), 1–7. https://doi.org/10.1080/14681366.2012.649421.
Gherardi, S., & Perrotta, M. (2014). Between the hand and the head: How things get done, and how in doing the ways of doing are discovered. Qualitative Research in Organizations and Management: An International Journal, 9(2), 134–150. https://doi.org/10.1108/QROM-06-2012-1079.
Gormley, G., Hodges, B. D., McNaughton, N., & Johnston, J. L. (2016). The show must go on? Patients, props and pedagogy in the theatre of the OSCE. Medical Education, 50(12), 1237–1240. https://doi.org/10.1111/medu.13016.
Harden, R. M., & Gleeson, F. A. (1979). Assessment of clinical competence using an objective structured clinical examination (OSCE). Medical Education, 13(1), 39–54. https://doi.org/10.1111/j.1365-2923.1979.tb00918.x.
Heal, C., D’Souza, K., Banks, J., Malau-Aduli, B. S., Turner, R., et al. (2019). A snapshot of current Objective Structured Clinical Examination (OSCE) practice at Australian medical schools. Medical Teacher, 41(4), 441–447. https://doi.org/10.1080/0142159X.2018.1487547.
Hodges, B. D. (2003). Validity and the OSCE. Medical Teacher, 25(3), 250–254. https://doi.org/10.1080/01421590310001002836.
Jackson, A. Y. (2013). Posthumanist data analysis of mangling practices. International Journal of Qualitative Studies in Education, 26(6), 741–748. https://doi.org/10.1080/09518398.2013.788762.
Kearney, G. P., Johnston, J. L., Hart, N. D., Corman, M. K., & Gormley, G. J. (2018). Protocol: Exploring the Objective Structured Clinical Examination (OSCE) using Institutional Ethnography. International Journal of Educational Research, 88, 42–47. https://doi.org/10.1016/j.ijer.2018.01.002.
Khan, K. Z., Gaunt, K., Ramachandran, S., & Pushkar, P. (2013). The Objective Structured Clinical Examination (OSCE): AMEE Guide No. 81. Part II: Organisation & Administration. Medical Teacher. https://doi.org/10.3109/0142159x.2013.818635
Lather, P., & St. Pierre, E. A. (2013). Post-qualitative research. International Journal of Qualitative Studies in Education, 26(6), 629–633. https://doi.org/10.1080/09518398.2013.788752
Mazzei, L.A., & Jackson, A.Y. (2008). Introduction: The limit of voice. (In A. Y. Jackson & L. A. Mazzei (Eds.), Voice in qualitative inquiry: Challenging conventional, interpretive, and critical conceptions in qualitative research. Abingdon Routledge).
Nicolini, D. (2009). Zooming in and zooming out: A package of method and theory to study work practices. In S. Ybema, D. Yanow, H. Weis, & F. H. Kamsteeg (Eds.), Organizational ethnography: Studying the complexity of everyday life (pp. 120–138). London: Sage).
Patrício, M. F., Julião, M., Fareleira, F., & Carneiro, A. V. (2013). Is the OSCE a feasible tool to assess competencies in undergraduate medical education? Medical Teacher, 35(6), 503–514. https://doi.org/10.3109/0142159X.2013.774330.
Pell, G., Fuller, R., Homer, M., & Roberts, T. (2010). How to measure the quality of the OSCE: A review of metrics – AMEE guide no. 49. Medical Teacher, 32(10), 802–811. https://doi.org/10.3109/0142159X.2010.507716.
Pierre, E. S. (2014). A brief and personal history of post qualitative research: Toward “post inquiry”. Journal of Curriculum Theorizing, 30(2), 1–19.
Pugh, D., Desjardins, I., & Eva, K. (2018). How do formative objective structured clinical examinations drive learning? Analysis of residents’ perceptions. Medical Teacher, 40(1), 45–52. https://doi.org/10.1080/0142159X.2017.1388502.
Rudland, J., Wilkinson, T., Smith-Han, K., & Thompson-Fawcett, M. (2008). “You can do it late at night or in the morning. You can do it at home, I did it with my flatmate”. The educational impact of an OSCE. Medical Teacher, 30(2), 206–211. https://doi.org/10.1080/01421590701851312.
Schatzki, T. R. (2002). The site of the social: A philosophical account of the constitution of social life and change. University Park: Penn State Press.
Schatzki, T.R. (2012). A primer on practices: Theory and research. (In J. Higgs, R. Barnett, S. Billett, M. Hutchins & F. Trede (Eds.), Practice-based education. (pp. 13–26). Brill Sense).
St. Pierre, E. A. (2018). Writing post qualitative inquiry. Qualitative Inquiry, 24(9), 603–608. https://doi.org/10.1177/1077800417734567
Tamblyn, R. M., Klass, D. J., Schnabl, G. K., & Kopelow, M. L. (1991). The accuracy of standardized patient presentation. Medical Education, 25(2), 100–109. https://doi.org/10.1111/j.1365-2923.1991.tb00035.x.
Wilby, K. J., Govaerts, M. J. B., Dolmans, D. H. J. M., Austin, Z., & van der Vleuten, C. (2019). Reliability of narrative assessment data on communication skills in a summative OSCE. Patient Education and Counseling, 102(6), 1164–1169. https://doi.org/10.1016/j.pec.2019.01.018.
Acknowledgements
We would like to acknowledge Bronwyn Beovich, Katie Priestley and Imogene Rothnie’s work on this study. We would also like to thank the participants for their generous sharing of their experiences.
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Bearman, M., Ajjawi, R., Bennett, S. et al. The hidden labours of designing the Objective Structured Clinical Examination: a Practice Theory study. Adv in Health Sci Educ 26, 637–651 (2021). https://doi.org/10.1007/s10459-020-10015-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10459-020-10015-w