Impact statements

  • Validated tools for assessing person-centred consultation skills are important for identifying competence of pharmacists as they move into more clinical roles.

  • Adopting the MR-CAT as the chosen consultation skills assessment tool for undergraduate and postgraduate pharmacist education in the UK would establish familiarity with advanced person-centred skills for pharmacists and patients making patient partnership and patient autonomy more commonplace.

  • Since completion of this study, the MR-CAT has been used in the practice setting to assess over 4000 pharmacists’ consultations as part of a vocational education programme in England. Using a validated assessment tool has contributed to establishing a robust and credible learning programme.

Introduction

The last decade has seen a transition in the scope of pharmacy practice with a significant shift towards clinical service provision [1,2,3,4,5,6,7]. In England, publication of the Long Term Plan in 2019 signalled a clinical future for pharmacy within the National Health Service (NHS) [8]. Today pharmacists are embedded within general practice and care home settings with approximately eleven percent qualified as independent prescribers in the UK [9]. Pharmacists in England play a key role in the provision of clinical services in the community. Examples include the New Medicine Service which involves a consultation with a patient and the NHS urgent care service, which makes it quicker and easier for patients to access advice or treatment [10, 11]. Although development of clinical roles for pharmacists in other countries varies, a common goal of enabling pharmacists to expand their scope of practice and work collaboratively across healthcare systems is emerging [1, 12, 13].

A cornerstone of global pharmacy strategy and policy is the need to develop effective person-centred professionals who are equipped to empower patients as partners in their healthcare and engage in shared-decision making [1, 12, 14,15,16]. Professional and regulatory bodies have recognised this culture change by placing person-centred care at the heart of pharmacy standards and policy [17,18,19,20].

Developing good consultation technique lies at the heart of person-centred care and may lead to improved health outcomes for patients [21,22,23,24,25]. Historically pharmacists were trained as experts in medicines taking a product-centric approach [26]. The concept of patient counselling in relation to medicines is now out-dated and a contradiction to person-centred practice [16, 27].

The need to develop effective person-centred consultation skills as part of pharmacists’ initial education and training and at postgraduate level has been recognised as a priority [28,29,30]. In England, the national Consultation Skills for Pharmacy Practice programme was developed in 2014 in response [31]. Although undergraduate level training has not kept pace, newly implemented initial education and training standards in the UK have a strong focus on person-centred care [32].

Assessment of consultation performance is noted as the most challenging element of training [33, 34]. Standardised assessment criteria in the form of assessment tools, which allow learners to practise and support constructive feedback, are key to skill development [33, 34]. A lack of validated pharmacy specific consultation assessment tools has led to the exploration of non-pharmacy tools [28, 35,36,37]. Although deemed suitable for use, many have not been validated for a pharmacy context and were developed prior to the increased emphasis on shared decision-making [28, 35, 36]. The Medication Related Consultation Framework (MRCF) which is widely used in pharmacy consultation skills development and assessment was developed in 2011 [37, 38]. A review of consultation tools to promote the delivery of person-centred consultations in pharmacy recommended revision of the MRCF to satisfy the multi-faceted elements of a true patient-centred consultation [39]. The overall conclusion from the review was that suitable tools are needed to meet the requirements of a holistic patient-centred consultation in a practice setting [39].

As a result, a new global assessment tool; the medicines related consultation assessment tool (MR-CAT) was developed by two members of the research team (referred to as the developers). The MR-CAT is designed to be used by healthcare professionals to assess pharmacists’ consultations; it is not designed as a tool, for patients to assess the consultation. The structure of the MR-CAT (See Table 1) is based on the Calgary-Cambridge model (Initiating the session, Gathering information, Explanation and planning, Closing the session) and an additional section on behaviours which underpin the consultation [21]. The inclusion of shared decision making in the MR-CAT in place of explanation and planning and related behaviours, recognises the shift towards personalised care [40]. Within each section of the MR-CAT there are three levels of practice: below expectations, competent and excellent.

Table 1 The descriptors for the five sections of the medicines related-consultation assessment tool (MR-CAT) and the overall global assessment of the consultation for the three levels of practice

The developers used the Consultation skills for pharmacy practice: practice standards for England [41] to identify key descriptors summarising the expected skills and behaviours for each of the five sections of the MR-CAT and the overall global assessment of the consultation for below expectations, competent and excellent (See Table 1).

Aim

This study aimed to test the validity and reliability of the MR-CAT using psychometric methods.

Ethics approval

The study was approved on 23.08.2018 by the University of Manchester Proportionate Research Ethics Committee reference number 2019–4620-11,787.

Method

Production of recordings of simulated consultations

The research team created video recordings of simulated pharmacists’ consultations. Medical actors played the part of patients. The simulated consultations focused on long-term conditions and acute presentations in a primary care setting, for example medication review in a care home, post-discharge asthma review, type-2-diabetes medication review, urinary tract infection and knee pain. The pre-recorded simulated consultations were independently assessed by two members of the research team using the five sections of the MR-CAT and an overall global assessment of the consultation was determined for each recording. Fifteen recordings were then used in the validation study covering the three levels of practice for the overall global assessment of the consultation: Below expectations (n = 5), competent (n = 5) and excellent (n = 5). Each recording was given a unique identification number. A further three simulated consultations were recorded to train participants in how to use the MR-CAT.

Participants and training

Educators involved in training pharmacists to work in advanced practice roles who had also completed prior consultation skills training were invited to participate in the study. These participants (now referred to as ‘raters’) completed training to familiarise them with the MR-CAT and to learn how to use the tool.

Following this, raters independently viewed and assessed the three simulated consultations which had been created for training purposes. Raters then attended a second session where they discussed their rating of each of the recordings and their rationale for a rating to ensure all raters understood how to use the MR-CAT.

Rating of pre-recorded simulated consultations using the MR-CAT

After completing the training, raters took part in a first round of data collection (January 2020). Raters independently assessed the 15 pre-recorded simulated consultations as below expectations, competent or excellent against the five sections of the MR-CAT and then assigned an overall global assessment of the consultation (below expectations, competent or excellent) based on their ratings for the five sections. Raters were blinded to the levels of practice assigned by the research team to each of the recordings. Raters submitted their ratings for the five sections of the MR-CAT and the overall global assessment of the consultation via an online survey tool. There was a separate survey for each recording to prevent raters from comparing recordings before submitting their ratings. Raters were prevented from accessing the survey after they had submitted ratings so that they could not change or view their ratings after they had been submitted.

The second round of data collection was performed eight weeks after the first round of data collection (March 2020). A sub-sample (n = 6) of the original 15 recordings was used to establish intrarater reliability (test–retest analysis). The six recordings (two below expectations, two competent and two excellent) were given a different unique identification number in the second round and raters were blinded to the levels of practice. Raters independently assessed each recording and submitted their ratings for the five sections of the MR-CAT and the overall global assessment of the consultation via an online survey tool as in the first round.

Data analysis

Data were downloaded from the survey platform and subsequently analysed using Statistical Package for Social Sciences (SPSS) version 25 database (SPSS Inc., Chicago IL) and STATA for statistics and data management version 14 (StataCorp, College Station, TX). A range of statistical tests were used to test discriminant validity, intrarater and interrater reliability for each of the five sections of the MR-CAT and for the overall global assessment of the consultation. Analysis also included internal consistency testing for the whole tool. The two-tailed p-value was considered significant at p-value < 0.05.

Initially, the Cronbach’s alpha test for scales was used to evaluate the internal consistency or how closely the five sections of the tool are related as a group. The overall global assessment of the consultation rating and each of the section ratings (below expectations, competent and excellent) of the MR-CAT were entered into the analysis. To determine internal consistency, a Cronbach’s alpha greater than 0.7, was taken to indicate high internal consistency.

To explore the extent to which the MR-CAT could discriminate between consultations that were below expectations, competent or excellent a Kruskal–Wallis test, with post-hoc Wilcoxon rank sum analysis, was used. This compared raters’ overall global assessment of consultations that had a priori been classified as below expectations, competent or excellent using the mean and standard deviation with the rating of the raters for statistical differences between grouped consultation types.

The degree to which raters awarded similar ratings when observing the same consultations was investigated (inter-rater reliability). Each rater’s ratings for the five sections of the MR-CAT and the overall global assessment of the consultation were ranked across the 15 simulated consultations. Kendall’s coefficient of concordance was calculated to assess the degree of agreement between raters’ ranked ratings at each level of practice.

The extent to which raters produced consistent ratings when applying the MR-CAT to the same simulated consultation at two time points was investigated (intrarater reliability). The test–retest values were produced at Time 1 (original test of tool) and Time 2 (eight weeks after the original test) and compared. Spearman’s correlation coefficients (rho) were calculated for each section of the MR-CAT using rank orders of ordinal data.

Results

13 pharmacy educators participated in the study.

Internal consistency

The Cronbach’s alpha was very good at 0.97, demonstrating internal consistency of the MR-CAT and very good correlation between the sections of the MR-CAT and the overall global assessment of the consultation (all rated at r = 1.00).

Discriminant validity

Analysis of the raters’ mean ratings for the overall global assessment of the consultation were found to discriminate between the three levels of practice (below expectations, competent and excellent) (Kruskal–Wallis Chi-square = 128.71; df = 2; p-value < 0.001) (See Fig. 1).

Fig. 1
figure 1

Differences between mean ratings on the global assessment level

The post-hoc Wilcoxon rank sum analysis revealed that there were significant differences between each level of practice (z score < -8.33; p-value < 0.001 for all three levels of practice). Variation in raters’ mean ratings for the overall global assessment of the consultation was biggest for the competent level of practice (SD 0.54); consensus was greatest regarding the below expectations (or fail) level of practice (SD 0.32).,

Interrater reliability

Interrater reliability for all five sections of the MR-CAT and the overall global assessment of the consultation was good (Kendall’s W 0.80–0.90) apart from the introduction section of the MR-CAT where there was moderate agreement (Kendall’s W 0.68). Agreement was highest for the overall global assessment of the consultation and for the shared-decision making and consultation behaviours sections of the MR-CAT and was significant at p-value < 0.05 and 14 degrees of freedom for all five sections of the MR-CAT and for the overall global assessment of the consultation (see Table 2).

Table 2 Interassessor reliability (Kendall's W score) and intraassessor (test—retest) reliability score (Spearman's rho) for each section of the MR-CAT

Intrarater (test–retest) reliability

There was moderate to high intrarater reliability for all five sections of the MR-CAT and for the overall global assessment of the consultation (rho 0.64–0.84) in the test–retest analysis. Two-tailed p-value was considered significant at p-value < 0.05 (see Table 2). Raters 8 and 13 did not take part in this analysis.

Discussion

Key findings

Using psychometric methods, this study sought to test the validity and reliability of the MR-CAT. Our findings suggest that the MR-CAT is a valid and reliable instrument, that is capable of discriminating between different levels of consultation practice (below expectations, competent and excellent). This differentiates the MR-CAT from previous tools. The PharmaCAT uses a rating scale ranging from poor to excellent practice, whilst the MRCF offers descriptors of key skills and behaviours with the option to select yes or no if observed in a consultation [36, 37]. The MR-CAT design includes descriptors with clear definition of specific skills and behaviours which demonstrate each level of practice, consistent with the format used in the consultation tool used in medical education [36].

Interrater reliability was highest for the shared decision making and consultation behaviours sections of the MR-CAT and the global assessment rating. While, on the other hand, interrater agreement was only moderate for the introduction section of the MR-CAT; this finding is similar to other studies [37].

Summative assessment: avoiding failure to fail

When considering discriminant validity and the three levels of practice there was highest agreement in rater ratings for the ‘below expectations’ level of practice. This is important when using a tool for summative assessment (assessment of learning) as it suggests greatest consensus over what is a fail. Failing to fail underperforming students is a well-documented problem, which has serious implications for patient safety and professional competence [42,43,44,45]. One of the barriers cited as contributing to failure to fail is a lack of certainty and/or clarity around expected standards of performance particularly if assessors are inexperienced or lack confidence [42,43,44,45]. The use of criterion-referenced frameworks is recommended to promote clarity and fair and consistent treatment of learners and has been incorporated within the MR-CAT [45]. The inclusion of descriptors within the MR-CAT provides clear expectations of performance for learners and assessors for different levels of practice which assures fairness and equity in the assessment process and has the potential to help address the issue of failing to fail.

Formative assessment: preparing to pass

Theories which support the development of communication skills include practice in the workplace and reflective theory [46,47,48,49]. The MR-CAT has the potential to facilitate such learning in a practice setting because the global structure and descriptors support learners and assessors to differentiate between levels of practice and conceptualise what good consultation skills and behaviours are. Reflection is promoted by identification of strengths and areas for development. This in turn facilitates timely and personalised feedback in the consultation, a key component of skill development [33, 36, 50, 51] and further supports the potential utility of the MR-CAT for formative assessment (assessment for learning). The advantage this brings is that methods of summative assessment which also have a formative role are better than those that do not [52].

Advantages of global assessment tool design

Our findings provide assurance of the MR-CAT design as a global assessment tool that is capable of discriminating between different levels of pharmacist’s consultation practice (below expectations, competent and excellent). Tools containing more detailed skill elements whilst supporting a good structure and understanding of the consultation may hinder their application in practice [39]. Validation of the MRCF identified some inconsistencies in rater ratings of individual consultation behaviours within the discrete elements of the tool [37]. Given evidence that global assessment tools should be used in preference to a checklist approach to develop competence and improve professional authenticity it is likely that MR-CAT will be useful for assessing capability and support personal consultation style which adapts to the needs of the individual patient [53].

Limitations of the study

Limitations of the study include its small sample size (n = 13). However, the participants were educators involved in training pharmacists to work in advanced practice roles who had also completed prior consultation skills training. Therefore, the sample was very homogenous. A similar study which used psychometric testing to validate the MRCF also used a small homogenous sample (n = 10) which has been shown to be acceptable [37].

There was moderate to high intrarater reliability for all five sections of the MR-CAT and the overall global assessment of the consultation. The authors acknowledge the potential for memory bias due to the raters being presented with six of the original 15 simulated recorded consultations for the test–retest analysis. Human memory is a complex and broad concept [54]. Eight-weeks was chosen for the test–retest analysis because this was the same length of time between the first and second round of data collection in the MRCF validation study which was previously accepted as a methodologically robust approach [37].

Internal consistency for the overall global assessment of the consultation was good (Cronbach’s alpha = 0.97). A high Cronbach’s alpha > 0.95 suggests that there is collinearity among the items tested and may suggest redundancy. The items tested are the descriptors for each of the five sections of the MR-CAT (see Table 1) which assess different sections of the consultation. On this basis, it is unlikely that the high Crohnbach's alpha suggests redundancy. The most likely explanation for the high Cronbach’s alpha is if a pharmacist performs well in one section of the consultation, they are more likely to perform well in the other sections of the consultation. This is observed by the standard deviation (SD) of the overall global assessment of the consultation ratings given for each pre-recorded simulated videoed consultation (see Fig. 1). Those rated excellent and below expectations have very small SDs, indicating little variation in the rating. However, those rated competent have a larger SD suggesting that there is more variation in the assessment of a competent consultation.

While this validation study supports MR-CAT’s psychometric properties, this tool has so far only been tested using simulated consultations. There remains a need to undertake further research to test the usability and utility of the MR-CAT in a practice setting, possibly using a similar approach to ours which is widely accepted [55, 56]. Moreover, the pre-recorded simulated video consultations were situated in simulated general practice and care homes settings and were conducted by pharmacists. Therefore, further research would be required to assure its utility across other sectors of practice and for pharmacy technicians’ consultations.

Conclusion

The need for all healthcare professionals to demonstrate a person-centred approach in a consultation is imperative to quality of care across the health system and to providing personalised care. Developed specifically for the assessment of pharmacists’ consultation skills by observation of practice, the MR-CAT has been validated as a summative assessment tool which is useful to support pharmacists in advanced patient-facing roles. The ability of the MR-CAT to discriminate well between levels of practice for all five sections of the MR-CAT and for the overall global assessment of the consultation with specific descriptors of practice enables pharmacists to identify strengths and areas for development. This strengthens MR-CATs position as a formative assessment tool by enabling constructive feedback and reflection on practice.