Validating Measures of Third Year Medical Students’ Use of Interpreters by Standardized Patients and Faculty Observers
- First Online:
- Cite this article as:
- Lie, D., Boker, J., Bereknyei, S. et al. J GEN INTERN MED (2007) 22(Suppl 2): 336. doi:10.1007/s11606-007-0349-3
- 662 Downloads
Increasing prevalence of limited English proficiency patient encounters demands effective use of interpreters. Validated measures for this skill are needed.
We describe the process of creating and validating two new measures for rating student skills for interpreter use.
Encounters using standardized patients (SPs) and interpreters within a clinical practice examination (CPX) at one medical school.
Students were assessed by SPs using the interpreter impact rating scale (IIRS) and the physician patient interaction (PPI) scale. A subset of 23 encounters was assessed by 4 faculty raters using the faculty observer rating scale (FORS). Internal consistency reliability was assessed by Cronbach’s coefficient alpha (α). Interrater reliability of the FORS was examined by the intraclass correlation coefficient (ICC). The FORS and IIRS were compared and each was correlated with the PPI.
Cronbach’s α was 0.90 for the 7-item IIRS and 0.88 for the 11-item FORS. ICC among 4 faculty observers had a mean of 0.61 and median of 0.65 (0.20, 0.86). Skill measured by the IIRS did not significantly correlate with FORS but correlated with the PPI.
We developed two measures with good internal reliability for use by SPs and faculty observers. More research is needed to clarify the reasons for the lack of concordance between these measures and which may be more valid for use as a summative assessment measure.
KEY WORDSinterpreter use evaluation of skills assessment validation
Recent emphasis on cultural competence curricula in medical education has highlighted the need for training medical students and residents in the effective use of interpreters.1-3 The knowledge and application of this particular skill set is distinctly described as a separate learning entity (domain V) in the AAMC’s tool for assessing cultural competency training (TACCT).3 Several recent guidelines have reported areas of consistency and agreement, as well as controversy, about the best practices for interpreter use from both interpreter and clinician perspectives.4-8
Despite the need to train learners to use interpreters effectively and the availability of guidelines describing core skills, validated measures for evaluating these skills are not available. The challenges of designing such measures for clinical encounters are considerable. They range from the diverse and varying skill levels of interpreters available in practice settings to the difficulty of isolating communication skills necessary for using an interpreter to the shortage of faculty fluent in the patient’s language to assess the verbal content of non-English language encounters. Partial provider/learner fluency in the patient’s language may hinder or help the encounter when an interpreter is present. In addition, the manner in which clinicians utilize interpreters,9 and clinician awareness of the health literacy of their limited English proficiency (LEP) patients are also important variables affecting physician–patient communication,10 length of visits,11,12 and the quality of health care and patient satisfaction.13,14 Few attempts to design assessment tools of interpreter use have been reported in the literature, and those that have are case-specific and not generalizable across other encounters using interpreters.15
In this study, we used an empirical expert-based consensus framework to create two interpreter use skill assessment measures for clinical encounters where an interpreter is involved. We designed a study to test the internal reliability of the two measures and their performance compared to a widely used communication assessment measure in the setting of a standardized clinical examination.
We used an iterative process to create two measures: the interpreter impact rating scale (IIRS), for use by standardized patients (SPs), and the faculty observer rating scale (FORS), for use by faculty. The rationale for creating a new SP measure to assess interpreter-mediated communication was that additional or different observable skills apply to an encounter when an interpreter is present as a third party, which may impact patient/SP perception of the provider and perceived effectiveness of the information transfer.
IIRS and FORS Means and Variances
Corrected item-scale correlation
Combined total IIRS scale (Cronbach’s α = 0.90)
24.65 ± 4.16
Trainee showed direct eye contact with me during the encounter instead of at the interpreter most of the time.
3.74 ± 0.62
Trainee directly addressed the issues translated that were of concern to me.
3.52 ± 0.67
The trainee acknowledges and responds to my beliefs, concerns, and expectations about my problems.
3.35 ± 0.83
The trainee asked me questions in the first person (example: “Do you feel...” rather than “interpreter, can you ask him if he...”)
3.70 ± 0.56
The trainee sat at a comfortable distance from me (not too close and not too far away).
3.61 ± 0.66
The trainee’s nonverbal body communication was reassuring (i.e.,—mannerisms, facial expressions, and body language).
3.48 ± 0.51
Rate your overall satisfaction with the encounter.
3.26 ± 1.25
Combined total FORS scale (Cronbach’s α = 0.88; mean intraclass correlation coefficient = 0.61 for 4 raters)
41.57 ± 4.55
The trainee adequately explained the purpose of the interview to the interpreter.
3.50 ± 0.73
The trainee explained the interpreter’s role to the patient at the beginning.
3.08 ± 0.87
The trainee asked the patient one question at a time.
4.09 ± 0.43
The trainee listened to the patient without unnecessary interruption.
4.18 ± 0.40
The trainee asked questions to clarify his/her own understanding of the patient’s answers.
3.54 ± 0.53
The trainee presented information at a pace that was easy to follow for both patient and interpreter; that is, information was given in “digestible chunks”.
4.01 ± 0.59
The trainee maintained direct eye contact with the patient (instead of the interpreter).
3.75 ± 0.65
The trainee addressed the patient in the first person and not as “he/she”.
4.07 ± 0.99
The trainee appropriately closed the encounter: at a minimum, asked the patient if he/she had any questions.
3.48 ± 0.66
To what extent did the trainee keep the interpreter on track within his/her assigned role?
3.92 ± 0.55
Global rating of trainee’s effectiveness in using the interpreter for the patient encounter.
3.95 ± 0.56
Prototype items were constructed, pilot tested, and refined from the analysis of videotapes of other CPX encounters using interpreters. As a result of this process, consensus among the panel produced 7 items for the IIRS (see Table 1, IIRS) and 11 items for the FORS (see Table 1, FORS). The last item of each measure was a global rating item. Each item on both scales used a 5-point Likert-type rating scale where a value of 1 represented “marginal/low” and 5 represented “outstanding” performance. Scale scores for the IIRS and FORS were computed by summing across obtained item ratings on each measure; higher scores were viewed as indicative of higher skill level. Percentage of maximum scale scores were computed to place both measures on the same metric, i.e., the denominator was a maximum score of 35 on IIRS and 55 on the FORS.
To perform the validation, we chose to compare the performance of the IIRS and the FORS with a commonly used scale for communication skills, the physician patient interaction (PPI) scale.20 The PPI is a 7-item rating form and uses a 6-point Likert scale (ranging from 1 = “unacceptable” to 6 = “outstanding”) to measure the competency of communication in the context of professional behavior. The 7 items are “appeared professionally competent”, “effectively gathered information”, “listened actively”, “established personal rapport”, “appropriately explored my perspective”, “addressed my feelings”, and “met my needs”. These items were not constructed to take into account the presence of a third party interpreter or a language barrier as do the IIRS and FORS.
SETTING AND PROCEDURES
To control for variables such as patient performance, interpreter skill, and case details, we chose to validate our 2 assessment tools in the setting of a clinical practice examination (CPX) at a California medical school whose Institutional Review Board approved the study. The school was located near a city with the highest proportion of monolingual Latinos in the US.21 Study participants were 92 third year medical students (MS3s) who participated in the required CPX after completing their required clinical clerkships. One of the cases in the CPX was a 15-minute station involving a monolingual Spanish-speaking patient. Spanish was chosen as the language because of local community demographics. Both the SP and the standardized interpreter (SI) (actors who were bilingual in Spanish and English) received 4, 2-hour training sessions as a pair to prepare for their roles. Two SP–SI pairs were used for the entire examination to standardize case performance. We standardized the rating performance of the SP–SI pairs by having SPs and SIs rate videotaped encounters before the examination. All encounters were recorded on DVD. The medical students had previously received instruction on the use of interpreters from a lecture, had practice interviewing with SP–SI pairs in their first year of medical school, and had experience using interpreters in clinical encounters during clerkships.
Clinical Case Scenario and Student Task
A smoking cessation case involving a monolingual Spanish-speaking Latino patient wishing to quit smoking was designed and pilot tested with 39 MS3 in the preceding class cohort in May 2005. SPs and professional Spanish-speaking interpreters were used at the 2005 administration. Based on the findings, it was determined that the clinical tasks required of the students by the case (history taking and advising the patient) were at an appropriate level of challenge and fit within the 15 minutes provided for the encounter. This case was then administered in the 8-station CPX in May 2006 to the 92 MS3s. To control for performance in students with some degree of Spanish language proficiency, students were instructed not to use Spanish and that their skill using an interpreter, not Spanish proficiency, was being assessed.
Raters and Data Collection
SPs rated all 92 students using the IIRS and the PPI during the encounter and submitted their scores online to a central databank. Subsequently, one-fourth (23 of 92) of the encounters were selected for further assessment by faculty raters using the FORS. The 23 encounters were selected by an independent trainer to generate a range of performances based on SP-generated IIRS student scores across the entire cohort. Students who self-reported any degree of Spanish fluency were excluded from the analysis.
Four faculty participated in the independent rating of the 23 encounters using DVD records. The students who were rated were not familiar or known to the faculty but had been taught by the faculty in lectures. One rater was a course director (DL), one was the content expert for Diversity (PL), one was a researcher (SB), and one was a SP trainer (CF). Only 1 rater was bilingual in Spanish and English. All 4 raters had some expertise in teaching students about the effective use of interpreters. Each rater reviewed 2 DVDs containing all 23 encounters in the same sequence with students identified by number only. After rating the 23 student encounters using the FORS, each rater submitted completed rating forms to a data entry administrator blinded to the student and faculty identities.
The process of designing and constructing the 2 measures was qualitatively analyzed to document the content and face validity of the measures. Item analysis was performed using conventional methods. Internal consistency reliability was assessed by Cronbach’s coefficient alpha (α). Interrater reliability of the FORS was examined by the intraclass correlation coefficient (ICC) for 4 independent raters. Overlap of the IIRS, FORS, and PPI items and scale scores was analyzed by Pearson product-moment correlation coefficients. The nominal criterion for statistical significance of scale and item correlation coefficients was p < 0.05. No adjustment was made to this criterion for computing multiple correlation coefficients. All statistical analyses were computed with SPSS version 14.0 for Windows (SPSS, Chicago, IL, USA).
Reliability and Validity
Item means for the IIRS ranged from 3.26 (“rate your overall satisfaction with the encounter”) to 3.74 (“trainee showed direct eye contact with me during the encounter instead of the interpreter most of the time”) with SDs ranging from 0.58 to 0.89. Item means for the FORS ranged from 3.08 (‘the trainee explained the interpreter’s role to the patient at the beginning’) to 4.18 (“the trainee listened to the patient without unnecessary interruption”) with SDs from 0.40 to 0.87 (Table 1). Corrected item score–total scale score correlations (i.e., point-biserial correlation coefficients computed after removing the contribution of each item to the total score) ranged from 0.58 to 0.89 for the IIRS and 0.22 to 0.94 for the FORS. Values >0.20 are generally taken as evidence that the items comprising a measure are, in fact, measuring a similar construct.
The Cronbach coefficient α reliability for the IIRS and FORS was 0.90 and 0.88, respectively. Each measure had internal consistency reliability that exceeded the conventional minimum standard of 0.75. The interrater agreement using the FORS was good with an observed mean ICC = 0.61 and median = 0.65 (range 0.20–0.86) across the 4 faculty raters (see Table 1).
Analysis at the individual item level showed no statistically significant correlations among separate IIRS and FORS item scores, including global rating items. In fact, no coefficients were significant, although approximately 4 significant coefficients would be expected by chance alone using the nominal α = 0.05 criterion. We found that the distribution of IIRS and FORS scores, when expressed as a percentage of the maximum, were similar; faculty were not more conservative and SPs were not more liberal in their respective ratings of student performance.
Correlation of IIRS and FORS with PPI
The means and variances of the IIRS and PPI scores in the 23 randomly selected cases for the FORS substudy were not statistically different from the nonselected cases.
IIRS and FORS Scale Scores: Correlations with PPI Scale Scores
n = 23
Pearson correlation coefficient
Pearson correlation coefficient
PPI 1: appeared professionally competent
PPI 2: effectively gathered information
PPI 3: listened actively
PPI 4: established personal rapport
PPI 5: appropriately explored my perspective
PPI 6: addressed my feelings
PPI 7: met my needs
Total PPI score
None of the PPI items correlated with the FORS scale score (range −0.42 to −0.01). In contrast, every PPI item was positively and significantly correlated (p < 0.001) with the IIRS scale score. The range of correlation coefficients with the total IIRS score was 0.65 (PPI 7: “met my needs”) to 0.85 (PPI 3: “listened actively”) (Table 2).
We created and tested two new empirically designed measures for assessing student skills using highly skilled interpreters, using a Spanish-speaking case scenario during a high stakes examination. Using a subset of 23 clinical encounters, we examined SP and faculty ratings of student performance utilizing interpreters. The results suggest that the IIRS and the FORS each have acceptable levels of tangible content and face validity. It is reassuring that the PPI and the IIRS correlate well, suggesting that the IIRS may be a valuable measure for assessing interpreter use skills. The 2 measures also show high internal consistency reliability, and the FORS demonstrates an acceptable level of concordance among faculty observers.
The finding of the absence of correlation between items on the IIRS and the FORS is a surprise and of some concern, given that the verbal and nonverbal behaviors we sought to assess were identified a priori to represent desirable qualities of communication when an interpreter is present in a clinical encounter. We considered the reasons for this discordance by examining the individual items within the IIRS, PPI, and FORS. Items in the IIRS primarily address the SP’s assessment of the student’s verbal (items 2, 3, and 4; see IIRS Table 1) and nonverbal (items 1, 5, and 6; see IIRS Table 1) communication skills when a trained interpreter is present. Although very different in wording and content from the PPI items, the behaviors measured by the IIRS and the PPI both depend on SP assessment of the professional conduct and demeanor of the student. The high degree of correlation is not surprising considering that both scales were completed by the same SP on the same student for each encounter. Conversely, the FORS focuses on verbal behavior (Table 1: FORS items 1, 2, 3, 5, 6, 8, 9, and 10) demonstrated by the student, and, to a lesser extent, nonverbal observable behaviors (Table 1: FORS items 4 and 7). Examining the individual behaviors assessed for the IIRS and the FORS, it is notable that only 2 items reflect exactly the same behaviors: “eye contact” and “addressing the patient in the first person”. The remaining 4 (observable behavior) items for the IIRS focus on positioning, nonverbal communication, and responding to the patient’s needs. The remaining 7 (verbal) items on the FORS focus on behaviors that the SPs were not trained to assess: explanation of the interview to the interpreter and patient, listening without interruption, closing the encounter appropriately, and keeping the interpreter on track. Furthermore, faculty rated the encounters as outside observers viewing a DVD, whereas SPs were participants in the encounter and subject to emotional, interpersonal, and other nuances. Although the psychometric properties of SP ratings are established,22 other data suggest that the lens for examining these skills should be widened beyond SP rating scales.23 A recent study on real clinical encounters suggested that the positive subjective experience of partnership in an encounter does not fully reflect communication skills as measured by the PPI, and articulates the need for the inclusion of affective relationship dynamics (such as trust and power) in the assessment of communication skills.24 Furthermore, other factors such as primary childhood language and patient–provider ethnicity may also impact SP assessments of student communication skills.25 In addition we speculate that faculty (compared to trained SPs) may rate according to global impression rather than individual behaviors.
Despite its surprising results, our study has several strengths. We used an iterative process based on the review of the literature and clinical expertise to design, pilot test, and refine the SP case and the IIRS and FORS measures. This process mimicked the plan–do–study–act (PDSA) model used in quality improvement.26 Second, we used a highly structured and standardized setting of a high stakes assessment to test the performance of the new measures. This allowed us to control for variability in patient and interpreter responses and clinical tasks. Third, we used a variety of faculty raters with 1 quality in common: expertise in teaching and assessing medical students’ skill in using interpreters in clinical encounters.
Our study has some limitations. It was conducted at 1 institution for 1 class of medical students using a relatively small number of encounters with 1 case scenario. We examined only a Spanish language encounter. However, by standardizing the language used and the training of the interpreter in this field study, we have now provided a foundation to test the FORS with patient encounters in different languages and settings.
Clinical application of these measures presents challenges. The IIRS is limited by a need for intensive SP training and may not be easily adapted to real patient encounters where patients are often monolingual and ad hoc interpreters are common. Effective use of the FORS also depends on the skills of the interpreter. Furthermore, it is unclear from this study whether SP or faculty ratings should be considered the ‘gold standard’ for assessing student communication skills for using an interpreter. These limitations need to be addressed before either scale can be used in actual practice situations. Further research will involve demonstrating the face and construct validity of the FORS as used by faculty.
In summary, this study provides 2 reliable measures that can be used in standardized settings and have the potential to be adapted for real patient encounters. At our institution, the FORS is used in a formative way to provide feedback to clinician trainees after an encounter, and also as a faculty development tool. However, the validity of the FORS in actual clinical settings has not yet been demonstrated as this study was conducted in a standardized examination setting. Future plans include the examination of the IIRS and FORS in different language encounters, use in different health professions learners (such as nursing), and validation of the FORS against other faculty rating scales.
The US is becoming increasingly linguistically and culturally diverse.21,27 Training future physicians to use interpreters according to accepted standards21,28,29 is essential to the development of good clinicians.5,19 As student learning is closely linked to assessment, this study represents a step toward the goal of setting objective educational standards in the use of interpreters.
This project was supported in part by grant awards from the National Institutes of Health (NIH), the National Heart, Lung and Blood Institute, K07 HL-04-012 (2004-9), and from the AAMC grant initiative “Enhancing Cultural Competence in Medical Schools” (2005–2008) supported by the California Endowment. The contents are solely the responsibility of the authors and do not necessarily represent the official views of the NIH or AAMC. The authors gratefully acknowledge the assistance of Sarah Pardee and Jennifer Encinas in the manuscript preparation and data presentation.
Conflict of Interest