Background

Medical curricula are increasingly integrating collaborative learning [1, 2]. When learning in groups and teams, in which individual students work together to achieve a common goal, such as in team-based learning (TBL) and problem-based learning (PBL), there is an expectation for students to be accountable to both their instructor and peers [3]. One way in which students are held accountable is through the utilization of peer feedback, also known as peer assessment or peer evaluation, which allows students to recognize areas of their strength and weakness as team members. Feedback is essential for learning; it can help students recognize their potential areas of deficiency in their knowledge, skills, or attitude. It is hoped that students use feedback to improve and become effective teammates [4].

There are many advantages to using peer feedback in the medical school curriculum. One advantage is that it can provide a valuable and unique perspective regarding the overall performance of students [5]. Compared to rare encounters with faculty, peers often work together for extended periods of time. Peer assessment is beneficial in assessing areas of proficiency based on multiple observations, rather than one encounter [6]. For this reason, compared to faculty, peers may have the ability to provide more accurate assessments of competencies such as teamwork, communication, and professionalism [7]. Additionally, according to Searby and Ewers, peer assessment may motivate students to produce high-quality work [8]. Overall, peer evaluation may help students improve metacognitive and reflection skills and develop a thorough understanding of coursework by identifying knowledge gaps and reinforcing positive behavior [6, 8,9,10,11].

While the literature suggests that peer evaluation can lead to positive outcomes, there are also limitations that must be addressed prior to implementation. Improper or poorly timed peer feedback may impair student relationships and disrupt team function [11]. Poor implementation of peer assessment may create an undesirable class environment that includes distrust, increased competition, or the tendency to exert less effort than they would working alone [3, 12]. Additionally, many students may feel uncomfortable providing peer feedback because of lack of anonymity, potential bias in scores due to interpersonal relationships, and lack of expertise in making assessments [6, 13,14,15]. Moreover, some students believe peers may be overly nice and not provide honest feedback [13].

Overall, it seems that peer evaluation in a collaborative learning environment has the ability to provide valuable feedback to medical students. It may provide skills necessary to effectively work on inter- and multidisciplinary teams as physicians. It may also increase student confidence and quality of work. The goal of this systematic review was to examine the utilization, effectiveness, and quality of peer feedback in a collaborative learning environment, specifically in undergraduate medical education. The objectives were to determine the role peer feedback plays in student learning and professional development, ascertain the impact peer feedback might have on team dynamics and success, and learn if and how the quality of peer feedback is assessed.

Methods

Data sources

A comprehensive literature search was conducted by one of the authors (M.M.). Databases searched included: PubMed, PsycINFO, Embase, Cochrane Library, CINAHL, ERIC, Scopus, and Web of Science. Search terms included index terms (MeSH terms or subject headings) and free text words (see Appendix for complete search strategies for PubMed): ((education, medical, undergraduate [mh] OR students, medical [mh] OR schools, medical [mh] OR “undergraduate medical education” OR “medical student” OR “medical students” OR “medical schools” OR “medical school”) AND (TBL OR “team-based learning” OR “team based learning” OR “collaborative learning” OR “problem-based learning” OR “problem based learning”) AND ((“peer evaluation” OR “peer feedback” OR “peer assessment”) OR ((peer OR peers OR team OR teams) AND (measur* OR assess* OR evaluat*))). The searches were limited to peer-reviewed English-language journal articles published between 1997 and 2017. Overall, the authors felt a 20-year limit was appropriate to ensure the data being evaluated was applicable in undergraduate medical education today.

Study selection

Duplicates of all articles retrieved were excluded and screened for full-text review if they were original research articles that assessed the use of peer feedback by medical students in a collaborative learning environment during medical school. Editorials, comments, general opinion pieces, letters, survey research studies, and reviews were excluded. All reference lists of selected articles for full-text screening were hand searched for additional relevant articles not discovered in the initial database searches.

Paired reviewers (S.L. and M.E.) screened titles and abstracts of retrieved articles independently. Citations with abstracts that seemed potentially relevant to the selection criteria were included as candidates for full-text screening. The same two reviewers screened these full-text articles independently and in duplicate based on the selection criteria. When a discrepancy arose in article selection between the two reviewers at the full-text article screening stage, the disagreement went to arbitration by a third reviewer (M.M.) who served as a tiebreaker.

Data extraction and quality assessment

The systematic review was guided by the Preferred Reporting Items for Systematic Reviews and Meta-Analysis Statement (PRISMA) [16] and the Best Evidence Medical Education (BEME) Guide No. 10, “The Effectiveness of Self-Assessment on the Identification of Learner Needs, Learner Activity, and Impact on Clinical Practice” [16, 17]. A standard extraction framework was developed and piloted with a small sample of included studies to abstract and code data from the selected studies. The two authors (S.L. and M.E.) read each article independently and used the extraction framework to extract data. Data was extracted on the country where the study took place, type of course, type of participants, sample size, type of collaborative learning environments (team-based learning, problem-based learning, etc.), and impact and outcomes of peer feedback being evaluated. This information can be found in Table 1. In addition, the types of studies, data sources for peer evaluation, peer evaluation grading criteria, and assessment methods for the quality of peer feedback were also extracted. A qualitative systematic review was conducted due to the heterogeneity of the selected studies in terms of research designs, types of peer feedback, types of student participants, settings, and outcome measures of the impact of peer feedback.

Table 1 Student Peer Feedback and Outcomes in the Context of Collaborative Learning Environment

We used an adapted Kirkpatrick evaluation model [45] to classify the effectiveness/impact of peer feedback in this review. There are six levels ranging from level 1 (Reaction - participants’ views on the learning experience) to level 4B (Results – improvement in student learning as a direct result of their educational intervention) [45].

We used a grading system, the Gradings of Strength of Findings of the Paper by Colthart et al. to score the strength of study findings on a scale of 1 to 5 [17]. Articles scored 4 or above on the strength of findings were considered to be higher quality papers. Articles scored 3 had conclusions that could probably be based on the results. Articles graded as 2 or 1 were regarded as, “results ambiguous, but there appear to be a trend,” or "No clear conclusions can be drawn (not significant). Two authors (S.L. and M.E.) independently carried out data extraction and quality appraisal. Three authors (S.L., M.E., and M.M.) reviewed and discussed discrepancies and came to a consensus.

Results

Retrieved studies

A total of 1301 articles were returned from the literature search. After removal of duplicates, 948 remained. A further 905 articles were excluded after title and/or abstract screening leaving 43 for full-text review. Of these, 26 articles were included in this review in addition to 5 further articles identified through hand searching of references (31 in total). Further details shown in PRISMA flowchart (Fig. 1).

Fig. 1
figure 1

Flowchart of Study Selection Process

Study characteristics

Of the included studies, the majority were completed in the United States (n = 14), followed by the Netherlands (n = 4), and Australia (n = 4), Canada (n = 3), and the United Kingdom (n = 3). Other countries included Bahrain, Brazil, Finland, and Lebanon.

The sample size for the studies ranged from 30 to 633 students. Most studies included first-year (n = 18) and second-year (n = 9) medical students. Peer feedback was evaluated through collaborative learning activities integrated into preclinical courses (n = 7) and clerkships (n = 2); although many studies did not provide clear information on where peer feedback was evaluated in the curriculum. PBL (n = 18) and TBL (n = 6) were used in the collaborative learning setting.

The research methodology of selected studies included 15 quantitative, 3 qualitative, and 13 mixed methods (defined as including both quantitative and qualitative data). Most studies utilized quantitative questionnaires to obtain data (n = 26). Other data sources included narrative comments (n = 14), focus groups (n = 5), open-discussions (n = 1) and individual interviews (n = 1). Many studies did not describe the grading criteria for the peer feedback (n = 17); of those that did, most were formative (ungraded) in nature (n = 8), while some courses included formative and summative (graded) peer feedback (n = 2) and some only included summative peer feedback (n = 4). Eleven studies reported that students received instruction on how to provide appropriate peer feedback, but no studies provided descriptions on whether or not the quality of feedback was evaluated by faculty.

A total of 17 studies evaluated the effect of peer feedback on professionalism in some manner. There were 12 studies that evaluated the effectiveness of peer feedback for the assessment of professionalism. Of those 12 studies, eight had positive results, two had mixed results, one had negative outcomes, and one was inconclusive. Eight studies evaluated the use of peer feedback for the development of professional behavior, in which seven had positive results and one had mixed results. Ten studies examined the effect of peer feedback on student learning, in which four had positive results, three had mixed results, one had negative outcomes, and three were inconclusive. In addition, there were six studies that examined the role of peer feedback on team dynamics. Of those studies, four had positive outcomes, whereas two had negative results. Table 1 contains more details about the selected articles, including sample size, participants, and setting.

Discussion

This systematic review examined the role of peer feedback in a collaborative learning environment in undergraduate medical education. It revealed a large number of variations in research design, approaches to peer feedback, and definitions of outcome measures. Due to the heterogeneity of these studies, it was difficult to assess the overall effectiveness and utility of peer feedback in collaborative learning for medical students. Despite these differences, the overall outcomes for most of the studies were positive.

Assessment of professionalism

The professional development of medical students is an essential aim of the medical school curriculum [20]. Peer feedback may provide reliable and valid assessment of professionalism. In the systematic review, several studies reported positive outcomes for the assessment of several aspects of professionalism. Chen and colleagues reported peers were able to accurately evaluate respect, communication and assertiveness of student leaders [1]. Dannefer and colleagues found that peer feedback was consistently specific and related to the professionalism behaviors identified by faculty as fundamental to practice. In addition, they found that peers often gave advice on how to improve performance, which is often considered a type of feedback that can result in positive changes in behavior [23, 46]. Emke and colleagues found that the use of peer evaluation in TBL may help identify students who may be at risk for professionalism concerns [25].

Although many studies had encouraging results in regards to the use of peer feedback for the assessment of professionalism, some studies did not show positive results. For example, Roberts and colleagues found that peer assessment of professional learning behavior was highly reliable for within-group comparisons, but poor for across-group comparisons, stating that peer assessment of professional learning behaviors may be unreliable for decision making outside a PBL group [37].

Development of professionalism

According to Emke and colleagues, professional behavior is a cornerstone of the physician-patient relationship, as well as the relationship between colleagues working together on a multidisciplinary team [25]. Many studies reported positive outcomes when assessing the development of professionalism. Nofziger and colleagues found that 65% of students reported important transformations in awareness, attitudes, and behaviors due to high quality peer assessment [31]. Papinczak and colleagues found that peer assessment strengthened the sense of responsibility that group members had for each other, in which several students were enthusiastic and committed to providing helpful and valid feedback to support the learning of their peers [33]. In addition, studies by Tayem and colleagues and Zgheib and colleagues reported improvements in communication skills, professionalism, and ability to work on a team [40, 44].

Student learning

Peer assessment may also be beneficial for student learning. Tayem and colleagues reported that a large percentage of students that participated in peer assessment in TBL felt that peer assessment helped increase their analytical skills as well as their ability to achieve their learning objectives and fulfill tasks related to the analysis of problems [40]. Zgheib and colleagues noted that students learned how to provide peer evaluations that were specific and descriptive, and expressed in terms that were relevant to the recipient’s needs, preparing them for their role in providing feedback as future physicians [44]. While peer assessment may be an important tool to improve student learning, some outcomes were mixed. For example, Bryan and colleagues stated that students were capable of commenting on professional values, but may lack the insight to make accurate evaluations. They recommended that peer-assessment should be used as a training tool to help students learn how to provide appropriate feedback to others [19].

Team dynamics

Students must learn how to effectively work on a team to become successful physicians [47]. Peer feedback may be valuable in the development of medical students in becoming effective team members. For example, Tayem and colleagues reported that a large proportion of participants agreed or strongly agreed that their respect towards the other group members and desire to share information with them had improved. The students also agreed that they had become more dependable as a result of peer assessment [40].

Grading criteria

In many cases, the grading criteria were not clearly described. Of the studies that described the grading criteria, the majority felt that peer feedback was appropriate to use in a formative, or ungraded, manner. Although most studies did not describe the details of their grading criteria, the literature supports the use of formative and summative assessments for peer feedback evaluations. According to Cestone and colleagues, peer assessment in TBL can provide formative information to help individual students improve team performance over time. Formative peer feedback can also aid in the development of interpersonal and team skills that are very important for success in future endeavors [3]. According to Cottrell and colleagues, one evaluation is not adequate, in which implementing the peer assessment multiple times and across a variety of learning contexts lends students opportunities to make formative changes for improvement [20].

Instruction on how to provide high-quality peer feedback

Providing effective feedback for peers is a skill that should be developed early in medical and health professions education [1]. According to Burgess and colleagues, peer review is a common requirement among medical staff, but since formal training in providing quality feedback is not common in the medical school curriculum, physicians are often not well prepared for this task [48].

Out of 31 studies, 11 described the instruction that was provided to students on how to provide effective peer feedback. These studies stressed the important of training students how to provide feedback for strengths as well as areas needing improvement. Without any guidance or training on how to provide peer feedback, students may feel confused or not know how to assess peer properly. In the study by Garner and colleagues, they found that students were unclear about the purpose of peer assessment and felt the exercise was imposed upon them with little preparation or training, creating anxiety for individual students [26]. Nofziger stated that students should receive training to provide specific, constructive feedback and that the institutional culture should emphasize safety around feedback, while committing to rewarding excellence and addressing concerning behaviors [31]. Better instruction on how to provide appropriate feedback may make the goals of peer feedback clearer as well as decrease student anxiety when performing assessments of their peers.

Evaluation of feedback quality

This systematic review also evaluated whether or not the quality of peer feedback was assessed by students or faculty. Most included studies failed to address the importance of the evaluation of peer feedback by students, including the potential importance of educators reviewing the feedback quality. Faculty should be trained on how to evaluate the quality of feedback to make sure students are effectively assessing their peers, as well as help remediate students who are not performing to the best of their abilities.

Limitations

Our systematic review was limited to published studies in English. As a result, potential publication bias and language bias may have been introduced to the review. Other reporting biases in these selected studies may contribute to biased conclusion of the study reports. These biases could potentially present a threat to the validity of any type of review including this systematic review. The presence of negative and inconclusive studies and small effects do not support publication bias. A full analysis for publication bias is not feasible and thus, cannot be ruled out. An advantage is that this review is one of the first to apply standard methods of evaluating both study outcomes and quality of the existing literature related to peer feedback during collaborative learning in undergraduate medical education. The descriptive methods inherently limit potential conclusions that may be drawn from reported results. These proved to be challenging to apply because the majority of the studies were descriptive in nature.

With regards to strength of methods, an overwhelming majority of studies received lower ratings. This pattern was often due to outcomes not specified a priori, no power discussion, unclear primary study objective, or discordant conclusion with originally stated study aim. A few studies seemed to use validated surveys to measure student perceptions but then reported outcomes measures without interpretation of results.

The strongest studies were conducted by Chen, Cottrell, Kamp, Parikh, and Roberts [1, 4, 20, 28, 37]. The strength of Cottrell 2006 was the report of internal consistency of the rubric and generalizability of the results [20]. Kamp 2014 had a sound pre-post design. However, they did not report a power calculation [28].

Finally, the most common study outcomes were evaluated at levels 1 and 2 according to the modified Kirkpatrick Evaluation Model [45]. These corresponded mostly to the feasibility of doing peer feedback in a variety of learning contexts. The second most common outcome was student perceptions of peer feedback. Overall, students had favorable perceptions. Concordance between self and others’ feedback was considered level 1. Lastly, professionalism was the most common level 3 outcome reported.

Conclusions

The objectives of this systematic review were to determine the role peer feedback plays in student learning and professional development, ascertain the impact peer feedback might have on team dynamics and success, and learn if and how the quality of peer feedback is assessed. Our review highlights the heterogeneity of the current literature regarding the use of peer feedback in undergraduate medical education. Overall, peer feedback in a collaborative learning environment may be a reliable assessment for professionalism and aid in the development of professional behavior. Many studies felt that peer feedback was appropriate to use in a formative manner. Most studies do not address the importance of the quality of peer feedback provided by students. Due to the wide variations in the outcomes defined by these studies, it may be beneficial to have more standardized definitions for student learning, team-dynamics, and professionalism. Despite the large variety of contexts and outcomes studied, there seems to be a consistent message. Peer feedback in collaborative learning is feasible and may be useful. Steps to ensure success include training faculty and students on peer feedback methods and purpose. Because developing and implementing peer feedback systems takes significant energy and resources, further studies should increase in methodologic reporting rigor and seek to expand outcomes to include, but not be limited to, quality of peer feedback (including the effectiveness of providing faculty and student training), the effect on academic performance, institutional culture, and benefits to future employers and patients.