Background

The study of medicine is demanding and puts a significant strain on the mental and physical health of medical students, who perceive medical education as an anxiety-inducing and stressful field of study [1]. This perception has been mirrored by the high level of stress in medical students. Several systematic reviews and metanalyses [2,3,4], as well as local [5, 6] and multicenter studies [7, 8] have found a significant prevalence of stress among medical students ranging from 21 to 56%. Consequently, burnout and depression among medical students have increased [4]. Other negative effects of stress on medical students have also been reported, including poor clinical performance, poor decision-making, poor peer interaction, interpersonal conflict, academic dishonesty, and sleep problems [4, 9]. Stress has also been associated with suicide, alcoholism, and drug abuse [10,11,12,13]. These negative effects eventually jeopardise optimal medical care and impact patient health negatively [14, 15]. While medical students have identified a number of stressors, research indicates that examinations are the most frequently reported sources of stress [14, 16,17,18,19,20,21]. As a result, there is a growing body of research on how to improve the mental health of medical students and promote what is currently known as resilience.

Resilience is a psychological construct that refers to the characteristics needed to adapt to the dynamic changes of life and maintain mental well-being [22]. The topic of resilience is of interest in many disciplines, including developmental psychology, sociology, education and, in particular, health professions education (HPE) [23,24,25,26,27]. In psychology, resilience generally refers to the status of an individual who is adapting to significant adversity while maintaining good mental and physical well-being [28]. Alva [29] defined academically resilient students as those ‘who sustain high levels of achievement motivation and performance despite the presence of stressful events and conditions that place them at risk of doing poorly in school and ultimately dropping out of school’ [29]. Further, academically resilient students are able to maintain mental agility and continue growing and developing despite academic and life setbacks [30,31,32,33]. In a recent meta-analysis of resilience constructs across 21 resilience scales, Wadi et al. [34] identified four primary resilience characteristics: 1) control: maintaining composure and control in the face of stressful adversity; 2) involvement: being committed to overcoming adversity; 3) resourcefulness: using available resources for appropriate solutions to overcome adversity; and 4) growth: continuing to grow and rebounding stronger from adversity. These four constructs provide a solid foundation for the implementation of interventions fostering resilience.

Studies on HPE have shown that resilience is positively correlated with compassion, satisfaction, and patient care and negatively associated with burnout, secondary stress, anxiety, intolerance to ambiguity, and poor communication [35]. Numerous health institutions have implemented interventions based on these studies [15, 36, 37]. Common intervention guidelines include training workshops focussed on psychosocial skills, such as mindfulness, Stress Management and Resilience Training (SMART), and narrative and simulation training [38,39,40,41]. Although these interventions have been shown to have some positive effects [42], they lack a solid theoretical foundation and do not focus on assessment, the primary source of student stress. Resilience theory must be integrated with the assessment context to provide a solid basis to guide intervention strategies and maintain the quality of assessment. Interestingly, a recent paper explored the intersection between resilience and curricula. The authors presented a principle-based approach to curriculum design to foster resilience as an integral part of the curriculum in higher education [43]. Although this approach sheds light on the philosophical approach to building a curriculum to create resilient graduates, it does not address the exact link between assessment and students’ resilience. To the best of our knowledge, no study has incorporated resilience into the assessment process. Accordingly, the purpose of this study was to develop a framework to promote resilience in assessment planning and practice. In this way, assessment would serve as a source of resilience and promote the development of resilience-improving characteristics among students.

Methods

The researchers followed the guidelines for developing a theoretical model proposed by Whetten [44]. These guidelines include four essential questions for model development: (1) What are the constructs/factors that should be considered to explain the model? (2) How are these constructs/factors related to each other? (3) Why is the proposed relationship represented by this portray? (4) What are the implications of this model for research and practice? Each question refers to a developmental phase of the theoretical model. Accordingly, the researchers developed the model in these four phases. This involved establishing the foundation (literature review and focus group discussion) and triangulating the findings [45] through a content validation and response process. Figure 1 presents a flow chart summarising the study phases, questions, research areas, methods, and outcomes of each phase.

Fig. 1
figure 1

Flow chart summarizing study phases, questions, research areas, methods, and outcomes of each phase

Phase I

Identifying resilience constructs

In this phase, we aimed to answer the following question: ‘What are the constructs/factors that should be considered in developing the model?’ Accordingly, we first identified the research areas related to the model development: academic resilience, assessment, and test anxiety. Then, we examined narrative and scoping reviews and focus group discussions to find evidence in these areas.

Narrative reviews

We conducted three concurrent narrative reviews to collect sufficient scientific research for the model synthesis [46]. The first review identified the theoretical foundations and factors influencing academic resilience. The second review delineated the theoretical foundations of assessment in HPE. The third review identified the theories behind test anxiety. The articles included in this review were compiled into a table of evidence synthesis [47] (Appendix I) to extract key information regarding the theoretical foundations of academic resilience, test anxiety, and assessment systems and match them with the four resilience constructs [34].

Scoping review

Four of the authors (MW, MSBY, AFA, and NZ) conducted a scoping review to identify factors affecting test anxiety [48] following Arksey and O'Malley [49] stages of scoping reviews. Six electronic databases were used: PubMed, CINAHL, PsychINFO, ERIC (through EBSCOHST), SCOPUS, and ProQuest. The Preferred Reporting of Items for Systematic Reviews and Meta- Analyses (PRISMA) Statement [50] was followed to report the scoping review steps. Based on the factors identified in each study, codes from all studies were compiled to generate subthemes and overarching themes [48]. Appendix II contains a detailed description of the scoping review.

Focus group discussion (FGD)

Four of the authors (MW, MSBY, AFA, and NZ) conducted an FGD to elicit students’ knowledge, perspectives, and attitudes regarding test anxiety (TA) and their coping guidelines [51]. Appendix III contains a detailed description of the FGD steps and procedures.

Phase II

Relating the identified constructs and their factors with each other

Phase 2 answered Whetten’s second question: ‘How are these constructs/factors related to each other?, and based on the notion of triangulation [45], the authors performed three subsequent steps:

  • I. Based on the output of the scoping review and FGD, the factors decreasing TA were qualitatively analysed to generate guidelines for decreasing TA. Initially, similar factors were grouped together. Then, a suitable guideline was constructed capturing these factors. The initial guideline statements were revised to remove redundant statements and merge similar guidelines into a single statement. This iterative process was done until consensus was reached among the authors.

  • II. Guided by the assessment cycle [52] and the sociotechnical model of assessment [53], the guidelines were thematically categorised into five phases of assessment. Each phase was named and defined.

  • III. To evaluate the content validity of these guidelines, the authors utilised a structured tool called the Content Validity Index (CVI) [54, 55], which measures the proportional agreement when two or more expert panels independently evaluate the relevance of a model’s contents to the domain of interest. Ten experts, who were medical education and student assessment specialists, were invited to join the panels [54]. A four-point Likert scale was created in a Google form and sent to the experts via email. They were asked to evaluate the relevance of each guideline to its corresponding category (phase of assessment) [56]. An answer of 1 indicated the guideline was irrelevant, whereas an answer of 4 indicated that the item was extremely relevant.

  • IV. Three indices were calculated: item/guideline-level CVIs (I-CVIs), scale-level CVIs (S-CVI) over the universal agreement method (S-CVI/UA), and the average calculation method (S-CVI/Ave) [54, 55, 57]. In I-CVIs, the relevance of each guideline to its related phase of assessment was evaluated by experts. Using a dichotomous rating of relevance, experts’ ratings of 1 or 2, indicating non-relevance, were counted as 0, while ratings of 3 and 4, indicating relevancy, were counted as 1 [57]. In S-CVI/UA, the proportion of guidelines receiving a rating of 3 or 4 (relevant) from all expert panels was calculated. In S-CVI/Ave, the mean I-CVI score for all guidelines was summed [57]. Appendix V contains the full content validity protocol.

  • V. Based on these indices and the experts’ recommendations, five guidelines were removed, so the final set included 19 guidelines.

Phase III

Sorting and ranking guidelines by experts

This phase answered Whetten’s third question: ‘Why is the proposed relationship represented by this portray?’ The same experts who were invited for content validation were asked to sort and rank each guideline into the appropriate four resilience constructs [58]. Using the checkboxes grid on the Google form, all of the guidelines were listed in one column, and the resilience constructs were placed at the top of the four adjacent columns. A column headed ‘not applicable’ was added for any guidelines that were irrelevant to the resilience constructs.

The responses were analysed using Microsoft Excel’s sorting and ranking functions. For each resilience construct, an Excel-based graphical representation was created based on the consensus of 50% of the experts if they categorised this particular guideline under a specific resilience construct.

Finally, using the Draw.io website, a conceptual map was created to link the guidelines-based assessment phases and resilience constructs to the theoretical foundation of phase I.

Phase IV

Implication for practice

In phase IV, the authors aimed to answer Whetten’s final question: ‘What are the implications of this model for research and practice?’ We evaluated the guidelines from the users’ perspective based on the response process method [59]. Twenty [20] participants were invited via email [60]. Apart from the invitation, the email contained a description of the research objectives and the validation process. A link to a video describing the application of SAR in assessment practice was also included. A response process form was attached to the email. The medical teachers were asked to review all guidelines and rate each of them based on its clarity and comprehensibility using a four-point scale (1 = not clear and comprehensible, 2 = somewhat clear and comprehensible, 3 = clear and comprehensible, 4 = very clear and comprehensible). Additionally, the participants were asked to provide written feedback on open-ended questions about the feasibility and applicability of the guidelines. Appendix VI contains the full content validity protocol.

We calculated three FVI indices: item/guideline FVIs (I-FVIs), scale/model FVIs using the universal agreement method (S-FVI/UA), and scale/model FVIs using the average method (S-FVI/UA) (Ave). First, all ratings were converted to a dichotomous scale: not clear (ratings of 1 and 2) and clear (ratings of 3 and 4). Then, we calculated the percentage of medical teachers who gave each guideline a ‘clear’ rating (I-FVIs). The proportion of guidelines receiving a rating of 3 or 4 (clear) from all medical teachers was then determined in S-FVI/UA. The average score of all I-FVIs for all guidelines was calculated in S-FVI/Ave. Based on these indices and the recommendations of the medical educators, two guidelines were eliminated, resulting in a final set of 17 guidelines.

The final configuration of the framework

After all these phases, Microsoft Visio® was used to reshape the framework for the systematic assessment of resilience to make it more comprehensible and straightforward to implement. The four resilience constructs were placed in relation to the assessment phases, and the final list of phase-related guidelines was also placed.

Results

Phase I

The narrative reviews yielded relevant theoretical foundations in three areas: academic resilience, assessment, and test anxiety. These foundations were tabulated, and each relevant implication(s), through which resilience could be promoted, was identified (Table 1). In the first three columns, research areas, theoretical foundations/frameworks, and the implications of each study were presented, respectively. Guided by the integrated resilience model [34], every implication was matched to its suitable four resilience constructs: control, involvement, resourcefulness, and growth [34]. The matching was based on the agreement of the authors.

Table 1 Matching the implication of the identified theoretical frameworks with resilience constructs

The scoping review revealed that factors related to test anxiety can be categorised into two groups: those that increase TA and those that decrease TA (Appendix II). Likewise, the thematic analysis of the medical student focus group discussion yielded three major themes: students, academic resources, and examiners. Each theme was subdivided into subthemes that corresponded to an increase or decrease in TA [51] (Appendix III).

Phase II

Compiling and categorizing the generated guidelines

Based on the scoping review and FGD findings, the authors compiled a list of test anxiety-reducing guidelines (Appendix IV). The list was evaluated, and guidelines that were duplicated were eliminated. Guided by the assessment cycle [52] and the model of ‘assessment as a sociotechnical system’ [53], the authors categorised the generated guidelines into groups and subsequently named and defined each group. The following five phases of evaluation were identified:

  • 1. Assessment direction, which focuses on improving the candidate’s understanding of the assessment’s scope and procedure.

  • 2. Assessment preparation, which emphasises enhancing the candidate’s cognitive, mental, and psychomotor readiness to optimise assessment performance.

  • 3. Assessment experience, which enhances the formative assessment component.

  • 4. Examiner focus, which relates to improving examiner behaviour to improve the candidate’s performance and decrease the candidate’s anxiety.

  • 5. Student reflection, which encourages reflection by students.

Table 2 shows each phase and its related guidelines.

Table 2 Categorization of SAR guidelines based on stages and phases of assessment

Content validation of the guidelines

Six of the 10 expert panels responded to the invitation to participate in the content validation process. The expert panels had extensive experience in medical education, student assessment, and psychological well-being (Appendix V). Following the principles for calculating CVIs [55, 57], all items (guidelines) with an I-CVI of 0.83 or higher were deemed relevant and included in the response process study based on the CVI. A global acceptance level of 0.80 or higher for S-CVI/UA and S-CVI(Ave) indicates that all components are relevant to the framework. Before the subsequent step, modifications were made based on feedback, which resulted in the elimination of items scoring less than 0.83. Table 3 shows the list of guidelines based on the content validation.

Table 3 The CVIs of the guidelines after removing four items

Figure 2 illustrates the correlation and configuration of the relationship between the findings of phase I of the study. The right side of Fig. 2 presents the recommended guidelines for fostering resilience through five phases of assessment. They are connected to the four resilience constructs with arrows that point in both directions to illustrate the reciprocal relationship between guidelines and resilience constructs. The four resilience constructs were linked to the theoretical foundations of academic resilience and the assessment system, and they were shown to have an inverse relationship with test anxiety.

Fig. 2
figure 2

Mapping phase I output with the SAR framework and its guidelines

Phase III

Figures 3, 4, 5, and 6 display the results of the experts’ ranking and sorting. Each figure represents a resilience construct and its corresponding guidelines, on which 50% of the experts agreed.

Fig. 3
figure 3

Sorting and ranking of guidelines relating “self-control” construct

Fig. 4
figure 4

Sorting and ranking of guidelines relating “management of resources” construct

Fig. 5
figure 5

Sorting and ranking of guidelines relating “engagement” construct

Fig. 6
figure 6

Sorting and ranking of guidelines relating “growth” construct

Phase IV

The response process aimed to evaluate the use of SAR and its guidelines among medical teachers. Twelve [12] of the 20 invited panels responded to the invitation. The panels consisted of medical teachers from various disciplines and universities (Appendix VI). On the basis of the FVI [60], 17 items (guidelines) with an I-FVI of 0.83 or higher were retained, indicating their relevance, while two guidelines with scores below the threshold were eliminated (two guidelines). Table 4 presents the final list of guidelines. As a result, the overall framework S-FVI/UA and S-FVI (Ave) improved from 0.92 to 0.94, indicating the clarity and understandability of all framework components (Table 4).

Table 4 The FVI indices of SAR after removing some guidelines

The medical teachers also gave encouraging written insights about the guidelines and their clarity (Appendix VII), which are summarised as a word cloud in Fig. 7.

Fig. 7
figure 7

Word cloud of the most common comments made by medical educators

The SAR framework

The SAR framework (Fig. 8) incorporates the overarching relationship between the four resilience constructs [34], the five phases of the assessment process, and their relevant strategies for promoting resilience. The four constructs of resilience include 1) self-control, meaning that students should be able to govern themselves and face adversity, 2) management, which describes students’ ability to use available resources effectively to overcome obstacles, 3) engagement, which refers to students’ ability to be involved and committed to pursuing the challenge, and 4) growth, which reflects students’ ongoing development to face future challenges. The four constructs are part of an ongoing continuous cycle.

Fig. 8
figure 8

The Systematic Assessment for Resilience (SAR) framework

The ‘assessment experience’ phase is represented as the nucleus of the framework, as it is the core of the assessment process. In this phase, students’ resilience may be promoted through various strategies, such as increasing the frequency of formative assessment, encouraging targeted mock exams  [87,88,89,90,91], promoting collaborative assessment and open book exams, peer assessment, and introducing progress testing. In the ‘assessment direction’ phase, students’ knowledge of the assessment scope and process may be improved through various strategies, including the sharing of assessment mapping and the assessment rubric, briefing on the overall assessment coverage, establishing a briefing session before the exam, and familiarising students with the assessment methods. Such strategies will foster resilience, especially the self-control and growth constructs. In the ‘assessment preparation’ phase, students’ cognitive, mental, and psychomotor preparedness are improved to maximise their assessment performance. Various strategies may be used, such as advising students on study skills, time management, and exam skills, directing students to good materials for revision, and providing strategies for students to reduce test anxiety. These strategies will promote resilience, especially the self-control and management constructs. The ‘examiner focus’ assessment phase deals with examiners’ behaviour to maximise students’ performance and reduce their anxiety. This phase can be improved by establishing a non-threatening environment during examinations, for example, by smiling, engaging in welcoming and professional behaviour, building rapport, and showing a sense of humour. All of these will foster resilience, especially the engagement and management constructs. The ‘student reflection’ assessment phase promotes students’ reflection by providing space in the written assessment for students to offer feedback on the examiner. Such strategies will improve students’ resilience, particularly in the engagement and growth constructs.

Discussion

As the prevalence of pathological stress among medical and HPE students continues to rise [3,4,5,6,7,8], a number of interventional programmes have been designed to improve their mental health [15]. The primary criticism of these programmes is their emphasis on causes other than the primary cause, which is the assessment. This study provides a methodical approach for promoting resilience while practicing the assessment. A variety of strategies promote resilience through the five phases of assessment. Resilience is comprised of four constructs, which are shown in Fig. 8. In the following sections, we will discuss each assessment phase and how it can promote resilience.

Assessment experience

The assessment experience phase is purposively located at the centre of this framework, as it promotes the four resilience constructs. The assessment experience emphasises the frequency of formative assessment or any other assessment experience (e.g. mock exam), offering students an opportunity to engage in a simulated challenge similar to the real assessment. Such an experience creates a space for self-regulation and thus enhances self-efficacy and learning growth [92]. There is a growing body of research highlighting the crucial role of formative assessment in maximising learning [93, 94]. Formative assessment serves as a tool for practicing assessment, reflecting on performance, and identifying weak points and opportunities to improve actual future performance [68]. In this study, the strategies within the assessment experience were designed to maximise self-regulatory learning and evaluative judgment [92, 93]. In a nutshell, the more exposure to the assessment experience, the greater the options for strengthening evaluative judgment and hence self-regulation, which is expressed as ‘self-control’ in our framework. Furthermore, increasing ‘assessment experience’ will lead students to the focus on the prudent use of available resources, ‘management’, vividly experiencing the actual assessment, ‘engagement’, and self-esteem—in other words, on ‘growth’ (Fig. 8).

Assessment direction

The current assessment practice in the medical and HPE fields is competency-based assessment [95,96,97]. Consequently, it is crucial to communicate with students clearly about these competencies and how they will be evaluated [98]. The ‘assessment direction’ involves directing students towards assessment by providing them with clear instructions on how the assessment will be administered and what is expected of them. Knowing what is expected of them during assessment will enable them to shape their learning and direct their efforts to achieve these objectives [99]. A considerable body of studies has shown that defining the assessment expectations (objectives) will allow students to target their efforts to achieve them [98, 100]. However, there is an issue in terms of which assessment criteria and standards should be communicated to students. There is no research consensus on the suitable methods for communicating with students. In our framework, we provide a variety of strategies examiners could use to direct students toward assessment.

Articulating the assessment direction has benefits for both the examinee and examiner. The examinee can then tailor his or her efforts to the assessment direction. The examiner can select suitable assessment modalities to measure the desired outcomes [96]. Additionally, it reduces the burden on examiners related to answering examinees’ questions about exams. Moreover, examinees’ test anxiety will be reduced, and they will have more agency in meeting the expectations [101]. In summary, directing students toward assessment will enhance their self-control, supporting self-regulated learning [93], and empower them to grow and face future challenges [100].

Assessment preparation

When guiding students in the assessment direction, it is important to ensure that they have the proper tools to succeed. Hence, assessment preparation plays an important role in meeting the challenge of assessment. Studies have shown that helping students control their negative thoughts and advising them on learning skills and time management will maximise their learning [102, 103]. In our framework, we believe that teachers play the central role in maximising learners’ behaviour. In addition to preparing high-quality learning materials, teachers can also provide students with rich resources to improve their mental well-being and reduce the negative effects of assessment. This practice will enhance students’ self-control and shift their mindset so they can use the resources around them to face assessment challenges.

Examiner focus

This framework highlights the examiner’s conduct during the exam. The direct encounter between examiner and examinee has a psychological dimension, creating life-altering memories that can either destroy or reinforce the examinee’s self-esteem and, consequently, resilience. While the presence of an examiner in the examination room automatically causes test anxiety, the situation will become worse if there is a lack of proper communication or if the examiner chooses to fail students based on personal preferences or biases [51]. This negative situation has been dubbed ‘the hawk effect’ [104]. Our framework proposes techniques to mitigate this unintended consequence and foster an atmosphere conducive to reciprocal communication and learning, which automatically enhances students’ ‘engagement’ and prepares them for similar situations in the future by managing learning resources wisely and effectively.

Student reflection

Self-assessment (or self-reflection) has been proven to be an effective approach to support learning engagement. In self-reflection, students evaluate their performance related to both internally set goals and externally set criteria [105]. In this framework, teachers (assessors) provide systematised avenues for self-reflection to achieve the desired resilience construct: growth. While several studies have presented different approaches for promoting self-reflection, we encourage assessors to use the most common reflection method: feedback [100]. Nicol and Macfarlane‐Dick [68] described the most important aspect of high-quality feedback: ‘Good quality external feedback is information that helps students troubleshoot their own performance and self-correct; that is, it helps students take action to reduce the discrepancy between their intentions and the resulting effects’ [68]. Numerous studies on HPE have promoted the use of feedback [106], and it is key component in the programmatic assessment framework [107]. Both feedback and self-reflection support each other to maximise learning and hence ‘growth’ [108]. Through this feedback, students will receive constructive comments regarding their performance based on the teacher’s expectation or established criteria, which they can then use internally to redesign a suitable learning path to achieve their goals [100]. Studies have shown that students’ self-reflection leads to deep learning, self-efficacy, self-regulation, and personal growth [93, 106]. While some researchers argue that students should be trained in self-regulation, others contend that self-regulation is a spontaneous process that can be maximised by providing a suitable platform to practice it [109]. Consequently, the framework provides additional resources for assessors to promote self-reflection.

Limitations and future perspectives

During the literature review phase of developing the SAR framework, efforts were made to broaden the search of the scoping review to include literature in HPE rather than just medicine, and to conduct narrative reviews that considered higher education in general. However, the FGD only included one medical school. Another limitation of this research was that it only included articles written in English, which may introduce bias (also known as language bias [110]) and result in the omission of important cultural contexts and necessary details in data synthesis. However, the triangulation of the findings with those from the scoping review, other narrative reviews, and the FGD mitigated the aforementioned limitations [45].

Conclusions

Resilience is proven to be the desired construct for medical and health professions students. It fosters several characteristics graduates need to meet future challenges and adversities. The current study presents a systematic method for fostering student resilience through assessment practice. Based on rigorous methodological research and a theoretical foundation, the study provides a set of practical strategies for promoting resilience. Through five phases of assessment, namely, assessment direction, assessment preparation, assessment experience, examiner focus, and student reflection, the SAR framework aims to promote four resilience constructs: self-control, management, engagement, and growth.