Introduction

Assessment is an important tool to evaluate the teaching and learning process for improvement purposes (Nykowanna, 2019). In broader terms, assessment helps schools judge and decide about excellent conditions of student learning as well as the quality and effectiveness of instruction (Jones et al., 2009; Tosuncuoglu, 2018). In this context, Bloom devised a taxonomy used to develop assessments that consider each of the six levels of hierarchy in the cognitive domain (Ramirez, 2017). Lower-order thinking domains include knowledge (recalling details), comprehension (description in somebody’s words), and application (using existing knowledge to produce results). Higher-order thinking domains encompass analysis (finding out links between facts and concepts), synthesis (producing a new original work), and evaluation (judging and showing one’s position) (Deller, 2019; Freahat & Smadi, 2014).

Scholars used the two cognitive categories of Bloom’s taxonomy to analyze and determine the levels of questions asked in exams, and they established two types, namely lower-level questions (LLQs) and higher-level questions (HLQs) (Ramirez, 2017). The LLQs focus on the recall of fundamental and universal concepts and processes. On the other hand, the HLQs are well advanced and challenging because they require students to engage in deeper and analytical thinking processes (Assaly & Smadi, 2015; Pendergast & Swain, 2013; The Learning Center, 2020). The level of student thinking activities during both instruction and exams is critical for their intellectual abilities, higher performance, and professional success. As a result, exams should be designed to encourage students to give opinions about the exam questions, form creative answers, and relate the exam answers with their own experiences and real-life situations (Assaly & Smadi, 2015; Sydoruk, 2018).

Hence, writing high-quality exams that include both HLQs and LLQs plays a crucial role in helping students achieve the target learning outcomes and evaluate their level of proficiency in a specific course (Wachiuri et al., 2017). The HLQs help students deepen the learning materials and promote their critical thinking mindset and creativity. In this vein, Song (2019) claimed there is a positive relationship between teacher’s questions and students’ critical thinking. In the same line of thought, Wagner (2011) underscored that high-quality education should infuse the seven survival skills at higher-level thinking. Daggett (2016) echoed Wagner’s views and indicated that beyond knowledge and remote learning, schools must emphasize the higher-order skills to develop critical thinking citizens through higher-order materials.

Therefore, the present study seeks to examine the degree to which the questions of the Primary Leaving Examinations in the English subject prepared by the Rwanda Education Board include both higher and lower-order thinking levels. In the first place, it is essential to assess how well students master the information within the six levels of the taxonomy. In a step further, it is critical to analyze whether the exam questions of the national Primary Leaving Examinations in the English courses are based on both higher and lower-order thinking levels.

Context of the study

As a Belgian colony, Rwanda has used French as the medium of instruction and language of administration from the colonial period (1916) up to 1994 (Kayigema & Mutasa, 2014). In December 2008, the government declared that English would replace French as the teaching language, and the decision became effective in January 2009. In 2015, the government launched a competency-based curriculum in substitution of a knowledge-based curriculum (Ndihokubwayo et al., 2019; Ngendahayo & Askell–Williams, 2016).

The Rwanda Education Board (REB) viewed the new language of instruction and the curriculum reform as practical tools to provide Rwandan students with the skills and competencies they need to become entrepreneurial as well as locally and internationally innovative (Ndihokubwayo & Habiyaremye, 2018). Thus, the government believed that skilled citizens could help the country become an economic hub in central Africa and make a giant leap towards becoming an African Singapore (Krikorian, 2019; The World Bank & Government of Rwanda, 2018).

First, English would place Rwandan learners in a better position to exploit job and business opportunities at local and international levels (Euromonitor International, 2010). That is why the government decided to use English as the language of instruction from Primary four (fourth grade) onwards. Second, along with English as a classroom language in 2008, the REB launched the competency-based curriculum in 2015 because authorities believed it was conducive to promoting a skills-based economy (Government of Rwanda, 2020).

Third, the REB hoped the CBC would promote generic and intersecting competencies such as critical thinking, creativity, problem-solving, communication, interpersonal relations, life skills, and life-long learning. The REB thought such competencies would enable students to become life-long learners who can adapt to the fast-changing world and the uncertain future (Ngendahayo & Askell–Williams, 2016; Rwanda Education Board, 2015a).

Currently, many countries all over the world have adopted a competency-based curriculum (CBC) since its inception in the 1970s in the USA (Sifuna and Obonyo, 2019). Germany and Australia introduced the CBC in their education system in the 1980s and the 1990s, respectively, while England rolled it out in 2008–2010 (Anderson-Levitt, 2017; Nsengimana, 2021). The province of Quebec adopted competency-based approaches in 2001, and the Canadian central government reformed the K-12 curricula by incorporating competencies beginning in 2015 (Christensen & Lane, 2016). In Asia, Singapore implemented CBC in 2010, and China enacted the CBC in 2013 (Rajandiran, 2021; Wang, 2019). In Latin America, Mexico introduced the CBC in 2004, and Brazil initiated it in 2013 (Anderson-Levitt, 2017; Costin & Pontual, 2020).

The concept of CBC has recently also gained popularity in Africa, and this curriculum reform is implemented in different parts of the continent in an effort to attain quality education and ensure that learning leads to high standards and levels of performance (Ruth & Ramadas, 2019; Sifuna & Obonyo, 2019). Some African countries such as Botswana, Cameroon, Senegal, and South Africa rolled out the same curriculum that underscores authentic assessment strategies and promotes competencies and the use of knowledge and skills in real-life situations (Ruth & Ramadas, 2019; Sifuna & Obonyo, 2019). More specifically, in the East Africa Community where Rwanda is a member state, Tanzania and Kenya adopted the CBC in 2005 and 2017 respectively to better empower learners with skills and attitudes they need for job market competitiveness in the twenty-first century (Ondimu, 2018).

Several situational factors contribute to effective implementation of competency-based approaches such as sufficient resources, culture, and traditions, as well as equity and quality (Ryan et al., 2009). Other elements key to success are teacher awareness and attitudes on curriculum changes and collaboration among CBC implementors (Dlamini et al., 2018; Nsengimana, 2021; Scheopner Torres et al., 2018). Unfortunately, such laudable efforts seeking to take a similar economic and social development path as Rwanda faced multiple challenges. They include lack of teacher preparedness and training on the new curriculum, mismatching between the content and pedagogical practices, poor public participation and haphazard CBC implementation, and lack of approved textbooks for learners (Akala, 2021; Sifuna & Obonyo, 2019). Additionally, Ruth and Ramadas (2019) indicated that there has been lack of contextualization of CBC implementation, limited understanding of CBC objectives, and its implementation with fidelity (Kafyulilo et al., 2013; Komba & Mwandaji, 2015).

For proper mastery and implementation of the competency-based curriculum, the Government of Rwanda initiated the District Continuous Professional Development Committees (DCCs) in July 2016, which later on in January 2017 culminated in Sector/School-based Continuous Professional Development Committees (SCCs) (Rwanda Education Board, 2019). The ultimate mandate of SCCs is to enable teachers to develop professionally to achieve quality teaching and learning by improving classroom practice as stipulated in the new curriculum (Rwanda Education Board, 2019). It is claimed that the most significant opportunity for SCCs for teachers is at the school level, where they can frequently engage with each other to support improvement in practice (Rwanda Education Board, 2019).

Problem statement

The design and formulation of a curriculum as a policy is the initial phase of the process. The implementation with fidelity is another yet critical step that needs to look into the degree to which such high-standards goals in books are being attained in schools through effective instructional and assessment practices (Makunja, 2016; Tabaro, 2018). Hence, upon the implementation of the CBC, it is crucial to examine the extent to which primary leaving examinations are on the path towards competencies using the Bloom’s taxonomy approach.

More specifically, this study intends to analyze the Primary Leaving English Examinations (PLEEs) by scrutinizing the cognitive domain levels according to the Bloom’s taxonomy framework. The REB (2015b) emphasized that, within the CBC framework, examiners should give questions from the higher levels of Bloom’s taxonomy more weight than those from the knowledge and comprehension level. In addition, the REB recommended that examiners must ensure questions do not require memorization or recall answers only but test for broad competencies as stated in the syllabus (Gottipati & Shankararaman, 2017). In other words, the REB urged examiners to place more importance or weight on questions that indicate or reflect higher desired competencies (Rwanda Education Board, 2015b).

Research question

This study investigates the cognitive levels of the questions used in the English primary six leaving examinations for all schools in Rwanda administered nationwide from 2013 to 2019. We thus put forward the following research question: To what extent do English primary six leaving examination questions cover the lower and higher-order cognitive levels of Bloom’s taxonomy?

Needs and significance of the study

The purpose of the study is to identify the degree to which the PLEEs questions included both higher and lower-order cognition levels of Bloom’s taxonomy. The study is the first of its nature that investigates the cognitive level of the questions in primary six English exams in Rwanda. Most literature on assessment in Rwandan education has mainly focused on higher education (Bahati et al., 2016; Mugisha, 2010; Tabaro, 2017). A handful of studies have investigated assessment in secondary schools only (Ndihokubwayo et al., 2020; Ngendahayo & Askell-Williams, 2016; Sibomana, 2016). As mentioned earlier, Rwanda embarked on high-stakes educational reforms involving the switch from French to the English language as a classroom language in 2008 and the implementation of the CBC in 2015 (Ndihokubwayo et al., 2020). Since the inception of the two high-stakes policies, there has been no empirical study on assessment practices regarding teaching English as a foreign language (EFL) in primary schools.

Thus, this study will be the trailblazing empirical investigation of EFL assessment practices in upper primary schools. More specifically, the work will be helpful to a wide range of people, including the REB and other government decision-makers in education, school leaders, and teachers. The study will also benefit the research community interested in evaluating student learning outcomes in general and assessment practices regarding curricular reforms and language education policies involving teaching EFL in postcolonial contexts.

Literature review

Bloom’s Taxonomy of educational objectives

The most common taxonomy used in education is that of Bloom (Krathwohl, 2002; Wilson, 2016). He created this taxonomy in 1956, and it is a hierarchical listing of thinking. Among other countless uses, the taxonomy helps instructors to teach and students to learn. Bloom’s taxonomy is a popular tool used in educational services to develop assessments that measure the six cognitive levels (Deller, 2019). Bloom’s taxonomy has been used in several studies that aimed to assess the languages (Köksal & Ulum, 2018; Pikhart & Klimova, 2019), student learning outcomes and course exam questions (Welch et al., 2017), school curriculum (Koç & Öntaş, 2020), and school workbooks (Uğur, 2019). Bloom’s taxonomy has also been a tool to investigate the type and the cognitive level of the textbooks’ questions and a guide to classify questions (Zorluoglu et al., 2020).

Krathwohl (2002) revised the initial taxonomy by removing the “Synthesis” and including the “Creation” level as the highest category. He came up with six cognitive skills including remembering, understanding, applying, analyzing, evaluating, and creating. (Armstrong, 2010; Pendergast & Swain, 2013; Wilson, 2016). As indicated in Table 1, these revised Bloom’s taxonomy cognitive domain levels are divided into two levels. Higher-Order Thinking Skills (HOTS) comprise creating, evaluating, and analyzing. Lower-Order Thinking Skills (LOTS) include applying, understanding, and remembering. Table 1 visualizes the revised Bloom’s taxonomy levels ordered step by step from down to top.

Table 1 Cognitive domain levels of the revised Bloom’s taxonomy

The revised Bloom’s taxonomy has been used as a framework to assess language learning (Köksal & Ulum, 2018; Rosell-Aguilar, 2017), English curricula (Arvianto et al., 2020), and English examination or English question papers (Al-Khayyat, 2020). Hence, understanding the revised Bloom’s taxonomy is critical for writing high-quality English examination questions that test all required cognitive skills to reflect the target competencies.

Assessment and washback effect

Good questions do not only stimulate effective learning and assessment, but they must also align with curriculum and instruction because assessment greatly influences both learning and teaching. The phenomenon is described as wash-back effect, and it refers to the impact of testing on teaching and learning (Sundayana et al., 2018). This effect influences the attitudes, behaviors, and motivation of teachers, learners, and parents either in a positive or in a negative manner (Elshawa et al., 2016; Sundayana, et al., 2018). When this effect enhances teaching quality and student learning, the effect is positive, which is a sign of good teaching practices. Conversely, the negative washback is due to the structure of content or format of the course or exam designed in a narrow and partial way (Elshawa et al., 2016). This implies that proper planning of effective assessment is essential for effective teaching and learning (Tosuncuoglu, 2018). Therefore, drafting good exam questions requires critical attention to the effect of assessment on both teaching quality and student learning outcomes.

To achieve this goal, Pendergast and Swain (2013) recommended the alignment of the curriculum, instruction, and assessment. Based on constructive alignment principles, assessment tasks, teaching, and learning experiences must be linked to the desired curriculum learning outcomes. Therefore, Pendergast and Swain (2013) go on to urge that teaching and learning experiences should mirror the assessment tasks content in order to prepare leaners appropriately for what is anticipated in the exams.

Sustainable assessment for life-long learning

Traditional classroom assessment has long focused on evaluating the instructional quality and measurement of student learning outcomes (Andrade & Brookhart, 2020). Beyond this primary goal of assessment, Boud and Soler (2015) proposed the notion of sustainable assessment. It aims to prepare students to develop learning spirit for the future and empower them with knowledge and skills that will allow them to operate successfully in a complex and challenging society (Baird et al., 2017; Boud, 2000; James, 2017). Sustainable assessment contributes to training students to become self-managing people, who, in collaboration with others, “can draw on whatever they need to continue learning effectively beyond the end of the course and be able to make judgments about their learning outcomes” (Boud & Soler, 2015, p. 3). Sustainable assessment thus promotes self-reflection, judgment, and readiness for practice. It boosts evaluation and prepares students for life-long learning in an unknown and unpredictable future (Nguyen & Walker, 2016).

In a dynamic and demanding world, knowledge and information retrieval, as an outcome of education, are no longer sufficient to produce the type of citizens that countries need to cope with economic, social, and technological changes (Assaly & Smadi, 2015). Thus, sustainable assessment does not focus on content knowledge and generating grades but rather on ensuring that relevant and meaningful learning occurs (Bramwell-Lalor, 2018). In other words, assessments should prepare students for life-long learning (Nguyen & Walker, 2016). Hence, with rapid development in digital technologies that are growing, education is forced to shift from knowledge acquisition in class to empowering learners with HOTs to prepare them to compete globally and succeed in the twenty-first century (Dall et al., 2018; Pendergast & Swain, 2013; Ross, 2018).

Methods

For the purpose of this article, we used The Primary Leave English Examinations (PLEEs) as the study materials. We investigated the PLEEs that the REB has written and administered for 7 years, from 2013 to 2019. We collected data through the content analysis methodology. This approach is used to identify, organize, and analyze occurrences of specific messages included in texts. With the content analysis approach, the analysis unit is the smallest item on which researchers focus. It is supposed to bear minimal information responding to one or a series of research questions (Erlingsson & Brysiewicz, 2017; Yanovitzky & Weber, 2020). According to Titscher et al. (2012), “the units of analysis are the smallest components of texts in which the occurrence and the characterization of variables (properties, categories) are examined” (p. 58). An analysis unit can be a single word, a structure, a syntactic construction, or a theme (Titscher et al., 2012). In the framework of this study, the analysis units are action verbs used in association with one of the revised Bloom’s taxonomy six cognitive levels.

First, in efforts to answer the research question, we identified, listed, and analyzed question stems based on each of the cognitive level and action verbs referring to the lower or higher thinking order of the revised Bloom’s taxonomy. To do so, we developed a checklist based on Bloom’s taxonomy six cognitive domains (Assaly & Smadi, 2015; Atiullah et al., 2019). The checklist included a table with nine columns comprising the cognitive levels (higher and lower), the six cognitive domains, and action verbs used in each exam section for each of the 7 years of exam administration (2013–2019). Each exam paper included four areas, namely comprehension, grammar, vocabulary, and composition. We obtained a copy of the seven national PLEEs that we used to identify and analyze the questions. We found 574 questions that we scrutinized to investigate the degree to which the PLEEs addressed both higher and lower cognitive levels of the revised Bloom’s taxonomy.

Second, we combined the checklist with an explanatory table that contained the cognitive levels, the six cognitive domains, definitions, examples, and illustrative action verbs. We used the table to quickly and accurately decide the cognitive domains used in the exam questions. We coded these action verbs of the cognitive part for each level: (1) remembering; (2) understanding; (3) applying (4); analyzing (5); evaluating; (6) creating.

Third, after coding exam questions according to the six cognitive domains, we recorded and presented data in tables. We provided frequencies for each school year, cumulative summation of questions in each of the six cognitive domains over 7 years, and related percentages for the 7 years of exams. We employed Microsoft Excel to perform those computations and analyze cognitive levels. Finally, we drew the comparative information using tables indicating frequencies and percentages, as illustrated in Tables 25.

Table 2 Frequencies and percentages of the six levels of the cognitive domain in the revised Bloom’s taxonomy in comprehension questions

Findings

As captured in Tables 25, we displayed frequencies and percentages followed by sample questions to illustrate the pattern of HOTS and LOTS found in the four sections of the national PLEEs.

Revised Bloom’s taxonomy in comprehension exam questions

In Table 2, we note frequencies and percentages of cognitive levels employed in comprehension exam questions.

In this section, most questions are related to the lower thinking domain levels of remembering, understanding, and applying that make up the total percentage of 94.96%, 4.20%, and 0.84% respectively. As illustrated in Table 2, there was no question pertaining to the higher-order thinking levels. Sample questions from the comprehension section are presented below:

  • Answer these questions using full sentences.

  • Choose true or false.

  • How is this story important to you as a student?

  • Match the term with their meanings according to the passage.

  • What type of foods should people not eat?

  • Write three measures of controlling diseases in our community.

Revised Bloom’s taxonomy in grammar exam questions

Apart from five questions that emerged from the analyzing domain of the higher-order cognition level, Table 3 indicates that the majority of the questions in this section come from the three lower-order cognitive levels, namely remembering, understanding, and applying.

Table 3 Frequencies and percentages of the six levels of the cognitive domain in the revised Bloom’s taxonomy in grammar questions

As shown in Table 3, the trend from 2013 to 2019 indicates that grammar exam questions overwhelmingly covered the low order thinking levels of remembering (80.69%), applying (16.55%), and understanding (1.03%). A marginal percentage (1.72%) of questions came from one higher cognition level (analyzing). Moreover, the percentage of remembering questions (80.69%) is higher than that of applying level (16.55%) and almost five times greater than applying and understanding levels combined. Sample questions illustrating the four cognitive levels are listed below:

  • Fill in the gaps with the simple past tense.

  • I had breakfast then I went to school. (Rewrite using: ……… after ………).

  • Make the following sentences singular.

  • Organize each set of words into a meaningful sentence.

  • Write the correct word from those given below for each statement.

Revised Bloom’s taxonomy in vocabulary exam questions

Table 4 illustrates that there are only five questions pertaining to the applying domain of the lower-order cognition level. The rest of the questions in this section come from the lowest order level, that is remembering.

Table 4 Frequencies and percentages of the six levels of the cognitive domain in the revised Bloom’s taxonomy in vocabulary questions

As displayed in Table 4, vocabulary exam questions covered only the two low order thinking levels of remembering (96.93%) and applying (3.07%). Similar to the grammar part, the vocabulary section lacks higher-order thinking questions. Additionally, it is worth noting a large percentage of remembering questions (96.93%) against a minimal number (3.07%) of questions that emerged from the applying domain. Here are sample questions formulated as follows:

  • Choose the correct answers and complete the sentences.

  • Fill the following gaps with the correct word from the list below.

  • Give the opposites of the following adjectives.

  • Match the words in side A to their synonyms in side B using arrows.

  • Use the words in the box to complete the story.

  • Write down the abbreviations for the following words.

Revised Bloom’s taxonomy in composition exam questions

Two levels of thinking domains come out of six in the composition exam questions. Table 5 portrays the frequencies and related percentages of the identified cognitive levels.

Table 5 Frequencies and percentages of the six levels of the cognitive domain in the revised Bloom’s taxonomy in composition

The findings obtained for composition exam questions indicate the occurrence of one higher-order cognition level, namely the creating domain with only two questions. As can be noted from Table 5, there is no exam question pertaining to the other remaining five cognitive levels. The two questions exemplifying the creating themes are mentioned below:

  • Choose one topic and write a composition of about 150 to 200 words in the space below.

  • Choose one question and write a story of not more than 100 words.

Revised Bloom’s taxonomy in exam questions for 7 years

Table 6 presents the frequencies and percentages of cognitive levels in all sections of exam questions for seven years from 2013 to 2019.

Table 6 Frequencies and percentages of the six levels of the cognitive domain in the revised Bloom’s taxonomy in the exam questions for 7 years

As depicted in Table 6, a remarkable percentage of exam questions (98.79%) relates to the lower-order thinking skills in total. It is notable that remembering as the lowest thinking order level has the highest portion (88.00%) of the exam questions in the four exam sections. Both understanding and applying levels put together occupy the share of 10.79%. Conversely, as low as 1.21% of exam questions pertain to the higher-order cognition levels.

Discussion

The study aimed to investigate the degree to which the REB considered both lower and higher thinking levels in the questions found in the national Primary Leaving English Examinations (PLEEs) from 2013 to 2019. The analysis of 574 exam questions reveals the immense prevalence of questions that emerged from the three LOTS domains (remembering, understanding, and applying). It stands out that the remembering domain, as the lowest level, is more dominant in all sections of exam questions (88.00%) whereas both analyzing and creating domains of the HOTS levels occupy a very small number of exam questions (1.21%). Overall, as depicted in Table 5, across the four exam sections for 7 years, the national PLEE questions predominantly come from the LOTS rather than the HOTS domains. The reason behind the prevalence of the LOTS processes in the national English exam questions is likely due to three factors, namely general consideration of the learning objectives, the EFL context, and situational factors related to curricular reforms and language educational policies in Rwanda.

In general, there is a tendency of giving a paramount importance and special attention to the knowledge and remembering domains (Tikhonova & Kudinova, 2015; Tsaparlis, 2020). According to Krathwohl (2002), knowledge and comprehension are frequently considered the foundation to all the other education objectives (Chmielewska & Gilanyi, 2018). Comprehension and information retrieval are a pre-requisite for in-depth understanding of more complex thinking activities, which is probably the reason why the LOTS are the most frequent intellectual ability to be focused on in upper primary levels. In the same lines of thoughts, the individual’s knowledge increase is conducive to the development of the individual’s better grasp, interpretation, and evaluation of the world around them. As Zareian et al. (2015) claimed, “Higher-order processes such as evaluating and creating must be based upon previous knowledge of our realities, which is, what we remember” (p. 315). Hence, examiners and instructors may be inclined to include more LOTS than HOTS exams questions because they believe learners need to build stronger foundations to these lower-order thinking processes at an earlier age so that they can move on to schemata that are more complex later (Tikhonova & Kudinova, 2015).

Furthermore, as displayed in Table 7, in addition to this study, five published articles indicate a common trend of a high frequency of lower-order cognitive processes in EFL textbooks and assessment practices from the primary (elementary) to the university level because students’ proficiency levels might be low. As sample studies in Table 7 illustrate, there is an average of 78.7% of LOTS questions versus a small portion of 21.3% of HOTS exam questions in primary school. Similarly, there is higher predominance of LOTS in assessment practices and textbooks at the university level. Thus, it is assumed that students do not have a good command of the English language that would enable them to tackle HOTS activities (Lee & Wallace, 2017; Margana & Widyantoro, 2017).

Table 7 Level of use of HOTS and LOTS in exam questions and textbooks

Major factors underlying EFL students’ low English language proficiency include, but are not limited to, teachers’ training and proficiency, curriculum, student motivation and engagement, limited exposure to English, and lack of practice outside the classroom, as well as the social context and school environment (Alharbi, 2015; Al-Mahrooqi, 2012; Escudero et al., 2020; Hung, 2019). As a result, both instructors and national examination boards may tend to include more LOTS than HOTS exam questions. They suppose students’ low proficiency level may hinder them from attempting exam questions that require complex cognitive processes as described in the HOTS categories.

Implications for educational and assessment practices

According to the study findings, there is an overwhelming prevalence of the LOTS (98.79%) over the HOTS (1.21%) exam questions. Such an unbalance of questions based on the six cognitive domains in the national examinations may lead to the negative impact of the washback effect on instructional quality and student learning.

The impact of high-stake national exams on teaching and learning quality

Abundant research revealed assessment could influence teaching and learning quality (Kulasegaram & Rangachari, 2018; Wanner & Palmer, 2018). As Darling-Hammond and Rustique-Forrester (2005) pointed out, “tests can drive instruction in ways that mimic not only the content, but also the format and cognitive demands of tests” (p. 3). Scholarship indicates that well-organized and thoughtfully written assessments could enhance the quality of teaching and student learning. These types of assessment described above are likely to improve instruction because they offer challenging and thought-provoking tasks both for teachers to prepare and for students to aspire to. This type of assessment will also drive teacher professional learning in their continuous improvement efforts (Darling-Hammond & Rustique-Forrester, 2005; Mayes et al., 2020; Wanner & Palmer, 2018).

On the contrary, literature contends assessment might have negative effects on teaching quality and student learning and performance when instruction and assessment do not focus on complex reasoning, critical analysis, and creativity, as was the case in the national PLEEs. In the same vein, studies suggest that teachers tend to teach to the test under the pressure of accountability of showing high student achievement, and in some worst-case scenarios, they teach the test. They spend significant instructional time on exercises that look exactly like the exam items. They pay special attention to instructional methods such as remembering and recitation to better prepare students for national exams. National and high-stake exam-oriented accountability has been proven detrimental to instructional strategies that promote the teaching of complex reasoning, critical thinking, and problem-solving skills (Darling-Hammond & Rustique-Forrester, 2005; Ryan & Henderson, 2017; Wilson, 2018).

Additionally, when teachers are more concerned with national exams their students will take, teachers will adjust their instruction to cover specific exam items instead of considering the entire curriculum or at least the priority domains from which those exam questions come from (Darling-Hammond & Rustique-Forrester, 2005; Kuang, 2020). In this vein, when decision-makers use student scores for student promotions and teacher evaluation purposes, “teachers feel pressured to use test formats in their instruction and to teach in ways that contradict their ideas of sound instructional practice” (Darling-Hammond & Rustique-Forrester, 2005, p. 13).

Consequently, the fact that students spend more instruction time preparing for the exam or studying past exams that are highly LOTS-driven can impede the acquisition of survival skills, as described in the curriculum books. Thus, students will miss the mastery of complex reasoning skills as stipulated in the curriculum. When primary education does not consider critical thinking instruction and sustainable assessment, students are likely to face challenges in secondary and tertiary education as well as their personal and professional lives.

In the same vein, Pendergast and Swain (2013) underscored the importance of HOTS over LOTS because higher cognitive abilities challenge students to process, manipulate, and evaluate new information, which is critical for success in the twenty-first century. On the contrary, lower-order thinking focuses only on simple recall or identification of previously acquired knowledge such as reproducing and enumerating information previously memorized.

Limitations and future research

The qualitative study was the first empirical research to examine the assessment practices in primary education after the swift switch from French to English as the classroom language in 2008 and the implementation of the CBC in replacement of the knowledge-based curriculum in 2015. However, the study focused on the Primary Leaving English Examinations only. It would be better to look at national English examinations administered in the entire 12-year basic education system. This would deepen the understanding of the national education system regarding the effective and thoughtfully structured assessment approaches in alignment with the CBC content. Moreover, the study used a content analysis methodology that simply investigated sample national English examinations. Further inquiries should be extended to English textbooks used both in lower and upper primary education. Moreover, it would be useful to look into REB staff and teachers’ perceptions of the assessment practices in schools.

Conclusion

According to the findings, before the CBC implementation (2013–2015), it is noticeable that there were no exam questions categorized in the higher-order thinking levels. The rollout of the CBC from 2016–2019 does not seem to change the scenery because slight and insignificant changes were noticeable in the number of HOTS exam questions. There was a meager percentage of HOTS exam questions from the creating domain (0.34%).

Such findings do not support the government’s vision of developing a skilled citizenry through high-quality education driven by a competence-based curriculum. Scholarship in teaching English as a foreign language (EFL) suggested that the cognitive level used in assessment activities plays a pivotal role in linguistic and communicative competence development and mastery (Igbaria, 2013). Therefore, exam questions should be of various levels so that students can gain the competencies they need to succeed in school and life.

Undoubtedly, no one can claim that all questions on exams should be high level. Indeed, factual questions are important because students need to know certain basic information before engaging in higher-order thinking. Low-level cognitive questions increase the acquisition of accurate knowledge and pave the way for acquiring high-cognitive skills. However, high-level questions are practical tools for prompting and promoting thinking and improving other cognitive skills like problem solving and decision-making. Higher-order thinking skills are more challenging to learn, teach, and assess, but they are also vital because such skills are likely to be usable in novel situations. Beyond memorization and information retrieval, HOTS enable learners to become creative thinkers passionate about discovering, imagining, designing, inventing, and producing (Thomas & Thorne, 2009). Therefore, to meet this objective, Rwandan educational services should design teaching methods and assessment materials that reflect the outcomes as described in the curriculum and assess the expected standards (Alfauzan & Tarchouna, 2017; Pendergast & Swain, 2013; Sibomana, 2016).