Introduction

Cosmology as a field of inquiry has roots in mythology, philosophy, and religion. Cosmology pursues answers to some of the biggest and most fundamental questions about the Universe. As a precision observational science cosmology aims to tell the evolutionary history of the universe and predict its future. Cosmology topics are prevalent and consistent across most curricula at upper secondary level (Salimpour et al., 2020a). The cosmology curriculum, because of the fundamental questions raised about our place in a mysterious universe, has enormous potential to engage the curiosity of students and create rich discussions in the classroom. Simultaneously, these topics encompass complex space–time relations which are innately counterintuitive and require reasoning beyond everyday experiences (Salimpour et al., 2021b, in review).

Conceptions and Reasoning

The history of research into student alternative/misconceptions in astronomy is rich and diverse. Over the years, various misconceptions have been identified and various evidence-based interventions proposed to scaffold students towards conceptual change. The iSTAR database (Slater et al., 2016) contains 186 articles focussing on some aspect of misconceptions in astronomy. Some examples of alternative conception studies include topics such as night/day (e.g.: Vosniadou & Brewer, 1994), seasons (e.g.: Slater et al., 2018), moon phases (e.g.: Trundle et al., 2007), cosmology (e.g.: Prather et al., 2002), size/distances (e.g.: Miller & Brewer, 2010), and astrobiology (e.g.: Offerdahl et al., 2002). One of the underlying challenges in addressing student conceptions is that they are based on intuition grounded in everyday experiences and language (e.g.: Nussbaum & Novak, 1976; Vosniadou & Brewer, 1992, 1994), making them resistant to change (Driver & Easley, 1978). Furthermore, although studies have been extremely valuable in identifying alternative conceptions, there is much work that can be done in characterising the reasoning that underpins these conceptions.

Reasoning is one of the foundations of science, and one of the thrusts of science education has been to “instil the disciplinary habits of the mind of the scientist” (Kind & Osborne, 2017, p. 9). However, as argued by Kind and Osborne (2017), despite the importance and richness of reasoning in science, science education has yet to conceptualise and characterise scientific reasoning in the classroom in a way that is reflective of the epistemic practices of science. Following Driver and Easley (1978) in order to characterise alternative conceptions in a way that allows them to be addressed effectively requires an understanding of the underlying reasoning patterns. Alternative conceptions research has in essence been of two varieties –—nomothetic and ideographic (Driver & Easley, 1978). While nomothetic studies have their place and compare student understanding to a standard, ideographic studies can productively explore the reasoning underpinning students’ conceptions (e.g.: Vosniadou & Brewer, 1994).

Previous Work on Concept Inventories

Concept Inventories (CIs) have gained popularity as a basis for formative and summative assessment processes, particularly in astronomy education. CIs are diagnostic tests consisting of multiple-choice questions designed to explore a student’s understandings of a particular construct or a series of constructs/concepts (Bailey, 2009; Sadler et al., 2009; Wilson, 2005). There has been much research into the benefits of using CIs (Bailey, 2009; Wallace & Bailey, 2010), and the methodologies used to create them (Lindell et al., 2007). As of this reading, a range of validated CIs exist in astronomy (Table 1), which have mostly been of the nomothetic type aimed at undergraduate level and focus on levels in conceptual understanding and do not necessarily privilege reasoning.

Table 1 A collection of concept inventories in astronomy, together with their purpose and target level

Validated CIs focussing on cosmology education include those of Wallace (2011), which is aimed at undergraduate students in an introductory astronomy course. The CI consists of conceptual questions on very focussed topics in cosmology: expansion and evolution of the universe, the Big Bang, and the evidence for dark matter in spiral galaxies, which are aligned to the level the concepts are taught at undergraduate level. The questions do explore student reasoning informed by construct (concept) maps for each of the topics; however, that is not the primary focus. The CI is designed to be a pre-/post-test to measure student learning gains via Lecture-Tutorials in undergraduate introductory astronomy. The work of Aretz et al. (2016) focuses on exploring student pre-instructional ideas about the Big Bang Theory, while Aretz et al. (2017) used student responses and the work of Wallace (2011) to refine a construct map about the expansion of the universe. As such, there is no CI in cosmology developed specifically for high school students that privileges a progression in reasoning combined with conceptual and declarative knowledge. Given the complex and counterintuitive nature of much of cosmology, understanding and supporting students’ reasoning are important as it helps unpack the complex space–time relations, the epistemic practices of the discipline, and the rich narrative of how we have uncovered the mysteries of the Cosmos. Characterising hierarchies of reasoning can also provide teachers with guidance on how to frame their teaching to effectively address alternative conceptions, in a way that scaffolds students to unpack the concepts in cosmology.

In our previous work (Salimpour et al., 2021b, in review) using student responses from an open-ended survey, we identified alternative conceptions which aligned with previous studies (Aretz et al., 2016; Hansson & Redfors, 2006; Prather et al., 2002; Trouille et al., 2013; Wallace, 2011), but extended these to identify the underlying reasoning patterns. Looking deeper at the alternative conceptions revealed that they were linked to three fundamental reasoning challenges associated with the following:

  1. 1.

    Navigating spatial and temporal relations over enormous scales

  2. 2.

    The counterintuitive nature of cosmological concepts

  3. 3.

    Intuition based on everyday language and experience

Using the above reasoning challenges, we were able to construct a preliminary progression scale based on the SOLO taxonomy (Collis & Biggs, 1979), which combined with the quality of conceptions, provided a potential basis for a concept inventory. This current paper describes the development and validation of a concept inventory for cosmology in high school. The paper begins by introducing the research aim, then an explanation of the framing underpinning the study. Next, the paper highlights the methodological approach to the study, the results, and subsequent analysis. The paper concludes with a brief discussion of the findings, implications of the concept inventory, and concluding remarks.

Research Aim

Effective instruments to measure student understanding need to extend beyond declarative content knowledge or base-level comprehension to characterise the reasoning associated with deeper levels of conceptual knowledge. The aim of the research described in this paper is to develop and validate a concept inventory that can be used to appropriately characterise and monitor progression in student reasoning and learning in cosmology. The Cosmology Concept Inventory (CosmoCI) is a multi-dimensional concept inventory aimed at high school students, that moves beyond characterising lists of conceptions to enable exploration of progression in sophistication of student reasoning.

Framing This Study

Given this study aims to highlight the development and validation of a concept inventory for cosmology, we describe the framing of the study with regard to validity. The notion of validity is complex, and over many decades, there have been a range of debates about the various aspects of validity (e.g.: Cronbach, 1971; Cronbach & Meehl, 1955; Lissitz & Samuelsen, 2007; Messick, 1989; Sireci, 2007; Sireci & Parker, 2006). Traditionally, validity has been considered to be of three types: criterion, content, and construct. This has evolved into a unified theory of construct validity which has subsumed criterion and content validity as evidence for a more general framing of construct validity (Messick, 1989). What is considered valid depends on the context and aim and has to include expert judgement of the items in terms of their alignment with canonical/consensus ideas. Also relevant is the internal coherence across the levels, and alignment with students’ thinking.

The American Educational Research Association, American Psychological Association, and National Council on Measurement, in the most recent version of the Standards for Educational and Psychological testing (2014), define validity as the “degree to which evidence and theory support the interpretations of test scores for proposed uses of tests” (p. 11).

From an educational perspective, Lissitz and Samuelsen (2007) emphasise the importance of content validity; however, Sireci (2007) states that “A serious effort to validate use of an educational test should involve both subjective analysis of test content and empirical analysis of test score and item response data.” (p.481). Therefore, content validity on its own is not adequate. More recently, Sireci (2016) argues that the interpretation of test scores “is part of validation, and partly what validity refers to. Validating interpretations of test scores is a necessary component of any validation endeavour. However, it is not sufficient for defending the use of a test for a particular purpose” (p. 231).

This study aims to develop a concept inventory tool that will tap into students’ knowledge and reasoning in cosmology concepts, aligned with the four dimensions of cosmology developed by Salimpour et al. (2020b): size and scale, spacetime location, composition of the universe, and evolution of the universe. This tool is intended to.

  • act as a pre-test to explore students’ knowledge in relation to key concepts in cosmology

  • alert teachers to the key ideas in cosmology and provide them with an understanding of student thinking and reasoning

  • provide a stimulus for classroom discussion

  • provide a tool to monitor student learning

Methodology

The methodological approach in this study consists of two parts; the first is the process undertaken for developing the cosmology concept inventory (CosmoCI). The second involves using Rasch analysis in the form of a partial credit model (PCM) (Masters, 1982) to establish a scale, and evaluate the internal coherence and utility of the instrument.

Development of CosmoCI

The development of CosmoCI is underpinned by a Design-Based Research (DBR) framework (Anderson & Shattuck, 2012; Collins, 1992; Collins et al., 2004). The DBR cycle used in this study is visualised in Fig. 1.

Fig. 1
figure 1

Visualisation showing the different aspect of the DBR cycle employed in this study. As is noted, each step is iterative, allowing insights to be used in the next step

The first iteration of CosmoCI consisted of 23 open-ended and five multiple choice questions. The questions were developed by reviewing curriculum statements related to cosmology in 52 curricula, which covered the OECD countries, China, and South Africa (Salimpour et al., 2020a) and the International Baccalaureate (IB) Diploma programme, and using the curriculum statements to extract key concepts in cosmology that were prevalent across most curricula. These were then categorised into four overarching conceptual dimensions of cosmology: size and scale, spacetime location, composition of the universe, and evolution of the universe (Salimpour et al., 2020b). The questions drew on previous research carried out at an undergraduate level (Wallace, 2011), filtered through the research team’s experience in both cosmology and teaching cosmology. Mostly open-ended questions were used to allow students to express their reasoning without being restricted. The first iteration of CosmoCI was implemented with a pilot group of students (n = 75). The analysis provided preliminary insights into students’ knowledge and reasoning in relation to concepts across the four dimensions and allowed for the preliminary development of a universal rubric for characterising student responses (Salimpour et al., 2020b). On this basis, also, the questions were refined.

The second iteration used a slightly refined version in the wording of questions based on student responses. This is because student responses for some questions hinted that (a) some students may not have understood the question and (b) students were not sufficiently prompted to explain their reasoning. In addition, for the five multiple-choice questions, students were now asked to explain the reasoning for their choice. This second iteration was implemented with a larger group of participants (n = 286). The analysis of the student responses allowed for the refinement of the grading rubric to include finer level distinctions (initially 16 levels and then 14 levels) that attempted to capture the levels of student reasoning, while encompassing the variety of responses (Salimpour et al., 2021b, in review). The underlying framework for building the levels of student responses was based on the SOLO taxonomy (Collis & Biggs, 1979). This was a natural outcome that had developed during the first round of analysis, except finer levels within each SOLO level were included to capture the range of student responses. The finer-grained analysis, although helpful from a research perspective, we felt would be challenging for teachers to use to gain a big picture view of student reasoning at various levels of progression. Further, the patterns of responses across the questions and dimensions were extremely complex to analyse. Therefore, the finer levels were eventually collapsed into four of the five SOLO levels: relational, multi-structural, uni-structural, and pre-structural (Salimpour et al., 2021b, in review). This approach allowed broader patterns of student reasoning to be extracted, mapped to the SOLO levels. The reason the fifth SOLO level — extended abstract — was not included is that the nature of the questions being asked and the type of responses being canvassed did not prompt this extended form of reasoning.

The concept inventory idea is essentially based on the distractor-driven multiple-choice test (DDMC) (Sadler et al., 2009). The use of a multiple-choice format allows student understanding to be objectively assessed because it is practically efficient for teachers to use in the classroom environment. The task in this study was to take the variety of student responses for each question to construct multiple choice options that were exemplars of the responses at each of the four SOLO levels. That is, for each question, each option represented reasoning at a different SOLO level, constructed through thematic analyses of the open responses. Several cycles of refinement to the wording were made by the research team. The options were carefully designed not to be identifiable as higher level based on the use of abstracted disciplinary language and/or the length of the option. The mapping to SOLO levels provides teachers with a framework to characterise and appreciate the level of sophistication in student reasoning. An example of this categorisation approach is shown in Table 2

Table 2 Example of the categorisation of student responses into the four SOLO levels and multiple-choice option constructed based on student responses. Evo3 is the question numbering for the analysis. Evo3 denotes the third question in the dimension Evolution of Universe

Brief Overview of CosmoCI

The question structure of CosmoCI (Appendix) is based on four overarching conceptual dimensions (Fig. 2) each focussing on a particular aspect of Cosmology. Each dimension encompasses key concepts and discoveries in cosmology, for example the large-scale structure, dark energy, dark matter, and the cosmic microwave background, all of which are present in curricula and found in high school physics textbooks. CosmoCI encompasses both declarative knowledge and higher-level conceptual knowledge, placing the progression of student reasoning at the core of its structure. The content coverage is organised under the four overarching, fundamental dimensions described above (Fig. 2).

Fig. 2
figure 2

Characterisation of the four conceptual dimensions of cosmology

To illustrate the scope of CosmoCI and how the multiple-choice options are aligned to the SOLO taxonomy, Table 3 provides an example of a question from each dimension. It can be noted that at each SOLO level fundamental reasoning similarities can be seen which are explained in the “Essence of levels” as shown in Table 2.

Table 3 Example of a question for each dimension

Validation of CosmoCI

This third iteration of CosmoCI (Appendix) was implemented with a cohort of students in Australia and Sweden. The participants included high school students in grades 10 and 11 (n = 234). The sample selection was based on random opportunistic sampling (Newby, 2014). Australia and Sweden are used as instances, given that curriculum statements related to cosmology are relatively homogeneous across curricula (Salimpour et al., 2020a), and the authors have knowledge of the curriculum and access to schools in Australia and Sweden.

This study uses an item response theory (IRT) approach to validation. IRT is an item level approach and in essence aims to explore the relationship between test item (question) difficulty, and student ability (e.g.: Boone et al., 2014; Hambleton & Jones, 1993). Basically, easier questions are accessible by most students, and higher ability students are more likely to answer correctly difficult questions compared to a lower ability student. One aspect of IRT models is that results are independent of the participant group. IRT and its associated models (1-, 2-, 3-parameter) are independent of the participants taking the test. It should be emphasised that although the Rasch Model in some instances is referred to as being a type of IRT (specifically 1-parameter IRT model), there is a fundamental philosophical difference “…in that one model, IRT, is altered to fit data and one model, Rasch, is not altered to fit data” (Boone et al., 2014, p. 453). This current study uses Masters’ (1982) partial credit model (PCM), which is a development of the basic Rasch model, and shares the defining characteristics associated with other Rasch models. The PCM is used for polytomously scored items, which is to say that the distractors in a question are given different levels of credit because they demonstrate a particular level of knowledge or reasoning (Masters, 1988). Therefore, multiple-choice questions are not solely marked as “correct” or “incorrect.” For tests designed for the PCM model, the reasoning underpinning every multiple-choice option is distinct, with responses providing teachers with insights into the fundamental challenges that students face with regards to complex concepts (see for example, Briggs et al., 2006). This principle of identifying student reasoning is at the heart of this study, and thus using the PCM offers a framework for pursuing this line of inquiry.

The raw data from the CosmoCI was run through a custom Python script to clean the data, and format it for use with the QUEST software package (Adams & Khoo, 1993). QUEST is an analysis package that implements Rasch analysis on both dichotomous and polytomous data.

Results and Analysis

The statistical analysis reveals a reasonable fit to the Rasch model within the acceptable range set by literature (Adams & Khoo, 1993). Figure 3 shows that the item scores fit well within the “tram lines” (dotted lines).

Fig. 3
figure 3

Analysis of the item fit to the Rasch model. Dotted lines represent the boundaries of the model fit

One of the key outputs from the analysis is a set of Wright Maps (Wilson, 2005). Wright Maps allow researchers to quickly see the relationship between item difficulty and student ability, by placing both on the same measurement scale (Logit scale – log of the odds) (Fig. 4). This scale is a probabilistic measure and not an actual measure. Wright Maps were generated from the data analysis for each of the four conceptual dimensions in cosmology: size and scale, spacetime location, composition of the universe, and evolution of the universe (Salimpour et al., 2020b). The right side of the Wright Map lists the items, with the notation used being a three-letter word (Siz, Loc, Com, Evo), followed by ‘x.y’, where x is question number in that dimension and y is the multiple-choice level. Therefore, Siz1.2, refers to question 1, multiple-choice level 2 option (uni-structural) in the dimension size and scale. The order of items increases in difficulty (bottom to top). The left side of the Wright Map lists the students represented by X (which can be any number of students), in order of increasing ability (bottom to top). A student located in the same line as an item indicates that the student has a 50% chance of choosing that item option. Items above the student’s position are harder for that student (students have less than 50% chance of choosing that reasoning option), and items below that student’s position are easier (students have greater than 50% chance of choosing that reasoning option).

Fig. 4
figure 4

Wright map showing all the questions and choice options. Each X represents 3 students. The variation in font thickness differentiates between the four SOLO levels

Looking at the Wright Map (Fig. 4), it can be seen that the multiple-choice options for each question line up in the predicted SOLO reasoning order, providing a validation of the identification of the levels of reasoning underpinning the different choices in each question. There are, however, some outliers where the SOLO levels are inconsistent across questions which will be discussed below. Overall, the Wright Map also reveals that the multiple-choice options satisfactorily align with the ability of the students; i.e. the highest ability students are able to pick the multiple-choice options which are at a relational level (4).

There are some outliers in the choice options. For example, items Evo1.3 and Evo3.3 were both assigned to a multi-structural level in design of the instrument, yet Fig. 4 shows that they are among the least difficult items when implemented with students. Item Evo1.3 gives the age of the universe as “more than 100 billion years,” and item Evo3.3 states that the Big Bang Theory “is a theory for the origin/creation of the Universe, that proves an explosion of a tiny singularity led to the formation of the Universe.” The solution to this anomaly lies in the notion of student alternative conceptions. With regard to item Evo1.3, previous work shows that students reason that because the universe is so large, it must be extremely old (Salimpour et al., 2021b, in review). While for item Evo3.3, there is a prevalent alternative conception that the Big Bang is an explosion a point in space; this has been shown in other studies as well (Aretz et al., 2016; Wallace, 2011), and is perhaps owing to the way representations depict the Big Bang Theory (Salimpour et al., 2021a). Perhaps, it is not surprising that students find this option attractive due to its pervasiveness as a popular metaphor, distinct from choosing the response through the reasoning process it seems to represent. The reason these items are deemed multi-structural lies in the fact that students need to bring together different lines of reasoning which are sophisticated, albeit alternative.

In addition to the above, the outliers are also an indication of a more fundamental challenge when collapsing the SOLO levels. The fine-grained expanded (14 and 16) SOLO categories made distinctions that took into account the following aspects present in student responses:

  • the sophistication of reasoning;

  • whether students’ ideas were scientifically correct;

  • whether students were expressing alternative conceptions in a considered way;

  • the amount of detail/justification students providing in their responses; and

  • whether the multiple-choice component (if present) of a question was correct.

Some of these distinctions were lost in collapsing the data. However, this was justified since the complexity of the open responses represented by those categories masked the broader patterns of reasoning seen more clearly when the number of categories was reduced. We argue that the simplification offered by SOLO allows the instrument to capture patterns in students’ levels of thinking across questions and the four conceptual dimensions (size and scale, spacetime location, composition and evolution of the universe). Nevertheless, we acknowledge that the SOLO levels cannot capture all elements of students’ reasoning (Biggs & Collis, 1989; Watson et al., 1995).

Discussion

This study highlighted the process of developing and validating a cosmology concept inventory — CosmoCI — targeted at the high school level. The Rasch analysis in the form of a PCM shows a good alignment of the multiple-choice items in terms of increasing levels of sophistication and difficulty in line with student ability. For every question, the levels line up in the predicted order. The iterative coding of student responses naturally fit the SOLO taxonomy which provided a general framework for characterising the level of sophistication of the reasoning underpinning student responses. One of the key challenges of this study was unpicking knowledge and reasoning, and their interaction in framing student responses. The use of two iterations of open-ended questions allowed relations between “level of knowledge” and “pattern of thinking” to be explored (Salimpour et al., 2020b) and subsequently coordinated in CosmoCI.

The complexities in student responses at first warranted a fine-grained approach and so the SOLO taxonomy was expanded to capture these; however, to capture broad parameters of reasoning collapsing to the four SOLO levels provided a suitable characterisation. While the SOLO taxonomy theoretically represents distinct levels, in practice, these have blurred interfaces that allow characteristics from one level to manifest in adjoining levels. This is particularly evident in the clustering of multiple-choice options in the Wright map (Fig. 4). For example, among the level 4 options, there are some level 3 options with the same Logit score. Nevertheless, the consistencies in ordering of levels in the map demonstrate that the SOLO taxonomy as used in this study provides a valid and useful guide to anchor the progression in sophistication of student reasoning and knowledge in cosmology.

CosmoCI, through the situating of responses in a progression of reasoning, can be used as an assessment tool pre- and post-instruction. In addition, representation of the narrative of how scientists have come to know that the dark energy makes up the majority of the universe, that the universe is 13.8 billion years old, or that the universe is undergoing accelerated expansion, coupled with the innate interest it piques in students, means the CosmoCI can be a powerful stimulus to opening up rich discussions in the classroom. This latter aspect of instigating discussions was supported by the feedback from teachers who implemented CosmoCI in the classroom. The collection of questions can set the stage for learning to begin and support teachers to frame learning activities, as much as it can evaluate student conceptual understanding and reasoning in cosmology. This idea of determining what the student knows, so that teaching can be framed accordingly, echoes the views of Ausubel (1968).

It could be argued that sustaining such discussions in the classroom by teachers who may not be confident with the subject area could prove challenging. However, with support, teachers could extend discussions of the questions raised by CosmoCI to consider the epistemic practices of cosmology — the way that ideas are built on evidence. The discussion of ideas, views, theories, and hypotheses form a vital part of the epistemic practices of science. The questions in CosmoCI and their categorisation into the four conceptual dimensions of cosmology provide the scaffolding needed to focus and instigate such discussions. With that in mind, the authors of this study are in the process of developing a teaching sequence for cosmology that incorporates an extensive guide for teachers, which includes a guide on unpacking the CosmoCI instrument.

Given that CosmoCI is developed through a design-based research (DBR) approach, the next stage will be the refinement of CosmoCI over various classroom implementations and more structured feedback from teachers.

Conclusions

This study aimed to develop and validate a Cosmology Concept Inventory (CosmoCI) for high school that can be used to monitor progression in reasoning and learning in cosmology. The validation process, through the application of a partial credit model IRT, shows that it is possible to capture the reasoning levels of students in cosmology using a multiple-choice instrument. The SOLO taxonomy proved versatile in capturing levels of reasoning associated with conceptions in cosmology. The instrument has an educative purpose through this reasoning focus, allowing teachers to discern the types of reasoning associated with cosmological declarative knowledge and concepts, acting as a formative and summative assessment tool to capture student thinking. It can open up discussions for teachers and students, and the possibility of informed support for moving students’ thinking in cosmology forward. The inclusion within the instrument of aspects of the evidence bases for cosmological knowledge also aligns it with current thinking about the need to represent scientific practices in teaching and learning science, and with increasing attention to epistemic knowledge as an appropriate outcome. A beneficial outcome from implementing CosmoCI in the first author’s classroom and those of other teachers who were part of the study was the rich discussions instigated between students and the teacher, providing insights into their thinking and the opportunity to frame their learning.