Abstract
Using multiple external representations is advocated for learning in STEM education. This learning approach assumes that multiple external representations promote richer mental representations and a deeper understanding of the concept. In mathematics, the concept of function is a prototypical content area in which multiple representations are used. However, there are hardly any experimental studies investigating the effect of learning functional thinking with multiple representations compared to learning with only one form of representation. Therefore, this article reports on a quasiexperimental intervention study with students from Grade 7, using three measurement time points. The study compared the multirepresentational learning of functional thinking with both tables and graphs with monorepresentational learning with either tables or graphs. The results show that multirepresentational learning led to advantages in learning qualitative functional thinking. However, in quantitative functional thinking, learning with both graphs and tables did not result in higher learning gains than learning exclusively with graphs. Furthermore, students were better able to transfer their knowledge from graphs to tables than vice versa. The results also indicate that multirepresentational learning requires more time than monorepresentational learning but can lead to higher learning gains. In sum, the results show that the effect of learning with representations is a complex interaction process between learning content and the forms of representation.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Mathematics and science classes often deal with concepts (e.g., function, energy, atom) that are not directly accessible through the senses. Therefore, STEM learners are often provided with oral or written information about the concepts in question. The learning process consists of performing tasks that prompt the processing of the information given. The information is frequently provided in written form and combines different forms of external representation (e.g., texts, pictures, charts, graphs, tables, formulas). The learner’s goal in working with these external representations is to process them in their mind and generate adequate internal mental representations of the concept to be learned.
Consequently, over the last decades, the interplay between external and internal representations while learning has been an essential topic of educational psychology (e.g., Mayer, 2014; van Meter et al., 2020) as well as chemistry education (Gilbert & Treagust, 2009), physics education (Treagust et al., 2017), biology education (Treagust & Tsui, 2013), and mathematics education (e.g., Duval, 2006, 2017; Heinze et al., 2009). In this research, there is a consensus that learners should have rich mental representations that allow them to solve problems related to the concept flexibly.
However, the question of how external representations in learning foster or hinder the generation of elaborate and flexible mental representations is still open. It is often assumed that to develop rich mental representations, learning with multiple external representations is helpful. To justify this assumption, it is stressed that each form of representation highlights only certain aspects of a concept; therefore, multiple representations complement each other, and learning with multiple representations thus confers advantages (Cuoco, 2001; Gagatsis & Shiakalli, 2004; Kaput, 1992; Lesh et al., 1987). It is also argued that representations differ in terms of the information they contain or the actions that they support (e.g., Ainsworth, 1999). Thus, learners can gain from the complementarity of the benefits derived from different forms of representation in multirepresentational learning environments. As a result, students are not limited to the strengths and weaknesses of one particular form of representation (Ainsworth et al., 2002).
However, results from empirical research challenge the simple principle that two representations are better than one. Multiple representations can confuse the learner and can result in higher cognitive demand that hinders learning (e.g., Sweller et al., 1998). For example, the splitattention effect (see Schroeder & Cenkci, 2018) shows that the combination of different external representations (e.g., text and picture) is not always conducive to the learning process. It may even be that some representations are better suited to learning a particular concept than others (Cobb et al., 1992). Furthermore, it has become increasingly apparent that complex interactions occur between external representations and the mental representations constructed by learners (Schnotz, 2014). The potential effectiveness of multirepresentational learning materials is likely to be influenced by how the learning environment accommodates these complex interactions. Therefore, it is necessary to better understand the requirements for processing different forms of representation.
Theoretical background
In mathematics, learning with multiple representations plays an essential role when learning the concept of function. As an abstract mathematical concept, it is only accessible through external representations (Duval, 2006). Representing a function through an equation, a graph, a table, a verbal description, or a realworld situation are five common methods (Moschkovich et al., 1993). Because of the importance of the function concept for mathematics and its permeation of everyday life, one goal of mathematics education is the development of functional thinking in both primary education (Blanton & Kaput, 2011) and secondary education (Doorman et al., 2012; Günster & Weigand, 2020; Lichti & Roth, 2018). Functional thinking can be defined as “a way of thinking that is typical for dealing with functions” (Vollrath, 1989, p. 6, translated by the authors). A common approach that is used to further elaborate on functional thinking is to highlight three aspects of the function concept: correspondence, covariation, and mathematical object (e.g., Doorman et al., 2012; Günster & Weigand, 2020; Lichti & Roth, 2018; Vollrath, 1989). The aspect of covariation or covariational reasoning has received increasing attention in mathematics education research in recent years (e.g., Johnson, 2012, 2015; Thompson & Carlson, 2017; Wilkie, 2020). Although there are slightly different perspectives on covariation (see Thompson & Carlson, 2017), a common perspective on covariational reasoning for functions with one single variable is the investigation of the following question: What effect does varying the value of one variable (e.g., xvalue) have on the value of the other variable (e.g., the yvalue)?
One situational context proposed to foster functional thinking with a particular focus on covariation is the bottle problem (Carlson, 1998; Johnson et al., 2017; Swan, 1985; Thompson & Carlson, 2017). In this problem context, bottles (or other vessels that can contain water, such as vases) of different shapes are imagined, and water is poured in at a constant rate. The students then have to investigate the functional relationship between time and the bottle’s water level in various ways.
Figure 1 shows an exemplary task of a bottle problem in which the information about the filling process is given as a graph. Sometimes, a graph is considered to be especially suitable for developing functional thinking because it offers a view of the function as a whole (Vollrath, 1989) and it has the advantages of an iconic representation (Vogel, 2006). The task in Fig. 1 illustrates that functional thinking’s covariational aspect can be further divided into quantitative functional thinking and qualitative functional thinking (Rolfes et al., 2018). In subtask (a), the students have to determine the rate of change, which is an elementary concept of the covariational aspect (see Carlson et al., 2002). In this subtask, students could determine the rate of change by drawing a slope triangle and calculating rise over run (e.g., 3 cm/3 s = 1 cm/s). Calculating the rate of change requires quantitative functional thinking because the solution is based on precise function values that have to be inferred from the graph, that is, it deals with functional thinking in a quantitative way. To solve subtasks (b) and (c), on the contrary, is possible without inferring particular function values from the graph. For these tasks, it is sufficient to investigate the shape of the graph qualitatively. The steeper the graph, the higher the function values’ change (in this case: the water level’s height). Moreover, if the water level rises faster, the bottle has to become narrower. Therefore, subtasks (b) and (c) require qualitative functional thinking. As a previous empirical study with function tasks showed (Rolfes, 2018), the ability in quantitative functional thinking and the ability in qualitative functional thinking are connected but different abilities because a twodimensional model that assumed different psychometric dimensions for these two abilities fitted substantially better than a onedimensional model. The discriminatory validity of the twodimensional model was satisfactory, with a latent correlation between the two dimensions quantitative and qualitative functional thinking of 0.79.
In Fig. 1, the tasks are based on a graph. However, it would also be possible to present the task and its relevant information with a table (see Fig. 2). In this form of representation, students could analyze the aspect of covariation by calculating the difference values of the xvalues and the yvalues, as illustrated in Fig. 2. In almost the same way as in a graph with a slope triangle, the rate of change can be calculated in a table by calculating the difference quotient (e.g., 3 cm/3 s = 1 cm/s). Students can solve subtask (b) by realizing that the difference in the yvalues is higher between 6 and 12 s than between 0 and 6 s.
The possibility of conveying information in different forms of representation raises the question of how external representations can be compared. According to Larkin and Simon (1987), two representations are computationally equivalent if the relevant information can be obtained from both representations just as quickly and easily. Thus, a representation A is computationally more efficient than a representation B if representation A makes it possible to solve the task more easily and faster than representation B. Therefore, the representation with the highest computational efficiency is the most advantageous representation for a task (Larkin & Simon, 1987). In the tradition of cognitive psychology, computational efficiency describes the interaction between the form of representation and the task regarding the human mind in general.
The computational efficiency of forms of representation depends on which type of functional thinking has to be performed (Rolfes et al., 2018). For example, for quantitative functional thinking (e.g., numerically calculating a rate of change), previous research has shown that a table had greater computational efficiency than a graph (Rolfes et al., 2018). However, for qualitative functional thinking (e.g., evaluating in which part of a piecewise linear function the increase is the fastest), a graph turned out to be more computationally efficient than a table because one could visually solve the task by assessing the slope of the graph (Rolfes et al., 2018). However, computational efficiency primarily refers to the performance on a task: Which form of representation is best for using students’ existing cognitive structures? Up until now, no statement has been made in previous research about the advantages of forms of representation in learning. Learning involves the experiential and relatively permanent modification of the cognitive structure (Anderson, 1995). Studies examining performance dependent on different forms of representation do not measure a change in cognitive structures. Hence, such studies provide only limited information about the effect of different forms of representation on learning procedures and concepts.
Therefore, the question about which influence the forms of representation have on learning functional thinking remains open. In addition to the term computational efficiency (Larkin & Simon, 1987), we refer to the effect of representational forms on the learning process as learning efficiency: A representation A is more learning efficient than a representation B if learning with representation A more efficiently modifies or extends cognitive structures for a particular requirement than learning with representation B does. That is, when learning within a certain amount of time with a representation A has a higher effect on learning a specific content than when learning with a representation B in the same amount of time, representation A has more learning efficiency than representation B.
Present study
Tables and graphs are two forms of representation very commonly used for teaching the concept of function. Therefore, this empirical study aimed to compare tables’ and graphs’ learning efficiency for learning functional thinking. Additionally, we examined the efficiency of multirepresentational learning (i.e., learning with both tables and graphs) compared with monorepresentational learning (i.e., learning with either tables or graphs).
Research question: When learning functional thinking, does the learning efficiency differ depending on whether students learn with tables only, with graphs only, or with both tables and graphs?
There were three different, in part competing, hypotheses to our research question:

Hypothesis 1 (Multiple representation hypothesis): Functional thinking can be learned more efficiently in a multirepresentational than in a monorepresentational manner. That is, it is more learningefficient to learn with both graphs and tables than to learn only with graphs or only with tables.
Justification: It is often assumed from a theoretical perspective that learning with multiple representations brings advantages. This is based on the assumption that using multiple representations broadens the spectrum of cognitive possibilities and therefore enhances the learning outcome.

Hypothesis 2 (Representational congruence hypothesis): Functional thinking should be learned with the same representation that the students later use for solving problems that require functional thinking. That is, graphs are the most efficient representation for learning graphbased functional thinking; tables are the most efficient representation for learning tablebased functional thinking.
Justification: Every representation requires different cognitive schemata. Therefore, students need to learn how to deal with each representation separately. For example, if a person wants to work out the piecewise linear function’s section with the highest rate of change (see Figs. 1 and 2), the procedures will differ between graphs and tables. When using a graph, the slopes of the lines can be compared visually. When using a table, a visual strategy is not possible and, therefore, changes must be numerically examined by, for example, calculating difference values.

Hypothesis 3 (Master representation hypothesis): Functional thinking can be learned more efficiently with graphs than with tables.
Justification: The representation that better supports the development of suitable cognitive schemata for functional thinking is more efficient for learning. Applying this argument to graphs and tables, it can be assumed that when learning with graphs, more diverse and flexible cognitive schemata can be developed than when learning with tables. Graphs contain many global features (e. g., general shape, intervals of increase or decrease) that can be interpreted (Leinhardt et al., 1990). For example, it is possible to analyze change behavior with a graph numerically (e.g., by calculating the rate of change) and qualitatively using the graph’s visual properties (e.g., from the graph’s slope). However, when using tables, the change behavior can only be analyzed numerically by calculating difference values or difference quotients. Therefore, the graph can be considered to be a superior form of representation to a table.
Method
Participants
The study included N = 331 students from 13 Grade 7 classes from five academictrack coeducational secondary schools (Gymnasium) in the German federal state of RhinelandPalatinate. The study took place in proximate learning in the first quarter of the school year. The participating students were recruited by the teachers who were willing to participate with their classes in the study. The students’ parents or legal guardians gave their informed consent for the students to participate in the research study. The teachers provided the information that the students had not yet been formally introduced to the concept of function (scheduled, according to the curriculum, to happen in Grades 7 or 8).
Design and materials
Design
The study consisted of three experimental groups and one control group (see Fig. 3). One experimental group learned exclusively with tables, another learned exclusively with graphs, and the third learned with both tables and graphs. The control group received instruction on a topic other than the concept of function (e.g., geometry). The study was quasiexperimental in that the nine classes willing to participate in the intervention were randomly assigned to an experimental group. The three experimental groups completed a pretest, a short intermediate test in the middle of the intervention, and a posttest at the end of the intervention. Four classes were willing to act as the control group, which only completed the pre and posttest to estimate the pretest’s testing effect.
Learning environment
During the intervention, the students in the experimental groups received instruction on essential covariational aspects of functional thinking. For this purpose, a lesson sequence (the complete learning environment is presented in Online Supplement 1) was administered according to direct instructions (see Hattie, 2009), which means the explicit teaching of content followed by an exercise phase. Direct instruction has been shown to be highly effective in empirical studies (Hattie, 2009). This teaching method was chosen in our study mainly because it made it possible to administer a highly standardized lesson sequence in each participating class. New content was typically developed in an introductory phase and then consolidated and deepened by individual exercises or exercises solved with a classmate. Throughout the lesson sequence, flower vases served as the context because bottlefilling problems have been proven fruitful for teaching functional thinking (see Introduction). Lessons in all three experimental groups followed a strictly parallel structure and were carried out in all classes by the first author of this article. The class’s regular mathematics teacher was present only as an observer during the lessons and was not involved in the intervention. The process of filling vases with water at a constant rate was demonstrated with four differently shaped vases at the beginning of the learning sequence in the whole learning group. In the three experimental groups, the learning materials for the lessons differed only in the form of representation used (see Fig. 4): the tableonly group used tables exclusively; the graphonly group used graphs exclusively; and in the tablegraph group, tables and graphs were used alternately for the introductory examples and tasks. The students in the tablegraph group had to work with both tables and graphs because the relevant information for an exercise was given either in one particular table or in one particular graph (see Online Supplement 1). Aside from the variation of the forms of representation, the teaching materials used were identical because the functional context, tasks, and layout were the same in all three experimental groups.
In the first half of the intervention, functional thinking was learned mainly through piecewise linear functions. For this purpose, according to the given functional relationship in the tables or graphs, change behavior was quantitatively analyzed either by determining rates of change or by extrapolation. To determine rates of change using the tables, the students were instructed to note the difference values beside the table (see Fig. 2).
Next, change behavior was qualitatively analyzed by producing qualitative descriptions (“uniformly increasing”), qualitatively comparing the change behavior (“water level increases quicker in the first section than in the second”) or inferring possible vase shapes from the rise in water level. At the end of the first half of the intervention, an intermediate test was performed containing questions on piecewise linear functions.
In the second half of the intervention, functional thinking was extended to nonlinear relationships, for example, by discussing the processes of filling spherical vases. For these nonlinear functions, similar tasks to those in the first half of the intervention were used (e.g., determining average rates of change, inferring possible vase forms, and qualitatively describing change behavior)..
Strictly speaking, the term monorepresentational learning does not accurately characterize the experimental groups that learned solely with graphs or solely with tables. This is because, in addition to the tables or graphs, the students also had to deal with verbal descriptions (e.g., “the water level increases faster”) and images (e.g., a crosssectional drawing of a vase). The learning environment’s central learning goal was even to translate between tables or graphs and verbal descriptions or images. However, because tables and graphs constituted the factorial variation of the quasiexperiment and the fundamental difference of the intervention, the learning process of the two groups that did not learn with both tables and graphs was nevertheless called monorepresentational learning.
Pre, intermediate, and posttest
Eightytwo items were developed to evaluate the learning effect of the intervention (typical test items are provided in Online Supplement 2). To calculate the ability on a common scale, a linked multimatrix design test (Frey et al., 2009) was administered (see Table 1). Six different booklets were developed from the item clusters for the pre, intermediate, and posttest. The item clusters were distributed to the pre, intermediate, and posttest booklets so that each student of the experimental groups received all 82 items, albeit at different time points. The pre and posttest each consisted of 36 items, in which piecewise linear and nonlinear functions were included. The intermediate test consisted of 10 items from the item clusters A, B, and C, which contained piecewise linear functions because only this function type had been addressed in the intervention up to this point. To adequately take account of differences between booklets, the item cluster L_{V} was used in all booklets of the pretest and thereby formed a link (Frey et al., 2009) to the pretest. Similarly, the item cluster L_{N} was included in all booklet variants of the posttest and served as a link to the posttest.
The items’ psychometric fit was evaluated by scaling the data with a twodimensional Rasch model (quantitative vs. qualitative functional thinking). Three items showed values for the INFIT greater than 1.2 and were therefore excluded from the following analyses. All remaining 79 items met the defined quality criteria (INFIT ≤ 1.20, itemtotal correlation ≥ 0.20).
Data analysis
Learning effect analysis
The item response theory (IRT) in the form of a linear logistic test model (LLTM) was used to measure the learning effect (see Fischer, 1995). In a Rasch model, the basic model of the IRT, a specific item difficulty is estimated for every item. As a Rasch model’s extension, an LLTM decomposes item difficulties into components. Where an LLTM is used to measure change, the difficulty parameter of an item at the second time point is estimated as the sum of the item’s difficulty parameter at the first time point and one or more groupspecific or global change parameters (Fischer, 1995). The basic idea is that, with no loss of generality, any change in a person’s ability can be regarded as a change in the item parameters (Mair & Hatzinger, 2007). A group contrast can be included to allow item difficulty to vary between groups (Mair & Hatzinger, 2007). For example, if \({\sigma }_{T}\) denotes the item difficulty of an item administered in the tableonly group in the posttest, the item difficulty of the same item administered in the graphonly group in the posttest can be formulated as σ_{G} = σ_{T} – δ. The parameter δ estimates the difficulty difference parameter between the two experimental groups and indicates the ability difference between the graphonly and the tableonly groups (see Fig. 5). When the graphonly group learned more than the tableonly group, an item was easier for the graphonly group than for the tableonly group, resulting in a positive difficulty difference parameter δ. Whether a difficulty difference parameter δ differed significantly from zero was tested by conducting a Gaussian Ztest because the parameters δ are asymptotically normally distributed.
In the IRT framework, persons’ abilities and item difficulties are estimated in logits. Figure 6 shows the relationship between the ability on the logit scale and the probability of solving a task with medium item difficulty. As can be seen there, a change of δ = 0.5 logits approximately corresponds to an increase in the solution probability P of 10 percentage points (∆P). Because values on the logit scale are independent of the population, they can be interpreted as effect measures. The δvalues can be roughly interpreted similar to values of Cohen’s d, for which a value of 0.2 can be considered a small effect, a value of 0.5 represents a medium effect, and values of 0.8 and above are large effects.
Twodimensional LLTM
First, a twodimensional LLTM was calculated (LLTM2D). With this model, the learning efficiency of the different experimental groups on quantitative and qualitative functional thinking was evaluated. The first 79 parameters σ_{1} to σ_{79} of the LLTM2D calculated the item difficulties in the tableonly group in the pretest as the baseline item difficulties. The other 20 parameters δ_{1} to δ_{20} estimated the group differences in the two dimensions at the various time points. Of these group difference parameters, the first six basic parameters (δ_{1} to δ_{6}) each estimated the ability difference between the four groups in the pretest. Six further parameters (δ_{7} to δ_{12}) estimated the learning gain from pretest to intermediate test, taking into account the ability difference of the groups in the pretest. Hence, it should be pointed out that it was not analyzed whether the ability level differed in the intermediate test, but whether the learning gain differed. The parameters δ_{13} to δ_{20} estimated the learning gain from pre to posttest while partitioning out the ability differences in the pretest.
Fourdimensional LLTM
Additionally, a fourdimensional model that took the form of representation (table or graph) into account was estimated (LLTM4D). This fourdimensional model (Dimension 1: quantitative table items; Dimension 2: qualitative table items; Dimension 3: quantitative graph items; and Dimension 4: qualitative graph items) enabled us to analyze students’ ability to transfer their knowledge from graphs to tables and vice versa. Overall, the fourdimensional model showed a satisfactory model fit and satisfactory EAP/PV reliabilities (Dimension 1: .786; Dimension 2: .724; Dimension 3: .807; Dimension 4: .648).
Assessment of model fit
The LLTM2D and the LLTM4D models assumed a particular twodimensional or, respectively, fourdimensional structure of the construct functional thinking. Therefore, confirmatory factor analyses were conducted with the pretest data to assess whether these modeling approaches were psychometrically reasonable. Because all items were dichotomously scored, a fullinformation item factor analysis (Bock et al., 1988) was performed.
As a baseline model, we estimated the onedimensional model 1D, in which all items loaded on one factor. This baseline model was compared with three different multidimensional models. In the model 2D–Quan × Qual, functional reasoning was divided into the two factors quantitative functional thinking (factor Quan) and qualitative functional thinking (factor Qual). The model 2D–T × G assumed the two latent dimensions tablebased functional thinking (factor T) and graphbased functional thinking (factor G). Furthermore, the model 4D–QuanT × QualT × QuanG × QualG split quantitative and qualitative functional thinking according to whether the tasks had to be solved with tables or graphs.
To assess which of these models fitted best, model comparisons were performed using the Akaike information criterion (AIC, Akaike, 1974) and the Bayesian information criterion (BIC, Schwarz, 1978). Furthermore, the latent correlations of the dimensions were estimated. Finally, the EAPPV reliability of the models was calculated.
Assessment of multilevel effect
The quasiexperiment had a designrelated multilevel structure (Goldstein, 1987), with students (Level 1 units) nested in classes (Level 2 units). One problematic effect of a multilevel structure can be that measurements violate the assumption of independence because measurements within a class might be more homogeneous than those between different classes. However, hierarchical linear models should only be applied with a high number of Level 2 units (i.e., school classes) (see Maas & Hox, 2005). Nevertheless, to investigate the possible influence of the multilevel structure on the results, the extent of nonindependence can be estimated using the intraclass correlation (Gelman & Hill, 2006). The higher the intraclass correlation, the higher the dependence within classes and the higher the multilevel effect. Therefore, for the quasiexperiment, intraclass correlations for the pretest results were calculated for each model and each dimension with 10 plausible values (Adams et al., 2007). If the intraclass correlation does not deviate significantly from zero, there is no evidence that the multilevel structure affects the results.
Software
The calculations were made using the software package R (R Core Team, 2020). Dimension analyses were performed using the package Mirt (Chalmers, 2012), and the IRT scaling and calculation of the LLTMs were performed using the package TAM (Robitzsch et al., 2020).
Results
Assessment of model fit
The data analyses were based on the models LLTM2D and LLTM4D. The LLTM2D was based on a twodimensional model, 2D–Quan × Qual, which assumed the two dimensions of quantitative and qualitative functional thinking. The LLTM4D additionally separated quantitative and qualitative functional thinking, depending on whether it was performed with tables or graphs. To assess the appropriateness of these modeling approaches, model comparisons were performed.
The model comparison with the information criteria AIC and BIC showed that the model 2D–Quan × Qual and the model 4D–QuanT × QualT × QuanG × QualG fitted best (see Table 2). The onedimensional model had substantially higher values for AIC and BIC and, therefore, fitted substantially worse. The twodimensional model, 2D–T × G, which separated tablebased and graphbased functional thinking, fitted even worse than the onedimensional model.
The analyses of the latent correlations (see Table 3) confirmed the results of the model comparisons. The models 2D–Quan × Qual and 4D–QuanT × QualT × QuanG × QualG showed the highest discriminatory validity. In contrast, the dimensions tablebased and graphbased functional thinking in the model 2D–T × G were highly correlated.
Overall, the models 2D–Quan × Qual and 4D–QuanT × QualT × QuanG × QualG showed a similar model fit. Furthermore, the EAP/PV reliabilities of the models were satisfactory (2D–Quan × Qual: .809 and .772, 4D–QuanT × QualT × QuanG × QualG: .786, .724, .807, and .648).
Assessment of multilevel effect
This study was a quasiexperiment because the random assignment to experimental conditions occurred on the level of the learning groups. Therefore, it is possible that a multilevel effect distorted the results. To evaluate the extent of this multilevel effect, intraclass correlations were calculated.
In the twodimensional model, 2D–Quan × Qual, the mean intraclass correlation in the first dimension was .015 with a mean 95% confidence interval [–.012, .099], and in the second dimension .014 with a mean 95% confidence interval [–.012, .098]. This means that in both dimensions, the intraclass correlation did not differ significantly from zero because zero was always included in the mean confidence interval. In the fourdimensional model, 4D–QuanT × QualT × QuanG × QualG, the mean intraclass correlations also did not differ significantly from zero in any of the dimensions: Dimension 1: 95% CI [–.021, .062]; Dimension 2: 95% CI [–.007, .120]; Dimension 3: 95% CI [–.013, .094]; and Dimension 4: 95% CI [–.011, .105].
These results show that a significant multilevel effect was not detected, either in the twodimensional model, 2D–Quan × Qual, or in the fourdimensional model, 4D–QuanT × QualT × QuanG × QualG. Hence, it was reasonable to neglect the effect of the multilevel structure of the data in the analyses.
Learning effect analysis
Overall, the experimental groups showed a substantial increase in performance. The average p value of the experimental groups concerning the items administered in the pretest was .34. This average p value rose in the posttest to .64. That is, on average, across all experimental groups and dimensions, there was an increase of 30 percentage points.
To provide a more detailed insight into the learning effects, the results of the data analyses using the two models LLTM2D (based on the twodimensional model 2D–Quan × Qual) and LLTM4D (based on the fourdimensional model 4D–QuanT × QualT × QuanG × QualG) are presented below. Supplementary tables with the parameter values of both LLTMs are available in Online Supplement 3.
LLTM2D: quantitative versus qualitative functional thinking
The LLTM2D estimated gain scores (in logits) for the two constructs, quantitative functional thinking and qualitative functional thinking.
Quantitative functional thinking
From pretest to intermediate test (see Fig. 7, lefthand side), the tableonly group showed a significant learning gain (δ_{7} = 1.93, z = 25.7, p < .001) in quantitative functional thinking. However, the graphonly group demonstrated an even (significantly) higher learning gain (δ_{9} = 0.55, z = 6.05, p < .001) than the tableonly group. Up to the intermediate test, the tablegraph group showed a significantly lower learning gain than the graphonly group (by δ_{11} = − 0.54, z = − 4.37, p < .001), but almost as much learning gain as the tableonly group.
From pre to posttest, the learning gain in quantitative functional thinking in the control group (i.e., the testing effect) was significant, δ_{13} = 0.74, z = 21.07, p < .001. The tableonly group displayed a significantly higher learning gain than the control group, δ_{15} = 1.19, z = 27.9, p < .001. At the intermediate test stage, the learning gain in the graphonly group was significantly higher than the learning gain in the tableonly group, with a difference in learning gain of δ_{17} = 0.59, z = 11.3, p < .001, even though the magnitude of the intermediate test result was almost identical. However, there was no significant difference (δ_{19} = − 0.10, z = − 1.4, p = .15) in the learning gain from pre to posttest between the tablegraph group and the graphonly group.
Qualitative functional thinking
Overall, learning gains in qualitative functional thinking were lower than in quantitative functional thinking (see Fig. 7, righthand side). From pretest to intermediate test, the ability level in the tableonly group increased significantly by δ_{8} = 0.99, z = 11.6, p < .001. The graphonly group showed a significantly higher learning gain than the tableonly group, δ_{10} = 0.31, z = 3.0, p = .003. However, there was no significant difference between the tablegraph group and the graphonly group in qualitative functional thinking up to the intermediate test (δ_{12} = 0.16, z = 1.1, p = .27).
From pre to posttest, ability in qualitative functional thinking in the control group increased significantly, by δ_{14} = 0.23, z = 6.9, p < .001. The tableonly group showed a significantly higher learning gain than that of the control group, δ_{16} = 0.90, z = 21.7, p < .001, while the graphonly group had a significantly higher increase in ability than the tableonly group, δ_{18} = 0.15, z = 2.9, p = .004. However, the highest learning gain was recorded for the tablegraph group, δ_{20} = 0.42, z = 5.7, p < .001, again significantly higher than that of the graphonly group.
LLTM4D: quantitative and qualitative functional thinking with graphs and tables
To evaluate whether the students were able to transfer their acquired knowledge in functional thinking from graphs to tables and vice versa, a fourdimensional model was estimated. However, the four dimensions could not be measured reliably in the intermediate test because it consisted of only 10 items; therefore, in the LLTM4D, unlike the LLTM2D, only learning gains from pre to posttest were calculated.
Quantitative functional thinking with tables and graphs
In quantitative functional thinking with tables (see Fig. 8, lefthand side), the control group had significantly higher ability values in the posttest than in the pretest, δ_{1} = 0.82, z = 16.2, p < .001, while the tableonly group showed a significantly higher learning gain than the control group, δ_{5} = 1.50, z = 24.0, p < .001. Moreover, the learning gain of the graphonly group did not differ significantly from that of the tableonly group, δ_{9} = 0.09, z = 1.1, p = .25; likewise, there was no significant difference between the learning gain of the tablegraph group and that of the graphonly group, δ_{13} = 0.03, z = 0.2, p = .81. The finding that the three experimental groups’ learning gains were approximately equal is strikingly illustrated in Fig. 8 (on the lefthand side) by the fact that the lines are nearly parallel.
Turning to quantitative functional thinking with graphs (see Fig. 8, righthand side), the control group showed a significant learning gain from pre to posttest, by δ_{3} = 0.64, z = 13.0, p < .001. The tableonly group showed a significantly higher learning gain than the control group, δ_{7} = 0.66, z = 11.1, p < .001, while the graphonly group demonstrated a significantly higher learning gain than the tableonly group, δ_{11} = 1.24, z = 17.0, p < .001. There was no significant difference in learning gain between the graphonly group and the tablegraph group, δ_{15} = − 0.08, z = − 0.8, p = .43. Visually, Fig. 8 (on the righthand side) shows that the lines for the tablegraph group and the graphonly group are nearly parallel, whereas the tableonly group’s line is obviously less steep, indicating a lower learning gain.
Qualitative functional thinking with tables and graphs
In qualitative functional thinking with tables (see Fig. 9, lefthand side), the control group showed a significant increase from pre to posttest by δ_{2} = 0.19, z = 4.2, p < .001. The learning gain of the tableonly group was significantly higher than that of the control group, δ_{6} = 0.96, z = 17.4, p < .001. The graphonly group had a significantly lower gain than the tableonly group, δ_{10} = − 0.26, z = − 3.9, p < .001. In turn, the tablegraph group showed a significantly higher learning gain than the graphonly group, δ_{14} = 0.24, z = 2.5, p = .01.
In qualitative functional thinking with graphs (see Fig. 7, righthand side), the control group’s ability increased significantly from pre to posttest, δ_{4} = 0.20, z = 4.8, p <.001. The learning gain in the tableonly group was significantly higher than in the control group, δ_{8} = 0.28, z = 5.3, p < .001, while the graphonly group showed a significantly higher learning gain than the tableonly group, δ_{12} = 0.54, z = 8.2, p < .001. Again, the tablegraph group recorded the highest learning gain. Its learning gain was significantly higher than that of the graphonly group, δ_{16} = 0.31, z = 3.4, p < .001.
Discussion
In the following, we discuss the results of our study. In the first section, the results on learning functional thinking with representations are summarized and evaluated with regard to the three hypotheses. We then infer what practical implications for learning functional reasoning could follow from the results. Finally, the limitations of the study are presented, and future research desiderata are formulated.
Learning functional thinking
Multiple representation hypothesis
In the multiple representation hypothesis (Hypothesis 1), we assumed that multirepresentational learning would be more efficient than monorepresentational learning. We found some support for this hypothesis in the finding that the tablegraph group had the significantly highest learning gain from pre to posttest in qualitative functional thinking. However, in quantitative functional thinking, there was no significant difference between monorepresentational learning using graphs exclusively and multirepresentational learning using both tables and graphs. In the fourdimensional model, multirepresentational learning with tables and graphs had the significantly highest learning gain for qualitative functional thinking. In contrast, the two dimensions of quantitative functional thinking with tables and graphs showed no significant difference between the monorepresentational experimental group that learned with graphs only and the multirepresentational experimental group that learned with both tables and graphs.
What is striking about the results is that the tablegraph group showed an apparent increase from intermediate to posttest, whereas the monorepresentational learning groups showed little change from intermediate to posttest. For example, up until the intermediate test, multirepresentational learning with tables and graphs was no more efficient than monorepresentational learning using graphs exclusively for either quantitative or qualitative functional thinking. Up until the intermediate test, the multirepresentational group even showed a lower learning gain in quantitative functional thinking than the group that learned with graphs only. In contrast, the increase in qualitative functional thinking between these two experimental groups did not differ significantly. Overall, the results indicate that multirepresentational learning might require more time than monorepresentational learning. The reason for this result could be that working with multiple representations requires initially higher cognitive demands since more information has to be processed simultaneously and held in working memory. Conversely, however, multirepresentational learning presents an opportunity to achieve higher learning gains than are possible with monorepresentational learning alone. The present study also shows that the added value of multirepresentational learning depends on both the learning object and the learning time. The generality of the assumption that multirepresentational learning always leads to higher learning gains was not confirmed in our study.
Representational congruence hypothesis
The representational congruence hypothesis (Hypothesis 2) assumed that it is best to learn with the form of representation that is used later in the assessment. Essentially, this hypothesis suggests that tables are best for learning tablebased functional thinking, whereas graphs are most effective for learning graphbased functional thinking. The empirical data from the quasiexperiment provides some evidence for this hypothesis, albeit with some limitations. The graph did prove more efficient than the table for learning graphbased quantitative and qualitative functional thinking. In the same manner, the table proved more efficient than the graph for tablebased qualitative functional thinking. However, concerning tablebased quantitative functional thinking, learning with tables was not more efficient than learning with graphs. Instead, tablebased quantitative functional thinking was learned with graphs as efficiently as it was with tables. In other words, the cognitive schemata constructed by learning quantitative functional thinking with a graph appear to be relatively easily applicable to a table.
Master representation hypothesis
In the master representation hypothesis (Hypothesis 3), we assumed that graphs are a superior form of representation and are more efficient than tables for functional thinking. Support for this hypothesis was found in the results for both the twodimensional and the fourdimensional models. Thus, with the twodimensional model, a graph was found to be more efficient than a table for learning quantitative functional thinking. Likewise, in the fourdimensional model, graphs were more efficient than tables for learning both tablebased and graphbased quantitative functional thinking. In graphbased quantitative functional thinking, the difference in learning between the tableonly group and the graphonly group was relatively high (1.24 logits). This finding supports the thesis that transferring procedures learned with a table to a graph poses considerable difficulty. Applying quantitative functional thinking to a table, on the other hand, can be learned with graphs as well as with tables. The transfer of graphbased schemata to a table in quantitative functional thinking seems to pose relatively little difficulty. Therefore, a graph seems to be a more suitable form of representation for learning quantitative functional thinking than a table.
For qualitative functional thinking, however, the graph was not always more learningefficient than the table. In the fourdimensional model, a table was more learningefficient than a graph for tablebased qualitative functional thinking, whereas, for graphbased qualitative functional thinking, a graph was more learningefficient. This result is also plausible from a subjectmatter perspective. Qualitative functional thinking, for example, includes the ability to assess in which section the increase is highest in a piecewise linear function. When using a graph, this problem can be solved by comparing the slope of the line. The same problem can be solved numerically with a table only by calculating the difference between the x and yvalues in each section and comparing the differences’ size. Thus, using a table requires different procedures than those required when using a graph.
Computational efficiency versus learning efficiency
Overall, one can surmise from the results that the transfer’s success depends on the underlying cognitive processes. If the necessary cognitive processes are very similar, a transfer from one representation to another is relatively easy. However, if the cognitive processes differ substantially between representations, a transfer is relatively difficult. The results also show a difference between the computational efficiency of a form of representation and its learning efficiency. Students learn quantitative functional thinking better with a graph than with a table, as this study shows. That is, a graph promotes the construction of mental representations for quantitative functional thinking better than a table does, and a graph is better suited to learn, for example, the concept of the rate of change than a table. However, if students have understood the concept of the rate of change, previous research showed (Rolfes et al., 2018) that a table is better suited than a graph for actually calculating the rate of change.
Practical implications
The present study’s findings appear to justify the dominance of graphs over tables in learning the concept of function. A graph is a richer representation than a table because, besides the numeric data, it provides a visual impression that a table lacks. Therefore, it is reasonable to assume that a graph offers more learning opportunities than a table. Accordingly, a graph is more efficient than a table for learning many facets of functional thinking. Only in tablebased qualitative functional thinking was a table more beneficial than a graph in the current study. Therefore, when planning lessons, teachers should reflect as much as possible on the benefit of a particular form of representation concerning the learning process. The selection of the form of representation during the learning process should be determined not only by the question of which particular form of representation provides the most straightforward way to solving the given problem; sometimes, a more cumbersome approach with another form of representation might provide more learning opportunities and might be more rewarding for the learning process.
Being able to connect different representations of functions is in itself already an important learning goal of mathematics education. Besides this rationale for using multiple representations, the present study identified further benefits. Multirepresentational learning can lead to higher learning gains than those achieved through monorepresentational learning. Put simply, one can learn more with multiple forms of representation. This result provides empirical evidence that adopting an approach that uses multiple representations is beneficial for learning functional thinking. For example, this study showed that it is better to learn qualitative functional thinking using a graph and a table rather than only a graph. This leads to the initially counterintuitive conclusion that teaching qualitative functional thinking should not exclusively use graphs. Instead, the way in which qualitative problems (e.g., “In which time period do the function values change the quickest?”) might be solved with tables should also be considered. In general, these results show that computational efficiency and learning efficiency do not always correspond. Therefore, students should also be taught with representations that are computationally less efficient for the task to be learned; that is, quantitative functional thinking should also be taught with graphs, and qualitative functional thinking should also be taught with tables.
However, the benefits of multirepresentational learning have limits. In general, multirepresentational learning takes more time than monorepresentational learning because students first have to learn how to deal with the different representations. This fact presents a representation dilemma (Rau, 2017) that not only the content but also the representation itself must be learned. Furthermore, multirepresentational learning was not superior to monorepresentational learning for all learning content. For quantitative functional thinking, for example, the effect of learning with both tables and graphs was lower than learning with graphs alone from the pretest to the intermediary test. However, the effect from pre to posttest was comparable for both groups. Therefore, it seems appropriate to initially begin the learning process with a representation that is particularly suitable for the particular content being learned. When the learning content has been mastered to a satisfactory degree with this “first” representation, a second and possibly third representation can be added. This approach can potentially maximize the learning success, as conferred by our results. However, further representations only provide additional value if they require a different approach to the learning content than that of the first representation. If a transfer of what has been learned with the first form of representation to other forms is relatively easy, the addition of further representations might not add any further value.
Thus, the simple action motto “The more, the better!” is only partially supported by the present study’s results. Working with additional representations requires additional learning time; therefore, whether an additional form of representation justifies the additional time required must be weighed up. In classroom practice, it is not always easy to determine whether another form of representation truly adds value to the learning process. Here, the insight and experience of the teacher are required to make the right decision. Hence, further empirical research on learning with representations would be desirable to support teachers with this difficult decision.
Limitations
In the study’s learning environment, the connection between tables and graphs was not explicitly highlighted in the multirepresentational learning group. Instead, the students had to implicitly link the two forms of representation by solving similar problems using a table and a graph. Kaput (1989), for example, highlights the importance of connecting the forms of representation while teaching and learning. Hence, learning multirepresentationally with tables and graphs in the current study may have resulted in even higher learning benefits had the connection between tables and graphs been explicitly discussed. Future experiments should investigate this possibility further.
Some of our results are limited for methodological reasons. The problem of our multilevel structure (students nested in classes) has already been pointed out, although the insignificant intraclass correlations suggest that the multilevel effect did not lead to a severe distortion of the results. Furthermore, the intermediate test did not have the same reliability as the pre and posttest because the intermediate test consisted of only 10 items on linear functional relationships. In contrast, the pre and posttest each consisted of 36 items on both piecewise linear and nonlinear functional relationships.
Finally, the learning environment and the test items focused on the bottle problem. Although this situational context is often recommended for fostering functional thinking, the empirical question of how representative the bottle problem is for measuring functional thinking remains open.
Research desiderata
The present study considered learning efficiency as a characteristic of a form of representation with regard to the human mind in general. However, interindividual differential effects of the efficiency of mono or multirepresentational learning are possible and should be investigated further in detail. For example, against the backdrop of the cognitive load theory (Sweller et al., 2011), a plausible assumption is that learning with multiple representations creates a higher cognitive burden than monorepresentational learning. Therefore, it could be hypothesized that students with considerable cognitive capacity (e.g., high workingmemory capacity, high intelligence) in particular would benefit from multirepresentational learning because they can handle the higher cognitive load. Students with somewhat low cognitive capacity, on the other hand, could be overwhelmed by multiple representations in the learning process. On that basis, it would be worthwhile to empirically investigate the interaction between cognitive capacity and learning gains in multirepresentational learning.
Furthermore, our study focused on the quantitative measurement of the effects of learning with different forms of representations. Based on our hypotheses, which were derived from theoretical considerations and empirical evidence from previous studies, we were able to gain valuable insights into learning with representations. However, a followup study could explore the findings via qualitative research. Such a study could provide more information on students’ thinking and how students specifically deal with external representations such as tables and graphs in the learning process.
Finally, tables and graphs are only two of several forms of representation of functional relationships. In addition, the learning environment and the test items on the bottle problem focused on one particular context. Therefore, studies on the efficiency of learning with other forms of representation (e.g., symbolic algebra, verbal descriptions, reallife situations) and in other situational contexts should be conducted. For example, the symbolicalgebraic representation of functional relationships forms a central subject in mathematics education at the secondary school level. Although studies exist on the translation between, for example, a graph and a symbolicalgebraic form of representation, future studies should investigate which particular aspects of functional thinking can be learned using the symbolicalgebraic form of representation.
References
Adams, R. J., Wu, M. L., & Carstensen, C. H. (2007). Application of multivariate Rasch models in international largescale educational assessments. In M. von Davier & C. H. Carstensen (Eds.), Multivariate and mixture distribution Rasch models: Extensions and applications (pp. 271–280). Springer.
Ainsworth, S. (1999). The functions of multiple representations. Computers & Education, 33(2–3), 131–152. https://doi.org/10.1016/S03601315(99)000299
Ainsworth, S., Bibby, P., & Wood, D. (2002). Examining the effects of different multiple representational systems in learning primary mathematics. Journal of the Learning Sciences, 11(1), 25–61. https://doi.org/10.1207/S15327809JLS1101_2
Akaike, H. (1974). A new look at the statistical model identification. IEEE Transactions on Automatic Control, 19(6), 716–723. https://doi.org/10.1109/TAC.1974.1100705
Anderson, J. R. (1995). Learning and memory: An integrated approach. Wiley.
Blanton, M. L., & Kaput, J. J. (2011). Functional thinking as a route into algebra in the elementary grades. In J. Cai & E. Knuth (Eds.), Early algebraization: A global dialogue from multiple perspectives (pp. 5–23). Springer.
Bock, R. D., Gibbons, R., & Muraki, E. (1988). Fullinformation item factor analysis. Applied Psychological Measurement, 12(3), 261–280. https://doi.org/10.1177/014662168801200305
Carlson, M., Jacobs, S., Coe, E., Larsen, S., & Hsu, E. (2002). Applying covariational reasoning while modeling dynamic events: A framework and a study. Journal for Research in Mathematics Education, 33(5), 352–378. https://doi.org/10.2307/4149958
Carlson, M. P. (1998). A crosssectional investigation of the development of the function concept. In A. H. Schoenfeld, J. Kaput, & E. Dubinsky (Eds.), Research in Collegiate Mathematics Education (Vol. 3, pp. 114–162). Conference Board of the Mathematical Sciences.
Chalmers, P. (2012). mirt: A multidimensional item response theory package for the R environment. Journal of Statistical Software, 48(6), 1–29. Retrieved from http://www.jstatsoft.org/v48/i06/.
Cobb, P., Yackel, E., & Wood, T. (1992). A constructivist alternative to the representational view of mind in mathematics education. Journal for Research in Mathematics Education, 23(1), 2–33. https://doi.org/10.2307/749161
Cuoco, A. A. (Ed.). (2001). The roles of representation in school mathematics. National Council of Teachers of Mathematics.
Doorman, M., Drijvers, P., Gravemeijer, K., Boon, P., & Reed, H. (2012). Tool use and the development of the function concept: From repeated calculations to functional thinking. International Journal of Science and Mathematics Education, 10(6), 1243–1267. https://doi.org/10.1007/s1076301293290
Duval, R. (2006). A cognitive analysis of problems of comprehension in a learning of mathematics. Educational Studies in Mathematics, 61(1–2), 103–131. https://doi.org/10.1007/s106490060400z
Duval, R. (2017). Understanding the mathematical way of thinking – The registers of semiotic representations. Springer.
Fischer, G. H. (1995). Linear logistic models for change. In G. H. Fischer & I. W. Molenaar (Eds.), Rasch models: Foundations, recent developments, and applications (pp. 157–180). Springer.
Frey, A., Hartig, J., & Rupp, A. A. (2009). An NCME instructional module on booklet designs in largescale assessments of student achievement: Theory and practice. Educational Measurement: Issues and Practice, 28(3), 39–53. https://doi.org/10.1111/j.17453992.2009.00154.x
Gagatsis, A., & Shiakalli, M. (2004). Ability to translate from one representation of the concept of function to another and mathematical problem solving. Educational Psychology, 24(5), 645–657. https://doi.org/10.1080/0144341042000262953
Gelman, A., & Hill, J. (2006). Data analysis using regression and multilevel/hierarchical models. Cambridge University Press.
Gilbert, J. K., & Treagust, D. (2009). Multiple representations in chemical education. Springer.
Goldstein, H. (1987). Multilevel models in educational and social research. Griffin.
Günster, S. M., & Weigand, H.G. (2020). Designing digital technology tasks for the development of functional thinking. ZDM, 40(6), 999. https://doi.org/10.1007/s11858020011791
Hattie, J. (2009). Visible learning: A synthesis of over 800 metaanalyses relating to achievement. Routledge.
Heinze, A., Star, J. R., & Verschaffel, L. (2009). Flexible and adaptive use of strategies and representations in mathematics education. ZDM, 41(5), 535–540. https://doi.org/10.1007/s1185800902144
Johnson, H. L. (2012). Reasoning about variation in the intensity of change in covarying quantities involved in rate of change. The Journal of Mathematical Behavior, 31(3), 313–330. https://doi.org/10.1016/j.jmathb.2012.01.001
Johnson, H. L. (2015). Secondary students’ quantification of ratio and rate: A framework for reasoning about change in covarying quantities. Mathematical Thinking and Learning, 17(1), 64–90. https://doi.org/10.1080/10986065.2015.981946
Johnson, H. L., McClintock, E., & Hornbein, P. (2017). Ferris wheels and filling bottles: A case of a student’s transfer of covariational reasoning across tasks with different backgrounds and features. ZDM, 49(6), 851–864. https://doi.org/10.1007/s1185801708664
Kaput, J. J. (1989). Linking representations in the symbol systems of algebra. In S. Wagner & C. Kieran (Eds.), Research issues in the learning and teaching of algebra (pp. 167–194). Erlbaum.
Kaput, J. J. (1992). Technology and mathematics education. In D. A. Grouws (Ed.), Handbook of research on mathematics teaching and learning (pp. 515–556). Macmillan.
Larkin, J. H., & Simon, H. A. (1987). Why a diagram is (sometimes) worth ten thousand words. Cognitive Science, 11(1), 65–100. https://doi.org/10.1111/j.15516708.1987.tb00863.x
Leinhardt, G., Zaslavsky, O., & Stein, M. K. (1990). Functions, graphs, and graphing: Tasks, learning, and teaching. Review of Educational Research, 60(1), 1–64. https://doi.org/10.3102/00346543060001001
Lesh, R., Post, T., & Behr, M. (1987). Representations and translations among representations in mathematics learning and problem solving. In C. Janvier (Ed.), Problems of representation in the teaching and learning of mathematics (pp. 33–40). Erlbaum.
Lichti, M., & Roth, J. (2018). How to foster functional thinking in learning environments using computerbased simulations or real materials. Journal for STEM Education Research, 1(1–2), 148–172. https://doi.org/10.1007/s4197901800071
Maas, C. J. M., & Hox, J. J. (2005). Sufficient sample sizes for multilevel modeling. European Journal of Research Methods for the Behavioral and Social Sciences, 1(3), 86–92. https://doi.org/10.1027/16141881.1.3.86
Mair, P., & Hatzinger, R. (2007). Extended Rasch modeling: The eRm package for the application of IRT models in R. Journal of Statistical Software, 20(9), 1–20. https://doi.org/10.18637/jss.v020.i09
Mayer, R. E. (Ed.). (2014). The Cambridge handbook of multimedia learning (2nd ed.). Cambridge University Press.
Moschkovich, J., Schoenfeld, A. H., & Arcavi, A. (1993). Aspect of understanding: On multiple perspectives and representations of linear relations and connections among them. In T. A. Romberg, E. Fennema, & T. P. Carpenter (Eds.), Integrating research on the graphical representation of functions (pp. 69–100). Erlbaum.
R Core Team. (2020). R: A language and environment for statistical computing. R foundation for statistical computing [software]. Retrieved from http://www.Rproject.org/
Rau, M. A. (2017). Conditions for the effectiveness of multiple visual representations in enhancing STEM learning. Educational Psychology Review, 29(4), 717–761. https://doi.org/10.1007/s1064801693653
Robitzsch, A., Kiefer, T., & Wu, M. (2020). TAM: Test analysis modules. R package version 3.5–19 [software]. Retrieved from http://CRAN.Rproject.org/package=TAM
Rolfes, T. (2018). Funktionales Denken: Empirische Ergebnisse zum Einfluss von statischen und dynamischen Repräsentationen. [Functional thinking: Empirical results on the effect of static and dynamic representations]. Springer.
Rolfes, T., Roth, J., & Schnotz, W. (2018). Effects of tables, bar charts, and graphs on solving function tasks. Journal Für MathematikDidaktik, 39(1), 97–125. https://doi.org/10.1007/s131380170124x
Schnotz, W. (2014). Integrated model of text and picture comprehension. In R. E. Mayer (Ed.), The Cambridge handbook of multimedia learning (2nd ed., pp. 72–103). Cambridge University Press.
Schroeder, N. L., & Cenkci, A. T. (2018). Spatial contiguity and spatial splitattention effects in multimedia learning environments: A metaanalysis. Educational Psychology Review, 30(3), 679–701. https://doi.org/10.1007/s1064801894359
Schwarz, G. (1978). Estimating the dimension of a model. The Annals of Statistics, 6(2), 461–464. https://doi.org/10.1214/aos/1176344136
Swan, M. (1985). The language of functions and graphs: An examination module for secondary schools. Shell Center for Mathematical Education.
Sweller, J., Ayres, P., & Kalyuga, S. (2011). Cognitive load theory. Springer.
Sweller, J., van Merrienboer, J. J. G., & Paas, F. G. W. C. (1998). Cognitive architecture and instructional design. Educational Psychology Review, 10(3), 251–296. https://doi.org/10.1023/A:1022193728205
Thompson, P. W., & Carlson, M. P. (2017). Variation, covariation, and functions: Foundational ways of thinking mathematically. In J. Cai (Ed.), Compendium for research in mathematics education (pp. 421–456). NCTM.
Treagust, D. F., Duit, R., & Fischer, H. E. (Eds.). (2017). Multiple representations in physics education. Springer.
Treagust, D. F., & Tsui, C.Y. (2013). Multiple representations in biological education. Springer.
van Meter, P., List, A., Lombardi, D., & Kendeou, P. (Eds.). (2020). Handbook of learning from multiple representations and perspectives. Routledge.
Vogel, M. (2006). Mathematisieren funktionaler Zusammenhänge mit multimediabasierter Supplantation: Theoretische Grundlegung und empirische Untersuchung. [Mathematizing functional relationships with multimediabased supplantation: Theoretical rationale and empirical investigation.] Franzbecker.
Vollrath, H.J. (1989). Funktionales Denken [Functional thinking]. Journal Für MathematikDidaktik, 10(1), 3–37. https://doi.org/10.1007/BF03338719
Wilkie, K. J. (2020). Investigating students’ attention to covariation features of their constructed graphs in a figural pattern generalisation context. International Journal of Science and Mathematics Education, 18(2), 315–336. https://doi.org/10.1007/s10763019099556
Funding
Open Access funding enabled and organized by Projekt DEAL.
Author information
Authors and Affiliations
Contributions
This article is based on dissertation research (Rolfes, 2018) directed by the second and third authors of the article.
Corresponding author
Ethics declarations
Ethics statement
On behalf of all authors, the corresponding author states that there is no conflict of interest.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Supplementary Information
Below is the link to the electronic supplementary material.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rolfes, T., Roth, J. & Schnotz, W. Mono and MultiRepresentational Learning of the Covariational Aspect of Functional Thinking. Journal for STEM Educ Res 5, 1–27 (2022). https://doi.org/10.1007/s41979021000604
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s41979021000604