Introduction

Curiosity is considered a strong motivator of learning. While scientific interest in curiosity has persisted for decades (e.g., Berlyne, 1950, 1960; Gruber, Gelman, & Ranganath, 2014; Henderson & Moore, 1980; Hutt, 1970; Loewenstein, 1994; Kang et al., 2009), only recently has empirical work shown that curiosity is associated with better learning outcomes (e.g., Gruber et al., 2014; Kang et al., 2009; Reio & Wiswell, 2000; Stahl & Feigenson, 2015). Min Jeong Kang et al. (2009) found that the answers to high curiosity-inducing trivia questions were more likely to be recalled than for low curiosity-inducing trivia questions 2 weeks later. Additionally, curiosity was highest when learners had intermediate confidence about the answer to the question – in other words, maximal uncertainty about whether their guess was wrong or right. Functional magnetic resonance imaging (fMRI) results revealed an association between curiosity and increased activation of memory regions: When a participant had guessed an answer incorrectly and the correct answer was revealed, activation in the parahippocampal gyrus and left inferior frontal gyrus – regions related to long-term memory consolidation – were modulated by the individual’s level of curiosity. In a similar study, high-curiosity states facilitated learning of additional, interleaved stimuli (Gruber et al., 2014). Collectively, these results suggest that curiosity may globally enhance learning by heightening activation in memory regions of the brain. While these studies show a connection between curiosity and learning, why curiosity is associated with learning remains unclear. A clear understanding of this relationship is crucial to designing effective curiosity-based learning interventions, a hot topic right now in the education literature (e.g., Engel, 2011; Gordon, Breazeal, & Engel, 2015). We do not yet understand even the direction of causality in the relationship between curiosity and learning. Learning could appear enhanced because you are most curious about information that you are on the verge of learning. Thus, the apparent boost to learning may in fact be driven by one’s prior knowledge or there may be a feedback loop between the two factors rather than curiosity inherently boosting learning alone.

There’s no question that partial knowledge induces curiosity. Infants favor partially learned stimuli in preferential-looking paradigms (e.g., Aslin, 2007; Haith, 1980; Hunter & Ames, 1988; Roder, Bushnell, & Sasseville, 2000). Further, young children favor playing with toys for which they have an incomplete understanding (e.g., Bonawitz, van Schijndel, Friel, & Schulz, 2012; Cook, Goodman, & Schulz, 2011; Schulz & Bonawitz, 2007; Sobel et al., 2007). Curiosity is widely theorized to be triggered by the detection of incomplete knowledge (Loewenstein, 1994). This theory is bolstered by empirical results from the education literature that demonstrate relationships between domain-level knowledge and learning (e.g., Alexander, Kulikowich, & Schulze, 1994; Reio, 2004), as well as recent results that report relationships between adult learners’ confidence and their curiosity (e.g., Kang et al., 2009).

It is unclear, then, to what degree the apparent boosts to learning attributed to curiosity may in fact be driven by the partial knowledge itself inducing the curious state. It seems likely that there are bi-directional effects. These kinds of feedback loops are present in the kinds of algorithms used to guide learning in developmental robotics systems (e.g., Baranes & Oudeyer, 2013; Colas, Fournier, Sigaud, Chetouani, & Oudeyer, 2018; Forestier, Mollard, & Oudeyer, 2017), but the existence of such feedback loops has not been empirically tested in human learners.

Here, we aim to clarify the dynamic relationships between partial knowledge, curiosity, and learning using a modified version of a trivia question task. We investigate the degree to which partial knowledge is responsible for curiosity and the observed boosts in learning. We asked participants to take their best guess at trivia questions (Gruber et al., 2014), then provide a judgment of how close they believed their guess to be and a rating of their curiosity for the answer. We also had independent raters judge how close each guess was to the answer as a proxy for an objective measurement of conceptual proximity to the correct response. These data allow us to independently assess the degree to which curiosity and learning are driven by metacognitive estimates of informational uncertainty. Participants were then shown the answer, and later tested on their retention of the said answers. We used these data to comprehensively evaluate the factors that drive curiosity and learning, and the dynamics of these factors that result in boosts to learning.

Our results suggest that while learners’ curiosity is best predicted by how close they think they are, their actual learning is predicted by their actual prior knowledge, with a small apparent boost from being curious about the topic. The implication is that the mechanisms that drive learners’ curiosity are not identical to those that drive learning outcomes. These results also highlight the important role of metacognition in the processes that guide motivation and learning.

Method

We recruited 114 participants on Amazon Mechanical Turk based on the expected effect size for curiosity on trivia questions (as estimated via Kang et al., 2009 and Gruber et al., 2014) and how many participants we estimated we would need to exclude (e.g., due to inattention) based on previous laboratory experiments on Amazon Mechanical Turk. From this sample, we excluded 16 participants for providing the same curiosity rating on at least 90% of trials in the learning phase, and 11 for skipping 20% or more of trials.Footnote 1 We included the resulting 87 participants in the analyses reported below. Participants received a monetary payment at a rate of US$10.00 hourly in exchange for their participation, and the task lasted approximately 50 min.

Stimuli

We selected trivia questions from the stimulus set of Gruber et al. (2014), which was originally created to produce variability in curiosity responses across trivia questions and individuals (see Appendix A.1. for the specific questions and answers). We presented the questions in a randomized order to participants on Amazon Mechanical Turk through psiTurk (McDonnell et al., 2012).

Task

The task consisted of three phases: a learning phase, a distractor phase, and a testing phase. Participants were not told that they would be tested on the trivia questions until immediately before the testing phase.

In the learning phase, we presented participants with 100 trivia questions (e.g., “For which president was the blueberry jelly bean made?”), one question at a time (see Appendix A.1.). For each question, participants provided:

  1. (1)

    Their best guess,

  2. (2)

    Their estimate of how close their guess was to the correct answer on a scale of 1 (not close) to 7 (very close),

  3. (3)

    Whether they thought their guess was correct (yes or no), and

  4. (4)

    Their curiosity about the correct answer on a scale of 1 (not curious) to 7 (very curious).

After submitting their response to a trivia question, participants saw the answer to the trivia question for 5 s. Participants then indicated how surprised they were at seeing the answer, on a scale of 1 (not surprised at all) to 7 (very surprised). Participants were given the option to skip any trials that they did not want to answer. When participants chose to skip a trial, they saw the correct answer for 5 s and waited 5 more seconds before going to the next trial.

In the distractor phase, participants described their impressions of complex artistic images. Responses were collected but were not analyzed for content. The task served as a buffer task between the learning and testing phase.

In the testing phase, trivia questions were presented in the same order as in the training block. Participants guessed the answer to each question. Participants could skip any trials that they did not want to answer. When participants skipped a trial, they were taken to the next trial after a 10-s waiting period.

Similarity ratings

Three research assistants from the Rochester Baby Lab rated the similarity of each guess response in the learning and testing phase (n = 6,794) to the correct answer on a scale of 1 (not close at all) to 7 (very close). The raters were only provided with the trivia question, correct answer, and participant response (they did not see participant-reported curiosity ratings, participant-reported similarity ratings, or whether the responses came from the learning or testing phase). The raters had access to external aids (e.g., information on the internet) in order to make similarity judgments. All raters used the full scale of similarity ratings. The mean similarity rating was 2 out of 7 (1.9, 2.0, and 2.1 for each rater), with strong agreement within raters (Kw = 0.81, 0.81, and 0.84 for each two-rater comparison). To calculate the similarity rating for each item, ratings were z-scored by rater and averaged across raters for each item.

Accuracy ratings

The same research assistants who provided similarity ratings also provided accuracy ratings for all unique responses in the learning phase and test phase. Responses were scored as correct, incorrect, or ambiguous by each rater. Raters were allowed to use external aids (e.g., information on the internet) in order to judge whether a response was accurate or not. For analysis, accuracy of the response was decided based on the majority response. For example, responses were judged to be correct if at least two out of three raters scored the response as correct. There was strong agreement between the three raters (Fleiss’ kappa = 0.81).

Analyses and results

Each participant provided guesses in the learning and testing phase for 100 trivia questions, resulting in a total of 8,700 trials collected for analysis. The average accuracy was 18% for trials in the learning phase (Mdn = 0.16, range = 0.04–0.52) and 69% for trials in the testing phase (Mdn = 0.70, range = 0.27–0.95). Participants reported an average curiosity rating of 4.6 out of 7 (Mdn = 4.9, range = 1.2–6.7). Trials were excluded from analysis if the participant guessed correctly in the learning phase (n = 1,598/8,700, 18% of trials), was confident they got an initial guess correct in the learning phase but guessed incorrectly (n = 1,093/8,700, 13% of trials), skipped a response (n = 4 in learning phase and n = 17 in the testing phase) or did not provide an appropriate guess (e.g., wrote “I don’t know”; n = 15 in the learning phase and n = 17 in the testing phase, for a cumulative total of <1% of trials). The analyses reported below are completed on the remaining trials, where the participant indicated that they did not know the answer and guessed the answer incorrectly during the learning phase (n = 5,973/8,700, 69% of trials).Footnote 2 In our final dataset, participants reported an average curiosity rating of 4.4 out of 7 (Mdn = 4.6, range = 1.2–6.9). At test, participants were accurate on an average of 61% of trials (M = 0.61, Mdn = 0.62, range = 0.22–0.94).Footnote 3

Question 1: What influences curiosity?

First, we asked how curiosity related to learners’ sense of how close their guess was to a correct answer. A Pearson’s Product-Moment Correlation yielded a small, significant positive correlation between normalized curiosity and average, z-scored similarity rating that the participant provided for each initial guess, r(5971) = 0.23, p < 0.0001. Thus, people appear to be more curious about questions when they believe that their initial guess is closer to the correct answer (see Fig. 1). Next, we asked whether people were curious about answers that they were objectively closer to knowing, as judged by independent raters. A Pearson’s Product-Moment Correlation yielded a small, significant correlation, r(5971) = 0.09, p < 0.0001 (see Fig. 2); while the value is statistically significant, the variance explained by the relationship was negligible (<1% of variance). Prior knowledge judged by independent raters was significantly correlated to prior knowledge judged by the learners r(5971) = 0.25, p < 0.0001. To determine whether the observed difference in Pearson correlations between curiosity and our two prior knowledge measures was significant, we followed the recommended procedures in Wilcox (2009). First, we generated bootstrapped confidence intervals from 500 samples of 500 data points. We then calculated the correlations between normalized curiosity ratings, z-scored similarity ratings from participants, and z-scored similarity ratings from raters. Finally, we calculated the upper and lower bounds of a confidence interval with Zou’s method for overlapping correlations. This interval (0.05–0.20) did not contain 0, which indicates a significant difference in the magnitude of the two correlations. Collectively, these results show that perceptions of prior knowledge, rather than a more objective measure of prior knowledge, predict curiosity.

Fig. 1
figure 1

Curiosity and believed prior knowledge. There was a significant relationship between z-scored curiosity ratings (y-axis) and believed prior knowledge (x-axis; z-scored participant-provided similarity ratings). Participants were generally more curious about trivia questions when they believed that they were closer to knowing the answer. Each point in the figure represents a binned mean of 181 datapoints (binned r(31) = 0.97, p < 0.0001), for visual clarity; however, we note that the main analyses were conducted over the raw data (r(5971) = 0.23, p < 0.0001). Error bars denote 95% confidence intervals

Fig. 2
figure 2

Curiosity and independent measure of prior knowledge. There was a significant relationship between z-scored curiosity ratings (y-axis) and independent judgments of prior knowledge (x-axis; similarity ratings were provided by independent raters, z-scored, and averaged). Participants were generally more curious about trivia questions when they were closer to knowing the answer, as judged by three independent raters. Each point on the plot represents a binned mean of 181 data points (binned r(31) = 0.60, p < 0.001). However, main analyses are performed on the raw data (r(5971) = 0.09, p < 0.0001). Error bars denote 95% confidence intervals

Question 2: What influences learning?

Next, we asked whether the same mechanism that underlies curiosity – perceived prior knowledge – also underlies learning. To address this question, we constructed a mixed-effects logistic regression predicting test accuracy from z-scored curiosity ratings and z-scored similarity ratings with participant and trial terms included as random effects (see Table 1; estimated model R2 = 0.40). Controlling for prior knowledge (estimated by the participant in addition to three independent raters) and curiosity, there was a marginal effect of the participant’s own estimate of their prior knowledge on learning, β = 0.06, SE = 0.04, z = 1.7, p = 0.09 (see Fig. 3). Participants were marginally more likely to recall a trivia answer at test when they judged their initial guess as closer to the correct answer than when they judged their initial guess as distant from the correct answer. Controlling for the participant’s perceived prior knowledge, both curiosity and prior knowledge were significant predictors of test accuracy, β = 0.26, SE = 0.04, z = 7.10, p < 0.0001, and β = 0.27, SE= 0.06, z = 4.50, p < 0.0001, respectively. A 1 standard deviation (SD) increase (34.1%) in z-scored curiosity was associated with a 56% increase in the probability of getting the answer correct (see Fig. 4). Similarly, a 1 SD increase in z-scored objective prior knowledge was associated with a 57% increase in the probability of getting the answer correct (see Fig. 5). This analysis shows that while the learner’s perception of prior knowledge predicts the learner’s curiosity, it is a marginal predictor of learning outcomes. Rather, we find that both curiosity ratings and a more objective measure of prior knowledge predict learning outcomes.

Table 1 Regression coefficients from logistic regression predicting learning from curiosity ratings, rater- and participant-provided similarity judgments
Fig. 3
figure 3

Participant-estimated prior knowledge marginally predicts test accuracy. Left: A positive linear relationship between perceived prior knowledge (x-axis) and learning (y-axis). Trivia questions were marginally more likely to be recalled correctly at test if participants thought that their initial guess was close, controlling for independent measures of distance-to-learning and initial curiosity. Data points are binned means of 181 data points (binned r(31) = 0.72, p < 0.001). Right: The average similarity rating participants provided for questions that they recalled correctly (1) or incorrectly (0) at test are shown. In both figures, bars denote 95% confidence intervals. On average, participants thought they were closer to knowing the answers that they recalled correctly at test compared to answers that they recalled incorrectly at test

Fig. 4
figure 4

Learning is associated with greater curiosity. Left: A positive linear relationship between curiosity and test accuracy. Each data point represents the test accuracy (y-axis) and z-scored curiosity reported in the learning phase (x-axis) for a single trial. On average, when participants were more curious about a question in the learning phase, they were also more accurate at test. Data points are binned means of 181 data points (binned r(31) = 0.78, p < 0.0001). Right: The average curiosity rating for trivia questions that participants recalled correctly (1) and incorrectly (0) at test are shown. In both figures, bars represent 95% confidence intervals. On average, participants were initially more curious for questions recalled correctly at test compared to those recalled incorrectly. The effect of curiosity on learning was significant when controlling for the effects of prior knowledge and perceived prior knowledge

Fig. 5
figure 5

Rater-estimated prior knowledge predicts test accuracy. Left: A positive relationship between prior knowledge (x-axis) and learning (y-axis). Data points are binned means of 181 data points (binned r(31) = 0.65, p < 0.0001). Increases in prior knowledge, operationally defined as the similarity of an initial guess to the correct answer as determined by three raters, are associated with better test accuracy. Right: The average similarity rating (y-axis) for trivia answers that were recalled correctly (1) and incorrect (0) at test are shown. Bars denote 95% confidence intervals. On average, the raters judged initial guesses in the learning phase to be closer to the correct answer when the participant correctly recalled the trivia question at test. The effect of prior knowledge on learning was significant when controlling for the effects of perceived prior knowledge and curiosity

Results summary

The results of this experiment help clarify the relationship between curiosity and learning. While previous studies found that test accuracy for trivia questions is linked to a learner's curiosity for the questions (e.g., Gruber et al., 2014; Kang et al., 2009), it was previously unclear whether the learning effects attributed to curiosity would be better explained by curiosity-related variables, such as prior knowledge (e.g., Alexander et al., 1994). We find evidence for both prior knowledge and curiosity providing a boost to learning. Further, we find that curiosity is best predicted by the learner’s metacognitive estimate of their own knowledge, while learning is instead better predicted by an objective measure of existing knowledge. In other words, the cognitive mechanisms underlying curiosity and learning are distinct.

Discussion

Curiosity as a metacognitive signal

Why are learners curious about some things but not others? Our results suggest that learner’s metacognitive estimates of their own prior knowledge strongly drive their curiosity. This is an important insight for better understanding of how to motivate learning in an educational context, where previous work has noted that interest, motivation, and background knowledge may each contribute (e.g., Ainley, Hidi, & Berndorff, 2002; Reio, 2004). Important implications of this work are that learners’ curiosity should be piqued when they come to believe that they are close to knowing the answer, and that induced curiosity could boost learning beyond the advantage of just the partial information alone. Our data provide empirical evidence to bolster classic ideas by educators such as Maria Montessori, who argued that educators should aim to provide material that is just a bit beyond the level the learner has already mastered (e.g., the Montessori concept of “readiness”, 1964). Our work further builds on these ideas by highlighting the importance of the learner’s awareness of where they currently are in the process of knowledge acquisition – the metacognitive dimension of cognitive learning processes. Curiosity may serve as a metacognitive signal that indicates when there’s a match between the presented learning material and the learner’s readiness to encode it.

It follows that learning outcomes should then depend on individual differences in metacognitive monitoring and control – the abilities to monitor and act on one’s knowledge, respectively. The degree to which these individual differences matter is an important topic for future research, as well as how trait-level differences in curiosity impact learning dynamics.

What exactly are learners curious about?

These results leave some ambiguity about precisely what type of informational uncertainty is driving the effect. The target of learners’ curiosity may primarily be the contents of the perceived information gap, or simply the confirmation of whether or not their best guesses are correct. Our data suggest that learners care about both, and that the learning enhancements apply to the contents regardless. While our primary analyses were conducted only on trials in which individuals believed their guesses to be incorrect, subsequent analyses over all trials (see Appendix Fig. 6) reveal that participants who believed their guess was correct were more curious than those who believed their guess was incorrect. Most likely, both an interest in content and feedback contribute to a learner’s overall curiosity. In other words, it’s most likely that learners care both about the missing information itself as well as whether or not they guessed right.

Other factors that influence curiosity

Though the influence of perceived prior knowledge on curiosity is robust, the estimated effect size in our experiment is small, accounting for just 5% of the variance (r = 0.23). Thus, other factors may influence individuals’ curiosity in the world. Curiosity in real-world circumstances is most likely explained by a combination of multiple high-level factors (e.g., prior knowledge, but also conceptual novelty, topic familiarity, goal relevance) and low-level ones (e.g., visual complexity, salience). We note that curiosity is also influenced by attentional, affect, and mental states. For example, curiosity for science increases after learners watch awe-inducing videos (McPhetres, 2019). Likewise, a reduction in curiosity is one diagnostic criterion for detecting clinical conditions like depression (American Psychiatric Association, 2013) and Alzheimer's (Daffner, Scinto, Weintraub, Guinessey, & Mesulam, 1992). The interactions between these other factors and prior knowledge is an area ripe for future research.

The feedback loop between curiosity and learning

In our study, we found a bi-directional relationship between curiosity and learning. First, a learner’s perception of their own prior knowledge is related to greater curiosity for new information. Second, curiosity is associated with better learning outcomes. This is particularly interesting given previous work in developmental robotics that suggests that a curiosity-learning feedback loop arises out of a single curiosity drive and may maximize learning behaviors and outcomes (e.g., Oudeyer, 2004; Oudeyer & Kaplan, 2006). Our experiment provides the first empirical result in support of a bi-directional relationship between curiosity and learning in human learners.