Introduction

People tend to be overconfident in their own abilities and in their future performance on a variety of tasks. For example, adults overestimate their skills and performance on tasks related to reasoning, humor, and grammar (Kruger & Dunning, 1999); children overestimate their ability to remember pictures (Lipko, Dunlosky, & Merriman, 2009) and to perform physical tasks (Schneider, 1998); and consumers overestimate how easy it will be to learn to use a new product (Billeter, Kalra, & Loewenstein, 2011). Most relevant to the present research, students often overestimate how well they will perform on an upcoming test of their learning (e.g., Hacker, Bol, & Bahbahani, 2008; Miller & Geraci, 2011a, b; but see Griffin, Jee, & Wiley, 2009; Shanks & Serra, 2014), which can impair their study behaviors (Metcalfe & Finn, 2008). The reasons for such overconfidence, however, are not yet well understood. Toward this end, in the present studies we tested the hypothesis that students make overconfident predictions of their test performance because their predictions are related to—and potentially biased by—their desired level of performance, which is typically higher than their actual level of performance.

Persistent overconfidence

Researchers in a variety of domains have examined people’s self-assessments of their own abilities and predictions of their own future performance for both mental and physical tasks. In the context of human learning within formal-education settings, people’s thoughts about their thoughts (i.e., metacognition; see Briñol & DeMarree, 2012; Dunlosky & Metcalfe, 2009) play a major role in determining the efficacy of students’ study behaviors. For example, students utilize metacognitive knowledge when selecting strategies for how to study, engage in metacognitive monitoring when evaluating how well they already know what they are studying, and engage metacognitive control when using their knowledge and monitoring to decide for how long to continue to study, or whether to stop studying altogether (i.e., Nelson & Narens, 1990; for a review, see Dunlosky & Metcalfe, 2009). Critically, flaws in any of these components of metacognition can impair students’ study behaviors (Serra & Metcalfe, 2009).

In the present studies, we examined a common error in students’ monitoring of their learning and predictions of their future test performance: the persistent overconfidence of students’ judgments (e.g., Ehrlinger, Johnson, Banner, Dunning, & Kruger, 2008; Erev, Wallsten, & Budescu, 1994; Klayman, Soll, González-Vallejo, & Barlas, 1999; Koriat, Lichtenstein, & Fischhoff, 1980). Overconfidence in this regard (i.e., students’ predictions of their future test performance being significantly higher than their actual performance) can be problematic for students, because it can lead them to underprepare for exams (e.g., Bjork, Dunlosky, & Kornell, 2013; Dunlosky & Rawson, 2012; Grimes, 2002; Kornell & Bjork, 2008; Kornell & Metcalfe, 2006; but see Carpenter, Wilford, Kornell, & Mullaney, 2013) or make poor restudy decisions (cf. Finn, 2008; Metcalfe & Finn, 2008; Shanks & Serra, 2014).

Unskilled and unaware

Although it could prove problematic for any student to overestimate their learning (especially leading up to an exam), of further concern is the common observation that the lowest-performing (and, perhaps, lowest-ability) students are the ones most overconfident in their judgments, whereas the highest-performing students tend to be more accurate, or even demonstrate underconfidence (for a review, see Miller & Geraci, 2011b). This pattern is often referred to as “unskilled and unaware,” because it seems that the lowest-performing people in a domain not only lack the abilities or knowledge to perform well in that domain, but also lack metacognitive awareness of their deficits (e.g., Dunning, Johnson, Ehrlinger, & Kruger, 2003; Grimes, 2002; Kruger & Dunning, 1999; but see Griffin et al., 2009; Miller & Geraci, 2011b; Shanks & Serra, 2014). In formal-education settings, this pattern can be especially problematic. This pattern not only might lead the lowest-performing students to underprepare for exams within their courses, but also could suggest to these students that they are not capable of passing their courses, that they cannot achieve the level of performance they expected in those or other courses, and even that they are not at all capable of succeeding in that field of study (cf. Grimes, 2002). For example, Stinebrickner and Stinebrickner (2014) reported that students tend to change their major from a science to a nonscience field because their grades in their science courses were substantially lower than they initially expected.

To date, a number of researchers have attempted to uncover the factor(s) that produce the unskilled-and-unaware pattern (e.g., Ackerman, Beier, & Bowen, 2002; Burson, Larrick, & Klayman, 2006; Dunning et al., 2003; Ehrlinger et al., 2008; Kruger & Dunning, 2002; Krueger & Mueller, 2002; see also Miller & Geraci, 2011a, b). Although a variety of factors have been identified as contributing to this pattern (and some have been eliminated from contention), no one factor has been identified as being the major, or sole, cause of this outcome. In the present studies, we examined another potential contributing factor: students’ desired grades. Kruger and Dunning (1999) noted that motivational biases likely play a role in generating the unskilled-and-unaware effect (p. 1132), but no research that we are aware of has empirically examined this idea. To address this empirical deficiency, in the present studies we examined students’ desired exam grades and assessed whether those desired grades were associated with students’ exam grade predictions in the classroom.

Wishful thinking

The idea that students’ desired grades might impact their predictions of the grades they might actually earn—and cause these predictions to be overconfident—is supported by the empirical literature on wishful thinking. According to this account, people unintentionally base their predictions of future performance on how they want to perform rather than on how they expect to perform (e.g., Schneider, 1998; Stipek, Roberts, & Sanborn, 1984; Willard & Gramzow, 2009). For example, young children (i.e., preschoolers and kindergartners) continue to overestimate their future performance even when they have had experience with the target task and can remember their past performance accurately (e.g., Lipko, Dunlosky, & Merriman, 2009; Schneider, 1998). This persistent overconfidence seems to stem from wishful thinking rather than a general inability to factor past performance into predictions of future performance, since young children often do not demonstrate overconfidence when asked to make predictions of future performance for another child whose past performance they are aware of (e.g., Schneider, 1998; but see Lipko et al., 2009). Put differently, young children make better predictions of other children’s performance than of their own because their own predictions are biased by their own goals, which typically involve a desire for stellar performance.

Although college students’ grade predictions are not always accurate, in general adults do not show the very large and persistent effects of wishful thinking that children demonstrate (e.g., college students heavily factor past-test performance into their current learning predictions; see Ariel & Dunlosky, 2011; England & Serra, 2012; Finn & Metcalfe, 2007; Tauber & Rhodes, 2012). Nevertheless, we still propose that wishful thinking might exert an effect on adults’ performance predictions in a more subtle way, and there is some support for this prediction. Soderstrom and McCabe (2011) had college students study and then make judgments of learning (JOLs) for paired associates before taking a memory test over those items. Participants received between one and six points for recalling each item (point values were randomly assigned to the items). Most importantly, in one condition participants were told the number of points that each item would be worth in between the study and judgment of each item. Although the point value did not affect participants’ recall in this condition (because study had occurred before the point value was known), participants judged that their memory would be better for items worth more points than for items worth fewer points. Given that the participants in this condition could do nothing to influence their likelihood of recall at that point in time, it seems likely that their desire to earn more points led them to judge higher-point items to be more memorable than lower-point items—a wishful-thinking effect.

Soderstrom and McCabe (2011) offered compelling initial support for our prediction, but the context in which they examined the effects of wishful thinking was quite different from the one we proposed to examine, so there was no guarantee that desires would operate the same way in both situations. More specifically, whereas they examined item-by-item JOLs (made after studying each of several dozen items), our focus in the present studies was mostly on predictions of overall exam performance (single global performance predictions). It was not clear whether the effects obtained with item-by-tem memory predictions would also occur for global predictions of performance like those in the present studies; item-by-item predictions might be less prone to the biasing effects of desires than global predictions because the former are based heavily on the item-specific properties of those items (e.g., difficulty, see Koriat, 1997). Furthermore, we examined students’ predictions of performance that they made either before (Study 1) or after (Studies 2 and 3) the exams had occurred; it is not known whether the effects obtained with one of these types of prediction would also occur for the other type. For example, people’s metacognitive experience while learning or studying provides the major basis for performance predictions made before a test, whereas people’s metacognitive experience while taking the exam (i.e., during recall) provides the major basis for postexam performance predictions. Pretest judgments might be more susceptible to the effects of desires than posttest judgments are because people could base the latter type of judgment on their actual test experience. In addition, the “real-life” nature of the predictions that participants make for actual exams might involve greater thoughtfulness than predictions they make for a no-consequence laboratory exercise, and motivations might have greater relevance for “real-life” than for artificial predictions. Either of these possibilities might actually increase the likelihood of obtaining a relationship between desires and predictions for actual exams as compared to laboratory tasks.

The present studies

The purpose of the present studies was to examine the hypothesis that students’ exam predictions are typically overconfident because their predictions are heavily biased by their desired grade on the exam, which is fairly high. This explanation might be especially important for helping identify the factors that contribute to the unskilled-and-unaware pattern of performance predictions. To examine these issues, we had students enrolled in either a general psychology course (Study 1) or a research methods course (Studies 2 and 3) report their desired grades in addition to reporting the grades they actually expected to earn. If desired grades lead to inflated predictions of performance, then students’ predictions of their grades should be related not only to their actual performance, but also to their desired performance. Note that in each of our studies, students were asked to predict their absolute grades rather than their relative performance, to reduce the influence of people’s assessments of others’ performance on the unskilled-and-unaware phenomenon (Hartwig & Dunlosky, 2014).

Study 1

The purpose of Study 1 was to obtain some initial evidence to support our hypothesis that students’ desired grades lead to inflated predictions of their actual grades. To do so, we had students report their desired and expected grades for both the final exam and the overall course. The latter judgments would provide a potentially more conservative test of our hypothesis than the former, because students can presumably use a wealth of information acquired over the course of the semester to predict their final course grade. This study also allowed us to compare results involving students’ actual grade predictions in a classroom setting to similar studies obtained in the laboratory utilizing item-by-item judgments (i.e., Soderstrom & McCabe, 2011).

Method

Participants

The participants in the present sample were 485 college students enrolled in General Psychology at Texas Tech University, who participated for course credit. In this and all of these studies, the sample size was determined by the number of students enrolled in the course who were willing to provide data for the present purposes. More specifically, the sample was determined by asking all students in a given class to provide the information relevant to the present hypotheses. Although not all students chose to provide these estimates and desires, the vast majority of the students in each class did provide the information. This sample was 60% female and 40% male, with a modal age of 19 years old. Furthermore, including people who indicated more than one race or ethnic group, the sample was 73.4% White or European American, 20.6% Hispanic or Latino, 6.2% Black or African American, 3.5% Asian or Asian American, 1.6% Native American or Alaskan Native, and 2.1% some other race or group.

Procedure

Students completed the present questions as part of a larger end-of-semester course assessment. (Other measures related to course outcomes were collected at this time but were not directly relevant to the present purposes, and hence are not reported.) We collected all data online using the Qualtrics survey software during the last two weeks of the semester, before final exams were administered but after most other assignments and preceding exams in the course had been completed.

During the critical section of the assessment, students answered two questions about their expected performance on the final exam for the course: “What grade do you realistically want to earn on your final exam in this course?” (desired grade) and “What grade do you think you will actually earn on your final exam in this course?” (predicted grade). We included the word “realistically” in the phrasing of our desired-grade questions to prevent respondents from passively answering “100%.” We wanted them to indicate their desires, but to make predictions within the realm of their own perceived possibilities. For these questions, students were instructed to enter a whole number between 0 and 100 to indicate their expected grade out of 100% (and the survey software only accepted such values). Students then answered the same two questions, restated to be about their final grade in the course as a whole.

Note that although the exact format of the final exams (as well as the exams and assignments utilized during the semester that contributed to the final course grades) varied by instructor, for the most part there was high consistency across the sections of the course. In particular, the final exams were typically noncumulative and consisted primarily of multiple-choice questions.

Students were made aware that their responses would later be compared to their actual grades and were allowed to skip any of the questions if they chose to. Students also provided us with permission (or denied us permission) to access their final exam and course grades for the purpose of comparing them to their predictions. As such, the number of students contributing data to some of the analyses below varies slightly from measure to measure. In general, however, most students completed all of the present measures. An additional 26 students completed at least some of the course assessment but did not complete any of the present measures, and as such were excluded from the present sample (since they had no data to contribute). We then obtained final exam and course grades from the course instructors for comparison to the students’ predictions for those students who had completed at least some of the present measures and who provided us with permission to access these grades. After that point, the data were completely deidentified for analysis.

Results and discussion

For descriptive information only, we separated students into two sets of four groups based on (a) their actual scores on the final exam and (b) their actual final course grades, much as in Miller and Geraci (2011b). Both Table 1 and Fig. 1 present the means of students’ desired, predicted, and actual grades. Although both Table 1 and Fig. 1 present grouped data, all subsequent analyses used regression with continuous predictors and outcomes, unless otherwise reported. The values in the table and figure are provided mainly to illustrate the general pattern of results and for direct comparison to the results of other, similar studies (e.g., Miller & Geraci, 2011b). We also chose to group students on the basis of their actual grades rather than quartiles because grouping by quartile is inappropriate if one is to consider absolute accuracy or calibration at the level of actual grades. Although we grouped by grade for descriptive purposes (i.e., to more easily show that the unskilled-and-unaware pattern was obtained in the present studies), we conducted all analyses utilizing regression with continuous predictors.

Table 1 Descriptive statistics by group in Study 1
Fig. 1
figure 1

Mean desired, predicted, and actual final exam and final course grades by target-grade (final exam vs. course grade) and actual-grade grouping in Study 1. Error bars show the standard errors of the means (note that many are too small to be seen at this scale). These groupings are shown only for descriptive purposes and to aid comparison to other, related studies; all critical analyses (i.e., regression analyses) involved the analysis of continuous data

Final exam grade

To examine students’ potential projection of desires onto their grade predictions on the final exam, we simultaneously regressed the exam predictions on desired exam grade and actual exam grade. Students’ actual grades were related to their predicted grades (b = .113, se = .021, β = .190), t(477) = 5.32, p < .001, but their desired grade was a strong, significant predictor as well (b = .758, se = .047, β = .581), t(477) = 16.27, p < .001. Thus, both actual performance and desired performance were independent, significant predictors of grade predictions. This observation is anecdotally supported by the grouped data plotted in Fig. 1 (left panel), in which desired grades track grade predictions more closely than do actual grades. These results—which demonstrate a strong relationship between students’ desired and predicted grades—support our hypothesis that students’ desired grades lead to inflated predictions of their exam performance.Footnote 1

Final course grade

To examine students’ potential projection of desires onto their final course-grade predictions, we simultaneously regressed the course-grade predictions on desired course grade and actual course grade. Students’ actual grades were related to their predicted grades (b = .223, se = .023, β = .299), t(482) = 9.58, p < .001, but—as before—their desired grades were also strong predictors (b = .728, se = .037, β = .621), t(482) = 19.88, p < .001 (see Fig. 1, right panel). Again, desired grades tracked grade predictions more closely than did actual grades. As with the exam grades, these results support our hypothesis that students’ desired grades lead to inflated predictions of their final course performance.

The results pertaining to final course grades are, perhaps, more valuable than those pertaining to final exam grades, because students had a larger variety of information that they could have used to make their final course-grade predictions than to make their final exam predictions. Specifically, although past-test performance in the course might have provided some informative cues on which to base final exam predictions, that generalization would only be useful if students assumed that all exams in the course were of equal difficulty. In contrast, students’ final course grades were determined by the direct combination of their grades on exams, quizzes, and assignments throughout the semester; at the point at which we collected the present data, students should have been more or less aware of the final grade they’d likely earn in the course on the basis of their grades thus far. Nevertheless, students’ predictions of their course grades demonstrated the unskilled-and-unaware pattern and were highly related to—and therefore potentially biased by—their desired grades (see also Soderstrom & McCabe, 2011).

Study 2

Although Study 1 provided initial support for our predictions, it has a critical weakness. Specifically, because students made their predictions before they took the exam, their predictions could not be impacted by their experience in taking the exam, and may instead have been affected by overly optimistic plans to study that had been impacted by desired grades, leading to a potentially inflated impact of desired grades. In Study 2, we included the present measures in conjunction with an actual course examination. In this case, we had students indicate their desired grade just prior to taking the exam, but had them indicate their predicted grade immediately after completing the exam. Students’ postdictions of test performance are generally more accurate than their predictions (e.g., Glenberg & Epstein, 1987; Maki, 1998; Winne & Jamieson-Noel, 2002); thus, this study would offer a stronger test of the role that desired grades might play in students’ performance estimates. One possibility is that the subjective experience of taking the exam is particularly salient when students make their postexam predictions, and this could increase their ability to predict actual test performance free of the biasing influence of their desired grades. In contrast, another possibility is that students’ desires bias their interpretation of the metacognitive cues they use to make predictions of their performance, including their very recent experience of taking the exam (cf. Balcetis & Dunning, 2006, 2010; Dunning & Balcetis, 2013). If so, students’ grade predictions would continue to be biased by their desires.

Method

Participants

The participants in the present sample were 66 college students enrolled in a research methods course at Texas Tech University. They completed these measures as ancillary questions on their final exam. We did not collect demographic information for these students, since this analysis was not part of a larger course assessment, but the sample characteristics were likely similar to those in Study 1. Seven of the 66 students in the course did not complete all of the present measures, so the effective sample size was 59.

Procedure

Students completed the present questions as ancillary measures on their final examination. The final exam was noncumulative, covering the final quarter of a methods course. The exam consisted of 44 multiple-choice questions and two short-answer questions. Before students began the exam, they answered the question, “What percent grade (0–100%) do you (realistically) want to earn on this exam?” (desired grade). After completing the exam, students answered the question, “Now that you have taken this exam, what percent grade (0–100%) do you think you actually earned on this exam (not including extra credit)?” (predicted grade). For these questions, students were instructed to provide a whole number between 0 and 100. Students were made aware at the start of the exam that their responses would later be compared to their actual grades, and they were allowed to skip any of the questions if they chose to. In general, however, most students (59 of 66) completed all of the present measures. Once the exams had been graded, the numbers described above, along with students’ actual exam scores, were entered into a deidentified data file for analysis.

Results and discussion

For descriptive information, we separated students into four groups based on their actual scores on the final exam, as in Study 1 and previous studies (i.e., Miller & Geraci, 2011b). Both Table 2 and Fig. 2 present the means of students’ desired, predicted, and actual grades. Although both Table 2 and Fig. 2 present grouped data, all analyses reported below were based on regression unless otherwise reported.

Table 2 Descriptive statistics by group in Study 2
Fig. 2
figure 2

Mean desired, predicted, and actual final exam grades by actual-grade grouping in Study 2. Error bars show the standard errors of the means (note that only one student scored a 90 or higher, and hence those means have no error bars). These groupings are shown only for descriptive purposes and to aid comparison to other, related studies; all critical analyses (i.e., regression analyses) involved the analysis of continuous data

To examine students’ potential projection of desires onto grade predictions on the final exam, we regressed their exam predictions on desired exam grade and actual exam grade. Given that we had the students in Study 2 make their grade predictions after completing the exam, it is perhaps not surprising that their actual grades related to their predicted grades (b = .358, se = .085, β = .420), t(56) = 4.22, p < .001. Importantly, however, students’ desired grades were also related to their predicted grades (b = .532, se = .117, β = .451), t(56) = 4.53, p < .001. Thus, although students had just completed taking the exam—and thus their subjective experience on the exam should have been particularly salient—their predicted exam scores were still strongly related to their desired grades. These results provide further support for our hypothesis that students’ desired grades lead to inflated predictions of their exam performance.

Study 3

In our final study, we sought to replicate and extend the results of our previous studies by combining the strengths of the two studies into a single study. To do this, we had participants report their desired and predicted grades on each of four exams during a semester-long course. Because participants predicted their actual exam grades immediately following each exam (as in Study 2), recent experiential information was available to inform their predictions. But because participants made judgments over the span of a full semester, as the semester progressed they could accumulate knowledge of their performance in the course and on past exams on which to base their estimates (conceptually similar to Study 1). This allowed us to examine the extent to which the influence of people’s desires on their performance estimates persists as their opportunities for knowledge of their actual abilities increased.

Method

Participants

The participants in the present sample were 60 college students enrolled in a research methods course at Texas Tech University, who completed these measures as ancillary questions that appeared on all four exams given in the course during the semester. We did not collect demographic information for these students, since this analysis was not part of a larger course assessment, but the sample characteristics were likely similar to those in Study 1. The exams were noncumulative, carried equal weights in students’ grades, and were similar in structure to the one described in Study 2. Students were made aware at the start of the exam that their responses would later be compared to their actual grades, and they were allowed to skip any of the questions if they chose to. Because the lowest exam in this course was dropped from students’ final grades, many students (n = 40) opted not to take the fourth exam (i.e., they were satisfied with their grade to date and opted to drop a 0). (Note that the same results would be obtained if the fourth exam were completely excluded from the analyses.) Furthermore, not all students completed all measures each time, leaving observations for 171 occasions (see Table 3 for the statistics on each exam).

Table 3 Descriptive statistics by exam in Study 3

Procedure

The procedure was identical to that of Study 2 (i.e., students reported their desired grade at the start of each exam and their predicted grade at the completion of each exam), except that we collected data for all four exams in the course instead of only for the final exam.

Results and discussion

For comparison to the descriptive information from the earlier studies, we first separated students’ predictions into four groups based on their actual scores on each of the four exams. More specifically, Table 4 and Fig. 3 present the mean desired, predicted, and actual exam grades for any student scoring within a given range on any of the four exams in the course (i.e., most students contributed four sets of desired, predicted, and actual exam grades to these means). Although both Table 4 and Fig. 3 present grouped data, all analyses reported below were based on multilevel modeling across participants unless otherwise reported.

Table 4 Descriptive statistics by group in Study 3
Fig. 3
figure 3

Mean desired, predicted, and actual exam grades by actual-grade grouping in Study 3. Note that each student contributed up to four sets of exam grades and predictions. Error bars show the standard errors of the means. These groupings are shown only for descriptive purposes and to aid comparison to other, related studies; all critical analyses (i.e., regression analyses) involved the analysis of continuous data

Because we had multiple observations for each person (and thus the observations were not independent) and the number of observations per student varied, we employed multilevel modeling for these analyses. In this context, multilevel modeling has the advantages both of a repeated measures analysis of variance, in that it explicitly takes into account the nonindependence of the observations from a given unit (in this case, each student), and of regression, in that it allows for continuous predictors and provides parameter estimates that are analogous to regression coefficients. We used restricted maximum likelihood estimation and allowed the intercepts—but not the slopes—to vary across analyses (i.e., only the intercept was treated as a random effect). Analyses in which slopes were allowed to vary produced the same patterns of significance.

To examine participants’ projection of desires onto their grade predictions on each exam, we first conducted analyses parallel to those we had conducted in Studies 1 and 2, in which students’ predicted exam grades had been statistically predicted by their desired and actual exam grades using multilevel modeling. As in Studies 1 and 2, both students’ actual grades (b = .212, se = .054, β = .216), t(166) = 3.97, p < .001, and their desired grades (b = .543, se = .069, β = .389), t(152) = 7.86, p < .001, significantly predicted their predicted grades. Furthermore, as in Study 2, although students had just completed taking the exams, and thus their subjective experience on the exams should have been particularly salient, their predicted exam scores were still strongly related to their desired grades. Once again, the results support our hypothesis that students’ desired grades lead to inflated predictions of their exam performance.

We also examined whether the impacts of students’ actual and desired exam grades varied in their predictive utility across the semester. To do this, we added exam number (coded 1–4) and its interactions with actual and desired grades as linear predictors of predicted grade. This analysis did not alter the results reported above, since exam did not interact with either actual or desired grade, although both trends were such that the effects decreased as the semester progressed (ts < 1.4, ps > .18). Although students’ grade predictions did not decrease across the semester, their desired grades did (b = –1.98), t(124) = 3.99, p < .001. This finding suggests that students might have adjusted their desires to be within the realm of possibility across the semester as they learned more about their aptitude for the course and the level of difficulty of the exams. Therefore, the present data support the idea that people’s predictions of their grades are strongly related to—and therefore possibly biased by—their desired grades. These data do not, however, provide support for the notion that participants learn from their earlier mispredictions and make more accurate predictions as the semester progresses (although it is possible that this form of learning is relatively small, and that a larger sample size could detect such an effect).

General discussion

People’s predictions of their physical and mental performance are often overconfident (e.g., Billeter et al., 2011; Kruger & Dunning, 1999; Lipko et al., 2009; Schneider, 1998). Furthermore, the phenomenon known as “unskilled and unaware” is characterized by a pattern of performance predictions in which the lowest-performing people are the most overconfident in their performance predictions, but the highest-performing people tend to be accurate or to demonstrate underconfidence (e.g., Dunning et al., 2003; Grimes, 2002; Kruger & Dunning, 1999). In three studies, we examined the hypothesis that students’ predictions of their grades are persistently overconfident because their predictions are biased by their desired level of performance. The results of all three studies strongly supported this hypothesis: Students’ desired grades were strong predictors of their grade predictions, when students made their predictions either more than a week before the exam (Study 1) or immediately after completing the exam (Studies 2 and 3).

Overconfidence can have detrimental effects on students’ performance on an upcoming exam if they overestimate their learning and, in turn, underprepare for the exam (e.g., Bjork et al., 2013; Dunlosky & Rawson, 2012; Finn, 2008; Grimes, 2002; Kornell & Bjork, 2008; Kornell & Metcalfe, 2006; Metcalfe & Finn, 2008; Shanks & Serra, 2014). Furthermore, in the case of long-term or more general overconfidence (i.e., overconfidence on things in the longer-term future), students’ desires might also affect their intended plan of study. Students might anticipate that they will have ample time to study for an exam to earn a high grade on it, but end up overestimating how much time they will actually be willing or able to devote to study as the exam approaches (i.e., the planning fallacy; see Buehler, Griffin, & Ross, 2002). Future attempts to make students’ JOLs more accurate might address the potential role of desired grades examined in the present studies. We are not suggesting that researchers attempt to reduce the magnitude of students’ desired grades; rather, it may be worth exploring whether making students aware that their desires may bias their judgments could reduce these biases (i.e., keep students’ desired grades high but reduce students’ reliance on these desires when assessing the current state of their learning).

Might the unskilled be aware?

Despite the frequent occurrence of the unskilled-and-unaware pattern—and concern about its implications for the unskilled—evidence suggests that the lowest-performing people in a given domain have some realization that their estimates might not be accurate. Specifically, Miller and Geraci (2011b) had students rate how confident they were in the accuracy of their exam grade predictions. Despite obtaining the unskilled-and-unaware pattern of results, in which the lowest-performing students made highly overconfident grade predictions, these students were less confident in the accuracy of their predictions than were the highest-performing students. Why, then, did these students make high predictions that they were unsure about?

One possibility is based on the finding that discrepancies between a person’s desires and the actual experience while taking the exam can create a sense of conflict, ambivalence, confusion, or lack of clarity. For example, discrepancies between people’s actual and desired evaluations lead people to feel more ambivalent in that evaluation (DeMarree, Wheeler, Briñol, & Petty, 2014), and discrepancies between actual and desired levels of self-esteem undermine a person’s self-clarity (DeMarree & Rios, 2014). This presumably occurs because, when a person’s desired and actual evaluations are discrepant, their choices of behavioral responses to those two evaluations are likely to be in conflict. For example, when people want to be more positive or more negative than they currently are in their evaluation of a political candidate, their actual and desired evaluations might create conflicting approach/avoidance tendencies (DeMarree et al., 2014). In the present context, if people have a high desired grade for an exam but their experience taking the exam leads them to believe they have in fact performed poorly on it, they should hold their high grade prediction with less certainty than if they had desired a more moderate grade or experienced the exam as being easy. Put differently, the lowest-performing students may often experience a discrepancy between their goal grades and their experience with the exam, whereas the highest-performing students will not. This could lead the lowest-performing students to be relatively less confident in their predictions than the highest-performing students (cf. Händel & Fritzsche, 2016; Miller & Geraci, 2011b).

Another possibility is that students’ desired grades might serve as a psychological anchor, or mental starting point, for their grade predictions (e.g., England & Serra, 2012; Scheck, Meeter, & Nelson, 2004). After making an initial judgment (based largely on their desired grade), the lowest-performing students might do little to further evaluate or adjust their initial predictions. In contrast, the highest-performing students might also start with a desire-based anchor, but because of their knowledge of that domain, they might then be able to engage in further evaluation, deliberation, and adjustment of their predictions before arriving at a final value. Concordantly, the lowest-performing students should be relatively less confident in their predictions than the highest-performing students (cf. Händel & Fritzsche 2016; Miller & Geraci, 2011b), since different levels of deliberation went into their final evaluations (cf. Barden & Petty, 2008; Dunlosky, Serra, Matvey, & Rawson, 2005; Serra & England, 2012). Judgments made via more thoughtful anchoring and adjusting also tend to have greater impact and durability than less thoughtfully generated judgments (Wegener, Petty, Blankenship, & Detweiler-Bedell, 2010), which means that high-performing students might be more likely to act on the basis of (or in response to) their predictions than low-performing students. Importantly, although the present evidence and some prior published evidence (i.e., Händel & Fritzsche, 2016; Miller & Geraci, 2011b) are consistent with the idea that low-performing people do not engage in as much effort to adjust their judgments from their desired grade anchor, these findings do not elucidate why low-performing people might not adjust their judgments.

Finally, it is well-established that people’s JOLs are, at least in part, based on their interpretation of the available—and sometimes ambiguous—cues related to their learning (e.g., Koriat, 1997; Serra & Metcalfe, 2009). It is also well-established that people’s desires have a powerful influence over perception, leading them to see ambiguous stimuli as those that they are looking for (Balcetis & Dunning, 2006) or see desired objects as closer (Balcetis & Dunning, 2010), along with a number of other, similar perceptual biases (for a review, see Dunning & Balcetis, 2013). For example, when judging their current level of mastery, students who want to perform well on an exam or in a course might overemphasize cues about their learning that are in line with the desires, and ignore cues that suggest that they have not yet mastered the materials. More relevant to the present situation, students’ desires or goals might have biased their interpretation of the metacognitive cues they used to make their performance predictions. In the present studies, students’ desires to perform well on their exams might have led them to consider cues that supported the idea that they would perform well on the exam, leading to their optimistic performance predictions. The processes leading to this optimism, however, might not have led participants to be confident in that assessment. We do not yet know what factors influence participants’ confidence in their metacognitive judgments (e.g., Dunlosky et al., 2005; Serra, Dunlosky, & Hertzog, 2008; Serra & England, 2012); identifying such factors certainly warrants empirical examination in the future.

Closing remarks

To date, most discussion and study of the unskilled-and-unaware pattern has relied on the assumption that the pattern stems from at least two populations of people—the skilled and the unskilled—making performance predictions in two different ways. Unlike most current accounts of the unskilled-and-unaware pattern, the present account does not require the unskilled to be unaware of their abilities; instead, the present account utilizes the same mechanism—the influence of people’s desires on their performance predictions—to explain the predictions of both the skilled and the unskilled. The present results have clear implications for explaining overconfidence in general, and the unskilled-and-unaware pattern specifically, because most students (including the lowest-performing ones) wanted to earn high grades on the exams, and subsequently demonstrated marked overconfidence in their grade predictions. Although the projection of students’ desires is almost certainly not the only factor that drives the unskilled-and-unaware phenomenon, the present research suggests that these desires should be among the multitude of mechanisms considered when explaining this effect.