Skip to main content
Log in

Even after thirteen class exams, students are still overconfident: the role of memory for past exam performance in student predictions

  • Published:
Metacognition and Learning Aims and scope Submit manuscript

Abstract

Students often are overconfident when they predict their performance on classroom examinations, and their accuracy often does not improve across exams. One contributor to overconfidence may be that students did not have enough experience, and another is that students may under-use their knowledge of prior exam performance to predict performance on their upcoming exams. To evaluate the former, we examined student prediction accuracy across 13 exams in an introductory course on educational psychology. For the latter, we computed measures that estimate the extent to which students use the prior exam score when predicting performance and whether students should use the prior exam scores. Several outcomes are noteworthy. First, students were overconfident, and contrary to expectations, this overconfidence did not decline across exams. Second, students’ prior exam scores were not related to subsequent predictions, even though prior exam performance showed little bias with respect to predicting future performance. Thus, students appear to under-use prior performance despite its utility for improving prediction accuracy about future exam performance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2

Similar content being viewed by others

Notes

  1. As we detail in the discussion, such between-participant correlational analyses are indirect indicators of individual student’s judgment accuracy; so, this correlational evidence from Hacker et al. (2000) does not provide definitive support that MPE was being used by any given student.

  2. The weighted mean bias was calculated because the N for each exam was different as a result of some students missing some of the exams. Thus, the weighted mean bias was computed by multiplying the mean bias score for each exam by the N for that exam, summing the products together, and dividing this sum by the total N.

  3. To be included in the ANOVA, a student must have completed all 13 exams. Because students were allowed to drop one exam score from their overall course grade, many were missing at least one exam score. Multilevel models can be estimated when data are partially missing (Curran et al. 2010) whereas structural equations require some method to account for missing data, such as multiple imputation (Steele 2008).

  4. We thank an anonymous reviewer for suggesting this analysis.

References

  • Ariel, R., & Dunlosky, J. (2011). The sensitivity of judgment-of-learning resolution to past test performance, new learning, and forgetting. Memory & Cognition, 39, 171–184.

    Article  Google Scholar 

  • Bajšanski, I., Močibob, M., & Valerjev, P. (2014). Metacognitive judgments and syllogistic reasoning. Psychological Topics, 23, 143–165.

    Google Scholar 

  • Bastardi, A., Uhlmann, E. L., & Ross, L. (2015). Wishful thinking: belief, desire, and the motivated evaluation of scientific evidence. Psychological Science, 22, 731–732.

    Article  Google Scholar 

  • Bol, L., & Hacker, D. (2001). A comparison of the effects of practice tests and traditional review on performance and evaluation. The Journal of Experimental Education, 69, 133–144.

    Article  Google Scholar 

  • Bol, L., Hacker, D. J., O’Shea, P., & Allen, D. (2005). The influence of overt practice, achievement level, and explanatory style on calibration accuracy and performance. The Journal of Experimental Education, 73, 269–290.

    Article  Google Scholar 

  • Butler, D. L., & Winne, P. H. (1995). Feedback and self-regulated learning: a theoretical synthesis. Review of Educational Research, 65, 245–281.

    Article  Google Scholar 

  • Campbell, W. K., & Sedikides, C. (1999). Self-threat magnifies the self-serving bias: a meta-analytic integration. Review of General Psychology, 3, 23–43.

    Article  Google Scholar 

  • Curran, P. J., Obeidat, K., & Losardo, D. (2010). Twelve frequently asked questions about growth curve modeling. Journal of Cognitive Development, 11, 121–136.

    Article  Google Scholar 

  • Ditto, P. H., & Lopez, D. F. (1992). Motivated skepticism: use of differential decision criteria for preferred and nonpreferred conclusions. Journal of Personality and Social Psychology, 63, 568–584.

    Article  Google Scholar 

  • Dunlosky, J., & Hertzog, C. (2002). Updating knowledge about encoding strategies: a computational analysis of learning about strategy effectiveness from task experience. Psychology and Aging, 15, 462–474.

    Article  Google Scholar 

  • Dunlosky, J., & Rawson, K. A. (2012). Overconfidence produces underachievement: inaccurate self evaluations undermine students’ learning and retention. Learning and Instruction, 22, 271–280.

    Article  Google Scholar 

  • Dunlosky, J., Rawson, K. A., Marsh, E. J., Nathan, M. J., & Willingham, D. T. (2013). Improving students’ learning with effective learning techniques: promising directions from cognitive and educational psychology. Psychological Science In the Public Interest, 14, 4–58.

    Article  Google Scholar 

  • Finn, B., & Metcalfe, J. (2007). The role of memory for past test in the underconfidence with practice effect. Journal of Experimental Psychology. Learning, Memory, and Cognition, 33, 238–244.

    Article  Google Scholar 

  • Finn, B., & Metcalfe, J. (2008). Judgments of learning are influenced by memory for past test. Journal of Memory and Language, 58, 19–34.

    Article  Google Scholar 

  • Garavalia, L. S., & Gredler, M. E. (2002). An exploratory study of academic goal setting, achievement calibration and self-regulated learning. Journal of Instructional Psychology, 29, 221–230.

    Google Scholar 

  • Gluga, R., Kay, J., Lister, R., Kleitman, S. & Lever, T. (2012a). Coming to terms with Bloom: an online tutorial for teachers of programming fundamentals. Proceedings of the Fourteenth Australasian Computing Education Conference (ACE2012), Melbourne, Australia. CRPIT, 123. de Raadt, M. and Carbone, A. Eds., ACS., 147–156.

  • Gluga, R., Kay, J., Lister, R., Kleitman, S., & Lever, T. (2012b). Over-confidence and confusion in using Bloom for programming fundamentals assessment. Proceedings of the 43rd ACM technical symposium on Computer Science Education, SIGCSE ’12, 147–152.

  • Grimes, P. W. (2002). The overconfident principles of economics student: an examination of a metacognitive skill. The Journal of Economic Education, 33, 15–30.

    Article  Google Scholar 

  • Hacker, D. J., Bol, L., Horgan, D. D., & Rakow, E. A. (2000). Test prediction and performance in a classroom context. Journal of Educational Psychology, 92, 160–170.

    Article  Google Scholar 

  • Hacker, D. J., Bol, L., & Bahbahani, K. (2008). Explaining calibration accuracy in classroom contexts: the effects of incentives, reflection, and explanatory style. Metacognition and Learning, 3, 101–121.

    Article  Google Scholar 

  • Hertzog, C., Dixon, R. A., & Hultsch, D. F. (1990). Relationship between metamemory, memory predictions, and memory task performance in adults. Psychology and Aging, 5, 215–227.

    Article  Google Scholar 

  • Hertzog, C., Saylor, L. L., Fleece, A. M., & Dixon, R. A. (1994). Metamemory and aging: relations between predicted, actual and perceived memory task performance. Aging, Neuropsychology, and Cognition, 1, 203–237.

    Article  Google Scholar 

  • King, J. F., Zechmeister, E. B., & Shaughnessy, J. J. (1980). Judgments of knowing: the influence of retrieval practice. The American Journal of Psychology, 93, 329–343.

    Article  Google Scholar 

  • Kleitman, S., & Costa, D. S. J. (2014). The role of a novel formative assessment tool (stats-mIQ) and individual differences in real-life academic performance. Learning and Individual Differences, 29, 150–161.

    Article  Google Scholar 

  • Larson, J. R. (1977). Evidence for a self-serving bias in the attribution of causality. Journal of Personality, 45, 430–441.

    Article  Google Scholar 

  • Mezulis, A. H., Abramson, L. Y., Hyde, J. S., & Hankin, B. L. (2004). Is there a universal positivity bias in attributions? A meta-analytic review of individual, developmental, and cultural differences in the self-serving attributional bias. Psychological Bulletin, 130, 711–747.

    Article  Google Scholar 

  • Miller, T. M., & Geraci, L. (2011). Training metacognition in the classroom: the influence of incentives and feedback on exam predictions. Metacognition Learning, 6, 303–314.

    Article  Google Scholar 

  • Moreno, R., & Mayer, R. (2007). Interactive multimodal learning environments. Educational Psychology Review, 19, 309–326.

    Article  Google Scholar 

  • Nietfeld, J. L., Cao, L., & Osborne, J. W. (2006). The effect of distributed monitoring exercises and feedback on performance, monitoring accuracy, and self-efficacy. Metacognition and Learning, 1, 159–179.

    Article  Google Scholar 

  • Serra, M. J., & Ariel, R. (2014). People use the memory for past-test heuristic as an explicit cue for judgments of learning. Memory & Cognition, 42, 260–272.

    Article  Google Scholar 

  • Shaughnessy, J. J. (1979). Confidence-judgment accuracy as a predictor of test performance. Journal of Research in Personality, 13, 505–514.

    Article  Google Scholar 

  • Steele, F. (2008). Multilevel models for longitudinal data. Journal of the Royal Statistical Society: Series A (Statistics in Society), 171(1), 5–19.

    Google Scholar 

  • Stone, N. J. (2000). Exploring the relationship between calibration and self-regulated learning. Educational Psychology Review, 12, 437–475.

    Article  Google Scholar 

  • Tauber, S. K., & Rhodes, M. G. (2012). Multiple bases for young and older adults’ judgments of learning in multitrial learning. Psychology and Aging, 27, 474–483.

    Article  Google Scholar 

  • Thiede, K. W. (1999). The importance of monitoring and self-regulation during multi-trial learning. Psychonomic Bulletin & Review, 6, 662–667.

    Article  Google Scholar 

  • Winne, P. H. (2001). Self-regulated learning viewed from models of information processing. In B. J. Zimmerman, D. H. Schunk, & H. Dale (Eds.), Self-regulated learning and academic achievement: theoretical perspectives (pp. 153–189). Mahwah NJ: Lawrence Erlbaum.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Nathaniel L. Foster.

Ethics declarations

Conflict of interest

The authors declare no conflict of interest.

Appendix

Appendix

Table 4 Example Items from the pre-questionnaire, exam 1, and the post-questionnaire
Table 5 Parameter estimates for the three models examining bias in student exam predictions
Table 6 Parameter estimates for the three models examining memory for past exam performance in student exam predictions

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Foster, N.L., Was, C.A., Dunlosky, J. et al. Even after thirteen class exams, students are still overconfident: the role of memory for past exam performance in student predictions. Metacognition Learning 12, 1–19 (2017). https://doi.org/10.1007/s11409-016-9158-6

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11409-016-9158-6

Keywords

Navigation