Abstract
Students often are overconfident when they predict their performance on classroom examinations, and their accuracy often does not improve across exams. One contributor to overconfidence may be that students did not have enough experience, and another is that students may under-use their knowledge of prior exam performance to predict performance on their upcoming exams. To evaluate the former, we examined student prediction accuracy across 13 exams in an introductory course on educational psychology. For the latter, we computed measures that estimate the extent to which students use the prior exam score when predicting performance and whether students should use the prior exam scores. Several outcomes are noteworthy. First, students were overconfident, and contrary to expectations, this overconfidence did not decline across exams. Second, students’ prior exam scores were not related to subsequent predictions, even though prior exam performance showed little bias with respect to predicting future performance. Thus, students appear to under-use prior performance despite its utility for improving prediction accuracy about future exam performance.
Similar content being viewed by others
Notes
As we detail in the discussion, such between-participant correlational analyses are indirect indicators of individual student’s judgment accuracy; so, this correlational evidence from Hacker et al. (2000) does not provide definitive support that MPE was being used by any given student.
The weighted mean bias was calculated because the N for each exam was different as a result of some students missing some of the exams. Thus, the weighted mean bias was computed by multiplying the mean bias score for each exam by the N for that exam, summing the products together, and dividing this sum by the total N.
To be included in the ANOVA, a student must have completed all 13 exams. Because students were allowed to drop one exam score from their overall course grade, many were missing at least one exam score. Multilevel models can be estimated when data are partially missing (Curran et al. 2010) whereas structural equations require some method to account for missing data, such as multiple imputation (Steele 2008).
We thank an anonymous reviewer for suggesting this analysis.
References
Ariel, R., & Dunlosky, J. (2011). The sensitivity of judgment-of-learning resolution to past test performance, new learning, and forgetting. Memory & Cognition, 39, 171–184.
Bajšanski, I., Močibob, M., & Valerjev, P. (2014). Metacognitive judgments and syllogistic reasoning. Psychological Topics, 23, 143–165.
Bastardi, A., Uhlmann, E. L., & Ross, L. (2015). Wishful thinking: belief, desire, and the motivated evaluation of scientific evidence. Psychological Science, 22, 731–732.
Bol, L., & Hacker, D. (2001). A comparison of the effects of practice tests and traditional review on performance and evaluation. The Journal of Experimental Education, 69, 133–144.
Bol, L., Hacker, D. J., O’Shea, P., & Allen, D. (2005). The influence of overt practice, achievement level, and explanatory style on calibration accuracy and performance. The Journal of Experimental Education, 73, 269–290.
Butler, D. L., & Winne, P. H. (1995). Feedback and self-regulated learning: a theoretical synthesis. Review of Educational Research, 65, 245–281.
Campbell, W. K., & Sedikides, C. (1999). Self-threat magnifies the self-serving bias: a meta-analytic integration. Review of General Psychology, 3, 23–43.
Curran, P. J., Obeidat, K., & Losardo, D. (2010). Twelve frequently asked questions about growth curve modeling. Journal of Cognitive Development, 11, 121–136.
Ditto, P. H., & Lopez, D. F. (1992). Motivated skepticism: use of differential decision criteria for preferred and nonpreferred conclusions. Journal of Personality and Social Psychology, 63, 568–584.
Dunlosky, J., & Hertzog, C. (2002). Updating knowledge about encoding strategies: a computational analysis of learning about strategy effectiveness from task experience. Psychology and Aging, 15, 462–474.
Dunlosky, J., & Rawson, K. A. (2012). Overconfidence produces underachievement: inaccurate self evaluations undermine students’ learning and retention. Learning and Instruction, 22, 271–280.
Dunlosky, J., Rawson, K. A., Marsh, E. J., Nathan, M. J., & Willingham, D. T. (2013). Improving students’ learning with effective learning techniques: promising directions from cognitive and educational psychology. Psychological Science In the Public Interest, 14, 4–58.
Finn, B., & Metcalfe, J. (2007). The role of memory for past test in the underconfidence with practice effect. Journal of Experimental Psychology. Learning, Memory, and Cognition, 33, 238–244.
Finn, B., & Metcalfe, J. (2008). Judgments of learning are influenced by memory for past test. Journal of Memory and Language, 58, 19–34.
Garavalia, L. S., & Gredler, M. E. (2002). An exploratory study of academic goal setting, achievement calibration and self-regulated learning. Journal of Instructional Psychology, 29, 221–230.
Gluga, R., Kay, J., Lister, R., Kleitman, S. & Lever, T. (2012a). Coming to terms with Bloom: an online tutorial for teachers of programming fundamentals. Proceedings of the Fourteenth Australasian Computing Education Conference (ACE2012), Melbourne, Australia. CRPIT, 123. de Raadt, M. and Carbone, A. Eds., ACS., 147–156.
Gluga, R., Kay, J., Lister, R., Kleitman, S., & Lever, T. (2012b). Over-confidence and confusion in using Bloom for programming fundamentals assessment. Proceedings of the 43rd ACM technical symposium on Computer Science Education, SIGCSE ’12, 147–152.
Grimes, P. W. (2002). The overconfident principles of economics student: an examination of a metacognitive skill. The Journal of Economic Education, 33, 15–30.
Hacker, D. J., Bol, L., Horgan, D. D., & Rakow, E. A. (2000). Test prediction and performance in a classroom context. Journal of Educational Psychology, 92, 160–170.
Hacker, D. J., Bol, L., & Bahbahani, K. (2008). Explaining calibration accuracy in classroom contexts: the effects of incentives, reflection, and explanatory style. Metacognition and Learning, 3, 101–121.
Hertzog, C., Dixon, R. A., & Hultsch, D. F. (1990). Relationship between metamemory, memory predictions, and memory task performance in adults. Psychology and Aging, 5, 215–227.
Hertzog, C., Saylor, L. L., Fleece, A. M., & Dixon, R. A. (1994). Metamemory and aging: relations between predicted, actual and perceived memory task performance. Aging, Neuropsychology, and Cognition, 1, 203–237.
King, J. F., Zechmeister, E. B., & Shaughnessy, J. J. (1980). Judgments of knowing: the influence of retrieval practice. The American Journal of Psychology, 93, 329–343.
Kleitman, S., & Costa, D. S. J. (2014). The role of a novel formative assessment tool (stats-mIQ) and individual differences in real-life academic performance. Learning and Individual Differences, 29, 150–161.
Larson, J. R. (1977). Evidence for a self-serving bias in the attribution of causality. Journal of Personality, 45, 430–441.
Mezulis, A. H., Abramson, L. Y., Hyde, J. S., & Hankin, B. L. (2004). Is there a universal positivity bias in attributions? A meta-analytic review of individual, developmental, and cultural differences in the self-serving attributional bias. Psychological Bulletin, 130, 711–747.
Miller, T. M., & Geraci, L. (2011). Training metacognition in the classroom: the influence of incentives and feedback on exam predictions. Metacognition Learning, 6, 303–314.
Moreno, R., & Mayer, R. (2007). Interactive multimodal learning environments. Educational Psychology Review, 19, 309–326.
Nietfeld, J. L., Cao, L., & Osborne, J. W. (2006). The effect of distributed monitoring exercises and feedback on performance, monitoring accuracy, and self-efficacy. Metacognition and Learning, 1, 159–179.
Serra, M. J., & Ariel, R. (2014). People use the memory for past-test heuristic as an explicit cue for judgments of learning. Memory & Cognition, 42, 260–272.
Shaughnessy, J. J. (1979). Confidence-judgment accuracy as a predictor of test performance. Journal of Research in Personality, 13, 505–514.
Steele, F. (2008). Multilevel models for longitudinal data. Journal of the Royal Statistical Society: Series A (Statistics in Society), 171(1), 5–19.
Stone, N. J. (2000). Exploring the relationship between calibration and self-regulated learning. Educational Psychology Review, 12, 437–475.
Tauber, S. K., & Rhodes, M. G. (2012). Multiple bases for young and older adults’ judgments of learning in multitrial learning. Psychology and Aging, 27, 474–483.
Thiede, K. W. (1999). The importance of monitoring and self-regulation during multi-trial learning. Psychonomic Bulletin & Review, 6, 662–667.
Winne, P. H. (2001). Self-regulated learning viewed from models of information processing. In B. J. Zimmerman, D. H. Schunk, & H. Dale (Eds.), Self-regulated learning and academic achievement: theoretical perspectives (pp. 153–189). Mahwah NJ: Lawrence Erlbaum.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no conflict of interest.
Appendix
Appendix
Rights and permissions
About this article
Cite this article
Foster, N.L., Was, C.A., Dunlosky, J. et al. Even after thirteen class exams, students are still overconfident: the role of memory for past exam performance in student predictions. Metacognition Learning 12, 1–19 (2017). https://doi.org/10.1007/s11409-016-9158-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11409-016-9158-6