Abstract
As beliefs are well-known antecedents of teachers’ practices, including assessment practices, sound measurement of teacher beliefs is critical for scholarly research as well as practical purposes. The present study examined the validity of inferences derived from the Conceptions of Assessment III—Abridged (COA-IIIA) instrument with US PK-12 pre-service (N = 554) and in-service (N = 341) teachers. Prior research on the COA-IIIA in diverse contexts and with different teacher populations has yielded highly divergent findings concerning this instrument’s internal score structure. The study relies on confirmatory factor analyses to investigate the COA-IIIA’s score structure among US pre-service and in-service teachers and to test the measurement invariance of select COA-IIIA measurement models across these groups. Findings indicate poor fit of most prior COA-IIIA measurement models with US pre-service and in-service teachers, though we were able to specify several simplified models that were at least metrically invariant across these populations.
Similar content being viewed by others
Data availability
The data that support the findings of this study are available from the authors upon reasonable request.
Notes
Our review of the literature focuses only on peer-reviewed studies in which the COA-IIIA was unmodified in terms of item content or response scale (to the extent we were able to verify this). However, some studies reviewed did translate the language of the items.
This included post-baccalaureate programs and combined undergraduate and graduate teacher education programs.
Of the PST sample, 0.92% selected “other” for their intended school level.
School level taught was indeterminable for 1.44% of IST participants with a non-missing response for this variable.
Missing data for PST demographic and professional variables constituted between 0.18 and 72.02% of PST cases and between 0.29 and 18.48% of IST cases. Some of this is due to inconsistencies across studies and datasets in available non-COA-IIIA data. Some of these estimates should be considered very unreliable.
As noted by a reviewer, future studies of the COA-IIIA may consider bifactor modeling approaches, which have enjoyed some use in prior research on this instrument (Brown & Remesal, 2017).
References
Barnes, N., Fives, H., & Dacey, C. M. (2015). Teachers’ beliefs about assessment. In H. Fives & M. G. Gill (Eds.), International handbook of research on teachers’ beliefs (pp. 284–300). Routledge.
Barnes, N., Fives, H., & Dacey, C. M. (2017). US teachers’ conceptions of the purposes of assessment. Teaching and Teacher Education, 65, 107–116.
Bentler, P. M. (1990). Comparative fit indexes in structural models. Psychological Bulletin, 107, 238–246.
Bentler, P. M., & Bonett, D. G. (1980). Significance tests and goodness of fit in the analysis of covariance structures. Psychological Bulletin, 88, 588–606.
Brosseau-Liard, P. E., & Savalei, V. (2014). Adjusting incremental fit indices for nonnormality. Multivariate Behavioral Research, 49(5), 460–470.
Brosseau-Liard, P. E., Savalei, V., & Li, L. (2012). An investigation of the sample performance of two nonnormality corrections for RMSEA. Multivariate Behavioral Research, 47(6), 904–930.
Brown, G. T. L. (2004). Teachers’ conceptions of assessment: Implications for policy and professional development. Assessment in Education: Principles, Policy, & Practice, 11(3), 301–318.
Brown, G. T. L. (2006). Teachers’ conceptions of assessment: Validation of an abridged version. Psychological Reports, 99, 166–170.
Brown, G. T. L. (2011). Teachers’ conceptions of assessment: Comparing primary and secondary teachers in New Zealand. Assessment Matters, 3, 45–70.
Brown, G. T. L., & Michaelides, M. P. (2011). Ecological rationality in teachers’ conceptions of assessment across samples from Cyprus and New Zealand. European Journal of Psychology of Education, 26(3), 319–337.
Brown, G. T. L., & Remesal, A. (2012). Prospective teachers’ conceptions of assessment: A cross-cultural comparison. The Spanish Journal of Psychology, 15, 75–89.
Brown, G. T., & Remesal, A. (2017). Teachers’ conceptions of assessment: Comparing two inventories with Ecuadorian teachers. Studies in Educational Evaluation, 55, 68–74.
Brown, G. T. L., Kennedy, K. J., Fok, P. K., Chan, J. K. S., & Yu, W. M. (2009). Assessment for student improvement: Understanding Hong Kong teachers’ conceptions and practices of assessment. Assessment in Education: Principles, Policy & Practice, 16, 347–363.
Brown, G. T. L., Lake, R., & Matters, G. (2011). Queensland teachers’ conceptions of assessment: The impact of policy priorities on teacher attitudes. Teaching and Teacher Education, 27, 210–220.
Brown, G. T. L., Chaudhry, H., & Dhamija, R. (2015). The impact of an assessment policy upon teachers’ self-reported assessment beliefs and practices: A quasi-experimental study of Indian teachers in private schools. International Journal of Educational Research, 71, 50–64.
Brown, G. T. L., Gebril, A., & Michaelides, M. P. (2019). Teachers’ conceptions of assessment: A global phenomenon or a global localism. Frontiers in Education, 4(16). https://doi.org/10.3389/feduc.2019.00016
Cheung, G. W., & Rensvold, R. B. (2002). Evaluating goodness-of-fit indexes for testing measurement invariance. Structural Equation Modeling, 9(2), 233–255.
Clark, C., & Peterson, P. (1986). Teachers’ thought processes. In M. Wittrock (Ed.), Handbook of research on teaching (pp. 255–296). Macmillan.
Daniels, L. M., & Poth, C. A. (2017). Relationships between pre-service teachers’ conceptions of assessment, approaches to instruction, and assessment: An achievement goal theory perspective. Educational Psychology, 37(7), 835–853.
Daniels, L. M., Poth, C., Papile, C., & Hutchison, M. (2014). Validating the Conceptions of Assessment-III Scale in Canadian preservice teachers. Educational Assessment, 19(2), 139–158.
Datnow, A., & Hubbard, L. (2015). Teacher capacity for and beliefs about data-driven decision making: A literature review of international research. Journal of Educational Change, 17(1), 7–28.
Dimitrov, D. M. (2010). Testing for factorial invariance in the context of construct validation. Measurement and Evaluation in Counseling and Development, 43(2), 121–149.
Fan, X., & Sivo, S. A. (2007). Sensitivity of fit indices to model misspecification and model types. Multivariate Behavioral Research, 42(3), 509–529.
Gebril, A., & Brown, G. T. L. (2014). The effect of high-stakes examination systems on teacher beliefs: Egyptian teachers’ conceptions of assessment. Assessment in Education: Principles, Policy & Practice, 21(1), 16–33.
Hammer, J. H. (2016). Construct replicability calculator: A Microsoft Excel-based tool to calculate the Hancock and Mueller (2001) H index. Available from https://DrJosephHammer.com/. Accessed 13 Mar 2023.
Hancock, G. R., & Mueller, R. O. (2001). Rethinking construct reliability within latent variable systems. In R. Cudeck, S. du Toit, & D. Sorbom (Eds.) Structural equation modeling: Present and future—A Festschrift in honor of Karl Joreskog (pp. 195–216). Lincolnwood, IL: Scientific Software International.
Harris, L. R., & Brown, G. T. L. (2009). The complexity of teachers’ conceptions of assessment: Tensions between the needs of schools and students. Assessment in Education: Principles, Policy & Practice, 16(3), 365–381.
Hu, L. T., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55.
Jung, E., & Yoon, M. (2016). Comparisons of three empirical methods for partial factorial invariance: Forward, backward, and factor-ratio tests. Structural Equation Modeling: A Multidisciplinary Journal, 23(4), 567–584.
Kerr, K. A., Marsh, J. A., Ikemoto, G. S., Darilek, H., & Barney, H. (2006). Strategies to promote data use for instructional improvement: Actions, outcomes, and lessons from three urban districts. American Journal of Education, 112(4), 496–520.
Kraner, C., & Reeves, T. D. (2018, October). Conceptions of Assessment-III Abridged (COA-IIIA) score reliability and validity among pre-service teachers in the United States. Paper presented at the meeting of the Mid-Western Educational Research Association, Cincinnati, OH.
Leighton, J. P., Gokiert, R. J., Cor, M. K., & Heffernan, C. (2010). Teacher beliefs about the cognitive diagnostic information of classroom-versus large-scale tests: Implications for assessment literacy. Assessment in Education: Principles, Policy & Practice, 17(1), 7–21.
Little, R. J. (1988). A test of missing completely at random for multivariate data with missing values. Journal of the American Statistical Association, 83, 1098–1202.
MacCallum, R. C., Widaman, K. F., Zhang, S., & Hong, S. (1999). Sample size in factor analysis. Psychological Methods, 4, 84–99.
Mardia, K. V. (1970). Measures of multivariate skewness and kurtosis with applications. Biometrika, 57, 519–530.
Marsh, J. A. (2012). Interventions promoting educators’ use of data: Research insights and gaps. Teachers College Record, 114(11), 1–48.
McDonald, R. P. (2013) Test theory: A unified treatment. Mahwah, NJ: Lawrence Earlbaum Associates.
Meade, A. W., & Bauer, D. J. (2007). Power and precision in confirmatory factor analytic tests of measurement invariance. Structural Equation Modeling, 14(4), 611–635.
Meredith, W. (1993). Measurement invariance, factor analysis and factorial invariance. Psychometrika, 58(4), 525–543.
Mullis, I. V., Martin, M. O., Minnich, C. A., Stanco, G. M., Arora, A., Centurino, V. A., & Castle, C. E. (2012). TIMSS 2011 encyclopedia: Education policy and curriculum in mathematics and science. International Association for the Evaluation of Educational Achievement.
Olsson, U. H., Foss, T., Troye, S. V., & Howell, R. D. (2000). The performance of ML, GLS, and WLS estimation in structural equation modeling under conditions of misspecification and nonnormality. Structural Equation Modeling, 7(4), 557–595.
Pajares, M. F. (1992). Teachers’ beliefs and educational research: Cleaning up a messy construct. Review of Educational Research, 62(3), 307–332.
Pellegrino, J. W., Chudowsky, N., & Glaser, R. (Eds.). (2001). Knowing what students know: The science and design of educational assessment. National Academies Press.
Prenger, R., & Schildkamp, K. (2018). Data-based decision making for teacher and student learning: A psychological perspective on the role of the teacher. Educational Psychology, 38(6), 734–752.
Reeves, T. D., Summers, K. H., & Grove, E. (2016). Examining the landscape of teacher learning for data use: The case of Illinois. Cogent Education, 3(1).
Rosseel, Y. (2012). lavaan: An R package for structural equation modeling. Journal of Statistical Software, 48(2), 1–36.
Ruiz-Primo, M. A., Shavelson, R. J., Hamilton, L., & Klein, S. (2002). On the evaluation of systemic science education reform: Searching for instructional sensitivity. Journal of Research in Science Teaching, 39(5), 369–393.
Satorra, A., & Bentler, P. M. (2001). A scaled difference chi-square test statistic for moment structure analysis. Psychometrika, 66(4), 507–514.
Segers, M., & Tillema, H. (2011). How do Dutch secondary teachers and students conceive the purpose of assessment? Studies in Educational Evaluation, 37(1), 49–54.
Steiger, J. H., & Lind, J. C. (1980, June). Statistically-based tests for the number of common factors. Paper presented at the Annual Spring Meeting of the Psychometric Society, Iowa City, IA.
Tschannen-Moran, M., Hoy, A. W., & Hoy, W. K. (1998). Teacher efficacy: Its meaning and measure. Review of Educational Research, 68(2), 202–248.
Acknowledgements
The authors thank all pre-service and in-service teachers who participated in the study, as well as Amy Martin, Evelyn Comber, David Walker, Kelly Summers, and Thomas Smith for their feedback provided on an earlier version of this manuscript. We also thank Gavin Brown for his swift and helpful assistance in specifying several of the tested models, and Mary Sanderson for the data analytic support.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Reeves, T.D., Onder, Y. & Kraner, C. Validation and invariance of the Conceptions of Assessment-III Abridged (COA-IIIA) among pre-service and in-service teachers in the United States. Educ Asse Eval Acc 35, 419–447 (2023). https://doi.org/10.1007/s11092-023-09407-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11092-023-09407-4