Abstract
This theoretical piece discusses the concept of a teacher learning progression in an attempt to integrate teacher learning and assessment. From the authors’ perspective, the main features of the teacher learning progression are the longitudinal understanding of teacher knowledge and practice, and the opportunity to align teacher evaluations’ formative and summative purposes. Criteria to assess existing teacher learning progressions are proposed and used to examine examples of teacher assessment systems implemented in different parts of the world. The concept of teacher learning progression has national and international implications for teacher training, for teaching assessment and for the design and implementation of educational policies.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Notes
Cut scores vary by state as do passing rates, and to date state cut scores range from 35 to 41. For details, see Stanford Center for Assessment, Learning, and Equity (SCALE), 2018.
In addition to the Danielson’s framework, teacher assessments have included estimates of teacher value-added based on student performance on standardized tests (Millman, 1997). Value-added measures aim at capturing teacher-specific contribution to the learning process of students taking into consideration variables outside of teachers’ control such as students’ sociodemographic characteristics (Blazar et al. 2016; Close et al. 2020; Harris and McCaffrey 2010). Because of that, value- added measures (teacher effect) are considered fairer than other measures based exclusively on students’ performance level (McCaffrey et al. 2004). The discussion presented in this article, however, will not refer to estimates of teacher effects as they are not linked directly either to a concrete model nor to an operationalization of what teachers should know and be able to do. There is no theory behind the analysis of value-added models in terms of what the standard teacher should achieve and how, nor there are descriptions regarding their progress over time. In that sense, the value-added evaluations provide limited formative information. Furthermore, value-added measures have been criticized because the scales used are often not vertically scaled and narrowly assess basic skill (Darling-Hammond 2012), because of measurement error (Hill 2009), the difficulty to attribute learning gains to multiple teacher present in the classroom or to teachers who are only partially involved (Isenberg and Hock 2011), and the instability of their estimation (Newton et al., 2012). In addition, Rothstein (2010) reports that short-term value-added estimates are weakly related to teachers’ long-run effects. Additionally, the assumption of random assignment of teachers to schools and classrooms and the random assignment of children to schools seldom holds true (Rothstein 2010).
This conceptualization of curriculum content is then seen as the justification for the structure of the typical “standardized test” which is called “domain testing.” In domain testing, the aim is to estimate the proportion of a domain of items related to a concept that a student would get right in an ideal situation at a given point in time—hence, it is fundamentally a cross-sectional perspective. In fact, the student is only given a sample of that domain, and then the true proportion is seen as being estimated by the proportion correct in the sample. In the case of complicated school domains such as “maths” the domain is usually two-level—the coarser level being the overall domain—“maths”—and the finer level being composed of a domain for each of the sub-topics in that particular math curriculum (such as number, geometry, etc.). As the subject changes from year to year (say grade 5 to grade 6), the cross-sectional domain definition changes to map the expected student changes with time. Thus, the different domains for each grade can be seen as a series of cross-sectional perspectives.
They include explaining the material, eliciting student thinking, technology use, recognizing pattern of thinking, generating class conditions for learning, providing feedback and communication with parents and peers. They were developed based on the work of Grossman et al. (2009), and Ball and Forzani (2010).
References
Alonzo, A. (2011). Learning progressions that support formative assessment practices. Measurement, 9(2–3), 124–129. https://doi.org/10.1080/15366367.2011.599629
Alonzo, A. C. (2012). Learning progressions: Significant promise, significant challenge. Zeitschrift Für Erziehungswissenschaft, 15(1), 95–109.
Alonzo, A. C. (2018). An argument for formative assessment with science learning progressions. Applied Measurement in Education, 31(2), 104–112.
Alonzo, A., & Elby, A. (2019). Beyond empirical adequacy: Learning progressions as models and their value for teachers. Cognition and Instruction, 37(1), 1–37.
Australian Institute for Teaching and School Leadership [AITSL]. (2011). National Professional Standards for Teachers. Melbourne, Australia: AITSL.
Avalos, B. (2011). Teacher professional development in Teaching and Teacher Education over ten year. Teaching and Teacher Education, 27 (2011).
Avalos-Bevan, B. (2018). Teacher evaluation in Chile: Highlights and complexities in 13 years of experience. Teachers and Teaching, 24(3), 297–311. https://doi.org/10.1080/13540602.2017.1388228
Ball, D. L., & Forzani, F. M. (2009). The work of teaching and the challenge for teacher education. Journal of Teacher Education, 60(5), 497–511. https://doi.org/10.1177/0022487109348479
Ben-Peretz, M. (2011). Teacher knowledge: What is it? How do we uncover it? What are its implications for schooling? Teaching and Teacher Education, 27(1), 3–9. https://doi.org/10.1016/j.tate.2010.07.015
Black, P. (2013). Pedagogy in theory and in practice: Formative and summative assessments in classrooms and in systems. Valuing Assessment in Science Education: Pedagogy, Curriculum, Policy, pages 207–229.
Black, P., Wilson, M., & Yao, S. (2011a). Road maps for learning: A guide to the navigation of learning progressions. Measurement: Interdisciplinary Research and Perspective, 9(2–3), 71–123. https://doi.org/10.1080/15366367.2011.591654
Black, P., Wilson, M., & Yao, S. (2011b). Comments and thoughts. Measurement: Interdisciplinary Research and Perspective, 9(2–3), 169–172. https://doi.org/10.1080/15366367.2011.604275
Blazar, D., Litke, E., & Barmore, J. (2016). What does it mean to be ranked a “high” or “low” value-added teacher? Observing differences in instructional quality across districts. American Educational Research Journal, 53(2), 324–359. https://doi.org/10.3102/0002831216630407
Campbell, R. J., Kyriakides, L., Muijs, R. D., & Robinson, W. (2003). Differential teacher effectiveness: Towards a model for research and teacher appraisal. Oxford Review of Education, 29(3), 347–362. https://doi.org/10.1080/03054980307440
Clinton, J., Dinham, S., Savage, G., Aston, R., Dabrowski, A., Gullickson, A., Calnin, G., & Arbour, G. (2015). Evaluation of the implementation of the Australian Professional Standards for Teachers. The University of Melbourne, AITSL.
Close, K., Amrein-Beardsley, A., & Collins, C. (2020). Putting teacher evaluation systems on the map: An overview of states' teacher evaluation systems post-Every Student Succeeds Act. Education Policy Analysis Archives, 28, Article 58. https://doi.org/10.14507/epaa.28.5252
Corcoran, T., Mosher, F.A., & Rogat, A. (2009). Learning progressions in science: An evidence- based approach to reform. CPRE Research Reports. https://doi.org/10.12698/cpre.2009.rr63
Danielson, C. (1996). Enhancing professional practice: A framework for teaching. Alexandria, VA: Association for Supervision and Curriculum Development.
Danielson, C. (2007). Enhancing professional practice: A framework for teaching (2nd ed.). Association for Supervision and Curriculum Development.
Darling-Hammond, L. (2012). Powerful teacher education: Lessons from exemplary programs. JosseyBass.
Darling-Hammond, L. (2017). Teacher education around the world: What can we learn from international practice? European Journal of Teacher Education, 40(3), 291–309. https://doi.org/10.1080/02619768.2017.1315399
DeLuca, C., LaPointe-McEwan, D., & Luhanga, U. (2016). Teacher assessment literacy: A review of international standards and measures. Educational Assessment, Evaluation and Accountability, 28, 251–272. https://doi.org/10.1007/s11092-015-9233-6
DeLuca, C., Chapman-Chin, A., & Klinger, D. A. (2019). Toward a Teacher Professional Learning Continuum in Assessment for Learning. Educational Assessment, 24(4), 267–285. https://doi.org/10.1080/10627197.2019.1670056
Duckor, B., & Holmberg, C. (2019). Exploring how to model formative assessment trajectories of posing-pausing-probing practices: Toward a teacher learning progressions framework for the study of novice teachers. Journal of Educational Measurement, 56, 836–890. https://doi.org/10.1111/jedm.12239
Furtak, E., Circi, R., & Heredia, S. (2018). Exploring alignment among learning progressions, teacher-designed formative assessment tasks, and student growth: Results of a four-year study. Applied Measurement in Education, 31(2), 143–156. https://doi.org/10.1080/08957347.2017.1408624
Furtak, E. (2018). Supporting teachers’ formative assessment practice with learning progressions. Routledge. Taylor & Francis: London and New York.
Gitomer, D. H., Martínez, J. F., Battey, D., & Hyland, N. E. (2019). Assessing the assessment: Evidence of reliability and validity in the edTPA. American Educational Research Journal. https://doi.org/10.3102/0002831219890608
Grossman, P., Hammerness, K., & McDonald, M. (2009). Redefining teaching, re-imagining teacher education. Teachers and Teaching: Theory and Practice, 15(2), 273–289. https://doi.org/10.1080/13540600902875340
Halverson, R., Kelley, C., & Kimball, S. (2004). Implementing teacher evaluation systems: how principals make sense of complex artifacts to shape local instructional practice. In W. K. Hoy, & C. G. Miskel (Eds.), Research and theory in educational administration (Vol. 3. pp. Greenwich, CT: Information Age Press.
Harris, D., & McCaffrey, D. F. (2010). Valued-added: Assessing teachers contributions to student achievement. In M. Kennedy (Ed.), Teacher assessment and the quest for teacher quality: A handbook (pp. 251–282). Wiley.
Heneman III, H. G., Milanowski, A., Kimball, S. M., & Odden, A. (2006). standards-based teacher evaluation as a foundation for knowledge-and skill-based pay. CPRE Policy Brief, RB-45. http://www.cpre.org/index.php?option=com_content&task=view&id=68&Itemid=73
Henning, A. S. (2014). An argument-based validation study of the teacher performance assessment in Washington State, Durham theses, Durham University. Available at Durham E-Theses Online: http://etheses.dur.ac.uk/10728/
Heritage, H. (2011). Commentary on road maps for learning: A guide to the navigation of learning progressions. Measurement: Interdisciplinary Research and Perspectives, 9(2–3), 149–151. https://doi.org/10.1080/15366367.2011.599647
Hill, H. C. (2009). Evaluating value-added models: A measurement perspective. Journal of Policy Analysis and Management, 28, 702–709.
Huber, S. G., & Skedsmo, G. (2016). Teacher evaluation-accountability and improving teaching practices. Educational Assessment Evaluation and Accountability, 28, 105–109. https://doi.org/10.1007/s11092-016-9241-1
Isenberg & Hock (2011). Value-added methods in a high-stakes assessment system. Paper presented at the annual meeting of the American Educational Research Association in New Orleans, April 2011.
Ingvarson, L., & Rowe, K. (2007, Feb 5). Conceptualising and evaluating teacher quality: Substantive and methodological issues [Paper presentation]. Economics of Teacher Quality conference, Australian National University, Canberra, Australia. http://research.acer.edu.au/learning_processes/8
Jacob , B. & Lefgren, L. (2005). Principals as agents: Subjective performance measurement in education. NBER Working Paper, 11463. MA: National Bureau Economic Research.
Kane, T. J., & Staiger, D. O. (2012). Ensuring fair and reliable measures of effective teaching: culminating findings from the MET Project’s three-year study. Seattle, WA: Bill & Melinda Gates foundation.
Kennedy, M. A. (2016). Parsing the practice of teaching. Journal of Teacher Education, 67(1), 6–17. https://doi.org/10.1177/0022487115614617
Kyriakides, L., & Demetriou, D. (2007). Introducing a teacher evaluation system based on teacher effectiveness research: An investigation of stakeholders’ perceptions. Journal of Personnel Evaluation in Education, 20, 43–64. https://doi.org/10.1007/s11092-007-9046-3
Martínez, F., Taut, S., & Schaaf, K. (2016). Classroom observation for evaluating and improving teaching: An international perspective. Studies of Educational Evaluation, 49, 15–29.
Masters, G., & Forster, M. (1996). Progress maps. Assessment resource kit. Commonwealth of Australia.
McCaffrey, D. F., Lockwood, J. R., Koretz, D. M., Louis, T. A., & Hamilton, L. S. (2004). Models for value-added modeling of teacher effects. Journal of Educational and Behavioral Statistics, 29, 67–101.
Millman, J. (Ed.). (1997). Grading teachers, grading schools: Is student achievement a valid evaluation measure? Corwin Press.
National Assessment Governing Board [NAGB]. (2005). Science NAEP 2009: Science framework for the 2009 National Assessment of Educational Progress. WestEd and the Council of Chief State School Officers.
National Research Council [NRC]. (2001). Knowing what students know: The science and design of educational assessment. Committee on the Foundations of Assessment. J. Pellegrino, N. Chudowsky, & R. Glaser (Eds). Washington, DC: National Academy Press.
National Research Council [NRC]. (2006). Systems for state science assessment. Committee on test design for K-12 science achievement. M. Wilson & M. Bertenthal (Eds.), Board on Testing and Assessment, Center for Education. Division of Behavioral and Social Sciences and Education. Washington, D.C.: National Academy Press.
National Research Council [NRC]. (2007). Taking science to school: Learning and teaching science in grades K-8. Committee on Science Learning, Kindergarten through Eighth Grade. R. A. Duschl, H. A. Schweingruber, & A. W. Shouse (Eds.). Washington, D.C.: National Academy Press.
Newton, X., Darling-Hammond, L., Haertel, E. & Thomas, E. (2010). Valued-added modeling of teacher effectiveness: An exploration of stability across models and contexts. Education Policy Analysis Archives, 18(23). Retrieved from http://epaa.asu.edu/ojs/article/view/810.
Park, S., & Suh, J. K. (2015). From portraying toward assessing PCK. In A. Berry, P. Friedrichsen, & J. Loughran (Eds.), Re-examining Pedagogical Content Knowledge in Science Education (pp. 104–119). Routledge.
Pianta, R. C., Hamre, B. K., & Allen, J. P. (2012). Teacher-student relationships and engagement: Conceptualizing, measuring, and improving the capacity of classroom interactions. In S. L. Christenson, A. L. Reschly, & C. Wylie (Eds.), Handbook of research on student engagement (pp. 365–386). Boston, MA: Springer. https://doi.org/10.1007/978-1-4614-2018-7_17
Reagan, E. M., Schram, T., McCurdy, K., Chang, T., & Evans, C. M. (2016). Politics of policy: Assessing the implementation, impact, and evolution of the Performance Assessment for California Teachers (PACT) and edTPA. Education Policy Analysis Archives, 24(9), 1–27. https://doi.org/10.14507/epaa.24.2176
Roegman, R., Goodwin, A. L., Reed, R., & Scott-McLaughlin, R. M., II. (2016). Unpacking the data: An analysis of the use of Danielson’s (2007) Framework for Professional Practice in a teaching residency program. Educational Assessment Evaluation and Accountability, 28(2), 111–137. https://doi.org/10.1007/s11092-015-9228-3
Rothstein, J. (2010). Teacher quality in educational production: Tracking, decay, and student achievement. Quarterly Journal of Economics, 125(1), 175–214. https://doi.org/10.1162/qjec.2010.125.1.175
Sandilos, L. E., Sims, W., Norwalk, K. E. & Reddy, L. (2019). Converging on quality: Examining multiple measures of teaching effectiveness, Journal of School Psychology 74 (2019), 10–28.Author
Schneider, R. M., & Plasman, K. (2011). Science teacher learning progressions: A review of science teachers’ pedagogical content knowledge development. Review of Educational Research, 81(4), 530–565. https://doi.org/10.3102/0034654311423382
Shavelson, R. J, Moss, P., Wilson, M., Duckor, B., Baron, W., & Wilmot, D. (2010, May). The promise of teacher learning progressions: Challenges and opportunities for articulating growth in the profession [Paper presentation]. Teacher Learning Progressions symposium for Division D-Measurement and Research Methodology, American Education Research Association, Denver, CO, United States.
Shepard, L. A. (2018). Learning progressions as tools for assessment and learning. Applied Measurement in Education, 31(2), 165–174. https://doi.org/10.1080/08957347.2017.1408628
Shulman, L. S. (1986). Paradigms and research programs in the study of teaching: A contemporary perspective. In M. C. Wittrock (Ed.), Handbook of research on teaching (3rd ed., pp. 3–36). Macmillan.
Shulman, L. S. (1986). Those who understand: Knowledge growth in teaching. Educational Researcher, 15(2), 4–14. https://doi.org/10.3102/0013189X015002004
Stanford Center for Assessment, Learning, and Equity [SCALE] (2013). 2013 edTPA field test: Summary report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=827&ref=edtpa
Stanford Center for Assessment, Learning, and Equity [SCALE] (2015). Educative assessment & meaningful support: 2014 edTPA administrative report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=2183&ref=edtpa
Stanford Center for Assessment, Learning, and Equity [SCALE] (2016). Educative assessment and meaningful support: 2015 edTPA administrative report. Stanford, CA: Author. https://edtpa.aacte.org/news-area/1839.html
Stanford Center for Assessment, Learning, and Equity [SCALE] (2017). Educative assessment & meaningful support: 2016 edTPA administrative report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=3013&ref=edtpa
Stanford Center for Assessment, Learning, and Equity [SCALE] (2018). Educative assessment & meaningful support: 2017 edTPA administrative report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=3013&ref=edtpa
Taut, S., & Sun, Y. (2014). The development and implementation of a national, standards-based, multi-method teacher performance assessment system in Chile. Education Policy Analysis Archives, 22(71), 1–33. https://doi.org/10.14507/epaa.v22n71.2014
Taut, S., Santelices, M. V., & Stecher, B. (2012). Validation of a national teacher assessment and improvement system. Educational Assessment, 17(4), 163–199. https://doi.org/10.1080/10627197.2012.735913
Taut, S., Valencia, E., Palacios, D., Santelices, M. V., Jiménez, D., & Manzi, J. (2014). Teacher performance and student learning: Linking evidence from two national assessment programmes. Assessment in Education: Principles Policy and Practice. https://doi.org/10.1080/0969594X.2014.961406
Whittaker, A., Pecheone, R., & Stansbury, K. (2018). Fulfilling our educative mission: A response to edTPA critique. Education Policy Analysis Archives, 26(30), 1–20. https://doi.org/10.14507/epaa.26.3720
Wilson, M. (2005). Constructing measures: An item response modeling approach. Lawrence Erlbaum Associates.
Wilson, M. (2018). Making measurement important for education: The crucial role of classroom assessment. Educational Measurement: Issues and Practice Spring, 37(1), 5–20. https://doi.org/10.1111/emip.12188
Acknowledgements
This research was conducted in part thanks to the support of Conicyt through Project Fondecyt 1160871 and Millenium Nucleous “Experience of Students in Chile´s Higher Education: Expectations and Realities.” We thank the research assistance of Alicia Ibáñez.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare no competing interests.
Additional information
Publisher's note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Veronica Santelices, M., Wilson, M. Aligning teacher assessments and teacher learning through a teacher learning progression. Educ Asse Eval Acc 34, 509–532 (2022). https://doi.org/10.1007/s11092-022-09388-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11092-022-09388-w