Skip to main content

Advertisement

Log in

Aligning teacher assessments and teacher learning through a teacher learning progression

  • Published:
Educational Assessment, Evaluation and Accountability Aims and scope Submit manuscript

Abstract

This theoretical piece discusses the concept of a teacher learning progression in an attempt to integrate teacher learning and assessment. From the authors’ perspective, the main features of the teacher learning progression are the longitudinal understanding of teacher knowledge and practice, and the opportunity to align teacher evaluations’ formative and summative purposes. Criteria to assess existing teacher learning progressions are proposed and used to examine examples of teacher assessment systems implemented in different parts of the world. The concept of teacher learning progression has national and international implications for teacher training, for teaching assessment and for the design and implementation of educational policies.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Notes

  1. Cut scores vary by state as do passing rates, and to date state cut scores range from 35 to 41. For details, see Stanford Center for Assessment, Learning, and Equity (SCALE), 2018.

  2. In addition to the Danielson’s framework, teacher assessments have included estimates of teacher value-added based on student performance on standardized tests (Millman, 1997). Value-added measures aim at capturing teacher-specific contribution to the learning process of students taking into consideration variables outside of teachers’ control such as students’ sociodemographic characteristics (Blazar et al. 2016; Close et al. 2020; Harris and McCaffrey 2010). Because of that, value- added measures (teacher effect) are considered fairer than other measures based exclusively on students’ performance level (McCaffrey et al. 2004). The discussion presented in this article, however, will not refer to estimates of teacher effects as they are not linked directly either to a concrete model nor to an operationalization of what teachers should know and be able to do. There is no theory behind the analysis of value-added models in terms of what the standard teacher should achieve and how, nor there are descriptions regarding their progress over time. In that sense, the value-added evaluations provide limited formative information. Furthermore, value-added measures have been criticized because the scales used are often not vertically scaled and narrowly assess basic skill (Darling-Hammond 2012), because of measurement error (Hill 2009), the difficulty to attribute learning gains to multiple teacher present in the classroom or to teachers who are only partially involved (Isenberg and Hock 2011), and the instability of their estimation (Newton et al., 2012). In addition, Rothstein (2010) reports that short-term value-added estimates are weakly related to teachers’ long-run effects. Additionally, the assumption of random assignment of teachers to schools and classrooms and the random assignment of children to schools seldom holds true (Rothstein 2010).

  3. This conceptualization of curriculum content is then seen as the justification for the structure of the typical “standardized test” which is called “domain testing.” In domain testing, the aim is to estimate the proportion of a domain of items related to a concept that a student would get right in an ideal situation at a given point in time—hence, it is fundamentally a cross-sectional perspective. In fact, the student is only given a sample of that domain, and then the true proportion is seen as being estimated by the proportion correct in the sample. In the case of complicated school domains such as “maths” the domain is usually two-level—the coarser level being the overall domain—“maths”—and the finer level being composed of a domain for each of the sub-topics in that particular math curriculum (such as number, geometry, etc.). As the subject changes from year to year (say grade 5 to grade 6), the cross-sectional domain definition changes to map the expected student changes with time. Thus, the different domains for each grade can be seen as a series of cross-sectional perspectives.

  4. They include explaining the material, eliciting student thinking, technology use, recognizing pattern of thinking, generating class conditions for learning, providing feedback and communication with parents and peers. They were developed based on the work of Grossman et al. (2009), and Ball and Forzani (2010).

References

  • Alonzo, A. (2011). Learning progressions that support formative assessment practices. Measurement, 9(2–3), 124–129. https://doi.org/10.1080/15366367.2011.599629

    Article  Google Scholar 

  • Alonzo, A. C. (2012). Learning progressions: Significant promise, significant challenge. Zeitschrift Für Erziehungswissenschaft, 15(1), 95–109.

    Article  Google Scholar 

  • Alonzo, A. C. (2018). An argument for formative assessment with science learning progressions. Applied Measurement in Education, 31(2), 104–112.

    Article  Google Scholar 

  • Alonzo, A., & Elby, A. (2019). Beyond empirical adequacy: Learning progressions as models and their value for teachers. Cognition and Instruction, 37(1), 1–37.

    Article  Google Scholar 

  • Australian Institute for Teaching and School Leadership [AITSL]. (2011). National Professional Standards for Teachers. Melbourne, Australia: AITSL.

  • Avalos, B. (2011). Teacher professional development in Teaching and Teacher Education over ten year. Teaching and Teacher Education, 27 (2011).

  • Avalos-Bevan, B. (2018). Teacher evaluation in Chile: Highlights and complexities in 13 years of experience. Teachers and Teaching, 24(3), 297–311. https://doi.org/10.1080/13540602.2017.1388228

    Article  Google Scholar 

  • Ball, D. L., & Forzani, F. M. (2009). The work of teaching and the challenge for teacher education. Journal of Teacher Education, 60(5), 497–511. https://doi.org/10.1177/0022487109348479

    Article  Google Scholar 

  • Ben-Peretz, M. (2011). Teacher knowledge: What is it? How do we uncover it? What are its implications for schooling? Teaching and Teacher Education, 27(1), 3–9. https://doi.org/10.1016/j.tate.2010.07.015

    Article  Google Scholar 

  • Black, P. (2013). Pedagogy in theory and in practice: Formative and summative assessments in classrooms and in systems. Valuing Assessment in Science Education: Pedagogy, Curriculum, Policy, pages 207–229.

  • Black, P., Wilson, M., & Yao, S. (2011a). Road maps for learning: A guide to the navigation of learning progressions. Measurement: Interdisciplinary Research and Perspective, 9(2–3), 71–123. https://doi.org/10.1080/15366367.2011.591654

    Article  Google Scholar 

  • Black, P., Wilson, M., & Yao, S. (2011b). Comments and thoughts. Measurement: Interdisciplinary Research and Perspective, 9(2–3), 169–172. https://doi.org/10.1080/15366367.2011.604275

    Article  Google Scholar 

  • Blazar, D., Litke, E., & Barmore, J. (2016). What does it mean to be ranked a “high” or “low” value-added teacher? Observing differences in instructional quality across districts. American Educational Research Journal, 53(2), 324–359. https://doi.org/10.3102/0002831216630407

    Article  Google Scholar 

  • Campbell, R. J., Kyriakides, L., Muijs, R. D., & Robinson, W. (2003). Differential teacher effectiveness: Towards a model for research and teacher appraisal. Oxford Review of Education, 29(3), 347–362. https://doi.org/10.1080/03054980307440

    Article  Google Scholar 

  • Clinton, J., Dinham, S., Savage, G., Aston, R., Dabrowski, A., Gullickson, A., Calnin, G., & Arbour, G. (2015). Evaluation of the implementation of the Australian Professional Standards for Teachers. The University of Melbourne, AITSL.

    Google Scholar 

  • Close, K., Amrein-Beardsley, A., & Collins, C. (2020). Putting teacher evaluation systems on the map: An overview of states' teacher evaluation systems post-Every Student Succeeds Act. Education Policy Analysis Archives, 28, Article 58. https://doi.org/10.14507/epaa.28.5252

  • Corcoran, T., Mosher, F.A., & Rogat, A. (2009). Learning progressions in science: An evidence- based approach to reform. CPRE Research Reports. https://doi.org/10.12698/cpre.2009.rr63

  • Danielson, C. (1996). Enhancing professional practice: A framework for teaching. Alexandria, VA: Association for Supervision and Curriculum Development.

  • Danielson, C. (2007). Enhancing professional practice: A framework for teaching (2nd ed.). Association for Supervision and Curriculum Development.

    Google Scholar 

  • Darling-Hammond, L. (2012). Powerful teacher education: Lessons from exemplary programs. JosseyBass.

    Google Scholar 

  • Darling-Hammond, L. (2017). Teacher education around the world: What can we learn from international practice? European Journal of Teacher Education, 40(3), 291–309. https://doi.org/10.1080/02619768.2017.1315399

    Article  Google Scholar 

  • DeLuca, C., LaPointe-McEwan, D., & Luhanga, U. (2016). Teacher assessment literacy: A review of international standards and measures. Educational Assessment, Evaluation and Accountability, 28, 251–272. https://doi.org/10.1007/s11092-015-9233-6

    Article  Google Scholar 

  • DeLuca, C., Chapman-Chin, A., & Klinger, D. A. (2019). Toward a Teacher Professional Learning Continuum in Assessment for Learning. Educational Assessment, 24(4), 267–285. https://doi.org/10.1080/10627197.2019.1670056

    Article  Google Scholar 

  • Duckor, B., & Holmberg, C. (2019). Exploring how to model formative assessment trajectories of posing-pausing-probing practices: Toward a teacher learning progressions framework for the study of novice teachers. Journal of Educational Measurement, 56, 836–890. https://doi.org/10.1111/jedm.12239

    Article  Google Scholar 

  • Furtak, E., Circi, R., & Heredia, S. (2018). Exploring alignment among learning progressions, teacher-designed formative assessment tasks, and student growth: Results of a four-year study. Applied Measurement in Education, 31(2), 143–156. https://doi.org/10.1080/08957347.2017.1408624

    Article  Google Scholar 

  • Furtak, E. (2018). Supporting teachers’ formative assessment practice with learning progressions. Routledge. Taylor & Francis: London and New York.

  • Gitomer, D. H., Martínez, J. F., Battey, D., & Hyland, N. E. (2019). Assessing the assessment: Evidence of reliability and validity in the edTPA. American Educational Research Journal. https://doi.org/10.3102/0002831219890608

    Article  Google Scholar 

  • Grossman, P., Hammerness, K., & McDonald, M. (2009). Redefining teaching, re-imagining teacher education. Teachers and Teaching: Theory and Practice, 15(2), 273–289. https://doi.org/10.1080/13540600902875340

    Article  Google Scholar 

  • Halverson, R., Kelley, C., & Kimball, S. (2004). Implementing teacher evaluation systems: how principals make sense of complex artifacts to shape local instructional practice. In W. K. Hoy, & C. G. Miskel (Eds.), Research and theory in educational administration (Vol. 3. pp. Greenwich, CT: Information Age Press.

  • Harris, D., & McCaffrey, D. F. (2010). Valued-added: Assessing teachers contributions to student achievement. In M. Kennedy (Ed.), Teacher assessment and the quest for teacher quality: A handbook (pp. 251–282). Wiley.

    Google Scholar 

  • Heneman III, H. G., Milanowski, A., Kimball, S. M., & Odden, A. (2006). standards-based teacher evaluation as a foundation for knowledge-and skill-based pay. CPRE Policy Brief, RB-45. http://www.cpre.org/index.php?option=com_content&task=view&id=68&Itemid=73

  • Henning, A. S. (2014). An argument-based validation study of the teacher performance assessment in Washington State, Durham theses, Durham University. Available at Durham E-Theses Online: http://etheses.dur.ac.uk/10728/

  • Heritage, H. (2011). Commentary on road maps for learning: A guide to the navigation of learning progressions. Measurement: Interdisciplinary Research and Perspectives, 9(2–3), 149–151. https://doi.org/10.1080/15366367.2011.599647

    Article  Google Scholar 

  • Hill, H. C. (2009). Evaluating value-added models: A measurement perspective. Journal of Policy Analysis and Management, 28, 702–709.

    Article  Google Scholar 

  • Huber, S. G., & Skedsmo, G. (2016). Teacher evaluation-accountability and improving teaching practices. Educational Assessment Evaluation and Accountability, 28, 105–109. https://doi.org/10.1007/s11092-016-9241-1

    Article  Google Scholar 

  • Isenberg & Hock (2011). Value-added methods in a high-stakes assessment system. Paper presented at the annual meeting of the American Educational Research Association in New Orleans, April 2011.

  • Ingvarson, L., & Rowe, K. (2007, Feb 5). Conceptualising and evaluating teacher quality: Substantive and methodological issues [Paper presentation]. Economics of Teacher Quality conference, Australian National University, Canberra, Australia. http://research.acer.edu.au/learning_processes/8

  • Jacob , B. & Lefgren, L. (2005). Principals as agents: Subjective performance measurement in education. NBER Working Paper, 11463. MA: National Bureau Economic Research.

  • Kane, T. J., & Staiger, D. O. (2012). Ensuring fair and reliable measures of effective teaching: culminating findings from the MET Project’s three-year study. Seattle, WA: Bill & Melinda Gates foundation.

  • Kennedy, M. A. (2016). Parsing the practice of teaching. Journal of Teacher Education, 67(1), 6–17. https://doi.org/10.1177/0022487115614617

    Article  Google Scholar 

  • Kyriakides, L., & Demetriou, D. (2007). Introducing a teacher evaluation system based on teacher effectiveness research: An investigation of stakeholders’ perceptions. Journal of Personnel Evaluation in Education, 20, 43–64. https://doi.org/10.1007/s11092-007-9046-3

    Article  Google Scholar 

  • Martínez, F., Taut, S., & Schaaf, K. (2016). Classroom observation for evaluating and improving teaching: An international perspective. Studies of Educational Evaluation, 49, 15–29.

    Article  Google Scholar 

  • Masters, G., & Forster, M. (1996). Progress maps. Assessment resource kit. Commonwealth of Australia.

    Google Scholar 

  • McCaffrey, D. F., Lockwood, J. R., Koretz, D. M., Louis, T. A., & Hamilton, L. S. (2004). Models for value-added modeling of teacher effects. Journal of Educational and Behavioral Statistics, 29, 67–101.

    Article  Google Scholar 

  • Millman, J. (Ed.). (1997). Grading teachers, grading schools: Is student achievement a valid evaluation measure? Corwin Press.

    Google Scholar 

  • National Assessment Governing Board [NAGB]. (2005). Science NAEP 2009: Science framework for the 2009 National Assessment of Educational Progress. WestEd and the Council of Chief State School Officers.

    Google Scholar 

  • National Research Council [NRC]. (2001). Knowing what students know: The science and design of educational assessment. Committee on the Foundations of Assessment. J. Pellegrino, N. Chudowsky, & R. Glaser (Eds). Washington, DC: National Academy Press.

  • National Research Council [NRC]. (2006). Systems for state science assessment. Committee on test design for K-12 science achievement. M. Wilson & M. Bertenthal (Eds.), Board on Testing and Assessment, Center for Education. Division of Behavioral and Social Sciences and Education. Washington, D.C.: National Academy Press.

  • National Research Council [NRC]. (2007). Taking science to school: Learning and teaching science in grades K-8. Committee on Science Learning, Kindergarten through Eighth Grade. R. A. Duschl, H. A. Schweingruber, & A. W. Shouse (Eds.). Washington, D.C.: National Academy Press.

  • Newton, X., Darling-Hammond, L., Haertel, E. & Thomas, E. (2010). Valued-added modeling of teacher effectiveness: An exploration of stability across models and contexts. Education Policy Analysis Archives, 18(23). Retrieved from http://epaa.asu.edu/ojs/article/view/810.

  • Park, S., & Suh, J. K. (2015). From portraying toward assessing PCK. In A. Berry, P. Friedrichsen, & J. Loughran (Eds.), Re-examining Pedagogical Content Knowledge in Science Education (pp. 104–119). Routledge.

    Google Scholar 

  • Pianta, R. C., Hamre, B. K., & Allen, J. P. (2012). Teacher-student relationships and engagement: Conceptualizing, measuring, and improving the capacity of classroom interactions. In S. L. Christenson, A. L. Reschly, & C. Wylie (Eds.), Handbook of research on student engagement (pp. 365–386). Boston, MA: Springer. https://doi.org/10.1007/978-1-4614-2018-7_17

  • Reagan, E. M., Schram, T., McCurdy, K., Chang, T., & Evans, C. M. (2016). Politics of policy: Assessing the implementation, impact, and evolution of the Performance Assessment for California Teachers (PACT) and edTPA. Education Policy Analysis Archives, 24(9), 1–27. https://doi.org/10.14507/epaa.24.2176

    Article  Google Scholar 

  • Roegman, R., Goodwin, A. L., Reed, R., & Scott-McLaughlin, R. M., II. (2016). Unpacking the data: An analysis of the use of Danielson’s (2007) Framework for Professional Practice in a teaching residency program. Educational Assessment Evaluation and Accountability, 28(2), 111–137. https://doi.org/10.1007/s11092-015-9228-3

    Article  Google Scholar 

  • Rothstein, J. (2010). Teacher quality in educational production: Tracking, decay, and student achievement. Quarterly Journal of Economics, 125(1), 175–214. https://doi.org/10.1162/qjec.2010.125.1.175

    Article  Google Scholar 

  • Sandilos, L. E., Sims, W., Norwalk, K. E. & Reddy, L. (2019). Converging on quality: Examining multiple measures of teaching effectiveness, Journal of School Psychology 74 (2019), 10–28.Author

  • Schneider, R. M., & Plasman, K. (2011). Science teacher learning progressions: A review of science teachers’ pedagogical content knowledge development. Review of Educational Research, 81(4), 530–565. https://doi.org/10.3102/0034654311423382

    Article  Google Scholar 

  • Shavelson, R. J, Moss, P., Wilson, M., Duckor, B., Baron, W., & Wilmot, D. (2010, May). The promise of teacher learning progressions: Challenges and opportunities for articulating growth in the profession [Paper presentation]. Teacher Learning Progressions symposium for Division D-Measurement and Research Methodology, American Education Research Association, Denver, CO, United States.

  • Shepard, L. A. (2018). Learning progressions as tools for assessment and learning. Applied Measurement in Education, 31(2), 165–174. https://doi.org/10.1080/08957347.2017.1408628

    Article  Google Scholar 

  • Shulman, L. S. (1986). Paradigms and research programs in the study of teaching: A contemporary perspective. In M. C. Wittrock (Ed.), Handbook of research on teaching (3rd ed., pp. 3–36). Macmillan.

    Google Scholar 

  • Shulman, L. S. (1986). Those who understand: Knowledge growth in teaching. Educational Researcher, 15(2), 4–14. https://doi.org/10.3102/0013189X015002004

    Article  Google Scholar 

  • Stanford Center for Assessment, Learning, and Equity [SCALE] (2013). 2013 edTPA field test: Summary report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=827&ref=edtpa

  • Stanford Center for Assessment, Learning, and Equity [SCALE] (2015). Educative assessment & meaningful support: 2014 edTPA administrative report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=2183&ref=edtpa

  • Stanford Center for Assessment, Learning, and Equity [SCALE] (2016). Educative assessment and meaningful support: 2015 edTPA administrative report. Stanford, CA: Author. https://edtpa.aacte.org/news-area/1839.html

  • Stanford Center for Assessment, Learning, and Equity [SCALE] (2017). Educative assessment & meaningful support: 2016 edTPA administrative report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=3013&ref=edtpa

  • Stanford Center for Assessment, Learning, and Equity [SCALE] (2018). Educative assessment & meaningful support: 2017 edTPA administrative report. Stanford, CA: Author. https://secure.aacte.org/apps/rl/res_get.php?fid=3013&ref=edtpa

  • Taut, S., & Sun, Y. (2014). The development and implementation of a national, standards-based, multi-method teacher performance assessment system in Chile. Education Policy Analysis Archives, 22(71), 1–33. https://doi.org/10.14507/epaa.v22n71.2014

    Article  Google Scholar 

  • Taut, S., Santelices, M. V., & Stecher, B. (2012). Validation of a national teacher assessment and improvement system. Educational Assessment, 17(4), 163–199. https://doi.org/10.1080/10627197.2012.735913

    Article  Google Scholar 

  • Taut, S., Valencia, E., Palacios, D., Santelices, M. V., Jiménez, D., & Manzi, J. (2014). Teacher performance and student learning: Linking evidence from two national assessment programmes. Assessment in Education: Principles Policy and Practice. https://doi.org/10.1080/0969594X.2014.961406

    Article  Google Scholar 

  • Whittaker, A., Pecheone, R., & Stansbury, K. (2018). Fulfilling our educative mission: A response to edTPA critique. Education Policy Analysis Archives, 26(30), 1–20. https://doi.org/10.14507/epaa.26.3720

    Article  Google Scholar 

  • Wilson, M. (2005). Constructing measures: An item response modeling approach. Lawrence Erlbaum Associates.

    Google Scholar 

  • Wilson, M. (2018). Making measurement important for education: The crucial role of classroom assessment. Educational Measurement: Issues and Practice Spring, 37(1), 5–20. https://doi.org/10.1111/emip.12188

    Article  Google Scholar 

Download references

Acknowledgements

This research was conducted in part thanks to the support of Conicyt through Project Fondecyt 1160871 and Millenium Nucleous “Experience of Students in Chile´s Higher Education: Expectations and Realities.” We thank the research assistance of Alicia Ibáñez.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Maria Veronica Santelices.

Ethics declarations

Conflict of interest

The authors declare no competing interests.

Additional information

Publisher's note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Veronica Santelices, M., Wilson, M. Aligning teacher assessments and teacher learning through a teacher learning progression. Educ Asse Eval Acc 34, 509–532 (2022). https://doi.org/10.1007/s11092-022-09388-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11092-022-09388-w

Keywords

Navigation