Skip to main content
Log in

Interdisciplinary Science Assessment of Carbon Cycling: Construct Validity Evidence Based on Internal Structure

  • Published:
Research in Science Education Aims and scope Submit manuscript

Abstract

Growing interest in interdisciplinary (ID) understanding has led to the recent development of four ID assessments, none of which have previously been comprehensively validated. Sources of evidence for the validity of tests include construct validity, such as the internal structure of the test. ID tests may (and should) test both disciplinary (D) and ID understanding, and the internal structure can be examined to determine whether the theoretical relationship among D and ID knowledge is supported by empirical evidence. In this paper, we present an analysis of internal structure of ISACC, an ID assessment of carbon cycling, selected because the developers included both disciplinary and ID items and posited a relationship among these two types of knowledge. Responses from 454 high school and college students were analyzed using confirmatory factor analyses (CFA) with the Mplus software for internal structure focusing on dimensionality, as well as functioning by gender and race/ethnicity, and reliability. CFA confirmed that the underlying structure of the ISACC best matches a two-factor path model, supporting the developers’ theoretical hypothesis of the impact of disciplinary understanding on interdisciplinary understanding. No gender effect was found in the factor structure of the ISACC, indicating that females and males have similar performance. Race/ethnicity performance was similar to other science assessments, revealing possible ethnicity bias on the part of the instrument. The reliability coefficients for the two-factor path model were found to be sufficient. This study highlights the importance of the internal structure of test instruments as a source of validity evidence and models a procedure to assess internal structure.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  • Abraham, J. (2004). Multidisciplinary explorations: Bridging the gap between engineering and biology. Journal of College Science Teaching, 33(5), 27–31.

    Google Scholar 

  • American Educational Research Association (AERA), American Psychological Association (APA), & National Council on Measurement in Education (NCME). (2014). Standards for educational and psychological testing. Washington, DC: AERA.

    Google Scholar 

  • Australian Curriculum and Assessment Reporting Authority [ACARA] (2016). ACARA STEM connections project report. Retrieved from https://www.australiancurriculum.edu.au/media/3220/stem-connections-report.pdf.

  • Boix Mansilla, V., & Duraisingh, E. D. (2007). Targeted assessment of students' interdisciplinary work: An empirically grounded framework proposed. The Journal of Higher Education, 78(2), 215–237.

    Google Scholar 

  • Boone, W. J., Staver, J. R., & Yale, M. S. (2014). Rasch analysis in the human sciences. Dordrecht: Springer Netherlands.

    Google Scholar 

  • Boone, W. J. (2016). Rasch analysis for instrument development: why, when, and how?. CBE—Life Sciences Education, 15(4), rm4.

  • Chen, F. F., West, S. G., & Sousa, K. H. (2006). A comparison of bifactor and second-order models of quality of life. Multivariate Behavioral Research, 41(2), 189–225.

    Google Scholar 

  • Chi, M. T. H., & Ceci, S. J. (1987). Content knowledge: Its role, representation, and restructuring in memory development. In H. W. Reese (Ed.), Advances in child development and behavior (Vol. 20, pp. 91–142). New York: Academic Press.

    Google Scholar 

  • Clark, L. A., & Watson, D. (1995). Constructing validity: Basic issues in objective scale development. Psychological Assessment, 7(3), 309–319.

    Google Scholar 

  • Cronbach, L. J. (1951). Coefficient alpha and the internal structure of tests. Psychometrika, 16(3), 297–334.

    Google Scholar 

  • Dimitrov, D. M. (2010). Testing for factorial invariance in the context of construct validation. Measurement and Evaluation in Counseling and Development, 43(2), 121–149.

    Google Scholar 

  • DiStefano, C., Liu, J., Jiang, N., & Shi, D. (2018). Examination of the weighted root mean square residual: Evidence for trustworthiness? Structural Equation Modeling, 25(3), 453–466.

  • Doerschuk, P., Bahrim, C., Daniel, J., Kruger, J., Mann, J., & Martin, C. (2016). Closing the gaps and filling the STEM pipeline: A multidisciplinary approach. Journal of Science Education and Technology, 25(4), 682–695.

    Google Scholar 

  • Education Scotland. (2012). CfE Briefing Interdisciplinary Learning. Retrieved from https://education.gov.scot/Documents/cfe-briefing-4.pdf.

  • Erduran, S., Guilfoyle, L., Park, W., Chan, J., & Fancourt, N. (2019). Argumentation and interdisciplinarity: Reflections from the Oxford argumentation in religion and science project. Disciplinary and Interdisciplinary Science Education Research, 1(8), 1–10.

    Google Scholar 

  • Garner, J. K., Kaplan, A., Hathcock, S., & Bergey, B. (2020). Concept mapping as a mechanism for assessing science teachers’ cross-disciplinary field-based learning. Journal of Science Teacher Education, 31(1), 8–33.

  • Gipps, C. V., & Murphy, P. (1994). A fair test? Assessment, achievement and equity. Open University Press.

  • Hirschfeld, G., & Von Brachel, R. (2014). Improving Multiple-Group confirmatory factor analysis in R–A tutorial in measurement invariance with continuous and ordinal indicators. Practical Assessment, Research, and Evaluation, 19 (1), 7.

  • Holbrook, J. B. (2013). What is interdisciplinary communication? Reflections on the very idea of disciplinary integration. Synthese, 190, 1865–1879.

    Google Scholar 

  • Hu, L. T., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling, 6(1), 1–55.

    Google Scholar 

  • Hurd, P. D. (1991). Why we must transform science education. Educational Leadership, 49(2), 33–35.

    Google Scholar 

  • Kane, M. T. (2006). Validation. In R. L. Brennan (Ed.), Educational measurement (4th ed., pp. 17–63). Westport: Praeger Publishers.

    Google Scholar 

  • Khisty, L. L. (1995). Making inequality: Issues of language and meanings in mathematics teaching with Hispanic students. In W. G. Secada, E. Fennema, & L. B. Adajian (Eds.), New directions for equity in mathematics education (pp. 279–297). Cambridge: Cambridge University Press.

    Google Scholar 

  • Kline, R. (2015). Principles and practice of structural equation modeling (4th ed.). New York: Guilford Publications.

    Google Scholar 

  • Kochelmans, J. (1979). Why interdisciplinarity? Interdisciplinarity and higher education. University Park: Pennsylvania State University Press.

    Google Scholar 

  • Lattuca, L. R., Voigt, L. J., & Fath, K. Q. (2004). Does interdisciplinarity promote learning? Theoretical support and researchable questions. The Review of Higher Education, 28(1), 23–48.

    Google Scholar 

  • LoGerfo, L., Nichols, A., & Reardon, S. F. (2006). Achievement gains in elementary and high school. Washington, DC: Urban Institute.

    Google Scholar 

  • Lord, F. M., & Novick, M. R. (1968). Statistical theory of mental test scores. MA: Addison-Wesley.

    Google Scholar 

  • Marasco, E. A., & Behjat, L. (2013). Developing a cross-disciplinary curriculum for the integration of engineering and design in elementary education. In Proc. of the 2013 American society of engineering education annual conference and exposition, Atlanta, GA.

  • Marshall, J., Banner, J., & You, H. (2018). Assessing the effectiveness of sustainability learning. Journal of College Science Teaching, 47(3), 57-67.

  • Messick, S. (1989). Validity. In R. L. Lin (Ed.), Educational measurement. New York: American Council on Education, Macmillan Publishing.

    Google Scholar 

  • Ministry of Education. (2015). The 2015 revised science curriculum. Report no. 2015–74. Sejong: Ministry of Education.

    Google Scholar 

  • Moss, P. A. (1992). Shifting conceptions of validity in educational measurement: Implications for performance assessment. Review of Educational Research, 62(3), 229–258.

    Google Scholar 

  • Musu-Gillette, L., de Brey, C., McFarland, J., Hussar, W., Sonnenberg, W., & Wilkinson-Flicker, S. (2017). Status and Trends in the Education of Racial and Ethnic Groups 2017 (NCES 2017-051). Washington, DC; U.S. Department of Education, National Center for education statistics. Retrieved [Sep 16] from http://nces.ed.gov/pubsearch.

  • Muthen, L. K., & Muthen, B. O. (1998-2018). Mplus User’s Guide (8th ed.). Los Angeles: Muthén & Muthén.

    Google Scholar 

  • National Research Council. (2001). Knowing what students know: The science and design of educational assessment. In J. Pellegrino, N. Chudowsky, & R. Glaser (Eds.), Board on testing and assessment, Center for Education, division of behavioral and social science and education. Washington, DC: National Academy Press.

    Google Scholar 

  • National Research Council. (2012). A Framework for K-12 Science Education: Practices, Crosscutting Concepts, and Core Ideas. Committee on a conceptual framework for new K-12 science education standards. Board on science education, division of behavioral and social sciences and education. Washington, DC: The National Academies Press.

  • Neurath, O. (1996). Unified science as encyclopedic integration. In S. Sarkar (Ed.), Logical empiricism at its peak: Schlick, Carnap, and Neurath (pp. 309–335). Boston: Harvard University.

    Google Scholar 

  • NGSS Lead States. (2013). Next generation science standards: For states, by states. Washington, DC: National Academies Press.

    Google Scholar 

  • Nowacek, R. S. (2005). A discourse-based theory of interdisciplinary connections. The Journal of General Education, 54(3), 171–195.

    Google Scholar 

  • Raykov, T. (2001). Bias of coefficient a for fixed congeneric measures with correlated errors. Applied Psychological Measurement, 25(1), 69–76.

  • Raykov, T. (2009). Evaluation of scale reliability for unidimensional measures using latent variable modeling. Measurement and Evaluation in Counseling and Development, 42(3), 223–232.

    Google Scholar 

  • Reiska, P., Soika, K., & Cañas, A. J. (2018). Using concept mapping to measure changes in interdisciplinary learning during high school. Knowledge Management & E-Learning: An International Journal (KM&EL), 10(1), 1–24.

    Google Scholar 

  • Rennie, L., Wallace, J., & Venville, G. (2012). Integrating science, technology, engineering, and mathematics: Issues, reflections, and ways forward. In L. Rennie, G. Venville, & J. Wallace (Eds.), Exploring curriculum integration: Why integrate? (pp. 1–11). Routledge.

  • Schaal, S., Bogner, F. X., & Girwidz, R. (2010). Concept mapping assessment of media assisted learning in interdisciplinary science education. Research in Science Education, 40(3), 339–352.

    Google Scholar 

  • Schumacker, R. E., & Lomax, R. G. (2010). A beginners guide to structural equation modeling. New York: Routledge

  • Sedlacek, W. E. (1994). Issues in advancing diversity through assessment. Journal of Counseling & Development, 72(5), 549–553.

    Google Scholar 

  • Shen, J., Liu, O. L., & Sung, S. (2014). Designing interdisciplinary assessments in sciences for college students: An example on osmosis. International Journal of Science Education, 36(11), 1773–1793. https://doi.org/10.1080/09500693.2013.879224.

    Article  Google Scholar 

  • Sireci, S. G. (2007). On validity theory and test validation. Educational Researcher, 36(8), 477–481.

    Google Scholar 

  • Smith, E. V. (2001). Evidence for the reliability of measures and validity of measure interpretation: A Rasch measurement perspective. Journal of Applied Measurement, 2(3), 281–311.

    Google Scholar 

  • Stock, P., & Burton, R. J. (2011). Defining terms for integrated (multi-inter-transdisciplinary) sustainability research. Sustainability, 3(8), 1090–1113.

    Google Scholar 

  • Tate, W. F. (1994). Race retrenchment and reform of school mathematics. Phi Delta Kappan, 75(6), 477–480.

    Google Scholar 

  • Willingham, J. C., Pair, J. D., & Parrish, J. C. (2016). A framework for cross-disciplinary engineering projects. Science Scope, 39(7), 68.

    Google Scholar 

  • Wilson, M. (2005). Constructing measures: An item response modeling approach. Lawrence Erlbaum Associates Publishers.

  • Wolfe, C. R., & Haynes, C. (2003). Assessing interdisciplinary writing. Peer Review, 6(1), 126–169.

    Google Scholar 

  • Wright, B. D.,& Stone, M. H. (1979). Best test design: Rasch measurement. Chicago, IL: Mesa Press.

  • You, H. S. (2017). Why teach science with an interdisciplinary approach: History, trends, and conceptual frameworks. Journal of Education and Learning, 6(4), 66–77.

    Google Scholar 

  • You, H. S., Marshall, J. A., & Delgado, C. (2018). Assessing students' disciplinary and interdisciplinary understanding of global carbon cycling. Journal of Research in Science Teaching, 55(3), 377-398.

  • You, H. S., Marshall, J. A., & Delgado, C. (2019). Toward Interdisciplinary Learning: Development and Validation of an Assessment for Interdisciplinary Understanding of Global Carbon Cycling. Research in Science Education, 1-25.

  • Yu, C. Y. (2002). Evaluating cutoff criteria of model fit indices for latent variable models with binary and continuous outcomes (Unpublished doctoral dissertation). University of California, Los Angeles, CA.

  • Zimmerman, D. W. (1972). Test reliability and the Kuder-Richardson formulas: Derivation from probability theory. Educational and Psychological Measurement, 32(4), 939–954.

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Hye Sun You.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

You, H.S., Park, S., Marshall, J.A. et al. Interdisciplinary Science Assessment of Carbon Cycling: Construct Validity Evidence Based on Internal Structure. Res Sci Educ 52, 473–492 (2022). https://doi.org/10.1007/s11165-020-09943-9

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11165-020-09943-9

Keywords

Navigation