Skip to main content

Measuring Quality Technology Integration in Science Classrooms

Abstract

Researchers, evaluators, and practitioners need tools to distinguish between applications of technology that support, enhance, and transform classroom instruction and those that are ineffective or even deleterious. Here we present a new classroom observation protocol designed specifically to capture the quality of technology use to support science inquiry in high school science classrooms. We iteratively developed and piloted the Technology Observation Protocol for Science (TOP-Science), building directly on our published framework for quality technology integration. Development included determining content and face validity, as well as the calculation of reliability estimates across a broad range of high school science classrooms. The resulting protocol focuses on the integration of technology in classrooms on three dimensions: science and engineering practices, student-centered teaching, and contextualization. It uses both quantitative coding and written descriptions of evidence for each code; both are synthesized in a multi-dimensional measure of quality. Data are collected and evaluated from the perspective of the teacher’s intentions, actions, and reflections on these actions. This protocol fills an important gap in understanding technology’s role in teaching and learning by doing more than monitoring technology’s presence or absence, and considering its integration in the context of the Next Generation Science Standards science and engineering practices. Its applications include research, evaluation, and potentially peer evaluation.

Introduction

Today’s science teachers are challenged with immersing their students in the practices of science inquiry (DiBiase and McDonald 2015; Herrington et al. 2016). This includes having students ask testable questions, design related studies, collect and interpret data, engage in argumentation from evidence, connect findings to other research, and more. The importance of these practices is readily apparent in the Next Generation Science Standards (NGSS), which address eight science and engineering practices (SEPs) in the context of disciplinary core ideas and crosscutting concepts (National Research Council 2012).

At the same time, digital technology has become a commonplace learning and organizational tool in many schools. For example, according to a 2017 report on connectivity in education, over 40 governors invested resources in improving technology infrastructure in their states, resulting in over 30 million students in 70,000 schools gaining broadband access between 2013 and 2016 (Education Super Highway 2017). Indeed, technology applications have the potential to support and extend student application and learning of science practices. This support or extension can occur with the use of everyday applications, such as Internet searches and word processing, as well as with applications designed specifically for learning environments, such as classroom management systems (Herrington and Kervin 2007). It can even take advantage of the same applications used by science professionals, such as delineating watershed boundaries, mapping land use, and measuring water quality with GIS to address questions about factors impacting stream health (e.g., Stylinski & Doty 2013). These tools have the potential to support, enhance, and even transform the learning experience if used effectively. However, they can also be used in ineffective ways that confuse, distract, and interfere with learning.

Researchers and evaluators need instruments that will enable them to describe the effective integration of technology into science classrooms to understand this changing classroom landscape and better support teachers. Observation protocols provide researchers and evaluators with critical tools for examining pedagogical approaches and curricular resources, including technology hardware and software, and their relationship to student learning (e.g., Bielefeldt 2012). A recent review identified 35 observation protocols aimed at describing teachers’ instructional practice for science classrooms; however, less than a third accounted for technology use (Stylinski et al. 2018). Of these, most only documented the presence of technology. Fewer than half of those that included technology described the quality of technology use by looking at the integration of technology with pedagogical practices. None comprehensively captured the quality of technology use in the specific context of instruction focused on science disciplinary practices.

In this paper, we present the Technology Observation Protocol for Science (TOP-Science), a new classroom observation protocol designed specifically to capture the quality of technology use, to support science inquiry in high school science classrooms. We describe how we determined face validity and construct validity through (a) a careful alignment of the Protocol with our published framework of technology integration in science classrooms and (b) iterative pilot testing that included expert review and inter-rater reliability.

Theoretical Foundation of the Protocol

The Protocol builds on a published theoretical framework that consists of four dimensions: technology type, science and engineering practices (SEPs), student-centered teaching, and relevance to students’ lives (Parker, Stylinski, Bonney, Schillaci & McAuliffe 2015). These dimensions are critical to high-quality technology integration, and their inclusion in our observation protocol supports the instrument’s content validity. In this section, we describe each of these dimensions and their connection to the Protocol presented in this paper.

Technology Type

Digital technology in classrooms can be defined as computer hardware, software, data, and media that use processors to store or transmit information in digital format (Pullen et al. 2009). In our framework, we expanded on this definition by describing three types based on the use of technology. Specifically, some technology applications are primarily used in instructional settings, such as assessment tools (e.g., clicker software) and online course management (e.g., PowerSchool, Google Classroom). Other technology applications are most commonly used in STEM workplace settings (e.g., computer modeling, image data analysis). Finally, ubiquitous technology applications are used in many settings, including in the classroom, workplace, and elsewhere in daily life (e.g., word processing, Internet search engines). The designation of a technology application to the instructional, ubiquitous, or STEM workplace category may change over time as technologies evolve and are adapted to new settings. For example, Microsoft Excel was originally developed for businesses, universities, and other workplace environments; but it now appears regularly in many different settings such as K-12 schools and homes. In the Protocol, we treated technology type differently from the other three dimensions. As described later, we consider the integration of technology with each of the other three dimensions, and we simply had observers record technology type as a way to capture the diversity of technology applications in the classroom setting.

Authentic Science and Engineering Practices

Real-world technology applications offer the potential to help teachers align with Next Generation Science Standards (NGSS), which weave eight science and engineering practices with cross-cutting concepts and core ideas (NRC, 2012). These eight practices provide a thoroughly vetted description of science practices relevant for the classroom, and thus, they form one of three central dimensions of our observation protocol, which, as noted, are considered in the light of technology integration. Specifically, users consider the integration of technology with this description of authentic science inquiry.

Technology, regardless of type, can scaffold students’ understanding and learning during science-inquiry-based activities (e.g., Devolder et al. 2012; Rutten et al. 2012). Classroom-based technology applications, both those designed specifically for instruction and those drawing from science workplace settings, can help teachers engage students in real-world science and engineering pursuits—deepening their understandings of science inquiry. This linkage between science and technology expands options for classroom science inquiry such as allowing students to observe phenomena otherwise not possible in a classroom setting (D’Angelo et al., 2014; Smetana and Bell 2011). For example, computer simulations, such as those offered by SimQuest, scaffold students’ understanding by asking them to conduct multiple trials of an experiment to explore scientific principles and the effects of changing certain variables on outcomes of interest (De Jong 2006).

Student-Centered Teaching

Student-centered teaching involves instructional practices that allow students to have a more self-directed experience with their learning. The Nellie Mae Education Foundation (NMEF) defines student-centered teaching as encompassing personalized instruction, student autonomy or ownership of the learning process, and competency-based instruction (NMEF and Parthenon-EY 2015; Reif et al. 2016).Footnote 1 Our Protocol breaks student-centered teaching into three aspects closely aligned with the Nellie Mae definition: personalized instruction, autonomy, and competency-based teaching. Personalized instruction occurs when students’ individual skills, interests, and developmental needs are taken into account for instruction. Students have autonomy over their own learning when they have frequent opportunities to take ownership of their learning process and have a degree of choice. Competency-based teaching is defined as teaching that progresses based on students’ mastery of skills and knowledge, rather than being dictated by how much time has been spent on an activity or topic (NMEF and Parthenon-EY 2015).

Many teachers have capitalized on the increased availability of technology by using digital devices and technology resources to encourage student-centered learning through increased personalization, individual pacing, and student ownership (Reif et al. 2016). Maddux and Johnson (2006) note that technology implementation strategies are most successful when they are more student-centered and move away from teacher-directed instruction. The use of technology in student-centered instruction can allow students to learn in ways that would not otherwise be possible without the technology. For example, online modules may facilitate personalized learning by allowing students to learn at their own pace, using various tools and resources to support learning and providing an appropriate level of scaffolding (Patrick et al. 2013). In addition, a learning management system such as Blackboard or Moodle might facilitate competency-based learning by helping to document students’ progress toward demonstrating competencies, which might otherwise be too burdensome without technology (Sturgis and Patrick 2010). Technology might also support student-centered teaching such that students may have autonomy over what kind of technology to choose for an activity, allowing technology to be a “potential enabler” of student-centered learning (NMEF and Parthenon-EY 2015, p. 6).

Contextualizing Learning for Relevance

The use of technology in classrooms enables teachers to make learning more relevant to the world in which students live (Brophy et al. 2008; Hayden et al. 2011; Miller et al. 2011). Since the 1990s, there has been an emphasis on trying to make science more relevant to students (Fensham 2004). The proponents of making science content more relevant to students are generally those who argue that such instructional practices lead to student-centered learning and increase student motivation for learning science (Moeller and Reitzes 2011; Ravitz et al. 2000; Stearns et al. 2012).

One of the more common ways to increase the relevance of science to students includes making connections between the content and the real world in which students live. This idea of increasing relevance is also sometimes referred to as authentic learning (Lombardi 2007; Reeves et al. 2002). Grounding lessons in the local geographic context, making connections to youth culture, or matching classroom tasks and activities with those of professionals in practice are all ways in which learning can be made more relevant or authentic to students. Technology can support and facilitate this type of authentic learning via such tools as web resources and search engines, online discussions or emails, online simulated environments, digital photography, or voice recorders (Herrington and Kervin 2007). Technology can facilitate authentic learning by using animations, sounds, and images, allowing students to engage more easily with the material on a conceptual level. For example, high school students in Australia used web resources in conjunction with a field trip to the Sydney Olympic Park, to research a problem the park was having with mosquitoes and rats attracted by food scraps, standing water, and trash left behind by visitors. The “geography challenge” was presented to students using animated scenarios involving the mosquitoes and rats affecting visitors’ experiences at the park, and invited them, as “geography consultants,” to investigate and suggest recommendations on ways to mitigate the problems and “restore an ecological balance to the area” (Brickell and Herrington 2006, p. 536). Students conducted web research before visiting the site, and presented a report with their findings to “park authorities.”

Summary

While the three central dimensions of the Protocol—SEPs, student-centered teaching, and contextualization—provide a useful frame for describing the integration of technology with key pedagogical practices, they are not mutually exclusive. The SEPs encourage making science learning authentic; that is, contextualizing learning, through reflecting the real work of scientists and including questions of interest to the field of science. The SEPs also encourage autonomy, a key facet of student-centered teaching, particularly through opportunities to define researchable questions.

Development of the Protocol

The development and testing process of the TOP-Science Protocol reflects that undertaken by similar observation tools (e.g., Dringenberg et al. 2012; Sawada et al., 2002). As noted, we built the initial draft of the Protocol on the theoretical framework previously described and after review of similar observation protocols. We refined this first draft through preliminary piloting and based on feedback from several experts in instrument development. We then conducted four rounds of iterative pilot testing of the Protocol (Table 1), each followed by feedback-driven revisions as well as additional expert review.

Table 1 Protocol pilot testing

In the first two rounds, individual observers completed feedback sheets that were then analyzed by the team and used to revise the Protocol. In the third and fourth rounds, observer pairs completed observations, compared results, and provided feedback based on the comparisons. Our research team, along with additional trained observers, conducted 66 observations of diverse science classes in 26 high schools across seven states. The seven states included in our study represent a broad range of areas from suburban, urban, and rural locations, as well as from the west coast, New England, and Mid-Atlantic regions of the country. The classrooms observed ranged from ninth to twelfth grades (including mixed grades). They occurred in a diversity of school settings including low-resourced and high-resourced schools, public schools, and charter schools, and included some that self-identify as STEM schools.

The final Protocol consists of four parts (Appendix A): (1) pre-observation teacher questions to understand teachers’ intention for technology integration in their class (Fig. A1), (2) observation sheets where codes and field notes are recorded for the four framework dimensions (Fig. A2) in 10-min intervals, (3) post-observation teacher questions to gather the teacher’s reflection on their lesson (Fig A3), and (4) a code summary sheet that produces a multi-dimensional measure of technology integration quality (Fig. A4).

A detailed Reference Guide walks the observer through each step in the Protocol process, and provides all instructions needed to implement the Protocol. For each of the three categories, there is a detailed description of what observers should look for under each, with descriptors and examples for each category and technology level. Table 2 provides sample technology codes and evidence text for each category, demonstrating different levels of technology integration for similar lessons.

Table 2 An example from the Reference Guide that illustrates three levels of technology integration for the three central dimensions of the protocol

The pre- and post-observation teacher questions offer insight into teacher technology integration goals and reflections on implementation. The observation sheet provides a quantitative metric and qualitative description of technology integration focused on teacher practices. Observations occur in 10-min intervals; within each interval, observers complete codes for each dimension and include field notes providing evidence for each of the codes. The code summary sheet includes a template to synthesize results across the 10-min intervals, also incorporating teacher responses to the pre- and post-questions. Four key features emerged from the protocol development process. Each is described below.

Multi-Level Coding System Coupled with Descriptive Text

To address the challenging process of developing meaningful and user-friendly codes that could describe the presence of each of our framework dimensions, including the extent to which technology was integrated, we tested a range of options, beginning with the most basic check for presence or absence of the category and technology. As we had already found in our literature review, however, the presence of technology is not the same as quality integration of that technology. To address this, the next iteration applied the terms absent/low/medium/high for each category. Ultimately, across the four rounds of pilot testing, we found that observers found it easiest to describe each dimension as either not present, incidentally present, or embedded in the lesson, and the level of technology integration in each of the dimensions as minimally integrated, partially integrated, or fully integrated.

Our initial draft of the Protocol utilized tallies to indicate the frequency of observed technology use and relevant instructional practices. After the first round of pilot testing, we realized that tallies of codes did not paint a rich enough picture of the quality of technology integration. Therefore, we introduced field notes to complement these tallies (as written text alone would present too onerous a task for our 10-min observation intervals), and piloted this structure in subsequent rounds. Specifically, observers supply written text describing evidence for each chosen code. Later, they combine these data to describe the quality of technology integration in the classroom (see “Synthesis of Data into a Quality Measure” for more information). Thus, the Protocol does not limit observers to a checklist or set of codes, but instead couples quantitative and qualitative data. This pairing provides the necessary flexibility to capture multiple and diverse ways that the same technology can be used for different purposes. For example, after selecting the appropriate technology-practice integration that involves GIS (a workplace technology application), the observer could describe evidence of how the teacher used GIS to help students brainstorm environmental science research questions, collect geospatially based data, present geospatially explicit findings, develop maps to share with community members, and more.

Integration of Technology with the Three Dimensions

In our initial attempts to operationalize the four dimensions from our Framework into an observation protocol, we recorded the technology use category separately from the three other categories (SEPs, student-centered teaching, and contextualization). We quickly realized, however, that in order to truly assess the quality of technology use, we had to more clearly define technology integration within the three other dimensions. Coding technology use separately from the dimensions and then combining the codes post hoc was artificial and did not accurately reflect a true measure of integration. To address this challenge, as described earlier, we replaced it with a system of coding the degree of technology integration with each of the components, not just technology use. For example, in one pilot classroom, students studied chemical reactions by choosing a reaction of interest and designing a demonstration that would illustrate this reaction during an open house event. With Chromebook laptop computers, they researched chemical reactions and potential risks and identified required supplies. Some students also used Google Docs to record this information and plan their demonstration in collaboration with other students. With our coding system, we were able to document the extent to which the activity was student-centered (i.e., embedded student centeredness, due to the level of autonomy students had over their proposal and research methods), and to what extent technology use was integrated in those student-centered practices (i.e., partially integrated, because students had options about the kind of technology to use in the activity, and the technology enhanced the student-centeredness of their demo designs). For this particular example, it would not be considered fully integrated because, while the use of technology enhanced the student-centeredness of the activity, the level of student centeredness was not necessarily contingent on the use of technology. Students would still have been able to choose reactions to study, design their demonstration, and collaborate with other students without technology.

Focus on Teacher Intentions, Actions, and Reflections

The Protocol explicitly focuses on observing the teacher and his/her actions and statements during the class period, and does not include codes for student actions. While relevant student actions and statements can be recorded in the evidence section, the observer does not attempt to capture students’ responses, as it is difficult to follow all of the students and know how they are responding without talking with them. In addition, their responses may be varied and confounded with factors unrelated to the teacher’s actions or intentions.

This focus on the teacher requires that the Protocol also include an opportunity for the teacher to describe their intentions and goals before class and their reflections after class, which provides context for the observation. Pairing these insights with the teacher’s actions during the lesson offers a more encompassing perspective for the observer to consider when evaluating the quality of technology integration.

Synthesis of Data into a Quality Measure

At the conclusion of the observation, the observer has a collection of 10-min “interval sheets” with codes for the three categories and evidence (written text) of each code. In order to aggregate these data into a meaningful picture of the quality of technology integration across the full lesson, we developed a post-observation summary that combines the quantitative and qualitative information from each interval sheet, as well as information from the pre- and post-observation teacher questions. This summary consists of an average and the highest-level technology codes for each category (science and engineering practices, student-centered teaching, and contextualized teaching) that is supported by relevant qualitative evidence from field notes.

Examining Validity and Reliability

We examined the Protocol’s validity throughout the development process, and inter-rater reliability of the quantitative codes was calculated using results from the fourth round of iterative piloting.

Instrument Validity

We addressed content validity of the Protocol by aligning it with an existing published theoretical framework and its literature base, and comparing the Protocol with similar protocols. The development of the Protocol consisted of an iterative process of collecting data from a variety of classrooms, reviewing interpretations and definitions of items and constructs, and reflecting on the literature and previously developed protocols. In this way, we established categories and indicators that reflect the range of practices of technology integration (American Educational Research Association 2014). We established face validity of the Protocol with an expert panel who reviewed and submitted written feedback on three drafts of the Protocol and supporting materials. In particular, the expert review confirmed the categories and ratings of technology integration and their feedback resulted in definitions that were more closely aligned with understandings of technology across a broader range of classrooms. We also made refinements to the reference guide and other support materials after each review.

Inter-Rater Reliability

We calculated inter-rater reliability using results of our fourth implementation (see Table 1). We observed 24 teachers in 17 schools in six states (MA, RI, CA, MD, WV, and OR). Grade levels ranged from 8th to 12th grade; schools were in urban, suburban, and rural areas, with a range of demographic characteristics.

Eight observers completed a total of 41 classroom observations, with two to three observers in each classroom. Each observer was observed from three to 10 total class periods. To the extent possible (given geographic limitations), different sets of partners observed different classrooms. After each observation, individuals completed the Protocol and recorded their results online. The pairs then compared results using a worksheet that helped to reconcile codes and identify feedback for further revisions. Inter-relater reliability was conducted on the 10-min intervals within each observed lesson, resulting in a total of 215 paired intervals. Krippendorf’s alpha was used to determine reliability; Krippendorf’s alpha is appropriate for this analysis because it “generalizes across scales of measurement; can be used with any number of observers, with or without missing data” (Hayes and Krippendorf 2007, p. 78). Alpha values ranged from 0.40 (student engagement) to 1.0 (STEM workplace technology, which had 100% agreement) (Table 3).

Table 3 Krippendorf’s alpha values for Protocol codes

Given these results, we made a series of adjustments to the Protocol and to the definitions and guidance included in the Reference Guide. Specifically, the code for student engagement had the lowest alpha estimate (0.40). That code was removed rather than revised, in order to retain the focus on teacher activities rather than student activities. Student-level observations, including engagement as well as academic achievement, were not within the scope of this Protocol.

The four codes for technology type (instructional, ubiquitous, STEM instructional, and STEM workplace) had alpha estimates ranging from 0.46 to 1.0. Given this range and elimination of technology type from the post-observation code summary, the four codes were removed and replaced with a space for a qualitative description of the technology type. The Reference Guide provided additional support on classifying technology types into each category.

We retained six codes in the final Protocol with Krippendorf’s alpha estimates from 0.51 to 0.78. Four codes (SEP level, SEP technology, student-centered level, student-centered technology) had acceptable values above 0.68, while contextualization level and contextualization technology were below 0.65. Given that the values were lower than desired and that the post-observation checks between observers resulted in recommendations for clarifications to the Reference Guide, we revised the Guide to address the recommendations, including refining descriptions for each code.

Discussion and Conclusion

The TOP-Science Protocol successfully addresses the gap around observing and documenting the quality of technology integration in high school classrooms (Bielefeldt 2012; Stylinski et al. 2018). In particular, it offers a tool that captures how teachers utilize technology to support teaching and learning of scientific pursuits and practices. The Protocol’s focus on integration of technology in three key dimensions expands other observation protocols by moving beyond simply registering technology use, and by targeting key categories that are specifically relevant for science classrooms. Its use of a limited number of descriptors, a straightforward three-level coding system, and the integration of descriptive text justifying each code choice provides a manageable frame for each 10-min observation interval. Together, these strategies direct the observer’s attention on teacher intentions, actions, and reflections without overwhelming them with trying to interpret student actions and perspectives. The final step of summarizing interval data by category offers an overall quality measure of technology use, while still providing the option to substitute any other synthesis calculation.

The Protocol has some limitations that should be considered in any application and could be addressed in future revisions. First, its structure suggests that full integration of technology in all three categories points to the highest quality. However, such extensive integration across an entire class period is likely not realistic or even desirable, as technology is not appropriate for all learning outcomes and activities. Our Protocol does not account for this variation in appropriate use of technology. That is, it is effective at identifying high-quality technology use, as well as incidental applications of technology, but does not elucidate when technology might or might not be the most appropriate tool to support learning. Thus, further work is needed to explore how to determine technology use relative to technology benefits and affordances. Second, the inter-rater reliability scores in some areas were lower than desired; revisions were made to both the coding structure and the Reference Guide to address the issues we identified, and additional inter-rater reliability using the updated Protocol and Reference Guide is needed to examine additional evidence of the Protocol’s reliability. Finally, the Protocol does not explicitly examine the impact of observed teacher practices on student outcomes. Our Protocol allows a view into teachers’ intention and actions; however, the final arbiter of high-quality instruction must always be student outcomes. A separate evaluative tool is needed to identify or measure these outcomes—ideally in ways that reflect the instructional dimensions relevant for science classrooms (i.e., science practices, student-centered teaching, and contextualization).

Despite these limitations, our Protocol successfully frames the quality of technology integration from the perspective of the teacher and in the context of today’s science classrooms and is being used in research (Marcum-Dietrich et al. 2019). As described above, the Protocol was developed with care to determine both face validity and construct validity, and adequate reliability has been demonstrated. Additional studies of inter-rater reliability using the final version of the Protocol will serve to further assess its strength as a measure of the quality of technology integration. The inclusion of a post-observation code summary adds an important component not included in many other protocols, by helping researchers to synthesize the observations of what is observed in the classrooms using the Protocol’s theoretical framework.

Notes

  1. 1.

    Nellie Mae’s definition includes a fourth component of student-centered teaching—“learning can happen anytime, anywhere”—however, because this component addresses learning that happens outside the classroom, it extends beyond the scope of the Protocol.

References

  1. American Educational Research Association, American Psychological Association, National Council on Measurement in Education, & Joint Committee on Standards for Educational and Psychological Testing. (2014). Standards for Educational and Psychological Testing. Washington, DC: AERA.

    Google Scholar 

  2. Bielefeldt, T. (2012). Guidance for technology decisions from classroom observation. Journal of Research on Technology in Education (International Society for Technology in Education), 44(3), 205–223.

    Google Scholar 

  3. Brickell, G., & Herrington, J. (2006). Scaffolding learners in authentic, problem based e-learning environments: The geography challenge. Australas J Educ Technol, 22(4), 531–547.

    Article  Google Scholar 

  4. Brophy, S., Klein, S., Portsmore, M., & Rogers, C. (2008). Advancing engineering education in P-12 classrooms. J Eng Educ, 97(3), 369–387.

    Article  Google Scholar 

  5. D’Angelo, C., Rutstein, D., Harris, C., Haertel, G., Bernard, R., & Borokhovski, E. (2014, March). Simulations for STEM Learning: Systematic Review and Meta-Analysis Report Overview. Menlo Park, CA: SRI Education.

  6. De Jong, T. (2006). Technological advances in inquiry learning. Science, 312(5773), 532–533.

    Article  Google Scholar 

  7. Devolder, A., van Braak, J., & Tondeur, J. (2012). Supporting self-regulated learning in computer-based learning environments: Systematic review of effects of scaffolding in the domain of science education. J Comput Assist Learn, 28(6), 557–573.

    Article  Google Scholar 

  8. DiBiase, W., & McDonald, J. R. (2015). Science teacher attitudes toward inquiry-based teaching and learning. The Clearing House, 88(2), 29–38.

    Article  Google Scholar 

  9. Dringenberg, E., Wertz, R. E. H., Purzer, S., & Strobel, J. (2012). Development of the science and engineering classroom learning observation protocol. Presented at the 2012 American Society for Engineering Education National Conference. Retrieved August 8, 2015 from http://www.asee.org/public/conferences/8/papers/3324/view.

  10. Education Super Highway. (2017). The 2016 state of the states. Retreived July 30, 2018, https://s3-us-west-1.amazonaws.com/esh-sots-pdfs/2016_national_report_K12_broadband.pdf.

  11. Fensham, P. J. (2004). Increasing the relevance of science and technology education for all students in the 21st century. Sci Educ Int, 15(1), 7–26.

    Google Scholar 

  12. Hayden, K., Ouyang, Y., Scinski, L., Olszewski, B., & Bielefeldt, T. (2011). Increasing student interest and attitudes in STEM: Professional development and activities to engage and inspire learners. Contemporary Issues in Technology and Teacher Education, 11(1), 47–69.

    Google Scholar 

  13. Hayes, A. F., & Krippendorf, K. (2007). Answering the call for a standard reliability measure for coding data. Commun Methods Meas, 1(1), 77–89.

    Article  Google Scholar 

  14. Herrington, J., & Kervin, L. (2007). Authentic learning supported by technology: 10 suggestions and cases of integration in classrooms. Educational Media International, 44(3), 219–236.

    Article  Google Scholar 

  15. Herrington, D. G., Bancroft, S. F., Edwards, M. M., & Schairer, C. J. (2016). I want to be the inquiry guy! How research experiences for teachers change beliefs, attitudes, and values about teaching science as inquiry. J Sci Teach Educ, 27(2), 183–204.

    Article  Google Scholar 

  16. Lombardi, M. M. (2007). Authentic learning for the 21st century: An overview. In D. G. Oblinger (Ed.) Educause Learning Initiative, ELI Paper 1: 2007. Retrieved October 19, 2017 from: https://www.researchgate.net/profile/Marilyn_Lombardi/publication/220040581_Authentic_Learning_for_the_21st_Century_An_Overview/links/0f317531744eedf4d1000000.pdf.

  17. Maddux, C. D., & Johnson, D. L. (2006). Information technology, type II classroom integration, and the limited infrastructure in schools. Comput Sch, 22(3–4), 1–5.

    Google Scholar 

  18. Marcum-Dietrich, N., Bruozas, M., & Staudt, C. (2019). Embedding computational thinking into a middle school science meteorology curriculum. Baltimore, MD: Interactive poster presented at the annual meeting of the National Association of Research in Science Teaching (NARST).

    Google Scholar 

  19. Miller, L. M., Chang, C.-I., Wang, S., Beier, M. E., & Klisch, Y. (2011). Learning and motivational impacts of multimedia science game. Computers and Education, 57(1), 1425–1433.

    Article  Google Scholar 

  20. Moeller, B., & Reitzes, T. (2011). Integrating technology with student-centered learning. Quincy, MA: Education Development Center, Inc., Nellie Mae Education Foundation.

    Google Scholar 

  21. National Research Council. (2012). A framework for K-12 science education: Practices, crosscutting concepts, and core ideas. Washington, DC: National Academies Press.

    Google Scholar 

  22. Nellie Mae Education Foundation, & Parthenon-EY. (2015). The connected classroom: Understanding the landscape of technology supports for student-centered learning.

  23. Parker, C. E., Stylinski, C. D., Bonney, C. R., Schillaci, R., & McAuliffe, C.  (2015). Examining the quality of technology implementation in STEM classrooms: Demonstration of an evaluative framework. J Res Technol Educ, 47(2), 105–121.

  24. Patrick, S., Kennedy, K., & Powell, A. (2013). Mean what you say: Defining and integrating personalized, blended and competency education. Vienna, VA: International Association for K-12 Online Learning.

    Google Scholar 

  25. Pullen, D. L., Baguley, M., & Marsden, A. (2009). Back to basics: Electronic collaboration in the education sector. In J. Salmons & L. Wilson (Eds.), Handbook of research on electronic collaboration and organization synergy (pp. 205–222). Hershey, PA: IGI Global.

    Chapter  Google Scholar 

  26. Ravitz, J., Wong, Y., & Becker, H. (2000). Constructivist-compatible and practices among US teachers. Irvine, CA: Center for Research on Information Technology and Organizations.

    Google Scholar 

  27. Reeves, T. C., Herrington, J., & Oliver, R. (2002). Authentic activities and online learning. Annual Conference Proceedings of Higher Education Research and Development Society of Australasia. Perth: Retrieved October 19, 2017, from http://researchrepository.murdoch.edu.au/id/eprint/7034/1/authentic_activities_online_HERDSA_2002.pdf.

  28. Reif, G., Schultz, G., & Ellis, S. (2016). A qualitative study of student-centered learning practices in New England high schools. Boston: Nellie Mae Education Foundation and University of Massachusetts Donahue Institute.

    Google Scholar 

  29. Rutten, N., van Joolingen, W. R., & van der Veen, J. T. (2012). The learning effects of computer simulations in science education. Comput Educ, 58(1), 136–153.

    Article  Google Scholar 

  30. Stylinski, C.D. & Doty, C.  (2013). The inquiring with GIS (iGIS) project: helping teachers create and lead local GIS-based investigations. In: Pages 161-190. In J. G. MaKinster, N. M. Trautman, & M. Barnett (Eds.), Teaching science and investigating environmental issues with geospatial technology: Designing effective professional development for teachers. Springer Publishing Co.

  31. Stylinski, C.D., DeLisi, J., Wong, J., Bonney, C., Parker, C.E. & Doty, C. (2018). Mind the Gap: Reviewing measures of quality and technology use in classroom observation protocols. Paper presented at NARST International Conference, Atlanta GA.

  32. Sawada, D., Piburn, M. D., Judson, E., Turley, J., Falconer, K., Benford, R., Bloom, I. (2002). Measuring Reform Practices in Science and Mathematics Classrooms: The Reformed Teaching Observation Protocol. Sch Sci Math, 102(6):245–253

    Article  Google Scholar 

  33. Smetana, L. K., & Bell, R. L. (2011). Computer simulations to support science instruction and learning: A critical review of the literature. Int J Sci Educ, 34(9), 1337–1370.

    Article  Google Scholar 

  34. Stearns, L. M., Morgan, J., Capraro, M. M., & Capraro, R. M. (2012). A teacher observation instrument for PBL classroom instruction. Journal of STEM Education: Innovations and Research, 13(2), 7–16.

    Google Scholar 

  35. Sturgis, C., & Patrick, S. (2010). When success is the only option: Designing competency-based pathways for next generation learning. Vienna, VA: International Association for K-12 Online Learning.

    Google Scholar 

Download references

Acknowledgments

We thank all of our study teachers who donated their time and opened their classrooms in support of this research.

Funding

This material is based upon work supported by the National Science Foundation under Grants #1438396 and 1438368. Any opinions, findings, and conclusions or recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation.

Author information

Affiliations

Authors

Corresponding author

Correspondence to Caroline E. Parker.

Ethics declarations

Ethical Approval

All procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the 1964 Helsinki declaration and its later amendments or comparable ethical standards.

Conflict of Interest

The authors declare that they have no conflict of interest.

Additional information

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A. Technology Observation Protocol for Science

Appendix A. Technology Observation Protocol for Science

Fig. A1
figure 1

Part 1. Pre-observation information (complete one for each observation)

Fig. A2
figure 2

Part 2. Interval sheet (complete one for every 10-min interval during the observation)

Fig. A3
figure 3

Part 3. Interval sheet (complete one for every 10-min interval during the observation)

Fig. A4
figure 4

Part 4. Post-observation summary

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Reprints and Permissions

About this article

Verify currency and authenticity via CrossMark

Cite this article

Parker, C.E., Stylinski, C.D., Bonney, C.R. et al. Measuring Quality Technology Integration in Science Classrooms. J Sci Educ Technol 28, 567–578 (2019). https://doi.org/10.1007/s10956-019-09787-7

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10956-019-09787-7

Keywords

  • Classroom observation protocol
  • Technology integration
  • High school science
  • Science inquiry
  • Evaluation