Abstract
This study contributed to our understanding of those factors that serve as predictors of science teachers’ selection and use of technologies and more specifically, how selection and usage was realized among teachers of different science disciplines. Notable descriptive statistics were examined, and we tested an explanatory model of how demographics, school context, pedagogical approaches and professional development (PD) influenced the likelihood of a teacher using a tool via a multilevel cross-classification-ordered logit analysis (Goldstein 1995). The findings revealed that science teachers were more likely to use hardware than software; more specifically, this included instructional tools (i.e., SMARTboards, clickers) and laboratory tools (probeware). Differences in teachers’ use of tools were largely due to differences in tools as opposed to differences in teacher characteristics. Use of a tool was more likely by teachers who taught physics, who taught via inquiry, or who had more PD with a tool. These findings have implications for how we conceptualize selection and usage of technologies that enter the science education pipeline; which tools become sustainable in the science classroom and how technological take-up differs across science disciplines.
Similar content being viewed by others
Notes
Probeware in this context interfaced computers and provided access to real-time data and provided opportunities for inquiry-based learning (Tinker, http://concord.org/sites/default/files/pdf/probeware_history.pdf).
See Basalla’s discussion of Joseph Needham analysis of the Chinese society and government (pp. 174–176).
While our study did not elicit specific uses of the SMARTboard, clickers, and probeware technologies, other studies have reported on the potential advantages of these tools. For example, Bell, Maeng and Binns (2013) reported that teachers used the SMARTboard™ activities to stir student attention, engagement, and interactions and to explain their thinking and convey their understanding of the content (p. 366). While Lopez et al. (2013) advocated using clickers to counter poor metacognitive and peer learning strategies, MacArthur and Jones’s (2008) review of 56 studies conducted in college-level science classrooms yielded mixed results regarding the clicker’s pedagogical benefits. Lastly, Yerrick et al. (2011) showed how widely used probeware can engage students with science.
References
Adadan E, Trundle KC, Irving KE (2010) Exploring grade 11 students’ conceptual pathways of the particulate nature of matter in the context of multirepresentational instruction. J Res Sci Teach 47:1004–1035
Baram-Tsabari A, Yarden A (2005) Characterizing children’s spontaneous interests in science and technology. Int J Sci Educ 27:803–826
Basalla G (1996) The evolution of technology. Cambridge University Press, New York, NY
Beare R (2007) Investigation into the potential of investigative projects involving powerful robotics telescopes to inspire interest in science. Int J Sci Educ 29:279–306
Bell RL, Trundle KC (2008) The use of a computer simulation to promote scientific conceptions of moon phases. J Res Sci Teach 45:346–372
Bell RL, Maeng JL, Binns IC (2013) Learning in context: technology integration in a teacher preparation program informed by situated learning theory. J Res Sci Teach 50:348–379
Benjamini Y, Krieger AM, Yekutieli D (2006) Adaptive linear step-up procedures that control the false discovery rate. Biometrika 93:491–507
Bryk AS, Raudenbush SW (1992) Hierarchical linear models. Sage, London
Campbell T, Zuwallack R, Longhurst M, Shelton BE, Wolf PG (2014). An examination of the changes in science teaching orientations and technology-enhanced tools for student learning in the context of professional development. International Journal of Science Education. doi:10.1080/09500693.2013.879622
Czerniak CM, Lumpe AT, Haney JJ, Beck J (1999) Teachers’ beliefs about using educational technology in the science classroom. Int J Educ Technol 1:1–17
FitzPatrick KA, Finn KE, Campisi J (2011) Effect of personal response systems on student perception and academic performance in courses in a health science curriculum. Adv Physiol Educ 35:280–289
Gallagher JJ (2007) Teaching science for understanding: a practical guide for middle and high school teachers. Merrill, Upper Saddle River, NJ
Gerard LF, Varma K, Corliss SB, Linn MC (2011) Professional development for technology-enhanced inquiry science. Rev Educ Res 81:408–448
Goldstein H (1995) Multilevel statistical models. Edward Arnold, Sydney
Graham JW (2009) Missing data analysis. Annu Rev Psychol 60:549–576
Greenberg R, Raphael J, Keller JL, Tobias S (1998) Teaching high school science using image processing: a case study of implementation of computer technology. J Res Sci Teach 35:297–327
Hall I, Higgins S (2005) Primary school students’ perceptions of interactive whiteboards. J Comput Assist Learn 21:102–117
Kennedy P (2008) A guide to econometrics. Blackwell, Cambridge
Kennewell S, Higgins S (2007) Introduction: special edition on interactive whiteboards. Learn Media Technol 32:207–212
King G, Zeng L (2001) Logistic regression in rare events data. Polit Anal 9:137–163
Konstantopoulos S (2008) The power of the test in three-level cluster randomized designs. J Res Educ Eff 1:66–88
Lawless KA, Pellegrino JW (2007) Professional development in integrating technology into teaching and learning: knowns, unknowns, and ways to pursue better questions and answers. Rev Educ Res 77:575–614
Little RJA, Rubin DB (2002) Statistical analysis with missing data. Wiley, New York
Lopez EJ, Nandagopal K, Shavelson RJ, Szu E, Penn J (2013) Self-regulated learning study strategies and academic performance in undergraduate organic chemistry: an investigation examining ethnically diverse students. J Res Sci Teach 50:660–676
Lotter C, Harwood WS, Bonner J (2007) The influence of core teaching conceptions on teachers’ use of inquiry teaching practices. J Res Sci Teach 44:1318–1347
MacArthur JR, Jones LJ (2008) A review of literature reports of clickers applicable to college chemistry classrooms. Chem Educ Res Pract 9:187–195
MacKinnon DP, Lockwood CM, Williams J (2004) Confidence limits for the indirect effect. Multivar Behav Res 39:99–128
Mistler-Jackson M, Songer NB (2000) Student motivation and internet technology: are students empowered to learn science? J Res Sci Teach 37:459–479
National Research Council (NRC) (2012) A framework for K-12 science education: practices, crosscutting concepts, and core ideas. National Academy Press, Washington, DC
Ng W, Gunstone R (2003) Science and computer-based technologies: attitudes of secondary science teachers. Res Sci Technol Educ 21:243–264
Odom AL, Settlage J, Pedersen JE (2002) Technology knowledge and use: a survey of science educators. J Sci Educ Technol 11:391–398
Passmore C, Stewart J (2002) A modeling approach to teaching evolutionary biology in high schools. J Res Sci Teach 39:185–204
Pedersen JE, Yerrick RK (2000) Technology in science teacher education: survey of current uses and desired knowledge among science educators. J Sci Teach Educ 11:131–153
Peugh JL, Enders CK (2004) Missing data in educational research. Rev Educ Res 74:525–556
Pituch KA, Stapleton LM, Kang JY (2006) A comparison of single sample and bootstrap methods to assess mediation in cluster randomized trials. Multivar Behav Res 41:367–400
Rasbash J, Woodhouse G (1995) MLn command reference. Multilevel Models Project, Institute of Education, London
Roth WM, Woszczyna C, Smith G (1996) Affordances and constraints of computers in science education. J Res Sci Teach 33:995–1017
Scheffer J (2002) Dealing with missing data. ResLett Inf Math Sci 3:153–160
Scheuren F (2004) What is a survey? Booklet
Slavin R (2005) Educational psychology. Allyn and Bacon, New York
Sorensen P, Twidle J, Childs A, Godwin J (2007) The use of internet in science teaching: a longitudinal study of developments in use by student–teachers in England. Int J Sci Educ 29:1605–1627
Stylianidou F, Boohan R, Ogborn J (2005) Science teachers’ transformations of the use of computer modeling in the classroom: using research to inform training. Sci Educ 89:56–70
Tabachnick BG, Fidell LS (2006) Using multivariate statistics. Allyn and Bacon, Boston
Tanner H, Jones S (2007) How interactive is your whiteboard? Math Teach Inc Micromath 200:37–41
Wang Y (2002) From teacher-centredness to student-centredness: are preservice teachers making the conceptual shift when teaching in information age classrooms? Educ Media Int 39:257–265
Yerrick R, Hoving T (1999) Obstacles confronting technology initiatives as seen through the experience of science teachers: a comparative study of science teachers beliefs, planning, and practice. J Sci Educ Technol 8:291–307
Yerrick R, Schiller J, Reisfeld J (2011) Who are you callin expert?: using student narratives to redefine expertise and advocacy lower track science. J Res Sci Teach 48:13–36
Zacharia Z (2003) Beliefs, attitudes, and intentions of science teachers regarding the educational use of computer simulations and inquiry-based experiments in physics. J Res Sci Teach 40:792–823
Zhao Y, Pugh K, Sheldon S, Byers JL (2002) Conditions for classroom technology innovations. Teach Coll Rec 104:482–515
Author information
Authors and Affiliations
Corresponding author
Electronic supplementary material
Below is the link to the electronic supplementary material.
Appendices
Appendix 1: Power Analysis of Sample Size
Level | Effect size | |||
---|---|---|---|---|
0.1 | 0.2 | 0.3 | 0.4 | |
2) Teacher | 0.32 | 0.84 | 0.99 | 1.00 |
1) Tool | 0.94 | 1.00 | 1.00 | 1.00 |
Appendix 2: Below We Illustrate The Coding Scheme that was Used for Identified Variables of the Study
Teacher demographics variables | |
---|---|
Gender | 0: female 1: male |
Years of teaching experience | Manually inputted by participant |
Highest educational level | 0: college 1: teaching certificate 2: masters 3: doctorate |
Subjects taught | manually inputted by participant |
Grade levels taught | manually inputted by participant |
Tenured | 0: no 1: yes |
Type of school | 0: public school 1: charter school 2: private school 3: alternative school |
School location | 0: urban 1: suburban 2: rural |
Age | 0: 30 years and under 1: 31– 40 years 2: 41–50 years 3: 51+ years |
Teacher attitudes about their students coding key | |
---|---|
My students… | Coding key |
a) have high learning ability | 1: strongly disagree 2: disagree 3: neutral 4: agree 5: strongly agree |
b) are interested in the subject matter | |
c) are well behaved in class | |
d) work well in groups | |
e) have participated in inquiry lessons | |
f) have experience using technology | |
g) have access to learning technology at home |
Teachers top three tools per professional development | |
---|---|
Tool category | Coding key |
Left blank | 0 |
Software—internet websites/software such as word, PowerPoint | 1 |
Software—data analysis tools | 2 |
Software—modeling/simulation | 3 |
Hardware—learning tools (student use) | 4 |
Hardware—instructional tools (teacher use) | 5 |
Hardware—computers/tablets | 6 |
Appendix 3: Correlation–Variance–Covariance Matrix
Correlation–variance–covariance matrix of outcome variables and explanatory variables for word-level analysis. The correlations, variances and co-variances are along the lower left triangle, diagonal and upper right triangle of the matrix.
Variable | 1 | 2 | 3 | 4 | 5 | 6 | 7 | 8 | 9 | |
---|---|---|---|---|---|---|---|---|---|---|
1 Technological tool use by a teacher | 0.596 | 0.008 | 0.014 | 0.057 | 0.021 | 0.024 | 0.012 | 0.009 | 0.000 | |
2 Educational level: Masters | 0.038 | 0.083 | 0.015 | 0.013 | −0.001 | 0.000 | 0.000 | 0.001 | 0.000 | |
3 Subject taught: physics | 0.038 | 0.111 | 0.218 | −0.042 | 0.001 | 0.000 | 0.001 | 0.000 | 0.000 | |
4 Students have participated in inquiry lessons | 0.072 | 0.045 | −0.089 | 1.035 | 0.003 | −0.002 | 0.001 | 0.003 | 0.000 | |
5 Days of training: <1 | 0.112 | −0.007 | 0.007 | 0.011 | 0.062 | −0.002 | −0.001 | −0.001 | 0.000 | |
6 Days of training: 1–2 | 0.171 | 0.006 | −0.001 | −0.012 | −0.049 | 0.032 | 0.000 | 0.000 | 0.000 | |
7 Days of training: 3–5 | 0.142 | 0.015 | 0.019 | 0.005 | −0.030 | −0.021 | 0.013 | 0.000 | 0.000 | |
8 Days of training: 6–10 | 0.131 | 0.028 | −0.011 | 0.030 | −0.024 | −0.016 | −0.010 | 0.008 | 0.000 | |
9 Days of training: more than 10 | 0.035 | 0.006 | 0.007 | −0.002 | −0.005 | −0.003 | −0.002 | −0.002 | 0.000 |
Rights and permissions
About this article
Cite this article
Waight, N., Chiu, M.M. & Whitford, M. Factors that Influence Science Teachers’ Selection and Usage of Technologies in High School Science Classrooms. J Sci Educ Technol 23, 668–681 (2014). https://doi.org/10.1007/s10956-014-9493-9
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10956-014-9493-9