Abstract
Counseling instructors using evaluations made by their students has shown to be a fruitful approach to enhancing teaching quality. However, prior experimental studies are questionable in terms of external validity. Therefore, we conducted a non-experimental intervention study in which all of the courses offered by a specific department at a German university were evaluated twice with a standardized student evaluation questionnaire (HILVE-II; overall 44 instructors, 140 courses, and 2,546 student evaluations). Additionally, twelve full time instructors received counseling after the first measurement point. Long-term effects over a period of 2 years and transfer effects to other courses were analyzed using multi-level analyses with three levels. Possible influences by bias and unfairness variables were controlled for. Our results indicate a moderate to large effect of counseling on teaching quality. In conclusion, if students’ evaluations are accompanied by counseling based on the evaluation results, they present a useful method to assure and increase teaching quality in higher education.
Similar content being viewed by others
Notes
An analogous workshop including counseling instructors was also conducted following the second evaluation.
References
Abrami, P. C., d’Apollonia, S., & Rosenfield, S. (2007). The dimensionality of student ratings of instruction: What we know and what we do not. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 385–456). Dordrecht: Springer.
Bray, J. H., & Howard, G. S. (1980). Methodological considerations in the evaluation of a teacher-training program. Journal of Educational Psychology, 72, 62–70.
Brinko, K. T. (1990). Instructional consultation with feedback in higher education. Journal of Higher Education, 67, 65–83.
Brinko, K. T., & Menges, R. J. (Eds.). (1997). Practically speaking: A sourcebook for instructional consultants in higher education. Stillwater, OK: New Forum.
Centra, J. A. (2003). Will teachers receive higher student evaluations by giving higher grades and less course work? Research in Higher Education, 44, 495–518.
Cohen, P. A. (1991). Effectiveness of student ratings feedback and consultation for improving instruction in dental school. Journal of Dental Education, 55, 145–150.
Cranton, P., & Smith, R. A. (1990). Reconsidering the unit of analysis: A model of student ratings of instruction. Journal of Educational Psychology, 82, 207–212.
Erickson, G. R., & Erickson, B. L. (1979). Improving college teaching. An evaluation of a teaching consultation procedure. Journal of Higher Education, 50, 670–683.
Feldman, K. A. (2007). Identifying exemplary teachers and teaching: Evidence from student ratings. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 93–143). Dordrecht: Springer.
Foersterling, F. (1985). Attributional retraining: A review. Psychological Bulletin, 98, 495–512.
Greenwald, A. G. (1997). Validity concerns and usefulness of student ratings of instruction. American Psychologist, 52, 1182–1186.
Hampton, S. E., & Reiser, R. A. (2004). Effects of a theory-based feedback and consultation process on instruction and learning in college classrooms. Research in Higher Education, 45, 497–527.
Knapper, C., & Piccinin, S. (Eds.). (1999). Using consultants to improve teaching. San Francisco: Jossey-Bass.
l’Hommedieu, R. L., Menges, R. J., & Brinko, K. T. (1990). Methodological explanations for the modest effects of feedback from student ratings. Journal of Educational Psychology, 82, 232–241.
Marsh, H. W. (1982). SEEQ: A reliable, valid and useful instrument for collecting students’ evaluations of university teaching. British Journal of Educational Psychology, 52, 77–95.
Marsh, H. W. (2007a). Students’ evaluations of university teaching: Dimensionality, reliability, validity, potential biases and usefulness. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 319–383). Dordrecht: Springer.
Marsh, H. W. (2007b). Do university teachers become more effective with experience? A multilevel growth model of students’ evaluations of teaching over 13 years. Journal of Educational Psychology, 99, 775–790.
Marsh, H. W., & Hocevar, D. (1991). Students’ evaluations of teaching effectiveness: The stability of mean ratings of the same teacher over a 13-year period. Teaching and Teacher Education, 7, 303–341.
Marsh, H. W., & Roche, L. A. (1993). Effects of grading leniency and low workload on students’ evaluations of teaching. Journal of Educational Psychology, 92, 202–228.
Marsh, H. W., & Roche, L. A. (1997). Making students’ evaluations of teaching effectiveness effective. American Psychologist, 52, 1187–1197.
McKeachie, W. J. (2007). Good teaching makes a difference—and we know what it is. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 457–474). Dordrecht: Springer.
McKeachie, W. J., Lin, Y., Daugherty, M., Moffett, M., Neigler, C., Nork, J., et al. (1980). Using student ratings and consultation to improve instruction. British Journal of Educational Psychology, 50, 168–174.
Menges, R. J., & Brinko, K. T. (1986). Effects of student evaluation feedback: A meta-analysis of higher education research. Paper presented at the meeting of the American Educational Research Association, San Francisco.
Murray, H. G. (1997). Effective teaching behaviors in the college classroom. In R. P. Perry & J. C. Smart (Eds.), Effective teaching in higher education: Research and practice (pp. 171–204). New York: Agathon Press.
Murray, H. G., & Lawrence, C. (1980). Speech and drama training for lecturers as a means of improving university teaching. Research in Higher Education, 13, 73–90.
Penny, A. R., & Coe, R. (2004). Effectiveness of consultation on student ratings feedback: A meta-analysis. Review of Educational Research, 74, 215–253.
Perry, R. P., & Smart, J. C. (Eds.). (2007). The scholarship of teaching and learning in higher education: An evidence-based perspective. Dordrecht: Springer.
Piccinin, S., Cristi, C., & McCoy, M. (1999). The impact of individual consultation on student ratings of teaching. International Journal for Academic Development, 4, 75–88.
Raudenbush, S. W., Bryk, A. S., Cheong, Y. F., & Congdon, R. T. (2001). HLM 5 (Version 5.04). Chicago: Scientific Software International.
Richardson, J. T. E. (2005). Instruments for obtaining student feedback: A review of literature. Assessment & Evaluation in Higher Education, 30, 387–415.
Rindermann, H. (1996). Zur Qualität studentischer Lehrveranstaltungsevaluationen [On the quality of student evaluations of college courses]. German Journal of Educational Psychology, 10, 129–145.
Rindermann, H. (2009). Lehrevaluation – Einführung und Überblick zur Forschung und Praxis der Lehrveranstaltungsevaluation an Hochschulen [Evaluation of teaching—an introduction and overview of research and practice] (2nd ed.). Landau, Germany: VEP.
Rindermann, H., & Heller, K. A. (2005). The benefit of gifted classes and talent schools for developing students' competences and enhancing academic self-concept. German Journal of Educational Psychology, 19, 133–136.
Rindermann, H., & Kohler, J. (2003). Lässt sich die Lehrqualität durch Evaluation und Beratung verbessern? Überprüfung eines Evaluations-Beratungs-Modells [Does evaluation and consulting improve quality of instruction? Test of an evaluation-consulting-model]. Psychologie in Erziehung und Unterricht, 50, 71–85.
Rindermann, H., Kohler, J., & Meisenberg, G. (2007). Quality of instruction improved by evaluation and consultation of instructors. International Journal for Academic Development, 12, 73–85.
Rindermann, H., & Schofield, N. (2001). Generalizability of multidimensional student ratings of university instruction across courses and teachers. Research in Higher Education, 42, 377–399.
Rotem, A. (1978). The effects of feedback from students to university instructors: An experimental study. Research in Higher Education, 9, 303–318.
Snijders, T. A. B., & Bosker, R. J. (2002). Multilevel analysis. An introduction to basic and advanced multilevel modeling. London: Sage.
Spiel, C., & Gössler, P. M. (2000). Zum Einfluss von Biasvariablen auf die Bewertung universitärer Lehre durch Studierende [On the influence of bias variables on assessments of university teaching by students]. German Journal of Educational Psychology, 14, 38–47.
Ting, K.-F. (2000). Cross-level effects of class characteristics on students’ perceptions of teaching quality. Journal of Educational Psychology, 92, 818–825.
Webler, W.-D. (1996). Qualitätssicherung in Lehre und Studium an deutschen Hochschulen [Quality Assurance in Teaching and Academic Studies at German Universities]. Zeitschrift für Sozialisationsforschung und Erziehungssoziologie, 16, 119–148.
Wendorf, C. A., & Alexander, S. (2005). The influence of individual- and class-level fairness-related perceptions on student satisfaction. Contemporary Educational Psychology, 30, 190–206.
Will, H., & Blickhan, C. (1987). Evaluation als Intervention [Evaluation as an intervention]. In H. Will, A. Winteler, & A. Krapp (Eds.), Evaluation in der beruflichen Aus- und Weiterbildung [Evaluation in Vocational Training] (pp. 43–59). Heidelberg: Sauer.
Wilson, R. C. (1986). Improving faculty teaching. Effective use of student evaluations and consultants. Journal of Higher Education, 57, 196–211.
Acknowledgment
We would like to thank all of the students and faculty at the university involved for their willingness to participate in the course evaluations.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
Estimation of the Bias Model in Four Steps
In the multi-level estimation of the distorting influences of potential bias and unfairness variables, the null model was expanded to a bias model in four steps described below. This takes into consideration bias and unfairness variables on all three levels (see Table 4) in accordance with recommendations made by Snijders and Bosker (2002, p. 86).
Step 1 (Bias Variables on the Student Level)
In the first step all six potential bias variables associated with the student level were added to the model (centered on the grand mean) and variations of the six regression coefficients between courses were allowed for (random slopes). The regression equation on the student level of the null model was therefore expanded to Y ijk = π0jk + Σ p π pjk ·X pijk + e ijk , whereby X p represents the pth of the six predictors. For each of the p regression weights π pjk , the equation π pjk = βp0k + r pjk with R p = Var(r pjk ) was additionally formulated on the course level in order to permit predictors to vary between courses. The estimation revealed that the effects of some variables vary between courses. The corresponding effects were retained as random effects, all remaining effects were fixed (i.e. setting r pjk = 0). Variables which revealed no significant relationship to the criterion and which effect did not vary between courses were removed from the model.
Step 2 (Bias Variables on the Course Level)
The second step comprised the block-wise addition of the five structural course characteristics, the four course mean scores and the two course distributions into the model as predictors of the intercept (centered on the grand mean; variations between instructors were permitted). Formally speaking, the parameters were estimated using the equation π0jk = β00k + Σ q β0qk ·X qjk + r 0jk on the course level (with X q as the qth term of the 11 predictors) as well as the eleven equations β0qk = γ0q0 + u 0qk on the instructor level with U 0q = Var(u 0qk ). The influences of all course characteristics were constant across instructors; consequently, all course level effects were fixed (i.e. setting u 0qk = 0). Course-level variables which did not predict the criterion variable significantly were therefore removed from the model.
Step 3 (Cross-Level Interactions Between Bias Variables on Course and Student Level)
In the third step, a test was made to determine whether the between course variations in the regression coefficients π pjk for three student variables (gender, perceived topic relevance and prior knowledge) could be explained by characteristics of the courses themselves. To this end, the regression coefficients for the three bias variables named were treated as dependent variables (slopes as outcomes). In this case, a significant effect indicated a cross-level interaction in which the effect of a variable on the student level is moderated by a variable on the course level. Formally, these effects are represented by three regression equations π pjk = βp0k + Σ l β plk ·X ljk + r pjk for the pth of the three student bias variables, whereby X l depicts the lth of the course characteristics included (coefficients β plk fixed on instructor level). Results confirmed two cross-level interactions.
Step 4 (Bias Variables on the Instructor Level)
In the fourth step, a final expansion incorporated the three potential bias variables at the instructor level. They were introduced in the intercept regression equation (grand mean centered): β00k = γ000 + Σ m γ00m ·X mk + u 00k with X m as the mth of the three predictors. None of them significantly biased student evaluations of teaching quality. They were therefore removed from the model.
Rights and permissions
About this article
Cite this article
Dresel, M., Rindermann, H. Counseling University Instructors Based on Student Evaluations of Their Teaching Effectiveness: A Multilevel Test of its Effectiveness Under Consideration of Bias and Unfairness Variables. Res High Educ 52, 717–737 (2011). https://doi.org/10.1007/s11162-011-9214-7
Received:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11162-011-9214-7