Skip to main content

Advertisement

Log in

Counseling University Instructors Based on Student Evaluations of Their Teaching Effectiveness: A Multilevel Test of its Effectiveness Under Consideration of Bias and Unfairness Variables

  • Published:
Research in Higher Education Aims and scope Submit manuscript

Abstract

Counseling instructors using evaluations made by their students has shown to be a fruitful approach to enhancing teaching quality. However, prior experimental studies are questionable in terms of external validity. Therefore, we conducted a non-experimental intervention study in which all of the courses offered by a specific department at a German university were evaluated twice with a standardized student evaluation questionnaire (HILVE-II; overall 44 instructors, 140 courses, and 2,546 student evaluations). Additionally, twelve full time instructors received counseling after the first measurement point. Long-term effects over a period of 2 years and transfer effects to other courses were analyzed using multi-level analyses with three levels. Possible influences by bias and unfairness variables were controlled for. Our results indicate a moderate to large effect of counseling on teaching quality. In conclusion, if students’ evaluations are accompanied by counseling based on the evaluation results, they present a useful method to assure and increase teaching quality in higher education.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1

Similar content being viewed by others

Notes

  1. An analogous workshop including counseling instructors was also conducted following the second evaluation.

References

  • Abrami, P. C., d’Apollonia, S., & Rosenfield, S. (2007). The dimensionality of student ratings of instruction: What we know and what we do not. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 385–456). Dordrecht: Springer.

    Chapter  Google Scholar 

  • Bray, J. H., & Howard, G. S. (1980). Methodological considerations in the evaluation of a teacher-training program. Journal of Educational Psychology, 72, 62–70.

    Article  Google Scholar 

  • Brinko, K. T. (1990). Instructional consultation with feedback in higher education. Journal of Higher Education, 67, 65–83.

    Article  Google Scholar 

  • Brinko, K. T., & Menges, R. J. (Eds.). (1997). Practically speaking: A sourcebook for instructional consultants in higher education. Stillwater, OK: New Forum.

    Google Scholar 

  • Centra, J. A. (2003). Will teachers receive higher student evaluations by giving higher grades and less course work? Research in Higher Education, 44, 495–518.

    Article  Google Scholar 

  • Cohen, P. A. (1991). Effectiveness of student ratings feedback and consultation for improving instruction in dental school. Journal of Dental Education, 55, 145–150.

    Google Scholar 

  • Cranton, P., & Smith, R. A. (1990). Reconsidering the unit of analysis: A model of student ratings of instruction. Journal of Educational Psychology, 82, 207–212.

    Article  Google Scholar 

  • Erickson, G. R., & Erickson, B. L. (1979). Improving college teaching. An evaluation of a teaching consultation procedure. Journal of Higher Education, 50, 670–683.

    Article  Google Scholar 

  • Feldman, K. A. (2007). Identifying exemplary teachers and teaching: Evidence from student ratings. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 93–143). Dordrecht: Springer.

    Chapter  Google Scholar 

  • Foersterling, F. (1985). Attributional retraining: A review. Psychological Bulletin, 98, 495–512.

    Article  Google Scholar 

  • Greenwald, A. G. (1997). Validity concerns and usefulness of student ratings of instruction. American Psychologist, 52, 1182–1186.

    Article  Google Scholar 

  • Hampton, S. E., & Reiser, R. A. (2004). Effects of a theory-based feedback and consultation process on instruction and learning in college classrooms. Research in Higher Education, 45, 497–527.

    Article  Google Scholar 

  • Knapper, C., & Piccinin, S. (Eds.). (1999). Using consultants to improve teaching. San Francisco: Jossey-Bass.

    Google Scholar 

  • l’Hommedieu, R. L., Menges, R. J., & Brinko, K. T. (1990). Methodological explanations for the modest effects of feedback from student ratings. Journal of Educational Psychology, 82, 232–241.

    Article  Google Scholar 

  • Marsh, H. W. (1982). SEEQ: A reliable, valid and useful instrument for collecting students’ evaluations of university teaching. British Journal of Educational Psychology, 52, 77–95.

    Article  Google Scholar 

  • Marsh, H. W. (2007a). Students’ evaluations of university teaching: Dimensionality, reliability, validity, potential biases and usefulness. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 319–383). Dordrecht: Springer.

    Chapter  Google Scholar 

  • Marsh, H. W. (2007b). Do university teachers become more effective with experience? A multilevel growth model of students’ evaluations of teaching over 13 years. Journal of Educational Psychology, 99, 775–790.

    Article  Google Scholar 

  • Marsh, H. W., & Hocevar, D. (1991). Students’ evaluations of teaching effectiveness: The stability of mean ratings of the same teacher over a 13-year period. Teaching and Teacher Education, 7, 303–341.

    Article  Google Scholar 

  • Marsh, H. W., & Roche, L. A. (1993). Effects of grading leniency and low workload on students’ evaluations of teaching. Journal of Educational Psychology, 92, 202–228.

    Article  Google Scholar 

  • Marsh, H. W., & Roche, L. A. (1997). Making students’ evaluations of teaching effectiveness effective. American Psychologist, 52, 1187–1197.

    Article  Google Scholar 

  • McKeachie, W. J. (2007). Good teaching makes a difference—and we know what it is. In R. P. Perry & J. C. Smart (Eds.), The scholarship of teaching and learning in higher education: An evidence-based perspective (pp. 457–474). Dordrecht: Springer.

    Chapter  Google Scholar 

  • McKeachie, W. J., Lin, Y., Daugherty, M., Moffett, M., Neigler, C., Nork, J., et al. (1980). Using student ratings and consultation to improve instruction. British Journal of Educational Psychology, 50, 168–174.

    Article  Google Scholar 

  • Menges, R. J., & Brinko, K. T. (1986). Effects of student evaluation feedback: A meta-analysis of higher education research. Paper presented at the meeting of the American Educational Research Association, San Francisco.

  • Murray, H. G. (1997). Effective teaching behaviors in the college classroom. In R. P. Perry & J. C. Smart (Eds.), Effective teaching in higher education: Research and practice (pp. 171–204). New York: Agathon Press.

    Google Scholar 

  • Murray, H. G., & Lawrence, C. (1980). Speech and drama training for lecturers as a means of improving university teaching. Research in Higher Education, 13, 73–90.

    Article  Google Scholar 

  • Penny, A. R., & Coe, R. (2004). Effectiveness of consultation on student ratings feedback: A meta-analysis. Review of Educational Research, 74, 215–253.

    Article  Google Scholar 

  • Perry, R. P., & Smart, J. C. (Eds.). (2007). The scholarship of teaching and learning in higher education: An evidence-based perspective. Dordrecht: Springer.

    Google Scholar 

  • Piccinin, S., Cristi, C., & McCoy, M. (1999). The impact of individual consultation on student ratings of teaching. International Journal for Academic Development, 4, 75–88.

    Article  Google Scholar 

  • Raudenbush, S. W., Bryk, A. S., Cheong, Y. F., & Congdon, R. T. (2001). HLM 5 (Version 5.04). Chicago: Scientific Software International.

    Google Scholar 

  • Richardson, J. T. E. (2005). Instruments for obtaining student feedback: A review of literature. Assessment & Evaluation in Higher Education, 30, 387–415.

    Article  Google Scholar 

  • Rindermann, H. (1996). Zur Qualität studentischer Lehrveranstaltungsevaluationen [On the quality of student evaluations of college courses]. German Journal of Educational Psychology, 10, 129–145.

    Google Scholar 

  • Rindermann, H. (2009). LehrevaluationEinführung und Überblick zur Forschung und Praxis der Lehrveranstaltungsevaluation an Hochschulen [Evaluation of teaching—an introduction and overview of research and practice] (2nd ed.). Landau, Germany: VEP.

  • Rindermann, H., & Heller, K. A. (2005). The benefit of gifted classes and talent schools for developing students' competences and enhancing academic self-concept. German Journal of Educational Psychology, 19, 133–136.

    Google Scholar 

  • Rindermann, H., & Kohler, J. (2003). Lässt sich die Lehrqualität durch Evaluation und Beratung verbessern? Überprüfung eines Evaluations-Beratungs-Modells [Does evaluation and consulting improve quality of instruction? Test of an evaluation-consulting-model]. Psychologie in Erziehung und Unterricht, 50, 71–85.

    Google Scholar 

  • Rindermann, H., Kohler, J., & Meisenberg, G. (2007). Quality of instruction improved by evaluation and consultation of instructors. International Journal for Academic Development, 12, 73–85.

    Article  Google Scholar 

  • Rindermann, H., & Schofield, N. (2001). Generalizability of multidimensional student ratings of university instruction across courses and teachers. Research in Higher Education, 42, 377–399.

    Article  Google Scholar 

  • Rotem, A. (1978). The effects of feedback from students to university instructors: An experimental study. Research in Higher Education, 9, 303–318.

    Article  Google Scholar 

  • Snijders, T. A. B., & Bosker, R. J. (2002). Multilevel analysis. An introduction to basic and advanced multilevel modeling. London: Sage.

    Google Scholar 

  • Spiel, C., & Gössler, P. M. (2000). Zum Einfluss von Biasvariablen auf die Bewertung universitärer Lehre durch Studierende [On the influence of bias variables on assessments of university teaching by students]. German Journal of Educational Psychology, 14, 38–47.

    Google Scholar 

  • Ting, K.-F. (2000). Cross-level effects of class characteristics on students’ perceptions of teaching quality. Journal of Educational Psychology, 92, 818–825.

    Article  Google Scholar 

  • Webler, W.-D. (1996). Qualitätssicherung in Lehre und Studium an deutschen Hochschulen [Quality Assurance in Teaching and Academic Studies at German Universities]. Zeitschrift für Sozialisationsforschung und Erziehungssoziologie, 16, 119–148.

    Google Scholar 

  • Wendorf, C. A., & Alexander, S. (2005). The influence of individual- and class-level fairness-related perceptions on student satisfaction. Contemporary Educational Psychology, 30, 190–206.

    Article  Google Scholar 

  • Will, H., & Blickhan, C. (1987). Evaluation als Intervention [Evaluation as an intervention]. In H. Will, A. Winteler, & A. Krapp (Eds.), Evaluation in der beruflichen Aus- und Weiterbildung [Evaluation in Vocational Training] (pp. 43–59). Heidelberg: Sauer.

  • Wilson, R. C. (1986). Improving faculty teaching. Effective use of student evaluations and consultants. Journal of Higher Education, 57, 196–211.

    Article  Google Scholar 

Download references

Acknowledgment

We would like to thank all of the students and faculty at the university involved for their willingness to participate in the course evaluations.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Markus Dresel.

Appendix

Appendix

Estimation of the Bias Model in Four Steps

In the multi-level estimation of the distorting influences of potential bias and unfairness variables, the null model was expanded to a bias model in four steps described below. This takes into consideration bias and unfairness variables on all three levels (see Table 4) in accordance with recommendations made by Snijders and Bosker (2002, p. 86).

Step 1 (Bias Variables on the Student Level)

In the first step all six potential bias variables associated with the student level were added to the model (centered on the grand mean) and variations of the six regression coefficients between courses were allowed for (random slopes). The regression equation on the student level of the null model was therefore expanded to Y ijk  = π0jk  + Σ p π pjk ·X pijk  + e ijk , whereby X p represents the pth of the six predictors. For each of the p regression weights π pjk , the equation π pjk  = βp0k + r pjk with R p  = Var(r pjk ) was additionally formulated on the course level in order to permit predictors to vary between courses. The estimation revealed that the effects of some variables vary between courses. The corresponding effects were retained as random effects, all remaining effects were fixed (i.e. setting r pjk  = 0). Variables which revealed no significant relationship to the criterion and which effect did not vary between courses were removed from the model.

Step 2 (Bias Variables on the Course Level)

The second step comprised the block-wise addition of the five structural course characteristics, the four course mean scores and the two course distributions into the model as predictors of the intercept (centered on the grand mean; variations between instructors were permitted). Formally speaking, the parameters were estimated using the equation π0jk  = β00k  + Σ q β0qk ·X qjk  + r 0jk on the course level (with X q as the qth term of the 11 predictors) as well as the eleven equations β0qk  = γ0q0 + u 0qk on the instructor level with U 0q  = Var(u 0qk ). The influences of all course characteristics were constant across instructors; consequently, all course level effects were fixed (i.e. setting u 0qk  = 0). Course-level variables which did not predict the criterion variable significantly were therefore removed from the model.

Step 3 (Cross-Level Interactions Between Bias Variables on Course and Student Level)

In the third step, a test was made to determine whether the between course variations in the regression coefficients π pjk for three student variables (gender, perceived topic relevance and prior knowledge) could be explained by characteristics of the courses themselves. To this end, the regression coefficients for the three bias variables named were treated as dependent variables (slopes as outcomes). In this case, a significant effect indicated a cross-level interaction in which the effect of a variable on the student level is moderated by a variable on the course level. Formally, these effects are represented by three regression equations π pjk  = βp0k + Σ l β plk ·X ljk  + r pjk for the pth of the three student bias variables, whereby X l depicts the lth of the course characteristics included (coefficients β plk fixed on instructor level). Results confirmed two cross-level interactions.

Step 4 (Bias Variables on the Instructor Level)

In the fourth step, a final expansion incorporated the three potential bias variables at the instructor level. They were introduced in the intercept regression equation (grand mean centered): β00k  = γ000 + Σ m γ00m ·X mk  + u 00k with X m as the mth of the three predictors. None of them significantly biased student evaluations of teaching quality. They were therefore removed from the model.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Dresel, M., Rindermann, H. Counseling University Instructors Based on Student Evaluations of Their Teaching Effectiveness: A Multilevel Test of its Effectiveness Under Consideration of Bias and Unfairness Variables. Res High Educ 52, 717–737 (2011). https://doi.org/10.1007/s11162-011-9214-7

Download citation

  • Received:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11162-011-9214-7

Keywords

Navigation