Research in Higher Education

, Volume 11, Issue 2, pp 153–165 | Cite as

The stability of teacher performance in the same course over time

  • Philip L. Smith


Traditional reliability studies report the stability of teaching performance, as measured by student course/instructor evaluations, as ranging from .40 to .90, depending upon the study. Recently, generalizability theory has been applied to these rating data to more specifically pinpoint sources of error in student course evaluations. The purpose of the present study is to (1) familiarize the reader with some of the work that has been done with student rating data using generalizability theory and (2) to illustrate the use of the theory on a problem that in the past has been examined using traditional reliability estimates. Specifically, the paper addresses the stability of teacher performance across time in the same course.

Key words

teacher performance student course evaluation 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.


  1. Aleamoni, L.J., and Spencer, R.E. The Illinois Course Evaluation Questionnaire: A description of its development and a report of some of its results.Educational and Psychological Measurement 1973,33 669–684.Google Scholar
  2. Bauswell, R.B., Schwartz, S., and Purohit, A. An examination of the conditions under which various student rating parameters replicate across time.Journal of Educational Measurement, 1975,12 273–280.Google Scholar
  3. Costin, F., Greenough, W.T., and Menger, R.J. Student ratings of college teaching: reliability and validity.Review of Education Research 1977,41 511–535.Google Scholar
  4. Cronbach, L.J., Gleser, G., Nanda, H., and Rajaratnam, N.The dependability of behavioral measurements New York: Wiley, 1972.Google Scholar
  5. Doyle, K.O. Construction and evaluation of scales for rating college instructors. Unpublished doctoral dissertation, University of Minnesota, 1972.Google Scholar
  6. Doyle, K.O.Student evaluation of instruction. Lexington, Mass: Lexington Books, D.C. Heath, 1975.Google Scholar
  7. Gillmore, G.M., Kane, M.T., and Naccarato, R.W. The generalizability of student ratings of instruction,Journal of Educational Measurement 1978,15 1–15.Google Scholar
  8. Hildebrand, M., Wilson, R.C., and Dienst, E.R.Evaluating university teaching. Berkeley: Center for Research and Development in Higher Education, University of California, Berkeley, 1971.Google Scholar
  9. Hogan, T.P. Similarity of student ratings across instructors, courses, and time.Research in Higher Education 1973,1 149–154.Google Scholar
  10. Kane, M.T., Gillmore, G.M., and Crooks, T.J. Student evaluations of teaching: the generalizability of class means.Journal of Educational Measurement 1976,13 171–184.Google Scholar
  11. Kane, M.T., and Brennen, R. The generalizability of class means.Review of Educational Research 1977,47 267–292.Google Scholar
  12. Kohlan, R.G. A comparison of faculty evaluations early and late in the course.The Journal of Higher Education 1973,44 587–595.Google Scholar
  13. Meyers, J.M.Fundamentals of experimental design (2nd ed.. Boston: Allyn and Bacon, 1972.Google Scholar
  14. Smith, P.L. The generalizability of student ratings of courses: asking the right questions.Journal of Educational Measurement 1979,16 77–87.Google Scholar
  15. Spencer, R.E., and Aleamoni, L.M. A student course evaluation questionnaire.Journal of Educational Measurement 1970,7 209–210.Google Scholar

Copyright information

© Agathon Press, Inc. 1979

Authors and Affiliations

  • Philip L. Smith
    • 1
  1. 1.Department of Educational PsychologyUniversity of WisconsinMilwaukee

Personalised recommendations