•  
  •  
 

Abstract

This article interrogates prevailing student evaluation instruments in higher education, assessing their construct validity through exploratory factor analysis of a campus wide rating scale. The analysis reveals that a single ambiguous factor explains negligible variance, suggesting ratings measure student attraction rather than pedagogical competence. Situating findings within accountability discourse and psychometric theory, this article urges replacing item aggregation strategies with a theory driven model of teaching effectiveness. It outlines an iterative procedure for item generation, confirmatory analysis, and context sensitive deployment, differentiating between global and specific, descriptive and evaluative prompts, and recognizing curricular variation across service, undergraduate, and graduate courses. By charting these methodological considerations, this article offers administrators and researchers a blueprint for constructing valid, reliable instruments that support equitable faculty assessment and meaningful instructional improvement.

Share

COinS
 

Accessibility Statement

This item was created or digitized prior to April 24, 2026, or is a reproduction of legacy media created before that date. It is preserved in its original, unmodified state specifically for research, reference, or historical recordkeeping. In accordance with the ADA Title II Final Rule, the University Libraries provides accessible versions of archival materials upon request. To request an accommodation for this item, please submit an accessibility request form.