TQF Frequently Asked Questions.
Please note: this document is currently in development.It is intended to provide guidance and resources to those interested in improving teaching assessment in higher education. The recommendations and philosophy of assessment described below will continue to evolve with additional resource-gathering and research. If you have additional questions or would like to share relevant literature, please do so on our FAQ Feedback page.
This document's development has been supported by funding from the National Science Foundation (DUE-1725959), the American Association of Universities, and the University of Colorado Boulder. Any opinions, conclusions, and recommendations expressed in this material are those of the author(s) and do not necessarily reflect the views of those organizations.
A pdf version is available here(last updated May 31, 2022).
Teaching Evaluations (General Issues and TQF Philosophy)
At many research universities, current evaluation systems for tenure and promotion poorly assess and/or undervalue the importance of teaching effectiveness, relying instead on research success (). Current practices at most research universities have led to limited value being placed on effective teaching at best, and — at worst — actually reward poor teaching ().
Normative practices are neither as effective as we want nor do they give pathways for improvement; that is, they often lack systematic processes for formative development es and can provide inaccurate summative information. For more information on this, see the section on SETs.
There is growing interest at the national level (; ) to improve the quality of teaching. Research on institutional change, effective teaching practices, and the scholarship of teaching and learning now enable us to develop a research-based evaluation framework. For a more thorough justification, see the TQF white paper (Finkelstein et al., 2017).
There is a need to balance teaching growth of all instructional faculty against the need to evaluate instructors for reappointment, tenure and/or promotion. In order to achieve this balance we recommend that evaluation processes include both formative and summative assessment. Summative assessment results in a determination of whether an instructor is effective or not, while formative assessment results in recommendations for improving teaching. Summative evaluation results in a product (a grade of teaching effectiveness), while formative evaluation continues a process (the refinement of teaching).
A recent study () that interviewed faculty on an evaluation system involving multiple methods of evaluation (student evaluations of teaching, faculty portfolios, and self-reflection) found that faculty felt that the evaluations were disproportionately summative, lacking sufficient formative outcomes. This study provides a good example of how formative evaluation can be used to refine the evaluation systems of faculty.
Peer voice allows for instructor evaluations to be more holistic, as instructors are qualified to evaluate teaching elements such as the instructor’s content understanding and course goals. Additionally, peer observers who are involved in the education research community are qualified to evaluate if the instructor is using up-to-date instructional strategies.
Faculty themselves are uniquely able to describe their justifications for teaching goals and course design, their efforts for professional development, the learning outcomes of their students, and their engagement with the scholarship of teaching.
Students can provide feedback on areas such as engagement, clarity, and their satisfaction with the course. They can provide comparisons between courses on difficulty, time required for the course, as well as information about themselves such as their perceived efforts relative to other students or confidence with the content before taking the course. Lastly, they can demonstrate awareness of learning goals and objectives (e.g. critical thinking skills, mastery of the material, ability to work on a team, and engage in independent learning, etc.).
There are some questions that cannot be adequately answered by only one of the three voices (students, peers, or the faculty member being evaluated) or individual measures (e.g., student ratings, faculty teaching practices inventories vs. teaching portfolios). Just as students are ill-equipped to judge a faculty’s understanding of content or course design, peer evaluators are unable to effectively comment on student interactions with the instructors. By using multiple measures, particularly when they include the three voices, we can effectively “span the space” of information that can be gathered about an instructor’s teaching effectiveness.
Bias is inherent in any evaluation, and one of the best ways to mitigate bias is to have multiple sources for the evaluation. For more information on this, see What bias is present in evaluations? in the Evaluation Systems section of the FAQ.
Most institutions require multiple measures. For example, the University of Colorado requires a minimum of three measures of faculty evaluation (). As a note, Colorado’s only required measure is using student evaluations of teaching.
The department has been identified as the key unit of sustained change in research-extensive universities, like the University of Colorado (Ƶ) ().
Our approach adopts the Departmental Action Team (DAT) model for engaging departments (). DATs are working groups with 4-6 faculty and 1-2 facilitators, designed to involve members of the department in the development and implementation of change efforts to make sustainable reform.
While TQF is attentive to the department as a key unit of change, it is not the only location of work. A systems view of teaching evaluation at the university will involve attention to: individual faculty, professional development programs, college and administrative levels, as well as discipline-based professional societies and extramural stakeholders. For example, recent efforts at discipline-based societies (e.g., ), professional organizations (e.g., AAU, ), accreditors (e.g., ), and legislative bodies (e.g., , relating to K-12 teaching assessment) represent a broader trend in attention to teaching evaluation by extramural stakeholders.
Evaluation Systems
Know what your goals are and what tools meet those goals. Some questions on end-of-term student ratings (SETs/FCQs) adequately measure student satisfaction, perception of workload, etc. Other tools, such as observations of classroom practices (e.g., COPUS), are helpful for understanding how classroom time is spent and what teaching practices are employed.
Consider the use of a rubric for looking at the multiple dimensions and ensuring that the measures provide a holistic view of the instructor. The Teaching Quality Framework Assessment Rubric and other resources are available here.
Use existing services and tools rather than making them up. There is extensive literature on using measures for teaching evaluation (e.g., ). In addition, existing tools can be adapted to suit local environments. Many units at Ƶ Boulder have worked with the TQF to develop tools and processes for teaching evaluations; a repository of publicly available examples can be found here. Additional institutional teaching evaluation resources can be found here. Please contact us if you are looking for a particular resource that cannot be found on our site.Additional Ƶ Boulder resources include:
- CTL Programs & Services
- ASSETT (A&S) services for class observations and more.
Have faculty engage in self-reflection (reviewing their own practices) in a structured and systematic matter.
When using materials, focus on the engagement of faculty in using measures to assess their own practices with an emphasis on formative assessment.
To answer the requisite question that most evaluation tools don’t distinguish between good and excellent instructors, we cite which finds that over 90% of instructors receive the top two ratings on a point scale for evaluation, with less than 1% receiving an unsatisfactory mark. When instructors were asked to rate themselves on a scale of 1 to 10, over 85% rated themselves an 8 or above. The same report offers suggestions such as a comprehensive evaluation plan, training administration in conducting evaluation, and rewarding good teaching.
One evaluation system that has found success is TAP. From the National Institute for Excellence in Teaching, report on the TAP system that was able to produce a normal distribution of instructor evaluation ratings, differentiating good instructors from exceptional instructors. A key point for their success is the use of measurement across multiple (4) dimensions with multiple ratings for each dimension that are averaged out.
As an alternative to “grading” instructors on a fixed-point scale, instructors can be promoted to higher roles reflecting their cumulative capabilities as an instructor. An example of this is the Royal Academy of Engineering Career Framework for University Teaching ().
Promotion and merit decisions at most research universities rely on instructors reaching a set level of aggregate contributions through research, teaching, and service. This is described as the one-bucket system, where all three factors contribute to fill up a bucket, where exemplary research can balance out poor instruction, or vice-versa. However, often the level required for promotion can be reached solely on research efforts, disincentivizing teaching and service. To value teaching and service, the University of California (UC) Irvine has begun transitioning to a three-bucket system. In the three-bucket system, research, teaching, and service each have their own buckets and levels must be reached for each factors. For detailed descriptions on how UC Irvine uses the three-bucket system for promotion and merit decisions, see the Criteria for Appointment, Promotion, and Appraisal section of their .
Departmental definition of assessment measures and relative weights of research, teaching, and service is essential in defining faculty expectations and provides a roadmap for faculty members’ continuous improvement of their teaching.
The traditional approach to evaluating teaching examines only the behavior of the instructor in the classroom and uses limited evidence, typically only student evaluations of teaching. The University of Kansas (KU) developed a to measure 7 dimensions in teaching to provide a holistic view of the instructor as an educator. The TQF has a modified version of this rubric available on the resources page of the TQF website.
For more information on rewarding teaching, see .
Ensure that the evaluation measures lead to actionable outcomes. The measures should be reliable (repeatable across different settings), valid (measures should judge instructors in a reasonable way, like the don’t judge a fish on its ability to climb a tree adage), fair (the measures minimize bias), and avoid undesirable consequences of using the measure (e.g. instructors making a class easier due to the discredited belief that students rate easy classes better).
Fairness and reliability are most easily addressed by using multiple measures. See the section for more information.
Disciplinary action and approaches for remediation for unsatisfactory evaluation results should be defined.
While drawing from evidence-based tools, be sure to consider how these tools are used locally, which are the best-suited tools and approaches to a given context, and how they should be adapted to fit local needs.
While adaptation is encouraged, there should be enough commonality across the different units that measures used in the evaluation system are coherent across the campus.
It can be easier to establish a new approach to evaluation by starting with low-stakes improvements, with evaluations used for merit rather than reappointment, promotion, and tenure. Additionally, value participation in the new evaluation system; that is, reward faculty for having trust and being willing to engage in new evaluation measures.
When considering formative feedback, comments should be specific and focus on behaviors and practices rather than the instructor. Comments should be also be provided to the instructor soon after the evaluation measurement ().
Expectations for instruction should be explicit and clear to instructors. This offers several benefits, including guiding expectations for faculty publicly and framing resources for improvement, such as professional development programs.
There are a few different types of bias. The easiest to control is bias in administration procedures. This bias is seen most easily in SETs. For example, de-anonymizing SETs or having the instructor present during administration of the SETs both lead to artificial inflation in ratings. For example, found that offering chocolate to students led to an increase in evaluation scores. To best control this form of bias, departments should create a standardized method of administering SETs, and in conducting instructor evaluations in general.
Another form of bias is implicit bias, for example gender and race bias. Regarding gender bias, there is little agreement in results. Studies that have falsified instructor gender or created fictitious instructor profiles have found significant differences in ratings based on gender (), while others that have studied real SETs have found no significant difference (). The discrepancies in researchers’ conclusions may be due to variation in bias across fields/disciplines and other contextual factors.
Bias exists in SETs between disciplines, with ratings significantly lower for STEM/quantitative disciplines than non-STEM/non-quantitative disciplines. For a review of the effects of instructor gender and discipline on student ratings, see IDEA Research Report #10 ().
Raise awareness of implicit bias
Make framework/criteria for evaluation explicit and public
Use multiple measures of teaching effectiveness.
Role of Peers in Evaluation: Peer Voice
Peer observation is the most common method of engaging peer voice in evaluation. There are many well-researched observation protocols available; many of these have been adapted by units here at Ƶ Boulder. Examples can be found here. Please contact us if you are looking for a particular resource that cannot be found on our site.
The University of Colorado, through the Arts & Sciences Support of Education Through Technology (ASSETT) program, offers the Visualizing Instructional Practices (VIP) service. Upon request, ASSETT staff will conduct a classroom observation using the Classroom Observation Protocol for Undergraduate STEM (COPUS).
Course Portfolios / Review ()
[In development...others to be added]
Role of Faculty/Instructors in Their Own Evaluation: Faculty Voice
Role of Students in Evaluation: Student Voice (incl. SETs)
A recent meta-analysis by found that there is little to no correlation between student learning gains and SETs. The study concludes that students are not likely to learn more from an instructor with better SET ratings. They follow with claims as to why this is the case, citing that student’s interest, motivation, prior knowledge, and other factors lead to SETs being unreliable. Another study by found that SETs were nearly six standard deviations lower for quantitative courses than non-quantitative courses.
A common issue with past SETs is that students are asked to rate instructors on areas they are ill-equipped to judge. For more information, see the Evaluation Systems section of this FAQ.
One well-researched student evaluation form is the Teaching Behavior Checklist (TBC) (). A factor analysis of the TBC found that the qualities measured can be grouped into two distinct categories: the instructor's professional competency/communication skills, and caring/support provided to students. Similarly, another resource is the .
Questions should be focused on the student perspective, such as their interactions with faculty, their awareness of course goals, and the demands on them in and out of class.
Questions can prompt narrative responses rather than multiple choice, Likert-scale style questions. An example of this approach can be seen in Yale’s end-of-term course evaluation system (Yale , Report of the Teaching and Learning Committee 2015-16; ).
Historically, online response rates have been significantly lower than in-class response rates. The gap in response rates decreased as class size increases, but at best online response rate is 15% worse than in-class. For a large-N review, see .
found that online response rate could match in-class response rate for research based assessments when sufficiently motivated. In practice, to get high response rates for online administration of SETs, instructors should: send multiple email reminders, give multiple in-class reminders, and provide participation credit for completion. It is important to note that online response rates were lower (though not statistically significantly lower) for students who had lower course grades as compared to in-class response rates.
Mid-semester feedback (MSF) can be helpful in engaging student voice for instructor evaluation. Historically, MSF has focused entirely on student experience, and thus has not been plagued with many of the reliability and validity issues present in SETs. Rather than rate faculty effectiveness, in MSF, students have the opportunity to discuss what has and hasn’t worked for them so far and provides faculty with the opportunity to more readily see the effects of making adjustments. For a review of the benefits of MSF and a thorough guide on effectuating MSF, see the Midterm Student Feedback Guide ().
Ƶ offers MSF in the form of the Classroom Learning Interview Process (CLIP). CLIPs are focus-group style interviews with students during usual class time, conducted by peer faculty.
For an example of a written, easy-to-use, single-page MSF form, see the Stop-Go-Change Evaluations (, accessed December 2018), available at PhysPort.
A common concern is that students will rate classes poorly if they are challenging, and rate classes higher if they think they’ll do better. found that a majority of instructors believe that difficulty, workload, and stringency of grading all negatively bias student ratings. argues that the valuing of SETs has led to grade inflation based on studies that show expected grades correlate with student ratings, examined using manipulated grades. However, when real grades and student grade expectations are compared, found that expected grades don’t affect SETs. Regarding course difficulty, Centra found there is a middle ground for class difficulty. Classes that were not only too difficult, but also ones that were too easy, received lower ratings than classes that found an appropriate balance.
Teaching Practices
One of the most impactful changes that can be made in the classroom is engaging students. Interactive engagement has shown to be far more effective regarding student understanding of content than traditional lecturing (). Additionally, failure rates are lowered and student success is increased in active learning environments as compared to traditional lecture ().
Engaging in non-traditional classroom practices such as collaborative learning and conceptual problem solving led to increased student self-efficacy ().
For STEM courses, the National Academies’ Consensus Study Report on STEM Education () and the White House’s PCAST Report () have described a variety of interactive practices and approaches that support enhanced student learning.
Identifying learning goals and sharing them with the students is useful for course development and contributes to department-wide knowledge of what students will learn after taking a course. Additionally, students often rely on them and find them helpful once learning goals becomes a recurring feature in courses.
Although there are many potential learning goals, some will be more important to the instructor and the purpose of the course than others. For example, learning to think ethically may be more valued in a journalism course than an intro-level physics lab, whereas the opposite may be true for the goal of improving quantitative literacy. By focusing in on key learning goals, the instructor can use teaching methods that are shown to be particularly effective at attaining those goals. For a list of teaching methods associated with learning objectives, see IDEA Technical Report #19 (), with a summary figure .
The Science Education Initiative (SEI) Handbook () provides a guide of questions to ask when developing learning goals, and addresses concerns such as whether learning goals lead to “teaching to the test” or if they restrict what an instructor can teach.
A strong baseline to consider while developing learning goals is the American Association of Colleges and Universities (AAƵ) which identify several essential learning outcomes:
Content understanding, with a focus on big questions
Intellectual and practical skills, such as critical thinking and communication skills
Personal and social responsibility, including ethical reasoning and establishing the foundations for lifelong learning
Applied learning, including the ability to synthesize knowledge across domains