SlideShare a Scribd company logo
Moving Beyond Student
Ratings to Evaluate Teaching
Vicki L. Wise, PhD
1
Where we are going in this presentation
• What is effective teaching
• Why collect evidence of effective teaching
• What evidence to collect
• How to bring evidence together for decision making
• One form of evidence: Student Ratings of Instruction (SRIs)
• Recognizing and addressing faculty concerns
2
What is effective teaching
Arreola (2007) defined five broad skill dimensions of college teaching: content expertise,
instructional design skills, instructional delivery skills, instructional assessment skills, and course
management skills.
L. Dee Fink (2012) distilled Ken Bain’s research on What the Best College Teachers Do down to four
fundamental tasks of teaching: Knowledge of subject matter, course design, course
management, and interactions with students.
The work of these researchers overlaps and reflects much of what is used in higher education.
3
Why collect evidence of effective teaching
1. To improve teaching
2. To improve the student learning experience
3. To meet expectations for continuous employment
4
What evidence to collect
Student ratings are only one source of data; they must be combined with
additional evidence so that administrators can make an informed
judgment about teaching quality.
-Hoyt and Pallett (1999). Appraising Teaching Effectiveness: Beyond Student Ratings.
No one piece of evidence can stand on its own. Evidence of teaching quality needs to take into
account multiple sources, as teaching is multidimensional. Moreover, the likelihood of obtaining
reliable and valid data and making appropriate judgments are increased with more evidence.
5
A portfolio (collection) of teaching effectiveness evidence could include at
least three forms of evidence.
Formative assessment or evaluation during a course, program or service provides information
useful in improving learning or teaching while it is still occurring. Examples are student reflection
exercises or teacher self reflections.
Summative assessment or evaluation is conducted at the end of a program, service, or experience
to make determinations of quality, worth, and meeting targeted outcomes. An example is a final
grade or end of course evaluations.
Direct methods of collecting information require a display of knowledge and skills. Indirect
methods require a reflection on learning, behavior, and attitudes rather than to demonstrate.
6
Evidence
Direct or
Indirect
Student Ratings I
Observations of Teaching D
Student Interviews I
Peer Ratings D
Teacher Self Evaluations I
Personal Goals for Teaching Improvement D
Teaching Curriculum Map D
Exit and Alumni Ratings I
Employer Ratings I
Teaching Scholarship D
Student Learning Outcomes Results D
Teaching Awards D
Philosophy of Teaching D
Innovative and Creative Teaching Techniques D
Course Materials D
Sample Student Work D
Mentorship of Students and Faculty D
Participation in Faculty Development Activities/ Consultations on Teaching D
Scholarship of Teaching and Learning D
7
How to bring evidence together for decision
making
Help, I have collected multiple forms of evidence and need to make
sense of it all for decision making.
• If a portfolio system is used, it could follow a basic template, and faculty
would be asked to provide evidence of meeting specific criteria. Department
chairs would use a rubric (checklist or scale) to rate the quality of evidence in
pre-determined categories.
• A rubric can also be used on each piece of evidence. The scale would specify
what constitutes as “meeting expectations”.
8
One Form of Evidence: Student Ratings of
Instruction (SRIs)
• Plenty of research supports that overall the evaluation of teaching generally
provides reliable and valid evidence of teaching effectiveness. (meta-analytic
review Gravestock & Gregor-Greenleaf, 2008)
• Research does not support using teaching evaluation as a indicator of
student learning though (Uttl, White, & Wong Gonzalez, 2016. Meta-analysis of
faculty's teaching effectiveness: Student evaluation of teaching ratings and student
learning are not related).
• Significant concerns about the effective use of such evaluation systems
continue to exist.
9
Concerns about SRIs
• Faculty: Anxiety provoking, outright hostility, suspicious of use, evaluations are
biased, students not competent evaluators, ratings influenced by grade expectations,
to name a few.
• Administrators: Generally positive. Concerns about validity. Administrators’ own
ideas of teaching and learning may bias their judgments of teaching effectiveness.
• Students: Generally positive. Believe process is valid and they can be good
evaluators. Unclear if and how data are used.
10
11
Faculty
Concerns in
Detail
 Overly prominent role they play in personnel decisions (tenure, post-tenure review, etc.).
 Reliance on them to make hiring, firing, and promotion decisions, rather than as
performance improvement tools.
 Initial poor ratings by students who later feel: "I-hated-you-for-forcing-us-to-learn-this-
material-but-now-I'm-glad-you-did-thank-you-and-keep-doing-what-you're-doing"
phenomenon.
 Reliance on summative evaluation, rather than student ongoing reflection throughout the
term. Not capturing instructional successes and failures in real time.
 Problem of small numbers
 Problem of response scale (inappropriate, not understood, not well defined).
 Teaching effectiveness not well defined. Not measuring what is valued
 Measure of student satisfaction not student learning
 Not used for learning about value added to the student from being in the course.
 Assessing popularity and/or easiness of instructor.
 Not looking at visible outcomes of a course (e.g., increase in student conference
presentations as a result of a course).
 Bias in scores:
o Gender bias in student scores, especially towards female instructors.
o More difficult classes tend to get lower scores.
o Younger appearing professors tend to get lower scores.
o Introductory classes tend to get lower scores.
Addressing Concerns
1. Follow best practices in instrument development
2. Standardize administration and policy
3. Plan for use of findings
12
1. Follow best practices in instrument
development
• Clearly define effective teaching as well as the components
• Develop questions that measure these components
• Measure behaviors/practices that are observable
• Develop response scale
• Develop two forms: one for formative assessment and one for summative
• Test questions with the intended audience and use to modify each form
• Determine how to analyze data appropriately, based on scale and use
• Determine how results will be reported and used
• Determine who will provide results and in what form
• Collect contextualizing information, especially if comparisons are made (discipline, yrs. taught, …, course
elective or required, course level)
13
2. Standardize administration and policy
• Standardize administration process:
• Set stage for importance (and use) with students and faculty; use a core set of questions; use
motivators rather than incentives (consider equity in this online vs face-to-face); ensure anonymity;
standardize method of delivery.
• Suggest response rate needed: at least 10 students per course or 2/3 of class whichever is
higher. May need to combine scores across terms for the same course.
• Standardize analyses based on data type (i.e., treat ordinal variables appropriately).
• Provide data interpretation and distribution guidelines
• Determine audience for results and level of information needed.
• Determine policy around data storage and access
14
3. Plan for use of findings
• Develop recommendations for formative use to improve teaching.
• Develop guidelines for summative use with other evidence to improve
decision making.
15

More Related Content

PPTX
Assessment ppt
PPTX
Evaluating Teaching in Higher Education
PPTX
Types of assessment and studies based on them
PPTX
Assessment literacy
PPTX
Assessment literacy
PPTX
Plan Your Next Plan - The Assessment Plan!
PPTX
Types of assessment updated
PPTX
Sedere Recent Trends in Assessment in Higher Education
Assessment ppt
Evaluating Teaching in Higher Education
Types of assessment and studies based on them
Assessment literacy
Assessment literacy
Plan Your Next Plan - The Assessment Plan!
Types of assessment updated
Sedere Recent Trends in Assessment in Higher Education

What's hot (20)

PPTX
Classroom assessment
PPT
SCC 2007 Student Learning Outcomes
PPT
Principles of Assessment - Best Practice
PPT
The challenges of Assessment and Feedback: findings from an HEA project
DOCX
DISCUSSION ISSUES ON ASSESSMENT
PPT
DU CTLAT Presentation Assessing Student Learning Outcomes Educational Program...
PPTX
Preparation of Classroom Assessment (SLP-B @ BISCAST)
PPTX
Using assessment data for improving teaching practice acer conference 2009 ppt
PPTX
Formative assessment
PPTX
Types of assessment
PPTX
Formative and summative assessments
PPT
Classroom assessment
PPTX
The reading teacher as classroom researcher
PPT
Presentation Issues on Assessment
PPT
classroom assessment
PPT
Exploring formative assessment
PDF
Lorna Earl Rethinking Classroom Assessment With Purpose In Mind
PPTX
Basics of Assessment
PPTX
Assessment process
PPSX
Diagnostic assessment report pptshow
Classroom assessment
SCC 2007 Student Learning Outcomes
Principles of Assessment - Best Practice
The challenges of Assessment and Feedback: findings from an HEA project
DISCUSSION ISSUES ON ASSESSMENT
DU CTLAT Presentation Assessing Student Learning Outcomes Educational Program...
Preparation of Classroom Assessment (SLP-B @ BISCAST)
Using assessment data for improving teaching practice acer conference 2009 ppt
Formative assessment
Types of assessment
Formative and summative assessments
Classroom assessment
The reading teacher as classroom researcher
Presentation Issues on Assessment
classroom assessment
Exploring formative assessment
Lorna Earl Rethinking Classroom Assessment With Purpose In Mind
Basics of Assessment
Assessment process
Diagnostic assessment report pptshow
Ad

Viewers also liked (18)

PDF
Why health IT needs storytelling
PDF
Interpreting Behavioral Analytics With Reliable Metrics
PPTX
Умные инвестиции в умных роботов
PDF
federal reserve
PDF
federal reserve
PDF
federal reserve
PDF
AstriCon 2016 - Using Asterisk and XMPP to provide greater tools to your cust...
PDF
Service encounters, emotional labor, and mindfulness
DOC
resume_brooklyn+08-16-16
PPTX
Influenza general intro. (symptoms and prevention )
PPT
Investor allstars talk 29 sept 2016
PDF
How to Talk Like a Graphic Designer
 
PPTX
Colección Mejores Lecciones de Liderazgo | Septiembre 2016 | Parte 2 de 2
PPTX
Techkon USA - SpectroDens
PDF
Microservices: Lessons Learned
PPTX
Amy Driscoll: Signature Assignments
PDF
Bill Gates in sporna semenska banka
Why health IT needs storytelling
Interpreting Behavioral Analytics With Reliable Metrics
Умные инвестиции в умных роботов
federal reserve
federal reserve
federal reserve
AstriCon 2016 - Using Asterisk and XMPP to provide greater tools to your cust...
Service encounters, emotional labor, and mindfulness
resume_brooklyn+08-16-16
Influenza general intro. (symptoms and prevention )
Investor allstars talk 29 sept 2016
How to Talk Like a Graphic Designer
 
Colección Mejores Lecciones de Liderazgo | Septiembre 2016 | Parte 2 de 2
Techkon USA - SpectroDens
Microservices: Lessons Learned
Amy Driscoll: Signature Assignments
Bill Gates in sporna semenska banka
Ad

Similar to Moving Beyond Student Ratings to Evaluate Teaching (20)

PPTX
Examination and Evaluation-ppt presentation.pptx
PDF
Tqf day 2 - assessment and feedback
PPTX
Assessment, measurement and Evaluation .pptx
PPTX
Classroom Assessment
PDF
Handouts-ClassroomAssessment-IP3.pdf
PPTX
Assessment for higher education (for biology faculty seminar)
PPTX
Seminar on trends and issues in assessment.pptx
PPTX
3RD REPORTER Evaluating Best Practices in Teaching and Learning.pptx
PDF
K to 12 classroom assessment ppt
DOC
article
PPT
An Introduction to Formative Assessment for School Teachers.ppt
PPTX
Assessment and Evaluation
PPTX
Modern Trends in Classroom Assessment.pptx
PPTX
Optimizing assessment systems to enhance teaching and learning
PPTX
2015 PGDT 423 (1).pptx
PPT
Appendix 4 Quality Of Assessment Practices Presentation
PPTX
Awsp Seattle Hilton May 2013
PDF
Test Development and Evaluation
PPT
Assessment (1)
Examination and Evaluation-ppt presentation.pptx
Tqf day 2 - assessment and feedback
Assessment, measurement and Evaluation .pptx
Classroom Assessment
Handouts-ClassroomAssessment-IP3.pdf
Assessment for higher education (for biology faculty seminar)
Seminar on trends and issues in assessment.pptx
3RD REPORTER Evaluating Best Practices in Teaching and Learning.pptx
K to 12 classroom assessment ppt
article
An Introduction to Formative Assessment for School Teachers.ppt
Assessment and Evaluation
Modern Trends in Classroom Assessment.pptx
Optimizing assessment systems to enhance teaching and learning
2015 PGDT 423 (1).pptx
Appendix 4 Quality Of Assessment Practices Presentation
Awsp Seattle Hilton May 2013
Test Development and Evaluation
Assessment (1)

Recently uploaded (20)

PDF
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
PDF
VCE English Exam - Section C Student Revision Booklet
PPTX
Cell Types and Its function , kingdom of life
PDF
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
PDF
Insiders guide to clinical Medicine.pdf
PPTX
Pharma ospi slides which help in ospi learning
PPTX
GDM (1) (1).pptx small presentation for students
PPTX
human mycosis Human fungal infections are called human mycosis..pptx
PPTX
Lesson notes of climatology university.
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PDF
O7-L3 Supply Chain Operations - ICLT Program
PDF
Classroom Observation Tools for Teachers
PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
PDF
Complications of Minimal Access Surgery at WLH
PDF
Pre independence Education in Inndia.pdf
PDF
FourierSeries-QuestionsWithAnswers(Part-A).pdf
PDF
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
PPTX
PPH.pptx obstetrics and gynecology in nursing
PDF
STATICS OF THE RIGID BODIES Hibbelers.pdf
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
VCE English Exam - Section C Student Revision Booklet
Cell Types and Its function , kingdom of life
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
Insiders guide to clinical Medicine.pdf
Pharma ospi slides which help in ospi learning
GDM (1) (1).pptx small presentation for students
human mycosis Human fungal infections are called human mycosis..pptx
Lesson notes of climatology university.
Supply Chain Operations Speaking Notes -ICLT Program
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
O7-L3 Supply Chain Operations - ICLT Program
Classroom Observation Tools for Teachers
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
Complications of Minimal Access Surgery at WLH
Pre independence Education in Inndia.pdf
FourierSeries-QuestionsWithAnswers(Part-A).pdf
3rd Neelam Sanjeevareddy Memorial Lecture.pdf
PPH.pptx obstetrics and gynecology in nursing
STATICS OF THE RIGID BODIES Hibbelers.pdf

Moving Beyond Student Ratings to Evaluate Teaching

  • 1. Moving Beyond Student Ratings to Evaluate Teaching Vicki L. Wise, PhD 1
  • 2. Where we are going in this presentation • What is effective teaching • Why collect evidence of effective teaching • What evidence to collect • How to bring evidence together for decision making • One form of evidence: Student Ratings of Instruction (SRIs) • Recognizing and addressing faculty concerns 2
  • 3. What is effective teaching Arreola (2007) defined five broad skill dimensions of college teaching: content expertise, instructional design skills, instructional delivery skills, instructional assessment skills, and course management skills. L. Dee Fink (2012) distilled Ken Bain’s research on What the Best College Teachers Do down to four fundamental tasks of teaching: Knowledge of subject matter, course design, course management, and interactions with students. The work of these researchers overlaps and reflects much of what is used in higher education. 3
  • 4. Why collect evidence of effective teaching 1. To improve teaching 2. To improve the student learning experience 3. To meet expectations for continuous employment 4
  • 5. What evidence to collect Student ratings are only one source of data; they must be combined with additional evidence so that administrators can make an informed judgment about teaching quality. -Hoyt and Pallett (1999). Appraising Teaching Effectiveness: Beyond Student Ratings. No one piece of evidence can stand on its own. Evidence of teaching quality needs to take into account multiple sources, as teaching is multidimensional. Moreover, the likelihood of obtaining reliable and valid data and making appropriate judgments are increased with more evidence. 5
  • 6. A portfolio (collection) of teaching effectiveness evidence could include at least three forms of evidence. Formative assessment or evaluation during a course, program or service provides information useful in improving learning or teaching while it is still occurring. Examples are student reflection exercises or teacher self reflections. Summative assessment or evaluation is conducted at the end of a program, service, or experience to make determinations of quality, worth, and meeting targeted outcomes. An example is a final grade or end of course evaluations. Direct methods of collecting information require a display of knowledge and skills. Indirect methods require a reflection on learning, behavior, and attitudes rather than to demonstrate. 6
  • 7. Evidence Direct or Indirect Student Ratings I Observations of Teaching D Student Interviews I Peer Ratings D Teacher Self Evaluations I Personal Goals for Teaching Improvement D Teaching Curriculum Map D Exit and Alumni Ratings I Employer Ratings I Teaching Scholarship D Student Learning Outcomes Results D Teaching Awards D Philosophy of Teaching D Innovative and Creative Teaching Techniques D Course Materials D Sample Student Work D Mentorship of Students and Faculty D Participation in Faculty Development Activities/ Consultations on Teaching D Scholarship of Teaching and Learning D 7
  • 8. How to bring evidence together for decision making Help, I have collected multiple forms of evidence and need to make sense of it all for decision making. • If a portfolio system is used, it could follow a basic template, and faculty would be asked to provide evidence of meeting specific criteria. Department chairs would use a rubric (checklist or scale) to rate the quality of evidence in pre-determined categories. • A rubric can also be used on each piece of evidence. The scale would specify what constitutes as “meeting expectations”. 8
  • 9. One Form of Evidence: Student Ratings of Instruction (SRIs) • Plenty of research supports that overall the evaluation of teaching generally provides reliable and valid evidence of teaching effectiveness. (meta-analytic review Gravestock & Gregor-Greenleaf, 2008) • Research does not support using teaching evaluation as a indicator of student learning though (Uttl, White, & Wong Gonzalez, 2016. Meta-analysis of faculty's teaching effectiveness: Student evaluation of teaching ratings and student learning are not related). • Significant concerns about the effective use of such evaluation systems continue to exist. 9
  • 10. Concerns about SRIs • Faculty: Anxiety provoking, outright hostility, suspicious of use, evaluations are biased, students not competent evaluators, ratings influenced by grade expectations, to name a few. • Administrators: Generally positive. Concerns about validity. Administrators’ own ideas of teaching and learning may bias their judgments of teaching effectiveness. • Students: Generally positive. Believe process is valid and they can be good evaluators. Unclear if and how data are used. 10
  • 11. 11 Faculty Concerns in Detail  Overly prominent role they play in personnel decisions (tenure, post-tenure review, etc.).  Reliance on them to make hiring, firing, and promotion decisions, rather than as performance improvement tools.  Initial poor ratings by students who later feel: "I-hated-you-for-forcing-us-to-learn-this- material-but-now-I'm-glad-you-did-thank-you-and-keep-doing-what-you're-doing" phenomenon.  Reliance on summative evaluation, rather than student ongoing reflection throughout the term. Not capturing instructional successes and failures in real time.  Problem of small numbers  Problem of response scale (inappropriate, not understood, not well defined).  Teaching effectiveness not well defined. Not measuring what is valued  Measure of student satisfaction not student learning  Not used for learning about value added to the student from being in the course.  Assessing popularity and/or easiness of instructor.  Not looking at visible outcomes of a course (e.g., increase in student conference presentations as a result of a course).  Bias in scores: o Gender bias in student scores, especially towards female instructors. o More difficult classes tend to get lower scores. o Younger appearing professors tend to get lower scores. o Introductory classes tend to get lower scores.
  • 12. Addressing Concerns 1. Follow best practices in instrument development 2. Standardize administration and policy 3. Plan for use of findings 12
  • 13. 1. Follow best practices in instrument development • Clearly define effective teaching as well as the components • Develop questions that measure these components • Measure behaviors/practices that are observable • Develop response scale • Develop two forms: one for formative assessment and one for summative • Test questions with the intended audience and use to modify each form • Determine how to analyze data appropriately, based on scale and use • Determine how results will be reported and used • Determine who will provide results and in what form • Collect contextualizing information, especially if comparisons are made (discipline, yrs. taught, …, course elective or required, course level) 13
  • 14. 2. Standardize administration and policy • Standardize administration process: • Set stage for importance (and use) with students and faculty; use a core set of questions; use motivators rather than incentives (consider equity in this online vs face-to-face); ensure anonymity; standardize method of delivery. • Suggest response rate needed: at least 10 students per course or 2/3 of class whichever is higher. May need to combine scores across terms for the same course. • Standardize analyses based on data type (i.e., treat ordinal variables appropriately). • Provide data interpretation and distribution guidelines • Determine audience for results and level of information needed. • Determine policy around data storage and access 14
  • 15. 3. Plan for use of findings • Develop recommendations for formative use to improve teaching. • Develop guidelines for summative use with other evidence to improve decision making. 15

Editor's Notes

  • #8: * All of these can be used both formatively and summatively
  • #11:   Before planning the process of implementing a student ratings of instruction (SRIs) process, consider assumptions and beliefs that faculty, administrators, and students bring to this process. Handout: Faculty Concerns
  • #13: Address concerns. Then consider processes that allow for inclusiveness of faculty, administrators, and students in planning and implementation.
  • #14: Address concerns. Then consider processes that allow for inclusiveness of faculty, administrators, and students in planning and implementation.
  • #15: Address concerns. Then consider processes that allow for inclusiveness of faculty, administrators, and students in planning and implementation.
  • #16: Address concerns. Then consider processes that allow for inclusiveness of faculty, administrators, and students in planning and implementation.