RELIABILITY AND
VALIDITY OF
ASSESSMENT
ITEM ANALYSIS
• Item analysis has to be done, before a meaningful
and scientific inference about the test can be made
in terms of its validity, reliability, objectivity and
usability.
• Process which examines student’s response to
individual test items inorder to assess the quality
of the items and of the test as a whole.
• The tools include :
 Item difficulty.
 Item discrimination.
 Item distractors.
THE PURPOSES OF ITEM
ANALYSIS
• Improve test items and identify unfair items.
• Reveal which questions were most difficult.
• If a particular distracter is the most often
chosen answer, the item must be examined.
• To identify common misconceptions among
students about a particular concept.
• To improve the quality of tests.
• If items are too hard, teachers can adjust the
way they teach.
Item Difficulty
• It is the percentage of students taking the test
who answered the item correctly.
• Higher the value, easier the item.
• D= R/N X 100
• R – Number of pupils who answered the item
correctly.
• N – Total number of pupils who tried them.
Example
• Number of pupils answered item correctly
= 40
Total number of pupils who tried them = 50
40/50 X 100 = 80 %
Ideal difficulty levels for multiple-
choice items
• Format IdealDifficulty
• Five-response multiple-choice 70
• Four-response multiple-choice 74
• Three-response multiple-choice 77
• True-false 85
Item Discrimination
• Ability of an item to differentiate among the
students on the basis of how well they know the
material being tested.
• A good item discriminates between those who do
well on the test and those who do poorly.
• Higher the discrimination index,better the item.
DI = RU – RL/1/2 X N
• RU – Number of correct responses from the upper
group
• RL – number of correct responses from lower
group
• N – total number of pupils who tried them.
Example
• Total score – 60
Total sample – 50
Upper group – 25
Lower group – 25
22 – 10/1/2 X 50 = 0.29
Interpretation
• 0.4 or higher – very good items.
• 0.3 to 0.39 - good items.
• 0.20 to 0.29 –fairly good items.
• 0.19 or less – poor items.
• So the item in the example, is a fairly good
item.
Distractors
• Analyzing the distractors (i.e., incorrect
alternatives) is useful in determining the
relative usefulness of the decoys in each item.
• The alternatives are probably totally
implausible and therefore of little use as
decoys in multiple choice items.
• One way to study responses to distractors is with
a frequency table that tells you the proportion of
students who selected a given distractor.
• Remove or replace distractors selected by a few
or no students because students find them to be
implausible.
RELIABILITY
• Reliability is the degree to which an
assessment tool produces stable and consistent
results.
TYPES OF RELIABILITY
• Test – retest reliability
• Parallel reliability
• Inter – rater reliability
• Internal consistency
• Form equivalence (Alternate form)
Test-retest reliability
• Obtained by administering the same test twice
over a period of time to a group of
individuals.
• Scores from Time 1 and Time 2 can then be
correlated to evaluate the test for stability.
• Also known as temporal stability.
Parallel forms reliability
• It is obtained by administering different
versions of an assessment tool to the same
group of individuals.
• Scores from the two versions can then be
correlated to evaluate the consistency of results
across alternate versions.
Inter-rater reliability
• Used to assess the degree to which different
judges or raters agree in their assessment
decisions.
• Useful because human observers will not
necessarily interpret answers the same way.
Internal consistency reliability
• It used to evaluate the degree to which
different test items that probe the same
construct produce similar results.
• Two types are
Average inter-item correlation
Split-half reliability
Average inter-item correlation
• Obtained by taking all of the items on a test
that probe the same construct , determining
the correlation coefficient for each pair of
items, and finally taking the average of all
of these correlation coefficients.
 Split-half reliability
• “Splitting in half” all items of a test to form
two “sets” of items.
• The total score for each “set” is computed.
• Determining the correlation between the two
total “set” scores to obtain split half
reliability.
Form equivalence (Alternate form)
• Also known as alternate form reliability.
• Two different forms of test, based on the same
content, on one occasion to the same
examinees.
• Reliability is stated as correlation between
scores of Test 1 and Test 2.
VALIDITY
• An indication of how well an assessment
actually measures what it is supposed to
measure.
• Refers to the accuracy of an assessment.
• It is the veracity of an assessment instrument.
TYPES OF VALIDITY
• Face validity
• Construct validity
• Content validity
• Criterion related validity
• Formative validity
• Sampling validity
Face Validity
• Measure of the extent to which an examination
looks like an examination in the subject
concerned and at the appropriate level.
• Candidates, teachers and the public have
expectations as to what an examination looks
like and how it is conducted.
Construct Validity
• The extent to which an assessment
corresponds to other variables, as predicted by
some rationale or theory.
• It is also known as theoretical construct.
Content Validity
• The extent to which a measure adequately
represents all facets of a concept.
• It is the extent to which the content of the test
matches the instructional objectives
Criterion-Related validity
• Degree to which content on a test (predictor)
correlates with performance on relevant
criterion measures (concrete criterion in the
"real" world?)
Formative Validity
• When applied to outcomes assessment it is
used to assess how well a measure is able to
provide information to help improve the
program under study.
Sampling Validity
• It is similar to content validity.
• It ensures that the measure covers the broad
range of areas within the concept under study.
FACTORS THAT CAN LOWER
VALIDITY
• Unclear directions
• Difficult reading vocabulary and sentence
structure
• Ambiguity in statements
• Inadequate time limits
• Inappropriate level of difficulty
Cont’d
• Poorly constructed test items.
• Test items inappropriate for the outcomes
being measured.
• Tests that are too short.
• Administration and scoring.
Cont’d
• Improper arrangement of items (complex to
easy?).
• Identifiable patterns of answers.
• Teaching.
• Students .
• Nature of criterion.
WAYS TO IMPROVE
VALIDITY AND RELIABILITY
IMPROVING RELIABILITY
• First, calculate the item-test correlations and
rewrite or reject any that are too low.
• Second, look at the items that did correlate
well and write more like them. The longer the
test, the higher the reliability up to a point
IMPROVING VALIDITY
• Make sure your goals and objectives are
clearly defined and operationalized.
• Expectations of students should be written
down.
• Match your assessment measure to your goals
and objectives.
Cont’d
• Have the test reviewed by faculty at other
schools to obtain feedback from an outside
party who is less invested in the instrument.
• Get students involved; have the students look
over the assessment for troublesome work.
RELATIONSHIP BEWEEN
RELIABILITY AND VALIDITY
• The two do not necessarily go hand-in-hand.
• We can illustrate it as follows.
o Reliable but not valid - an archer who always
hits about the same place but not near the
bullseye.
o Valid but not reliable - archer who hits various
places centered around the bullseye, but not
very accurately.
o Neither reliable nor valid - an archer who hits
various places all off to the same side of the
bullseye.
Cont’d
o Both reliable and valid - archer who hits
consistently close to the bullseye.
• A valid assessment is always reliable, but a
reliable assessment is not necessarily valid.
FACTORS IN RESOLVING CONFLICTS
BETWEEN VALIDITY AND RELIABILITY
• Validity is paramount.
• Validity will not damage educational
effectiveness but excessive concern for
reliability or costs may do so.
• Staff costs are limited by the credits in the
workload planning system being used.
Cont’d
• Student time costs are limited by the planned
learning hours allocated to them.
• Reliability cannot be 100% for any one
assessment and may need to be compromised.
• Between-marker reliability can be improved
by marker training and monitoring.
Cont’d
• Clear, detailed criteria will maximise examiner
reliability and validity.
• Educationally effective coursework
assessments are often simultaneously designed
to prevent plagiarism detection.
Cont’d
• Where each student produces a number of similar
assignments they can be randomly sampled.
• Self and peer assessment can reduce staff costs
and uses as a learning activity.
• High-reliability assessment is costly and so
should be used only where it is critical.
Cont’d
• Programme-wide design of assessment can
avoid the worst of the conflicts.
• Designing good assessments is a creative,
challenging task that demands expertise in the
teaching of the subject,time and is improved
by peer support and review.
QUESTIONS
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt
RELIABILITY AND VALIDITY OF ASSESSMENT.ppt

More Related Content

PPTX
Norm Reference Tests.pptx
PDF
SKILL OF NON VERBAL CUES1.pdf
PPTX
Concept of test , measurement and evaluation
PPT
Summative and formative evaluation
PPTX
Assessment for Creative Learners
PPTX
Blue print
PPT
Competency Based Teacher Education and METHODS Assessing Teacher Competence.
PPTX
Criterion Reference Tests.pptx
Norm Reference Tests.pptx
SKILL OF NON VERBAL CUES1.pdf
Concept of test , measurement and evaluation
Summative and formative evaluation
Assessment for Creative Learners
Blue print
Competency Based Teacher Education and METHODS Assessing Teacher Competence.
Criterion Reference Tests.pptx

What's hot (20)

PPTX
Constructivist approach
PPTX
Criterion-referenced assessment
PPTX
Classification of Assessment based on Nature of Interpretation-Norms Referen...
PPTX
continouis & comprehensive evaluation
PPTX
Learning by imitation
PPTX
Concept of Test, Measurement, Assessment and Evaluation
PPTX
Meaning and concept of test, testing, measurement, assessment and evaluation
PPTX
Rating scale
PDF
Curriculum transaction
PDF
Structure of Teacher Education System in India by Garima Tandon
PPTX
Teacher's professional ethics and accountability.pptx
PPTX
Performance test
PPTX
Assessment Approaches.pptx
PPTX
Administration/Conducting the Test
PPTX
Evaluation, different types of test items
PPTX
Presentation validity
DOCX
Internal versus external examination
PPTX
BASIC OF MEASUREMENT & EVALUATION
PPTX
Existentialism
PPTX
Life Oriented Curriculum And Interdisciplinary Curriculum
Constructivist approach
Criterion-referenced assessment
Classification of Assessment based on Nature of Interpretation-Norms Referen...
continouis & comprehensive evaluation
Learning by imitation
Concept of Test, Measurement, Assessment and Evaluation
Meaning and concept of test, testing, measurement, assessment and evaluation
Rating scale
Curriculum transaction
Structure of Teacher Education System in India by Garima Tandon
Teacher's professional ethics and accountability.pptx
Performance test
Assessment Approaches.pptx
Administration/Conducting the Test
Evaluation, different types of test items
Presentation validity
Internal versus external examination
BASIC OF MEASUREMENT & EVALUATION
Existentialism
Life Oriented Curriculum And Interdisciplinary Curriculum
Ad

Similar to RELIABILITY AND VALIDITY OF ASSESSMENT.ppt (20)

PDF
Qualities of Good Test.pdf
PPTX
item analysis.pptx education pnc item analysis
PPTX
Validity of Assessment Tools
PPTX
DEVELOPMENT AND EVALUATION OF SCALES/INSTRUMENTS IN PSYCHIATRY
PPTX
Characteristics of a Good Test
PPT
Chapter24
PPTX
QUALITIES OF GOod research instrument-WPS Office.pptx
PPTX
Review Refresher-Assessment-in-Learning.pptx
PPTX
Psychological Test Construction and its steps
PPTX
Test appraisal
PPTX
ESE444/544 - Types of Assessment
PPTX
Standardization of a test by Dr. Neha Deo
PPTX
Reliability and Validity.pptx
PPTX
Good test , Reliability and Validity of a good test
PPTX
Reliablity
PPTX
Item analysis with spss software
PPTX
Monika seminar
PPTX
Monika seminar
PPTX
Quantitative analysis
PPTX
Assessment of Learning
Qualities of Good Test.pdf
item analysis.pptx education pnc item analysis
Validity of Assessment Tools
DEVELOPMENT AND EVALUATION OF SCALES/INSTRUMENTS IN PSYCHIATRY
Characteristics of a Good Test
Chapter24
QUALITIES OF GOod research instrument-WPS Office.pptx
Review Refresher-Assessment-in-Learning.pptx
Psychological Test Construction and its steps
Test appraisal
ESE444/544 - Types of Assessment
Standardization of a test by Dr. Neha Deo
Reliability and Validity.pptx
Good test , Reliability and Validity of a good test
Reliablity
Item analysis with spss software
Monika seminar
Monika seminar
Quantitative analysis
Assessment of Learning
Ad

More from JayLagman3 (20)

PPTX
determiningthedependentandindependentvariablesautosaved-221125062001-e11bf218...
PPT
Mathematics-Fibonacci-sequence-and-Patterns.ppt
PPTX
math drill (operations on integers).pptx
PPTX
inductive reasoning (logic) 12345.pptx
PPT
counting principles (permutation combination).ppt
PPTX
COUNTING (permutation and combination).pptx
PPTX
discrete MATHematics 100 QUIZ .pptx
PPTX
Let’s get started with with discrete math.pptx
PPTX
Set Theory (Discrete Mathematics).pptx
PDF
Discrete Mathematics - Sequences .pdf
PPT
theories and philosophical foundations.ppt
PPT
theories and philosophical foundations.ppt
PPTX
the teaching of mathematics .pptx
PPT
Representation of Functions as Power Series.ppt
PPT
Mathematics - Number Theory 11235813.ppt
PPT
teaching number and number sense.ppt
PPTX
Mathematics in the Intermediate Grades.pptx
PPT
Taylor_and_Maclaurin_Series (2).ppt
PPTX
conceptual framework of mathematics.pptx
PPTX
the general shape of matatag curriculum.
determiningthedependentandindependentvariablesautosaved-221125062001-e11bf218...
Mathematics-Fibonacci-sequence-and-Patterns.ppt
math drill (operations on integers).pptx
inductive reasoning (logic) 12345.pptx
counting principles (permutation combination).ppt
COUNTING (permutation and combination).pptx
discrete MATHematics 100 QUIZ .pptx
Let’s get started with with discrete math.pptx
Set Theory (Discrete Mathematics).pptx
Discrete Mathematics - Sequences .pdf
theories and philosophical foundations.ppt
theories and philosophical foundations.ppt
the teaching of mathematics .pptx
Representation of Functions as Power Series.ppt
Mathematics - Number Theory 11235813.ppt
teaching number and number sense.ppt
Mathematics in the Intermediate Grades.pptx
Taylor_and_Maclaurin_Series (2).ppt
conceptual framework of mathematics.pptx
the general shape of matatag curriculum.

Recently uploaded (20)

PPTX
Share_Module_2_Power_conflict_and_negotiation.pptx
PDF
1.3 FINAL REVISED K-10 PE and Health CG 2023 Grades 4-10 (1).pdf
PDF
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
PDF
Τίμαιος είναι φιλοσοφικός διάλογος του Πλάτωνα
PDF
Hazard Identification & Risk Assessment .pdf
PDF
Complications of Minimal Access-Surgery.pdf
PPTX
Virtual and Augmented Reality in Current Scenario
PDF
CISA (Certified Information Systems Auditor) Domain-Wise Summary.pdf
PDF
semiconductor packaging in vlsi design fab
PDF
International_Financial_Reporting_Standa.pdf
PDF
Skin Care and Cosmetic Ingredients Dictionary ( PDFDrive ).pdf
PDF
AI-driven educational solutions for real-life interventions in the Philippine...
PDF
MICROENCAPSULATION_NDDS_BPHARMACY__SEM VII_PCI .pdf
PDF
Vision Prelims GS PYQ Analysis 2011-2022 www.upscpdf.com.pdf
PDF
MBA _Common_ 2nd year Syllabus _2021-22_.pdf
PDF
HVAC Specification 2024 according to central public works department
PDF
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
PPTX
ELIAS-SEZIURE AND EPilepsy semmioan session.pptx
PPTX
Introduction to pro and eukaryotes and differences.pptx
DOCX
Cambridge-Practice-Tests-for-IELTS-12.docx
Share_Module_2_Power_conflict_and_negotiation.pptx
1.3 FINAL REVISED K-10 PE and Health CG 2023 Grades 4-10 (1).pdf
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
Τίμαιος είναι φιλοσοφικός διάλογος του Πλάτωνα
Hazard Identification & Risk Assessment .pdf
Complications of Minimal Access-Surgery.pdf
Virtual and Augmented Reality in Current Scenario
CISA (Certified Information Systems Auditor) Domain-Wise Summary.pdf
semiconductor packaging in vlsi design fab
International_Financial_Reporting_Standa.pdf
Skin Care and Cosmetic Ingredients Dictionary ( PDFDrive ).pdf
AI-driven educational solutions for real-life interventions in the Philippine...
MICROENCAPSULATION_NDDS_BPHARMACY__SEM VII_PCI .pdf
Vision Prelims GS PYQ Analysis 2011-2022 www.upscpdf.com.pdf
MBA _Common_ 2nd year Syllabus _2021-22_.pdf
HVAC Specification 2024 according to central public works department
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
ELIAS-SEZIURE AND EPilepsy semmioan session.pptx
Introduction to pro and eukaryotes and differences.pptx
Cambridge-Practice-Tests-for-IELTS-12.docx

RELIABILITY AND VALIDITY OF ASSESSMENT.ppt

  • 3. • Item analysis has to be done, before a meaningful and scientific inference about the test can be made in terms of its validity, reliability, objectivity and usability. • Process which examines student’s response to individual test items inorder to assess the quality of the items and of the test as a whole.
  • 4. • The tools include :  Item difficulty.  Item discrimination.  Item distractors.
  • 5. THE PURPOSES OF ITEM ANALYSIS
  • 6. • Improve test items and identify unfair items. • Reveal which questions were most difficult. • If a particular distracter is the most often chosen answer, the item must be examined. • To identify common misconceptions among students about a particular concept.
  • 7. • To improve the quality of tests. • If items are too hard, teachers can adjust the way they teach.
  • 9. • It is the percentage of students taking the test who answered the item correctly. • Higher the value, easier the item.
  • 10. • D= R/N X 100 • R – Number of pupils who answered the item correctly. • N – Total number of pupils who tried them.
  • 11. Example • Number of pupils answered item correctly = 40 Total number of pupils who tried them = 50 40/50 X 100 = 80 %
  • 12. Ideal difficulty levels for multiple- choice items • Format IdealDifficulty • Five-response multiple-choice 70 • Four-response multiple-choice 74 • Three-response multiple-choice 77 • True-false 85
  • 14. • Ability of an item to differentiate among the students on the basis of how well they know the material being tested. • A good item discriminates between those who do well on the test and those who do poorly. • Higher the discrimination index,better the item.
  • 15. DI = RU – RL/1/2 X N • RU – Number of correct responses from the upper group • RL – number of correct responses from lower group • N – total number of pupils who tried them.
  • 16. Example • Total score – 60 Total sample – 50 Upper group – 25 Lower group – 25 22 – 10/1/2 X 50 = 0.29
  • 17. Interpretation • 0.4 or higher – very good items. • 0.3 to 0.39 - good items. • 0.20 to 0.29 –fairly good items. • 0.19 or less – poor items. • So the item in the example, is a fairly good item.
  • 19. • Analyzing the distractors (i.e., incorrect alternatives) is useful in determining the relative usefulness of the decoys in each item. • The alternatives are probably totally implausible and therefore of little use as decoys in multiple choice items.
  • 20. • One way to study responses to distractors is with a frequency table that tells you the proportion of students who selected a given distractor. • Remove or replace distractors selected by a few or no students because students find them to be implausible.
  • 22. • Reliability is the degree to which an assessment tool produces stable and consistent results.
  • 24. • Test – retest reliability • Parallel reliability • Inter – rater reliability • Internal consistency • Form equivalence (Alternate form)
  • 26. • Obtained by administering the same test twice over a period of time to a group of individuals. • Scores from Time 1 and Time 2 can then be correlated to evaluate the test for stability. • Also known as temporal stability.
  • 28. • It is obtained by administering different versions of an assessment tool to the same group of individuals. • Scores from the two versions can then be correlated to evaluate the consistency of results across alternate versions.
  • 30. • Used to assess the degree to which different judges or raters agree in their assessment decisions. • Useful because human observers will not necessarily interpret answers the same way.
  • 32. • It used to evaluate the degree to which different test items that probe the same construct produce similar results. • Two types are Average inter-item correlation Split-half reliability
  • 33. Average inter-item correlation • Obtained by taking all of the items on a test that probe the same construct , determining the correlation coefficient for each pair of items, and finally taking the average of all of these correlation coefficients.
  • 34.  Split-half reliability • “Splitting in half” all items of a test to form two “sets” of items. • The total score for each “set” is computed. • Determining the correlation between the two total “set” scores to obtain split half reliability.
  • 36. • Also known as alternate form reliability. • Two different forms of test, based on the same content, on one occasion to the same examinees. • Reliability is stated as correlation between scores of Test 1 and Test 2.
  • 38. • An indication of how well an assessment actually measures what it is supposed to measure. • Refers to the accuracy of an assessment. • It is the veracity of an assessment instrument.
  • 40. • Face validity • Construct validity • Content validity • Criterion related validity • Formative validity • Sampling validity
  • 42. • Measure of the extent to which an examination looks like an examination in the subject concerned and at the appropriate level. • Candidates, teachers and the public have expectations as to what an examination looks like and how it is conducted.
  • 44. • The extent to which an assessment corresponds to other variables, as predicted by some rationale or theory. • It is also known as theoretical construct.
  • 46. • The extent to which a measure adequately represents all facets of a concept. • It is the extent to which the content of the test matches the instructional objectives
  • 48. • Degree to which content on a test (predictor) correlates with performance on relevant criterion measures (concrete criterion in the "real" world?)
  • 50. • When applied to outcomes assessment it is used to assess how well a measure is able to provide information to help improve the program under study.
  • 52. • It is similar to content validity. • It ensures that the measure covers the broad range of areas within the concept under study.
  • 53. FACTORS THAT CAN LOWER VALIDITY
  • 54. • Unclear directions • Difficult reading vocabulary and sentence structure • Ambiguity in statements • Inadequate time limits • Inappropriate level of difficulty
  • 55. Cont’d • Poorly constructed test items. • Test items inappropriate for the outcomes being measured. • Tests that are too short. • Administration and scoring.
  • 56. Cont’d • Improper arrangement of items (complex to easy?). • Identifiable patterns of answers. • Teaching. • Students . • Nature of criterion.
  • 57. WAYS TO IMPROVE VALIDITY AND RELIABILITY
  • 58. IMPROVING RELIABILITY • First, calculate the item-test correlations and rewrite or reject any that are too low. • Second, look at the items that did correlate well and write more like them. The longer the test, the higher the reliability up to a point
  • 59. IMPROVING VALIDITY • Make sure your goals and objectives are clearly defined and operationalized. • Expectations of students should be written down. • Match your assessment measure to your goals and objectives.
  • 60. Cont’d • Have the test reviewed by faculty at other schools to obtain feedback from an outside party who is less invested in the instrument. • Get students involved; have the students look over the assessment for troublesome work.
  • 62. • The two do not necessarily go hand-in-hand. • We can illustrate it as follows. o Reliable but not valid - an archer who always hits about the same place but not near the bullseye.
  • 63. o Valid but not reliable - archer who hits various places centered around the bullseye, but not very accurately. o Neither reliable nor valid - an archer who hits various places all off to the same side of the bullseye.
  • 64. Cont’d o Both reliable and valid - archer who hits consistently close to the bullseye. • A valid assessment is always reliable, but a reliable assessment is not necessarily valid.
  • 65. FACTORS IN RESOLVING CONFLICTS BETWEEN VALIDITY AND RELIABILITY
  • 66. • Validity is paramount. • Validity will not damage educational effectiveness but excessive concern for reliability or costs may do so. • Staff costs are limited by the credits in the workload planning system being used.
  • 67. Cont’d • Student time costs are limited by the planned learning hours allocated to them. • Reliability cannot be 100% for any one assessment and may need to be compromised. • Between-marker reliability can be improved by marker training and monitoring.
  • 68. Cont’d • Clear, detailed criteria will maximise examiner reliability and validity. • Educationally effective coursework assessments are often simultaneously designed to prevent plagiarism detection.
  • 69. Cont’d • Where each student produces a number of similar assignments they can be randomly sampled. • Self and peer assessment can reduce staff costs and uses as a learning activity. • High-reliability assessment is costly and so should be used only where it is critical.
  • 70. Cont’d • Programme-wide design of assessment can avoid the worst of the conflicts. • Designing good assessments is a creative, challenging task that demands expertise in the teaching of the subject,time and is improved by peer support and review.