4
Most read
5
Most read
9
Most read
CHAPTER 6
ITEM ANALYSIS AND VALIDATION
TOPICS
Validation and Validity Reliability
Items Analysis: Difficulty
Index and Discrimination
Index
LEARNING OUTCOMES
At the end of the lesson, you should be able
to:
1. Explain the meaning of item analysis, item validity,
item difficulty, and discrimination index.
2. Determine the validity and reliability of given test
items.
3. Determine the quality of a test by its difficulty
index, discrimination index and the plausibility of
options.
Topic 1: ITEM ANALYSIS: DIFFICULTY INDEX AND
DISCRIMINATION INDEX
There are two importance characteristic of an item
that will be of interest to the teacher. These are: a) item
difficulty and b) discrimination index.
The difficulty of an item or item difficulty is defined
as the number of students who are able to answer the item
correctly divided by the total number of students.
Item difficulty = Number of student with correct answer
Total number of students
The item difficulty is usually expressed in percentage.
Example: What is the item difficulty index of an item if 25 students are unable to
answer it correctly while 75 answered it correctly?
Here, the total number of students is 1oo, hence the item difficulty index is 75/100 or
75%.
The following arbitrary rule is often used in the literature:
Range of Difficulty Index Interpretation Action
0 – 0.25 Difficulty Revise or discard
0.26 – 0.75 Right Difficulty Retain
0.76 – above Easy Revise or Discard
Index of discrimination = DU – DL
Where DU = Difficulty index of the upper 25% of the class
DL = Difficulty index of the lower 25% of the class
Example: Obtain the index of discrimination of an item if the upper 25% of the
class had a difficulty index of 0.60 (i.e 60% of the upper 25% got the correct
answer) while the lower 25% of the class had a difficulty index of 0.20.
Here, DU = 0.60 while DL = 0.20,
thus index discrimination = 0.60 – 0.20 = 0.40
Discrimination index is the difference between the proportion of the top
scorers who got an item correct and the proportion of the lowest scores who
got the item right. The discrimination index range is between -1 and +1.
Theoretically, the index of discrimination can range from -1.0 (when DU=O
and DL=1) to 1.0 (when DU=1 and D=O)
Example: Consider a multiple choice type of test of which the following
data were obtained.
Item Options
1 A B* C D
0 40 20 20 Total
0 15 5 0 Upper 25%
0 5 10 5 Lower 25%
Index Range Interpretation Action
-1.0 - -0.50 Can discriminate but item is
questionable
Discard
-0.51 – 0.45 Non-discriminating item Revise
0.46 – 1.0 Discriminating item Include
The correct answer is B. Let us compute the difficulty index and index of
discrimination:
Difficulty index = no. of students getting correct response
Total
= 40/100 = 40%, within a range of “good item”
The discrimination index can similarly be computed:
DU = no. of students in upper 25% with correct response
no. students in the upper 25%
= 15/20 = 0.75 or 75%
DL = no. of students in lower 25% with correct response
no. of students in the lower 25%
= 5/20 = 0.25% or 25%
Discrimination index = DU – DL = 0.75 – 0.25 = 0.50 or 50%
Thus, the item also has a “good discriminating power.”
More Sophisticated Discrimination Index
Item discrimination refers to the ability of an item to differentiate
among students on the basis of how well they know the material being
tested. Various hand calculation procedures have traditionally been used to
compare item responses to total test scores using high and low scoring
groups of students.
A good item is one that has good discriminating ability and has
sufficient level of difficult (not too difficult nor too easy).
At the end of the Item Analysis report, test items are listed according
to their degrees of difficulty (easy, medium, hard) and discrimination
(good, fair, poor). These distributions provide a quick overview of the test,
and can be used to identify items which are not performing well and which
can perhaps be improved od discarded.
The Item-Analysis Procedure
for Norm-provides the following
information:
1. The difficulty of the item
2. The discriminating power of
the item
3. The effectiveness of each
alternative
Some benefits derived from
Item Analysis are:
1. It provides useful
information for class
discussion of the test.
2. It provides data which help
students improve their
learning.
3. It provides insights and skills
that lead to the preparation
for better tests in the future.
TOPIC 2
VALIDATION AND VALIDITY
VALIDATION
Is the process of collecting and analyzing evidence to support the
meaningfulness and usefulness of the test.
VALIDITY
Is the extent to which a test measures what it purports to measure or as
referring to the appropriateness, correctness, meaningfulness, and
usefulness of the specific decisions a teacher makes based on the test
results.
TOPIC 2
VALIDATION AND VALIDITY
Content related evidence of validity
Refers to the content and format of the instrument.
Criterion-related evidence of validity
Refers to the relationship between scores obtained using the instrument
and scores obtained using one or more other tests (often called criterion).
Construct-related evidence of validity
Refers to the nature of the psychological construct or characteristic
being measured by the test.
TOPIC 3: REALIBILITY
Reliability refers to the consistency of the scores
obtained how consistent they are for reach individual
from one administration of an instrument to another
and from one set of items to another.
TOPIC 3: REALIBILITY
Reliability and validity are related concepts. If an
instrument is unreliable, it cannot get valid outcomes.
As reliability improves, validity may improve (or it
may not). However, if an instrument is shown
scientifically to be valid the it is almost certain that it
is also reliable.
TOPIC 3: REALIBILITY
Predictive validity compares the question with an outcome
assessed at a later time. An example of predictive validity is a
comparison of scores in the comparison of scores in the
National Achievement Test (NAT) with first semester grade
point average (GPA) in the college. Do NAT scores predict
college performance?
TOPIC 3: REALIBILITY
Construct validity refers to the ability of a test to
measure what it is supposed to measure. As
researcher, you intend to measure depression but you
actually measure anxiety so your research gets
comprised.
TOPIC 3: REALIBILITY
Reliability Interpretation
0.90 and above Excellent Reliability; at the level of the best standardized
tests
0.80 – 90 Very good for classroom test
0.70 – 0.80 Good for classroom test; in the range of the most. There
are probably a few items which could be improved.
0.60 – 0.70 Somewhat low. This test needs to be supplemented by
other measures (e.g. more tests) to determine grades.
There are probably some items which could be improved.
0.50 – 0.60 Suggests need for revision of test, unless it is quite short
(ten or fewer items). The test definitely needs to be
supplemented by other measures (e.g. more tests) for
grading.
0.50 or below Questionable reliability. This test should not contribute
heavily to the course grade, and it needs revision.
The following
table is a
standard
followed
almost
universally in
educational
test and
measurement.
Thank you

More Related Content

PPTX
Item analysis and validation
PPTX
Chapter 7 Performance-based assessment
PPTX
Controlling-Management Functions-Organisation Structure & Process
PPTX
The-Information-Age.pptx
PDF
Flexible Learning Environment Lesson 1: Online Distance Education and Commun...
PPTX
Lesson 3 Making Schools Inclusive.pptx
PPTX
Post-1986 Agrarian Reform.pptx
PPTX
Ausubel
Item analysis and validation
Chapter 7 Performance-based assessment
Controlling-Management Functions-Organisation Structure & Process
The-Information-Age.pptx
Flexible Learning Environment Lesson 1: Online Distance Education and Commun...
Lesson 3 Making Schools Inclusive.pptx
Post-1986 Agrarian Reform.pptx
Ausubel

What's hot (20)

PPTX
AUTHENTIC AND ALTERNATIVE ASSESSMENT METHODS
PPTX
Grading and reporting
PDF
Criteria in Choosing Appropriate Assessment Tool
PPTX
Assessment of learning1
PPTX
performance based -product oriented assessment
PPTX
Alternative-Response Test
PPTX
Affective Assessment
PPT
Lesson Planning for DepEd Teachers
PPTX
Chapter 2- Authentic assessment
PPTX
Chap.4 PROCESS-ORIENTED PERFORMANCE-BASED ASSESSMENT
PDF
Measurement Assessment and Evaluation in Outcomes-based Education (02 K12_HOT...
PPTX
Traditional assessment and non traditional assessment
PPTX
TTSCUR-GROUP-6-CURRICULUM-EVALUATION-FINAL (1).pptx
PPT
Interpretation of Assessment Results
PPTX
Item analysis2
PPTX
Asl 2 lecture 3
DOC
Kinds Of Learning Targets
PPTX
Norm or criterion referenced grading
PPTX
Norm referenced grading system
PPTX
Assessment of student learning 1
AUTHENTIC AND ALTERNATIVE ASSESSMENT METHODS
Grading and reporting
Criteria in Choosing Appropriate Assessment Tool
Assessment of learning1
performance based -product oriented assessment
Alternative-Response Test
Affective Assessment
Lesson Planning for DepEd Teachers
Chapter 2- Authentic assessment
Chap.4 PROCESS-ORIENTED PERFORMANCE-BASED ASSESSMENT
Measurement Assessment and Evaluation in Outcomes-based Education (02 K12_HOT...
Traditional assessment and non traditional assessment
TTSCUR-GROUP-6-CURRICULUM-EVALUATION-FINAL (1).pptx
Interpretation of Assessment Results
Item analysis2
Asl 2 lecture 3
Kinds Of Learning Targets
Norm or criterion referenced grading
Norm referenced grading system
Assessment of student learning 1
Ad

Similar to CHAPTER 6 Assessment of Learning 1 (20)

PPTX
Item Analysis 2023 for education in any area.pptx
PPTX
ANALYSIS & VALIDATION.pptxA broken clock that’s always 5 minutes late — it’s ...
PPTX
A Presentation for item analysis and validation
PPTX
my report for assessment in learning 1.pptx
PPTX
430660906-Item-Analysis.pptx
PPTX
ITEM-ANALYSIS-AND-VALIDATION-in-assessment-in-learning.pptx
PPTX
item analysis.pptx education pnc item analysis
PPTX
PPTX
CHAPTER-6-Assessment-in-Learning. .pptx
PPTX
Item Analysis
PPT
TEST ITEM ANALYSIS PRESENTATION 2022.ppt
PPT
Analyzing-Test-items-of-assessmentsg.ppt
PPTX
Item Analysis (a.k.a. Test Question Analysis) is a useful means of discoverin...
PPT
Analyzing and using test item data
PPT
Analyzing and using test item data
PPT
Analyzing and using test item data
PPT
Analyzing and using test item data
PPTX
Item analysis and validation
PDF
evaluations Item Analysis for teachers.pdf
PDF
Development of pyschologica test construction
Item Analysis 2023 for education in any area.pptx
ANALYSIS & VALIDATION.pptxA broken clock that’s always 5 minutes late — it’s ...
A Presentation for item analysis and validation
my report for assessment in learning 1.pptx
430660906-Item-Analysis.pptx
ITEM-ANALYSIS-AND-VALIDATION-in-assessment-in-learning.pptx
item analysis.pptx education pnc item analysis
CHAPTER-6-Assessment-in-Learning. .pptx
Item Analysis
TEST ITEM ANALYSIS PRESENTATION 2022.ppt
Analyzing-Test-items-of-assessmentsg.ppt
Item Analysis (a.k.a. Test Question Analysis) is a useful means of discoverin...
Analyzing and using test item data
Analyzing and using test item data
Analyzing and using test item data
Analyzing and using test item data
Item analysis and validation
evaluations Item Analysis for teachers.pdf
Development of pyschologica test construction
Ad

Recently uploaded (20)

PDF
David L Page_DCI Research Study Journey_how Methodology can inform one's prac...
PDF
My India Quiz Book_20210205121199924.pdf
PPTX
Module on health assessment of CHN. pptx
PDF
medical_surgical_nursing_10th_edition_ignatavicius_TEST_BANK_pdf.pdf
PDF
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
PDF
Skin Care and Cosmetic Ingredients Dictionary ( PDFDrive ).pdf
PPTX
Introduction to pro and eukaryotes and differences.pptx
PPTX
What’s under the hood: Parsing standardized learning content for AI
PDF
International_Financial_Reporting_Standa.pdf
PDF
Race Reva University – Shaping Future Leaders in Artificial Intelligence
PDF
Τίμαιος είναι φιλοσοφικός διάλογος του Πλάτωνα
PDF
FOISHS ANNUAL IMPLEMENTATION PLAN 2025.pdf
PDF
BP 704 T. NOVEL DRUG DELIVERY SYSTEMS (UNIT 1)
PPTX
Unit 4 Computer Architecture Multicore Processor.pptx
PDF
Mucosal Drug Delivery system_NDDS_BPHARMACY__SEM VII_PCI.pdf
PDF
LIFE & LIVING TRILOGY - PART (3) REALITY & MYSTERY.pdf
PDF
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
PDF
Empowerment Technology for Senior High School Guide
PDF
Complications of Minimal Access-Surgery.pdf
PDF
What if we spent less time fighting change, and more time building what’s rig...
David L Page_DCI Research Study Journey_how Methodology can inform one's prac...
My India Quiz Book_20210205121199924.pdf
Module on health assessment of CHN. pptx
medical_surgical_nursing_10th_edition_ignatavicius_TEST_BANK_pdf.pdf
BP 505 T. PHARMACEUTICAL JURISPRUDENCE (UNIT 1).pdf
Skin Care and Cosmetic Ingredients Dictionary ( PDFDrive ).pdf
Introduction to pro and eukaryotes and differences.pptx
What’s under the hood: Parsing standardized learning content for AI
International_Financial_Reporting_Standa.pdf
Race Reva University – Shaping Future Leaders in Artificial Intelligence
Τίμαιος είναι φιλοσοφικός διάλογος του Πλάτωνα
FOISHS ANNUAL IMPLEMENTATION PLAN 2025.pdf
BP 704 T. NOVEL DRUG DELIVERY SYSTEMS (UNIT 1)
Unit 4 Computer Architecture Multicore Processor.pptx
Mucosal Drug Delivery system_NDDS_BPHARMACY__SEM VII_PCI.pdf
LIFE & LIVING TRILOGY - PART (3) REALITY & MYSTERY.pdf
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
Empowerment Technology for Senior High School Guide
Complications of Minimal Access-Surgery.pdf
What if we spent less time fighting change, and more time building what’s rig...

CHAPTER 6 Assessment of Learning 1

  • 1. CHAPTER 6 ITEM ANALYSIS AND VALIDATION
  • 2. TOPICS Validation and Validity Reliability Items Analysis: Difficulty Index and Discrimination Index
  • 3. LEARNING OUTCOMES At the end of the lesson, you should be able to: 1. Explain the meaning of item analysis, item validity, item difficulty, and discrimination index. 2. Determine the validity and reliability of given test items. 3. Determine the quality of a test by its difficulty index, discrimination index and the plausibility of options.
  • 4. Topic 1: ITEM ANALYSIS: DIFFICULTY INDEX AND DISCRIMINATION INDEX There are two importance characteristic of an item that will be of interest to the teacher. These are: a) item difficulty and b) discrimination index. The difficulty of an item or item difficulty is defined as the number of students who are able to answer the item correctly divided by the total number of students.
  • 5. Item difficulty = Number of student with correct answer Total number of students The item difficulty is usually expressed in percentage. Example: What is the item difficulty index of an item if 25 students are unable to answer it correctly while 75 answered it correctly? Here, the total number of students is 1oo, hence the item difficulty index is 75/100 or 75%. The following arbitrary rule is often used in the literature: Range of Difficulty Index Interpretation Action 0 – 0.25 Difficulty Revise or discard 0.26 – 0.75 Right Difficulty Retain 0.76 – above Easy Revise or Discard
  • 6. Index of discrimination = DU – DL Where DU = Difficulty index of the upper 25% of the class DL = Difficulty index of the lower 25% of the class Example: Obtain the index of discrimination of an item if the upper 25% of the class had a difficulty index of 0.60 (i.e 60% of the upper 25% got the correct answer) while the lower 25% of the class had a difficulty index of 0.20. Here, DU = 0.60 while DL = 0.20, thus index discrimination = 0.60 – 0.20 = 0.40 Discrimination index is the difference between the proportion of the top scorers who got an item correct and the proportion of the lowest scores who got the item right. The discrimination index range is between -1 and +1.
  • 7. Theoretically, the index of discrimination can range from -1.0 (when DU=O and DL=1) to 1.0 (when DU=1 and D=O) Example: Consider a multiple choice type of test of which the following data were obtained. Item Options 1 A B* C D 0 40 20 20 Total 0 15 5 0 Upper 25% 0 5 10 5 Lower 25% Index Range Interpretation Action -1.0 - -0.50 Can discriminate but item is questionable Discard -0.51 – 0.45 Non-discriminating item Revise 0.46 – 1.0 Discriminating item Include
  • 8. The correct answer is B. Let us compute the difficulty index and index of discrimination: Difficulty index = no. of students getting correct response Total = 40/100 = 40%, within a range of “good item” The discrimination index can similarly be computed: DU = no. of students in upper 25% with correct response no. students in the upper 25% = 15/20 = 0.75 or 75% DL = no. of students in lower 25% with correct response no. of students in the lower 25% = 5/20 = 0.25% or 25% Discrimination index = DU – DL = 0.75 – 0.25 = 0.50 or 50% Thus, the item also has a “good discriminating power.”
  • 9. More Sophisticated Discrimination Index Item discrimination refers to the ability of an item to differentiate among students on the basis of how well they know the material being tested. Various hand calculation procedures have traditionally been used to compare item responses to total test scores using high and low scoring groups of students. A good item is one that has good discriminating ability and has sufficient level of difficult (not too difficult nor too easy). At the end of the Item Analysis report, test items are listed according to their degrees of difficulty (easy, medium, hard) and discrimination (good, fair, poor). These distributions provide a quick overview of the test, and can be used to identify items which are not performing well and which can perhaps be improved od discarded.
  • 10. The Item-Analysis Procedure for Norm-provides the following information: 1. The difficulty of the item 2. The discriminating power of the item 3. The effectiveness of each alternative Some benefits derived from Item Analysis are: 1. It provides useful information for class discussion of the test. 2. It provides data which help students improve their learning. 3. It provides insights and skills that lead to the preparation for better tests in the future.
  • 11. TOPIC 2 VALIDATION AND VALIDITY VALIDATION Is the process of collecting and analyzing evidence to support the meaningfulness and usefulness of the test. VALIDITY Is the extent to which a test measures what it purports to measure or as referring to the appropriateness, correctness, meaningfulness, and usefulness of the specific decisions a teacher makes based on the test results.
  • 12. TOPIC 2 VALIDATION AND VALIDITY Content related evidence of validity Refers to the content and format of the instrument. Criterion-related evidence of validity Refers to the relationship between scores obtained using the instrument and scores obtained using one or more other tests (often called criterion). Construct-related evidence of validity Refers to the nature of the psychological construct or characteristic being measured by the test.
  • 13. TOPIC 3: REALIBILITY Reliability refers to the consistency of the scores obtained how consistent they are for reach individual from one administration of an instrument to another and from one set of items to another.
  • 14. TOPIC 3: REALIBILITY Reliability and validity are related concepts. If an instrument is unreliable, it cannot get valid outcomes. As reliability improves, validity may improve (or it may not). However, if an instrument is shown scientifically to be valid the it is almost certain that it is also reliable.
  • 15. TOPIC 3: REALIBILITY Predictive validity compares the question with an outcome assessed at a later time. An example of predictive validity is a comparison of scores in the comparison of scores in the National Achievement Test (NAT) with first semester grade point average (GPA) in the college. Do NAT scores predict college performance?
  • 16. TOPIC 3: REALIBILITY Construct validity refers to the ability of a test to measure what it is supposed to measure. As researcher, you intend to measure depression but you actually measure anxiety so your research gets comprised.
  • 17. TOPIC 3: REALIBILITY Reliability Interpretation 0.90 and above Excellent Reliability; at the level of the best standardized tests 0.80 – 90 Very good for classroom test 0.70 – 0.80 Good for classroom test; in the range of the most. There are probably a few items which could be improved. 0.60 – 0.70 Somewhat low. This test needs to be supplemented by other measures (e.g. more tests) to determine grades. There are probably some items which could be improved. 0.50 – 0.60 Suggests need for revision of test, unless it is quite short (ten or fewer items). The test definitely needs to be supplemented by other measures (e.g. more tests) for grading. 0.50 or below Questionable reliability. This test should not contribute heavily to the course grade, and it needs revision. The following table is a standard followed almost universally in educational test and measurement.