SlideShare a Scribd company logo
Reliability and validity
Why do we need Reliability & Validity?
           (Measurement Error)
A participant’s score on a particular measure consists of 2
components:
  Observed score = True score + Measurement Error
True Score = score that the participant would have
obtained if measurement was perfect—i.e., we were able
to measure without error
Measurement Error = the component of the observed
score that is the result of factors that distort the score from
its true value
Factors that Influence
   Measurement Error

• Transient states of the participants:
  (transient mood, health, fatigue-level, etc.)
• Stable attributes of the participants:
  (individual differences in intelligence,
  personality, motivation, etc.)
• Situational factors of the research setting:
  (room temperature, lighting, crowding, etc.)
Characteristics of
      Measures and
      Manipulations
Precision and clarity of operational
definitions
Training of observers
Number of independent observations
on which a score is based (more is
better?)
Measures that induce fatigue or fear
Actual Mistakes

  Equipment malfunction
  Errors in recording behaviors by observers
  Confusing response formats for self-reports
  Data entry errors



Measurement error undermines the reliability
(repeatability) of the measures we use
Reliability

• The reliability of a measure is an
  inverse function of measurement error:
• The more error, the less reliable the
  measure
• Reliable measures provide consistent
  measurement from occasion to
  occasion
Estimating Reliability

Total Variance       =    Variance due     +   Variance due
in a set of scores        to true scores          to error

Reliability          =    True-score       /    Total
                          Variance             Variance

Reliability can range from 0 to 1.0
When a reliability coefficient equals 0, the scores reflect
nothing but measurement error
Rule of Thumb: measures with reliability coefficients of
70% or greater have acceptable reliability
Different Methods for
   Assessing Reliability


Test-Retest Reliability
Inter-rater Reliability
Internal Consistency Reliability
Test-Retest Reliability

Test-retest reliability refers to the
consistency of participant’s responses
over time (usually a few weeks, why?)
Assumes the characteristic being
measured is stable over time—not
expected to change between test and
retest
Inter-rater Reliability

If a measurement involves behavioral
ratings by an observer/rater, we would
expect consistency among raters for a
reliable measure
Best to use at least 2 independent
raters, ‘blind’ to the ratings of other
observers
Precise operational definitions and well-
trained observers improve inter-rater
reliability
Internal Consistency
          Reliability
• Relevant for measures that consist of more
  than 1 item (e.g., total scores on scales, or
  when several behavioral observations are
  used to obtain a single score)
• Internal consistency refers to inter-item
  reliability, and assesses the degree of
  consistency among the items in a scale, or
  the different observations used to derive a
  score
• Want to be sure that all the items (or
  observations) are measuring the same
  construct
Estimates of Internal
           Consistency

• Item-total score consistency
• Split-half reliability: randomly divide items
  into 2 subsets and examine the consistency
  in total scores across the 2 subsets (any
  drawbacks?)
• Cronbach’s Alpha: conceptually, it is the
  average consistency across all possible split-
  half reliabilities
• Cronbach’s Alpha can be directly computed
  from data
Estimating the Validity of a
          Measure

• A good measure must not only be reliable,
  but also valid
• A valid measure measures what it is intended
  to measure
• Validity is not a property of a measure, but an
  indication of the extent to which an
  assessment measures a particular construct
  in a particular context—thus a measure may
  be valid for one purpose but not another
• A measure cannot be valid unless it is
  reliable, but a reliable measure may not be
  valid
Estimating Validity
Like reliability, validity is not absolute
Validity is the degree to which variability
(individual differences) in participant’s
scores on a particular measure, reflect
individual differences in the
characteristic or construct we want to
measure
Three types of measurement validity:
          Face Validity
          Construct Validity
Face Validity

• Face validity refers to the extent to which a
  measure ‘appears’ to measure what it is
  supposed to measure
• Not statistical—involves the judgment of the
  researcher (and the participants)
• A measure has face validity—’if people think
  it does’
• Just because a measure has face validity
  does not ensure that it is a valid measure
  (and measures lacking face validity can be
  valid)
Construct Validity
Most scientific investigations involve
hypothetical constructs—entities that
cannot be directly observed but are
inferred from empirical evidence (e.g.,
intelligence)
Construct validity is assessed by
studying the relationships between the
measure of a construct and scores on
measures of other constructs
We assess construct validity by seeing
whether a particular measure relates as
it should to other measures
Self-Esteem Example

• Scores on a measure of self-esteem
 should be positively related to
 measures of confidence and optimism

• But, negatively related to measures of
 insecurity and anxiety
Convergent and
    Discriminant Validity

• To have construct validity, a measure
  should both:
• Correlate with other measures that it
  should be related to (convergent
  validity)
• And, not correlate with measures that it
  should not correlate with (discriminant
  validity)
Criterion-Related
•                   Validity
    Refers to the extent to which a measure
  distinguishes participants on the basis of a
  particular behavioral criterion
• The Scholastic Aptitude Test (SAT) is valid to the
  extent that it distinguishes between students that
  do well in college versus those that do not
• A valid measure of marital conflict should
  correlate with behavioral observations (e.g.,
  number of fights)
• A valid measure of depressive symptoms should
  distinguish between subjects in treatment for
  depression and those who are not in treatment
Two Types of Criterion-
      Related Validity
  Concurrent validity
      measure and criterion are assessed at the
      same time
  Predictive validity
      elapsed time between the administration
      of the measure to be validated and the
  criterion is a relatively long period
  (e.g., months or years)
Predictive validity refers to a measure’s ability
  to distinguish participants on a relevant
  behavioral criterion at some point in the future
SAT Example

• High school seniors who score high on
  the the SAT are better prepared for
  college than low scorers (concurrent
  validity)
• Probably of greater interest to college
  admissions administrators, SAT scores
  predict academic performance four
  years later (predictive validity)
Thank you.

More Related Content

PPTX
Reliability
PPTX
Monika seminar
PPT
Construct Validity
PPTX
Week 8 & 9 - Validity and Reliability
PDF
Validity & reliability
PPTX
Validity and Reliability
PPTX
Validity.pptx
PPTX
Reliability & Validity
Reliability
Monika seminar
Construct Validity
Week 8 & 9 - Validity and Reliability
Validity & reliability
Validity and Reliability
Validity.pptx
Reliability & Validity

What's hot (20)

PPTX
Classical Test Theory (CTT)- By Dr. Jai Singh
PPTX
PPT
Reliability
PPTX
RELIABILITY AND VALIDITY
PPTX
Reliability types
PPTX
QUASI EXPERIMENTAL DESIGN
PPTX
PPTX
Reliability (assessment of student learning I)
PPTX
Repeated Measures ANOVA
PPTX
Validity and its types
PPTX
Presentation on validity and reliability in research methods
PPTX
Correlation research
PPTX
Reliability & validity
PPT
Mann Whitney U test
PPT
Reliability and validity
PDF
Point Biserial Correlation - Thiyagu
PPTX
Item writing
PPTX
Reliability for testing and assessment
PPT
Standardized tests
Classical Test Theory (CTT)- By Dr. Jai Singh
Reliability
RELIABILITY AND VALIDITY
Reliability types
QUASI EXPERIMENTAL DESIGN
Reliability (assessment of student learning I)
Repeated Measures ANOVA
Validity and its types
Presentation on validity and reliability in research methods
Correlation research
Reliability & validity
Mann Whitney U test
Reliability and validity
Point Biserial Correlation - Thiyagu
Item writing
Reliability for testing and assessment
Standardized tests
Ad

Viewers also liked (20)

PPT
Presentation Validity & Reliability
PPT
Validity, its types, measurement & factors.
PDF
Reliability, validity, generalizability and the use of multi-item scales
PPT
Louzel Report - Reliability & validity
PPTX
Validity and reliability of questionnaires
PPTX
Validity, reliability & practicality
PPTX
Week 9 validity and reliability
PPTX
Validity & reliability seminar
PPTX
Validity & reliability an interesting powerpoint slide i created
PPT
Reliability and validity
PPTX
validity its types and importance
PPTX
Inter item reliability with surveys
PDF
8. validity and reliability of research instruments
PPTX
ALT Approaches for Reliability
PPT
Reliability and validity
PPT
The ph d and beyond the apprenticeship model of learning
PPT
Apprenticeship patterns
PPT
Lesson 03 chapter 6 sampling
PDF
Dynamic Factor Rotation
PPTX
Level of Measurement, Frequency Distribution,Stem & Leaf
Presentation Validity & Reliability
Validity, its types, measurement & factors.
Reliability, validity, generalizability and the use of multi-item scales
Louzel Report - Reliability & validity
Validity and reliability of questionnaires
Validity, reliability & practicality
Week 9 validity and reliability
Validity & reliability seminar
Validity & reliability an interesting powerpoint slide i created
Reliability and validity
validity its types and importance
Inter item reliability with surveys
8. validity and reliability of research instruments
ALT Approaches for Reliability
Reliability and validity
The ph d and beyond the apprenticeship model of learning
Apprenticeship patterns
Lesson 03 chapter 6 sampling
Dynamic Factor Rotation
Level of Measurement, Frequency Distribution,Stem & Leaf
Ad

Similar to Reliability & validity (20)

PPT
Characteristics of a good test
PPTX
unit 2.6.pptx
PPT
Evaluation of Measurement Instruments.ppt
PPT
Research Instrument Validity & reliability.ppt
PPTX
Research methodology measurement
PPTX
Research Methodology3_Measurement.pptx
PDF
Questionnaire and Instrument validity
PPTX
MEASUREMENT SCALE AND RELIABILITY IN RESEARCH
PPTX
MEASUREMENT METHODS AND SCALING FOR BUSINESS RESEARCH
PPTX
Reliability and validity- research-for BSC/PBBSC AND MSC NURSING
PPTX
PPTX
Validity
PPT
Research Methods.ppt Criterion for Good measurement. Goodness of measurement
PPT
Chapter 13 Measuremen and Scaling Concept Slides.ppt
PPTX
Measurement and scaling
PPT
251109 rm-c.s.-assessing measurement quality in quantitative studies
PDF
reliablity and validity in social sciences research
PPT
Topic 7 measurement in research
PPTX
Rep
PDF
RM-3 SCY.pdf
Characteristics of a good test
unit 2.6.pptx
Evaluation of Measurement Instruments.ppt
Research Instrument Validity & reliability.ppt
Research methodology measurement
Research Methodology3_Measurement.pptx
Questionnaire and Instrument validity
MEASUREMENT SCALE AND RELIABILITY IN RESEARCH
MEASUREMENT METHODS AND SCALING FOR BUSINESS RESEARCH
Reliability and validity- research-for BSC/PBBSC AND MSC NURSING
Validity
Research Methods.ppt Criterion for Good measurement. Goodness of measurement
Chapter 13 Measuremen and Scaling Concept Slides.ppt
Measurement and scaling
251109 rm-c.s.-assessing measurement quality in quantitative studies
reliablity and validity in social sciences research
Topic 7 measurement in research
Rep
RM-3 SCY.pdf

More from shefali84 (7)

PDF
100 short moral stories
DOCX
Doc1
PPT
11 routing
PPT
Mbti for mana 5338 abbrev. 06232010
PPT
Leadershipppt 090414175137-phpapp01
PPT
Strategic management
PPT
Basic concepts of measurement
100 short moral stories
Doc1
11 routing
Mbti for mana 5338 abbrev. 06232010
Leadershipppt 090414175137-phpapp01
Strategic management
Basic concepts of measurement

Recently uploaded (20)

PDF
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
PPTX
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
PPTX
PPT- ENG7_QUARTER1_LESSON1_WEEK1. IMAGERY -DESCRIPTIONS pptx.pptx
PDF
Supply Chain Operations Speaking Notes -ICLT Program
PDF
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
PPTX
Cell Structure & Organelles in detailed.
PPTX
Final Presentation General Medicine 03-08-2024.pptx
PPTX
History, Philosophy and sociology of education (1).pptx
PDF
RTP_AR_KS1_Tutor's Guide_English [FOR REPRODUCTION].pdf
PPTX
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
PDF
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
PDF
RMMM.pdf make it easy to upload and study
PDF
Complications of Minimal Access Surgery at WLH
PDF
01-Introduction-to-Information-Management.pdf
PDF
A GUIDE TO GENETICS FOR UNDERGRADUATE MEDICAL STUDENTS
PDF
Updated Idioms and Phrasal Verbs in English subject
PDF
A systematic review of self-coping strategies used by university students to ...
PPTX
Final Presentation General Medicine 03-08-2024.pptx
PDF
Computing-Curriculum for Schools in Ghana
PPTX
202450812 BayCHI UCSC-SV 20250812 v17.pptx
ChatGPT for Dummies - Pam Baker Ccesa007.pdf
1st Inaugural Professorial Lecture held on 19th February 2020 (Governance and...
PPT- ENG7_QUARTER1_LESSON1_WEEK1. IMAGERY -DESCRIPTIONS pptx.pptx
Supply Chain Operations Speaking Notes -ICLT Program
The Lost Whites of Pakistan by Jahanzaib Mughal.pdf
Cell Structure & Organelles in detailed.
Final Presentation General Medicine 03-08-2024.pptx
History, Philosophy and sociology of education (1).pptx
RTP_AR_KS1_Tutor's Guide_English [FOR REPRODUCTION].pdf
school management -TNTEU- B.Ed., Semester II Unit 1.pptx
Black Hat USA 2025 - Micro ICS Summit - ICS/OT Threat Landscape
RMMM.pdf make it easy to upload and study
Complications of Minimal Access Surgery at WLH
01-Introduction-to-Information-Management.pdf
A GUIDE TO GENETICS FOR UNDERGRADUATE MEDICAL STUDENTS
Updated Idioms and Phrasal Verbs in English subject
A systematic review of self-coping strategies used by university students to ...
Final Presentation General Medicine 03-08-2024.pptx
Computing-Curriculum for Schools in Ghana
202450812 BayCHI UCSC-SV 20250812 v17.pptx

Reliability & validity

  • 2. Why do we need Reliability & Validity? (Measurement Error) A participant’s score on a particular measure consists of 2 components: Observed score = True score + Measurement Error True Score = score that the participant would have obtained if measurement was perfect—i.e., we were able to measure without error Measurement Error = the component of the observed score that is the result of factors that distort the score from its true value
  • 3. Factors that Influence Measurement Error • Transient states of the participants: (transient mood, health, fatigue-level, etc.) • Stable attributes of the participants: (individual differences in intelligence, personality, motivation, etc.) • Situational factors of the research setting: (room temperature, lighting, crowding, etc.)
  • 4. Characteristics of Measures and Manipulations Precision and clarity of operational definitions Training of observers Number of independent observations on which a score is based (more is better?) Measures that induce fatigue or fear
  • 5. Actual Mistakes Equipment malfunction Errors in recording behaviors by observers Confusing response formats for self-reports Data entry errors Measurement error undermines the reliability (repeatability) of the measures we use
  • 6. Reliability • The reliability of a measure is an inverse function of measurement error: • The more error, the less reliable the measure • Reliable measures provide consistent measurement from occasion to occasion
  • 7. Estimating Reliability Total Variance = Variance due + Variance due in a set of scores to true scores to error Reliability = True-score / Total Variance Variance Reliability can range from 0 to 1.0 When a reliability coefficient equals 0, the scores reflect nothing but measurement error Rule of Thumb: measures with reliability coefficients of 70% or greater have acceptable reliability
  • 8. Different Methods for Assessing Reliability Test-Retest Reliability Inter-rater Reliability Internal Consistency Reliability
  • 9. Test-Retest Reliability Test-retest reliability refers to the consistency of participant’s responses over time (usually a few weeks, why?) Assumes the characteristic being measured is stable over time—not expected to change between test and retest
  • 10. Inter-rater Reliability If a measurement involves behavioral ratings by an observer/rater, we would expect consistency among raters for a reliable measure Best to use at least 2 independent raters, ‘blind’ to the ratings of other observers Precise operational definitions and well- trained observers improve inter-rater reliability
  • 11. Internal Consistency Reliability • Relevant for measures that consist of more than 1 item (e.g., total scores on scales, or when several behavioral observations are used to obtain a single score) • Internal consistency refers to inter-item reliability, and assesses the degree of consistency among the items in a scale, or the different observations used to derive a score • Want to be sure that all the items (or observations) are measuring the same construct
  • 12. Estimates of Internal Consistency • Item-total score consistency • Split-half reliability: randomly divide items into 2 subsets and examine the consistency in total scores across the 2 subsets (any drawbacks?) • Cronbach’s Alpha: conceptually, it is the average consistency across all possible split- half reliabilities • Cronbach’s Alpha can be directly computed from data
  • 13. Estimating the Validity of a Measure • A good measure must not only be reliable, but also valid • A valid measure measures what it is intended to measure • Validity is not a property of a measure, but an indication of the extent to which an assessment measures a particular construct in a particular context—thus a measure may be valid for one purpose but not another • A measure cannot be valid unless it is reliable, but a reliable measure may not be valid
  • 14. Estimating Validity Like reliability, validity is not absolute Validity is the degree to which variability (individual differences) in participant’s scores on a particular measure, reflect individual differences in the characteristic or construct we want to measure Three types of measurement validity: Face Validity Construct Validity
  • 15. Face Validity • Face validity refers to the extent to which a measure ‘appears’ to measure what it is supposed to measure • Not statistical—involves the judgment of the researcher (and the participants) • A measure has face validity—’if people think it does’ • Just because a measure has face validity does not ensure that it is a valid measure (and measures lacking face validity can be valid)
  • 16. Construct Validity Most scientific investigations involve hypothetical constructs—entities that cannot be directly observed but are inferred from empirical evidence (e.g., intelligence) Construct validity is assessed by studying the relationships between the measure of a construct and scores on measures of other constructs We assess construct validity by seeing whether a particular measure relates as it should to other measures
  • 17. Self-Esteem Example • Scores on a measure of self-esteem should be positively related to measures of confidence and optimism • But, negatively related to measures of insecurity and anxiety
  • 18. Convergent and Discriminant Validity • To have construct validity, a measure should both: • Correlate with other measures that it should be related to (convergent validity) • And, not correlate with measures that it should not correlate with (discriminant validity)
  • 19. Criterion-Related • Validity Refers to the extent to which a measure distinguishes participants on the basis of a particular behavioral criterion • The Scholastic Aptitude Test (SAT) is valid to the extent that it distinguishes between students that do well in college versus those that do not • A valid measure of marital conflict should correlate with behavioral observations (e.g., number of fights) • A valid measure of depressive symptoms should distinguish between subjects in treatment for depression and those who are not in treatment
  • 20. Two Types of Criterion- Related Validity Concurrent validity measure and criterion are assessed at the same time Predictive validity elapsed time between the administration of the measure to be validated and the criterion is a relatively long period (e.g., months or years) Predictive validity refers to a measure’s ability to distinguish participants on a relevant behavioral criterion at some point in the future
  • 21. SAT Example • High school seniors who score high on the the SAT are better prepared for college than low scorers (concurrent validity) • Probably of greater interest to college admissions administrators, SAT scores predict academic performance four years later (predictive validity)