SlideShare a Scribd company logo
Darlene Goldstein 29 January 2003
Receiver Operating Characteristic
Methodology
Outline
• Introduction
• Hypothesis testing
• ROC curve
• Area under the ROC curve (AUC)
• Examples using ROC
• Concluding remarks
Introduction to ROC curves
• ROC = Receiver Operating Characteristic
• Started in electronic signal detection
theory (1940s - 1950s)
• Has become very popular in biomedical
applications, particularly radiology and
imaging
• Also used in machine learning applications to
assess classifiers
• Can be used to compare tests/procedures
ROC curves: simplest case
• Consider diagnostic test for a disease
• Test has 2 possible outcomes:
– ‘postive’ = suggesting presence of disease
– ‘negative’
• An individual can test either positive or
negative for the disease
• Prof. Mean...
Hypothesis testing refresher
• 2 ‘competing theories’ regarding a population
parameter:
– NULL hypothesis H (‘straw man’)
– ALTERNATIVE hypothesis A (‘claim’, or
theory you wish to test)
• H: NO DIFFERENCE
– any observed deviation from what we
expect to see is due to chance variability
• A: THE DIFFERENCE IS REAL
Test statistic
• Measure how far the observed data are from
what is expected assuming the NULL H by
computing the value of a test statistic (TS)
from the data
• The particular TS computed depends on the
parameter
• For example, to test the population mean ,
the TS is the sample mean (or standardized
sample mean)
• The NULL is rejected fi the TS falls in a
user-specified ‘rejection region’
True disease state vs. Test result
not rejected rejected
No disease
(D = 0)

specificity
X
Type I error
(False +) 
Disease
(D = 1)
X
Type II error
(False -) 

Power 1 - ;
sensitivity
Disease
Test
Specific Example
Test Result
Pts with
Pts with
disease
disease
Pts without
Pts without
the disease
the disease
Test Result
Call these patients “negative” Call these patients “positive”
Threshold
Test Result
Call these patients “negative” Call these patients “positive”
without the disease
with the disease
True Positives
Some definitions ...
Test Result
Call these patients “negative” Call these patients “positive”
without the disease
with the disease
False
Positives
Test Result
Call these patients “negative” Call these patients “positive”
without the disease
with the disease
True
negatives
Test Result
Call these patients “negative” Call these patients “positive”
without the disease
with the disease
False
negatives
Test Result
without the disease
with the disease
‘‘
‘‘-’’
-’’ ‘‘
‘‘+’’
+’’
Moving the Threshold: right
Test Result
without the disease
with the disease
‘‘
‘‘-’’
-’’ ‘‘
‘‘+’’
+’’
Moving the Threshold: left
True
Positive
Rate
(sensitivity)
0%
100%
False Positive Rate
(1-specificity)
0% 100%
ROC curve
True
Positive
Rate
0
%
100%
False Positive Rate
0
%
100%
True
Positive
Rate
0
%
100%
False Positive Rate
0
%
100%
A good test: A poor test:
ROC curve comparison
Best Test: Worst test:
True
Positive
Rate
0
%
100%
False Positive
Rate
0
%
100
%
True
Positive
Rate
0
%
100%
False Positive
Rate
0
%
100
%
The distributions
don’t overlap at all
The distributions
overlap completely
ROC curve extremes
‘Classical’ estimation
• Binormal model:
– X ~ N(0,1) in nondiseased population
– X ~ N(a, 1/b) in diseased population
• Then
ROC(t) = (a + b-1
(t)) for 0 < t < 1
• Estimate a, b by ML using readings from
sets of diseased and nondiseased patients
ROC curve estimation with
continuous data
• Many biochemical measurements are in fact
continuous, e.g. blood glucose vs. diabetes
• Can also do ROC analysis for continuous (rather
than binary or ordinal) data
• Estimate ROC curve (and smooth) based on
empirical ‘survivor’ function (1 – cdf) in
diseased and nondiseased groups
• Can also do regression modeling of the test
result
• Another approach is to model the ROC curve
directlyas a function of covariates
Area under ROC curve (AUC)
• Overall measure of test performance
• Comparisons between two tests based on
differences between (estimated) AUC
• For continuous data, AUC equivalent to Mann-
Whitney U-statistic (nonparametric test of
difference in location between two
populations)
True
Positive
Rate
0
%
100%
False Positive
Rate
0
%
100
%
True
Positive
Rate
0
%
100%
False Positive
Rate
0
%
100
%
True
Positive
Rate
0
%
100%
False Positive
Rate
0
%
100
%
AUC = 50%
AUC = 90%
AUC = 65%
AUC = 100%
True
Positive
Rate
0
%
100%
False Positive
Rate
0
%
100
%
AUC for ROC curves
Interpretation of AUC
• AUC can be interpreted as the probability
that the test result from a randomly chosen
diseased individual is more indicative of
disease than that from a randomly chosen
nondiseased individual: P(Xi  Xj | Di = 1, Dj = 0)
• So can think of this as a nonparametric
distance between disease/nondisease test
results
Problems with AUC
• No clinically relevant meaning
• A lot of the area is coming from the range of
large false positive values, no one cares what’s
going on in that region (need to examine
restricted regions)
• The curves might cross, so that there might
be a meaningful difference in performance
that is not picked up by AUC
Examples using ROC analysis
• Threshold selection for ‘tuning’ an already
trained classifier (e.g. neural nets)
• Defining signal thresholds in DNA microarrays
(Bilban et al.)
• Comparing test statistics for identifying
differentially expressed genes in replicated
microarray data (Lönnstedt and Speed)
• Assessing performance of different protein
prediction algorithms (Tang et al.)
• Inferring protein homology (Karwath and King)
Homology Induction ROC
Concluding remarks – remaining
challenges in ROC methodology
• Inference for ROC curve when no ‘gold standard’
• Role of ROC in combining information?
• Incorporating time into ROC analysis
• Alternatives to ROC for describing test
accuracy?
• Generalization of positive/negative predictive
value to continuous test?
(+/-) predictive value = proportion of patients with
(+/-) result who are correctly diagnosed
= True/(True + False)

More Related Content

PPTX
Advance concept of screening_Nabaraj Paudel
PPTX
Group 1 ROC assignment explaining ROC and AUC
PDF
Advance concepts in screening
PPTX
Introduction to ROC Curve Analysis with Application in Functional Genomics
PPTX
ROC CURVE AND ANALYSIS.pptx
DOC
ch 18 roc.doc
PPTX
Roc curves
PPTX
ASSIGNMENT.pptx. explaining about Rock curve
Advance concept of screening_Nabaraj Paudel
Group 1 ROC assignment explaining ROC and AUC
Advance concepts in screening
Introduction to ROC Curve Analysis with Application in Functional Genomics
ROC CURVE AND ANALYSIS.pptx
ch 18 roc.doc
Roc curves
ASSIGNMENT.pptx. explaining about Rock curve

Similar to Receiver Operating Characteristics SIB-ROC.ppt (20)

PPT
Validity of a screening test
PPTX
ROC curve.pptx
PPT
7. evaluation of diagnostic test
PPTX
presentation Journal club notes 2021.pptx
PPTX
r_concepts
PDF
(20180524) vuno seminar roc and extension
PDF
Diagnosing a diagnostic april 08 2015
PPTX
PPTX
Diagnostic Tests for PGs
PDF
Side 2019 #8
PPTX
How to read a receiver operating characteritic (ROC) curve
PPT
Lecture11_ Evaluation Metrics for classification.ppt
PPTX
Tests of diagnostic accuracy
PPT
NY Prostate Cancer Conference - A. Vickers - Session 1: Traditional statistic...
PDF
3Assessing classification performance.pdf
PPTX
Prediction research: perspectives on performance Stanford 19May22.pptx
PPTX
Diagnostic test and agreement.pptx
PPTX
Screening of diseases
PPT
Diagnostic testing 2009
PPTX
clinical trial and study design copy 2.pptx
Validity of a screening test
ROC curve.pptx
7. evaluation of diagnostic test
presentation Journal club notes 2021.pptx
r_concepts
(20180524) vuno seminar roc and extension
Diagnosing a diagnostic april 08 2015
Diagnostic Tests for PGs
Side 2019 #8
How to read a receiver operating characteritic (ROC) curve
Lecture11_ Evaluation Metrics for classification.ppt
Tests of diagnostic accuracy
NY Prostate Cancer Conference - A. Vickers - Session 1: Traditional statistic...
3Assessing classification performance.pdf
Prediction research: perspectives on performance Stanford 19May22.pptx
Diagnostic test and agreement.pptx
Screening of diseases
Diagnostic testing 2009
clinical trial and study design copy 2.pptx
Ad

More from ssuserb53446 (15)

PPT
06-scheduling.ppt including multiple CPUs
PPT
02-archsupport.ppt Architecture Support for OPS
PPTX
Analog Communication Performance underNoise .pptx
PPTX
EFFECTS OF NOISE ON THE PERFOMANCE OF ANALOGUE.pptx
PPTX
SummerStudent17_HandsOn Data Cloud Computing.pptx
PPTX
Managed_ML_Systems_ sf Presentation.pptx
PPTX
Communications Systems Analogue Angle Modulation
PPT
Engineering Communication Systems Course
PPT
Embedded Systems ARM Computer Architecture
PPT
introduction to database systems ch2.ppt
PPTX
Asynchronous Transfer Mode Introductory Presentation
PPT
Database System Concepts Chapter 1: Introduction to Databases
PPTX
Entity relationship Models from Chapter 4
PPT
This is an introduction to Database Relational Model
PPT
Database System Introductory Concepts and All
06-scheduling.ppt including multiple CPUs
02-archsupport.ppt Architecture Support for OPS
Analog Communication Performance underNoise .pptx
EFFECTS OF NOISE ON THE PERFOMANCE OF ANALOGUE.pptx
SummerStudent17_HandsOn Data Cloud Computing.pptx
Managed_ML_Systems_ sf Presentation.pptx
Communications Systems Analogue Angle Modulation
Engineering Communication Systems Course
Embedded Systems ARM Computer Architecture
introduction to database systems ch2.ppt
Asynchronous Transfer Mode Introductory Presentation
Database System Concepts Chapter 1: Introduction to Databases
Entity relationship Models from Chapter 4
This is an introduction to Database Relational Model
Database System Introductory Concepts and All
Ad

Recently uploaded (20)

PPT
Predictive modeling basics in data cleaning process
PDF
Systems Analysis and Design, 12th Edition by Scott Tilley Test Bank.pdf
PDF
Votre score augmente si vous choisissez une catégorie et que vous rédigez une...
PPTX
Copy of 16 Timeline & Flowchart Templates – HubSpot.pptx
PDF
REAL ILLUMINATI AGENT IN KAMPALA UGANDA CALL ON+256765750853/0705037305
PPTX
STERILIZATION AND DISINFECTION-1.ppthhhbx
PPTX
(Ali Hamza) Roll No: (F24-BSCS-1103).pptx
PPTX
Topic 5 Presentation 5 Lesson 5 Corporate Fin
PDF
Microsoft Core Cloud Services powerpoint
PDF
How to run a consulting project- client discovery
PPTX
Introduction to Inferential Statistics.pptx
PPTX
sac 451hinhgsgshssjsjsjheegdggeegegdggddgeg.pptx
PPTX
QUANTUM_COMPUTING_AND_ITS_POTENTIAL_APPLICATIONS[2].pptx
PDF
Capcut Pro Crack For PC Latest Version {Fully Unlocked 2025}
PPTX
CYBER SECURITY the Next Warefare Tactics
PDF
Optimise Shopper Experiences with a Strong Data Estate.pdf
PPTX
Pilar Kemerdekaan dan Identi Bangsa.pptx
PPTX
Database Infoormation System (DBIS).pptx
PDF
annual-report-2024-2025 original latest.
PDF
168300704-gasification-ppt.pdfhghhhsjsjhsuxush
Predictive modeling basics in data cleaning process
Systems Analysis and Design, 12th Edition by Scott Tilley Test Bank.pdf
Votre score augmente si vous choisissez une catégorie et que vous rédigez une...
Copy of 16 Timeline & Flowchart Templates – HubSpot.pptx
REAL ILLUMINATI AGENT IN KAMPALA UGANDA CALL ON+256765750853/0705037305
STERILIZATION AND DISINFECTION-1.ppthhhbx
(Ali Hamza) Roll No: (F24-BSCS-1103).pptx
Topic 5 Presentation 5 Lesson 5 Corporate Fin
Microsoft Core Cloud Services powerpoint
How to run a consulting project- client discovery
Introduction to Inferential Statistics.pptx
sac 451hinhgsgshssjsjsjheegdggeegegdggddgeg.pptx
QUANTUM_COMPUTING_AND_ITS_POTENTIAL_APPLICATIONS[2].pptx
Capcut Pro Crack For PC Latest Version {Fully Unlocked 2025}
CYBER SECURITY the Next Warefare Tactics
Optimise Shopper Experiences with a Strong Data Estate.pdf
Pilar Kemerdekaan dan Identi Bangsa.pptx
Database Infoormation System (DBIS).pptx
annual-report-2024-2025 original latest.
168300704-gasification-ppt.pdfhghhhsjsjhsuxush

Receiver Operating Characteristics SIB-ROC.ppt

  • 1. Darlene Goldstein 29 January 2003 Receiver Operating Characteristic Methodology
  • 2. Outline • Introduction • Hypothesis testing • ROC curve • Area under the ROC curve (AUC) • Examples using ROC • Concluding remarks
  • 3. Introduction to ROC curves • ROC = Receiver Operating Characteristic • Started in electronic signal detection theory (1940s - 1950s) • Has become very popular in biomedical applications, particularly radiology and imaging • Also used in machine learning applications to assess classifiers • Can be used to compare tests/procedures
  • 4. ROC curves: simplest case • Consider diagnostic test for a disease • Test has 2 possible outcomes: – ‘postive’ = suggesting presence of disease – ‘negative’ • An individual can test either positive or negative for the disease • Prof. Mean...
  • 5. Hypothesis testing refresher • 2 ‘competing theories’ regarding a population parameter: – NULL hypothesis H (‘straw man’) – ALTERNATIVE hypothesis A (‘claim’, or theory you wish to test) • H: NO DIFFERENCE – any observed deviation from what we expect to see is due to chance variability • A: THE DIFFERENCE IS REAL
  • 6. Test statistic • Measure how far the observed data are from what is expected assuming the NULL H by computing the value of a test statistic (TS) from the data • The particular TS computed depends on the parameter • For example, to test the population mean , the TS is the sample mean (or standardized sample mean) • The NULL is rejected fi the TS falls in a user-specified ‘rejection region’
  • 7. True disease state vs. Test result not rejected rejected No disease (D = 0)  specificity X Type I error (False +)  Disease (D = 1) X Type II error (False -)   Power 1 - ; sensitivity Disease Test
  • 8. Specific Example Test Result Pts with Pts with disease disease Pts without Pts without the disease the disease
  • 9. Test Result Call these patients “negative” Call these patients “positive” Threshold
  • 10. Test Result Call these patients “negative” Call these patients “positive” without the disease with the disease True Positives Some definitions ...
  • 11. Test Result Call these patients “negative” Call these patients “positive” without the disease with the disease False Positives
  • 12. Test Result Call these patients “negative” Call these patients “positive” without the disease with the disease True negatives
  • 13. Test Result Call these patients “negative” Call these patients “positive” without the disease with the disease False negatives
  • 14. Test Result without the disease with the disease ‘‘ ‘‘-’’ -’’ ‘‘ ‘‘+’’ +’’ Moving the Threshold: right
  • 15. Test Result without the disease with the disease ‘‘ ‘‘-’’ -’’ ‘‘ ‘‘+’’ +’’ Moving the Threshold: left
  • 17. True Positive Rate 0 % 100% False Positive Rate 0 % 100% True Positive Rate 0 % 100% False Positive Rate 0 % 100% A good test: A poor test: ROC curve comparison
  • 18. Best Test: Worst test: True Positive Rate 0 % 100% False Positive Rate 0 % 100 % True Positive Rate 0 % 100% False Positive Rate 0 % 100 % The distributions don’t overlap at all The distributions overlap completely ROC curve extremes
  • 19. ‘Classical’ estimation • Binormal model: – X ~ N(0,1) in nondiseased population – X ~ N(a, 1/b) in diseased population • Then ROC(t) = (a + b-1 (t)) for 0 < t < 1 • Estimate a, b by ML using readings from sets of diseased and nondiseased patients
  • 20. ROC curve estimation with continuous data • Many biochemical measurements are in fact continuous, e.g. blood glucose vs. diabetes • Can also do ROC analysis for continuous (rather than binary or ordinal) data • Estimate ROC curve (and smooth) based on empirical ‘survivor’ function (1 – cdf) in diseased and nondiseased groups • Can also do regression modeling of the test result • Another approach is to model the ROC curve directlyas a function of covariates
  • 21. Area under ROC curve (AUC) • Overall measure of test performance • Comparisons between two tests based on differences between (estimated) AUC • For continuous data, AUC equivalent to Mann- Whitney U-statistic (nonparametric test of difference in location between two populations)
  • 22. True Positive Rate 0 % 100% False Positive Rate 0 % 100 % True Positive Rate 0 % 100% False Positive Rate 0 % 100 % True Positive Rate 0 % 100% False Positive Rate 0 % 100 % AUC = 50% AUC = 90% AUC = 65% AUC = 100% True Positive Rate 0 % 100% False Positive Rate 0 % 100 % AUC for ROC curves
  • 23. Interpretation of AUC • AUC can be interpreted as the probability that the test result from a randomly chosen diseased individual is more indicative of disease than that from a randomly chosen nondiseased individual: P(Xi  Xj | Di = 1, Dj = 0) • So can think of this as a nonparametric distance between disease/nondisease test results
  • 24. Problems with AUC • No clinically relevant meaning • A lot of the area is coming from the range of large false positive values, no one cares what’s going on in that region (need to examine restricted regions) • The curves might cross, so that there might be a meaningful difference in performance that is not picked up by AUC
  • 25. Examples using ROC analysis • Threshold selection for ‘tuning’ an already trained classifier (e.g. neural nets) • Defining signal thresholds in DNA microarrays (Bilban et al.) • Comparing test statistics for identifying differentially expressed genes in replicated microarray data (Lönnstedt and Speed) • Assessing performance of different protein prediction algorithms (Tang et al.) • Inferring protein homology (Karwath and King)
  • 27. Concluding remarks – remaining challenges in ROC methodology • Inference for ROC curve when no ‘gold standard’ • Role of ROC in combining information? • Incorporating time into ROC analysis • Alternatives to ROC for describing test accuracy? • Generalization of positive/negative predictive value to continuous test? (+/-) predictive value = proportion of patients with (+/-) result who are correctly diagnosed = True/(True + False)